Introduction
In the Federalist No. 68, Hamilton (Reference Hamilton1788/1961: 458) commends the Philadelphia Convention's plan for an Electoral College to choose the president, saying:
It was desirable that the sense of the people should operate in the choice of the person to whom so important a trust was to be confided. This end will be answered by committing the right of making it … to men chosen by the people for the special purpose…
It was equally desirable that the immediate election should be made by men most capable of analyzing the qualities adapted to the station, and acting under circumstances favourable to deliberation… A small number of persons, selected by their fellows from the general mass, will be most likely to possess the information and discernment requisite to so complicated an investigation.
This article aims not so much to test that conjecture as to calibrate it – to define the parameters within which, and the reasons for which, it may or may not be true. We go on to assess, in a similar manner, the epistemic consequences of bicameralism, of party government, and of representative government of two ideal types (trustee and delegate) and mixtures of them.
What we refer to as ‘the Federalist's conjecture’ is the proposition that a smaller group of more competent people is epistemically superior to a larger group of less competent ones. In this article, we focus on the two reasons that Hamilton offers for that proposition. (There may well be others, but we do not explore them here.Footnote 1) Hamilton's first reason has to do with the supposed fact that members of the smaller group will be more ‘capable’: they will have been selected because they ‘possess the information and discernment’ required for the task. Call that the Selection Effect (Vermeule, Reference Vermeule2009: 45–46). Hamilton's second reason has to do with the supposed fact that, as a smaller group, electors would be ‘acting under circumstances favorable to deliberation’. Call that the Deliberation Effect.
In Hamilton's own statement of the proposition, and in popular discussions both before and since, emphasis falls most heavily on the first factor, the Selection Effect. Condorcet (Reference Condorcet1785/1976: 50) was neither the first nor the last to say that ‘a very numerous [group] cannot be composed of very enlightened men’ and to commend a smaller and more select decision-making body in consequence.Footnote 2
However, it is simply not the case that a smaller group of more competent people will always be epistemically superior to a larger group of less competent ones. Ironically, that is the clear lesson of Condorcet's Jury Theorem (CJT). Whether the Federalist's conjecture is true in any given case depends on many things: the number of people in each group, large and small; the independence and individual competence of people in each group; how the agenda is set and voting is structured in each group; and how much deliberation contributes to the collective competence of each group.
In the section ‘The folly of the Selection Effect’, we offer some sample calculations to illustrate the implications of the CJT for the Federalist's conjecture. There we see that, given voting bodies of the size typical in modern polities and individual voter competence even modestly better than random, the Selection Effect alone is unlikely to suffice to vindicate the Federalist's conjecture. If that conjecture is to be true at all, that will instead probably be largely due to the Deliberation Effect, as we show in the section ‘Factoring in the Deliberation Effect’.
The CJT tells us that, all else held equal, smaller groups are epistemically less reliable than larger groups.Footnote 3 When a smaller group is included as a veto player in an overall decision-making procedure, its poorer epistemic performance compromises, to a greater or lesser extent, the epistemic competence of the overall process. Those small-group elements in a decision-making process constitute, in Vermeule's (Reference Vermeule2009: 50–53) felicitous phrase, ‘epistemic bottlenecks’. Examples include cases of smaller upper chambers in bicameral legislatures and small pivotal parties in parliaments. In the section ‘Institutional epistemic bottlenecks’, we provide sample calculations of the scale of epistemic costs (i.e. the probability of reaching an incorrect decision) that those particular epistemic bottlenecks might impose.
Representatives who take a ‘trusteeship’ attitude toward their role would constitute another epistemic bottleneck, insofar as a large number of voters elect a small number of trustee-style representatives. In the section ‘Delegate vs. trustee-style representatives’, we provide sample calculations to show what might be the epistemic costs of representatives taking a trustee-style attitude toward their tasks, compared to their taking a delegate-style attitude or to our making decisions by the direct vote of the electorate as a whole.
Actual representatives typically take a more mixed approach. In that same section, we provide sample calculations to show how many delegate-style representatives it would take in a mixed assembly to make that assembly's overall epistemic performance almost as good as that of an assembly composed of trustees alone. ‘Almost as good’ might be good enough, because once again trustees would be capable of benefiting from the Deliberation Effect in the assembly, in a way electorally bound delegates would not.
The Condorcet Jury Theorem
Assume that a large number of voters have to vote on two alternatives, of which one is the correct choice. The standard CJT model uses the following formula to calculate the probability PN(n, p) that a majority vote among n independent voters will yield the correct outcome,Footnote 4 when the probability of each of the voters voting for the correct outcome is p:
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20160309110001774-0834:S1755773911000245_eqnU1.gif?pub-status=live)
The CJT tells us that, when voting by simple majority rule, the group is more likely to vote for the correct outcome than is an individual voter, and larger groups perform better than smaller groups. (That is referred to as the ‘non-asymptotic’ result.) The CJT further tells us that the probability that the majority votes for the correct outcome approaches 1 for large groups (the ‘asymptotic’ result).Footnote 5 Furthermore, sample calculations show that that typically happens relatively quickly.
The CJT rests on the following assumptions:
1. there is a correct choice to be found;
2. each voter is competent, that is, each voter has the same probability p > 0.5 to detect the correct alternative;
3. the votes are probabilistically independent (conditional on the state of the world regarding the correct alternative), that is, voters are not influenced by each others’ votes, or common causes that introduce biases;
4. all voters vote sincerely for the alternative they believe to be correct.
We will discuss each assumption in turn. All can of course be criticized. However, our conclusion will be that, while obviously political life does not accord with those assumptions completely, the CJT framework captures much that is important in politics.
The ‘correct choice’ assumption
Some claim that politics is seldom, perhaps never, about making ‘correct’ choices (Black, Reference Black1958: 163; Miller, Reference Miller1992: 56). We shall offer here three reasons for proceeding with modeling on the assumption that political decisions can be correct or incorrect – or at least that there are truth-apt elements in most political decisions.
The first is ad hominem. The Federalist's own conjecture was predicated precisely on that same assumption. That is to say, the Federalist's conjecture assumed that there are facts that a smaller group of wiser people would be more likely to find. In assessing the soundness of Federalist's conjecture, it is only right that we should proceed from that same assumption.
Second, surely at least some of the claims that matter in politics can be either true or false (e.g. that man-made climate change is happening, that consumption taxes will be regressive in their impact). Not everything of political consequence is of that sort; there are other things that matter politically, which do not admit of truth or falsity. For ease of reference, let us call the former ‘facts’ and the latter ‘values’. Most actual political decisions will constitute a combination of both. Nevertheless, supposing that getting the facts right will lead to better outcomes, where there are facts to be gotten right, institutional designers ought to be interested in mechanisms for maximizing correct answers in the factual domain.
Third, let us just note in passing that there are also some far stronger claims that might be made here. The nature of value judgments is a subject of heated dispute within ethical theory. Some meta-ethicists are ‘moral realists’ and insist that there are facts of the matter about moral propositions. Others are ‘conventionalists’ about morality – but even they would agree that there is indeed a ‘truth of the matter about what is the convention around here,’ which is to say, what the convention of the society in question actually is. All that is strictly required to vindicate our use of the CJT is that there be a standard of ‘correctness’, that is independent from the votes of the people making the decision. As long as such a standard applies to a domain of questions relevant for political decision making, the CJT is of genuine political interest.
The ‘competence’ assumption
The classic CJT assumes, for mathematical convenience, a homogenous set of voters with an identical level of competence above p > 0.5 in two-option choices. But homogeneity is not strictly necessary. First, it has been shown that the CJT holds for voters of heterogeneous competence, so long as the distribution of competences within the group remains symmetric around a constant mean of $$\[-->$<>\bar{p}\: \gt \:0.5$$ as group size increases (Grofman et al., Reference Grofman, Owen and Feld1983). Second, it has been shown that at least the asymptotic result (group competence converges to 1 as the group size tends to infinity) holds even with asymmetric distributions of competence so long as
$\bar{p}\: \gt \:0.5$ (Owen et al., Reference Owen, Grofman and Feld1989; Dietrich, Reference Dietrich2008).Footnote 6
However, if people are incompetent, in the sense of p < 0.5, the CJT goes into reverse, and the larger the number of voters the more likely it will be that a majority vote among them will yield the wrong result. (We return to this issue in the penultimate section.) Of course, any given individual on any given occasion might well be wrong. That would be unsurprising. But under what conditions might a large group of individuals manifest judgment that is worse than random? On the face of it, it would seem that for them to be systematically worse than random, people would have to know what the right answer is and perversely conspire to vote against what they know to be the truth. Or they must all be influenced by a common factor that leads them astray. Both cases would be a violation of the ‘independence’ assumption (Sunstein, Reference Sunstein2009: 175). It is to that which we next turn.
The ‘independence’ assumption
The classic CJT formulation assumes that votes are probabilistically independent of one another, conditional on the state of the world regarding the correct alternative. That is to assume that the voters neither causally influence each other, nor are jointly influenced by the same causes. Since voters are typically part of a complex network of influences, the independence condition is unlikely to be met perfectly.Footnote 7
Politically, the most serious threat to the independence assumption is probably the existence of ‘opinion leaders’. If every one of the n voters slavishly votes exactly the same way as a single opinion leader, then we do not have n independent pieces of information being pooled in the vote: we have only one, that of the opinion leader. It is then not a case of ‘many minds’, but rather a case of ‘one mind, many mimics’. The probability that the majority vote is correct does not in that case increase with the number of voters; the probability that the majority of a group who slavishly vote the same way as a single opinion leader is correct is merely the probability that that opinion leader is correct.
While such extreme cases of opinion leadership are devastating to the CJT result, weaker cases are not. True, any degree of opinion leadership will always (to some greater or lesser extent) decrease the speed with which the probability of the majority being right approaches 1. But both the asymptotic and non-asymptotic CJT results still obtain, just so long as the probability π that any given voter will follow the opinion leader (i.e. vote the same way he does) is sufficiently small in relation to the probability p of the voter being right when acting independently from the opinion leader. Specifically, group competence PN(n, p) converges to 1 just so long as
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20160309110001774-0834:S1755773911000245_eqnU5.gif?pub-status=live)
(and it converges to the probability that the opinion leader himself chooses the correct outcome if that inequality is reversed). That proposition follows straightforwardly from Boland et al. (Reference Boland, Proschan and Tong1989; see Spiekermann and Goodin, forthcoming: Appendix A.2). Or again, if there are equal-and-opposed opinion leaders – or even just many uncorrelated opinion leaders – the effects of their opinion leadership on the majority vote will largely cancel out and the CJT results will still obtain (Spiekermann and Goodin, forthcoming).
Note, importantly, that jury theorems are still possible with weaker independence assumptions. Dietrich and List (Reference Dietrich and List2004) prove a new jury theorem where dependence is caused by shared evidence (see also Ladha, Reference Ladha1992, Reference Ladha1993, Reference Ladha1995). In that account, the CJT concerns the ability of individuals and groups to reach the same judgment as an ideal interpreter would on the basis of the shared evidence that is available to them. The CJT, thus understood, shows that the majority of a large group of independent voters (now independent conditional on the evidence) converges on the competence of an ideal interpreter who processes that evidence in the best possible way, and who chooses the option most likely to be true given that evidence. This way of modeling relaxes the independence assumption: votes no longer need to be independent conditional on the state of the world; all that is required is independence conditional on the available evidence, and correlation induced by shared evidence is no longer a problem.
This extension is plausible because the quality of group judgments naturally depends on the available evidence. It also explains why groups can be wrong even if the CJT's asymptotic result would suggest near infallibility: the probability of the majority's being correct is now upper-bounded by the probability that the evidence available to it, interpreted in the best possible way, is actually pointing to the correct result. We return to these issues in our discussion of the Deliberation Effect in the section ‘Mechanisms’ below.
A more general treatment of the effect of correlation and new jury theorems arising from this can be found in Kaniovski (Reference Kaniovski2010), Kaniovski and Zaigraev (Reference Kaniovski and Zaigraev2011) and Zaigraev and Kaniovski (Reference Zaigraev and Kaniovski2010). For purposes of modeling reported in this article, we proceed with the idealized assumption of independence, but arguments analogous to ours could be made with weaker independence conditions.
The ‘sincerity’ assumption
Finally, the CJT presupposes that voters vote for the alternative that they sincerely believe to be the correct one. But they might well be tempted to cast their vote ‘strategically’, because strategic voters realize that their votes only matter if they are pivotal. Strategic voters ponder the question: ‘would I vote for this if my fellow voters were exactly split on this matter?’ And often enough, the information derived from the assumption of a split distribution of votes overrules one's own private belief on the matter. In addition, if there are more than two options to choose from, there may also be strategic votes for second-best alternatives to prevent the victory of some option they think is even worse. If voters vote strategically rather than sincerely in that way, the CJT does not necessarily hold.Footnote 8
We do not deny that this can happen, particularly in small-scale committees and sometimes even among large electorates. But for strategic voting of that kind to be commonplace, the voters need to acquire a lot of information about their opponents and their reasoning processes, and compute this information adequately. One also needs to assume that voters care primarily about getting the overall outcome right, not about expressing their own sincere belief – an assumption far less obvious than is sometimes claimed. No matter how one assesses the plausibility of strategic voting in real-life settings, it is first worth seeing what would happen in ‘idealized’ circumstances where everyone always votes sincerely, as a baseline, before feeding those complications into any models.
The folly of the Selection Effect
The CJT tells us that it is perfectly possible for the majority among a large number of less competent (but still better-than-random) voters to be more likely correct than the majority among a small number of more competent voters. The Federalist's seemingly obvious conjecture is thus not necessarily true. It all depends on the competence of people in each group, and on the number of people in each group.
On the basis of formula 1, we can undertake some sample calculations to see how the Federalist's conjecture might fare under various assumptions. First let us look at how the CJT operates in the smaller-scale settings where its consequences are more easily calculated and hence to which it has most familiarly been applied. Table 1 presents results for various values of p and n.
Table 1 Probability that the majority of a group is correct, for given group sizes and competences
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary-alt:20160626112430-06815-mediumThumb-S1755773911000245_tab1.jpg?pub-status=live)
Focus on the case of 1001 people, each of whom is individually only p = 0.550 likely to be correct in a two-option choice. Suppose members of that group, following the Federalist's advice, decide to leave the choice to a small group of 11 persons who are more capable than themselves. How competent does each of those 11 persons have to be for that smaller group to outperform the larger group epistemically? Given 1001 persons with competence of p = 0.550, the probability that a majority among them is correct is PN = 0.999. For a majority vote among merely 11 people to be as reliable as that, those 11 people would have to be individually over p = 0.800 likely to be correct. Maybe this is not strictly impossible, but it certainly seems unlikely that those 1001 people can find 11 people who are that much more competent than themselves.
So far as the Selection Effect is concerned, the Federalist's conjecture becomes even more implausible once the number of voters grows very large, as is the case in real-world electorates.Footnote 9 (When we talk about applying our models to ‘real-world settings’, it is the number of people typically involved rather than any other aspect of the ‘real world’ to which we will typically be referring.)
Take for example a case modeled akin to the 2008 US election.Footnote 10 There, the president was elected indirectly by an Electoral College consisting of 539 members, instead of being elected directly by the 131,000,000 people who voted in that year's presidential election. For the sake of this example, suppose – contrary to fact, but in line with the Federalist's hopes and expectations – that members of the Electoral College exercise their own judgment rather than necessarily voting the same way as the citizens of the state they represent. And let us make the modest assumption that voters among the electorate as a whole are individually p = 0.510 competent. For a majority among the 539 members of the Electoral College to be epistemically as reliable as a majority among 131,000,000 such voters, members of the Electoral College would have to be individually p = 0.976 competent.Footnote 11 That seems inconceivable.
Thus, the Federalist's conjecture is unlikely to be true by virtue of the Selection Effect alone. Among groups the size of contemporary electorates, anyway, it seems unlikely that a small group of individuals can be chosen that is sufficiently more competent for its collective competence to exceed that of the electorate as a whole.
Factoring in the Deliberation Effect
The second prong of the Federalist's conjecture is that a smaller group might be more likely to be correct by virtue of its deliberative superiority. This possibility is less discussed among commentators on the Federalist or the Electoral College, but it is much discussed by a host of contemporary deliberative democrats (Bohman and Rehg, Reference Bohman and Rehg1997; Dryzek, Reference Dryzek2000; Luskin et al., Reference Luskin, Fishkin and Jowell2002; Goodin, Reference Goodin2008).
Mechanisms
There are various ways in which this might occur. Most straightforwardly, deliberation might be expected to increase the competence of individuals who are party to the deliberation. That would of course flow straightforwardly on into an increase of the collective competence of the deliberating group. But as we have seen, the gain in individual competence would typically have to be unrealistically large to merely catch up with the more numerous population. Without in any way minimizing the practical importance of increasing individual competence through deliberation, we set it aside for purposes of our subsequent discussion.
Beyond that, there are at least three additional mechanisms by which deliberation might make a small group collectively more competent, without altering the individual competence of the people comprising it:
1. In the course of deliberation, the group might uncover some new evidence.Footnote 12 On one plausible interpretation discussed in the section ‘The “independence” assumption’ above, all that the CJT proves is that the majority of a large group of individuals is almost certain to choose the best alternative given the evidence available to it (Dietrich and List, Reference Dietrich and List2004). That opens up possibilities for deliberation to improve the epistemic performance of the group. If deliberation unearths better evidence, the chance of the group being correct increases without an increase in individual competence.
2. Remember, too, that all the CJT proves is that the majority of a large group of independent voters who are better-than-random is almost certain to choose the best alternative from among the alternatives presented to them. The CJT model thus requires alternatives to be well defined. If the alternatives are not well defined, the correct alternative cannot be chosen. Insofar as deliberation among a smaller group leads them to an improved definition of the alternatives, and insofar as one of those alternatives is correct, the group is able to select it. That is another way in which deliberation among small groups can cause the collective competence of the group to increase, without the individual competence of its members increasing.Footnote 13
3. A third effect might be on processes of reasoning. Among large groups, decisions are inevitably taken by aggregating people's votes on the ‘bottom line’. Among smaller groups, there are more discursive opportunities for ‘premise-probing’ (Goodin, Reference Goodin2008: 87–92). That can obviously improve individual-level decision making, insofar as each internalizes the lessons learned from those probes of her own and others’ premises and revises her chain of reasoning accordingly. It can also improve group-level decision making, insofar as the decision process of the group incorporates votes on premises and not just votes on the ‘bottom line’ (List, Reference List2005; Bovens and Rabinowicz, Reference Bovens and Rabinowicz2006; Pigozzi, Reference Pigozzi2006). Incorporating premises is often (but not always) a better way to pool information and therefore reach better results.
Those are ways in which a Deliberation Effect might occur, independently of any Selection Effect.Footnote 14
Notice, however, that deliberation might also reduce the collective competence of the group, insofar as it leads to a violation of the independence assumption.Footnote 15 We do not deny the possibility that deliberation might have that negative effect on the collective competence of the group, as well as the positive ones outlined above. All those effects can be present, and the epistemic net effect of deliberation still be positive, so long as a reduction of independence is outweighed by other effects.
Consequences
Exactly how much those various Deliberation Effects might contribute to a smaller group's epistemic competence will vary with circumstances. Deliberation may increase or decrease correlation between voters. In the worst case, deliberation might induce dependence where there was none before, without increasing competence at all. But this is an unnecessarily pessimistic picture. In many cases, deliberation will reduce the overwhelming effects of opinion leaders, will make voters aware of a plurality of arguments, and will make it less likely that they just blindly adopt the opinions of their peers. Successful deliberation should bring us closer to the CJT's idealized independence assumption. In addition, it will often increase group competence along the pathways discussed.
We make no efforts to model these effects here in any formal way, so we cannot say just how big those effects are likely to be. Deliberative democrats offer many reasons to think that they may be large. But – and this is all that we need for present purposes – they do not have to be very big, in order to make the Federalist's conjecture true.
Let us demonstrate that by taking a deliberately low-ball estimate of the epistemic value of deliberation. Suppose that the superior deliberative circumstances of smaller representative assemblies will make them at least one percentage point more likely to reach the correct decision than the electorate as a whole. We can then calculate how individually competent each member of the representative assembly would have to be, in order for the assembly's collective competence to be within one percentage point of that of the electorate as a whole. In that way, we can calculate how big the Selection Effect needs to be in order for the Deliberation and Selection Effects combined to vindicate the Federalist's conjecture, even if the Deliberation Effect is no stronger than this very conservative assumption.
Return to a case akin to the 2008 US presidential election.Footnote 16 If we could have relied upon the Deliberation Effect to improve the smaller Electoral College's collective competence by just one percentage point, then voters who are p = 0.510 individually competent would only have had to choose representatives who are p = 0.551 individually competent in order for the collective competence of the 538 members of the Electoral College to exceed the collective competence of the electorate as a whole.Footnote 17 That would seem to be an eminently realistic ambition.
Thus, the Federalist's conjecture is plausibly true. But to be plausibly true, the conjecture has to rely heavily on the Deliberation Effect rather than upon the Selection Effect alone.
Institutional epistemic bottlenecks
One lesson of the CJT is that incorporating a smaller decisional body in the political process constitutes a bottleneck that undermines the epistemic reliability of the process overall. There are various familiar examples of such small decisional bodies in the institutional architecture of modern democracies. Smaller upper houses in bicameral legislatures are one example.Footnote 18 Small pivotal parties in parliaments are another. Just how much of an epistemic bottleneck might those be – assuming epistemic performance is one attribute (among many others) by which we should evaluate our political institutions?
Strong bicameralism
By ‘strong bicameralism’ we mean having two directly elected sets of representatives sitting in separate chambers of differing sizes, each of which has a veto on legislation.
We focus upon the particular example of the US Congress, comprising a larger House of Representatives (with 435 members) and a much smaller Senate (with 100 members). For the limiting case, suppose (again contrary to fact – but just to see what would happen in what, from the perspective of the CJT, would be the epistemically worst-case scenario) that each member takes a trustee-style attitude toward the representational task, voting purely on the basis of his or her own judgment. Further suppose that the competence of individuals in each chamber is p = 0.55. Finally, let us take the number of people in 2008 voting for members of the House of Representatives (112,588,380) as the number of people who would have voted in a referendum that year if direct democracy had been used instead.
With those assumptions in place, we are in a position to explore the epistemic consequences of giving the smaller upper chamber a veto on legislation passed by the larger lower chamber, compared with a unicameral assembly and with direct democracy. With an electorate the size of that in 2008, the probability of the majority among them coming to the correct conclusion is almost one, so long as voters are individually appreciably better than random at reaching the correct conclusion.Footnote 19 So the probability of a correct decision from a direct-democratic vote would be PN ≈ 1.0. If the US Congress consisted of the House of Representatives alone, the probability that a majority of 435 would be correct is, on the assumptions outlined above, PN = 0.982. That is not that far off direct democracy's mark. But the actual US Congress is strongly bicameral. For an enactment to become law, it must pass both the 435-member House and the 100-member Senate. On the assumptions above, the probability both chambers will reach the correct decision and thus enact it into law drops to 0.828.Footnote 20
Thus, on the assumptions above, requiring legislation also to be passed by the Senate rather than merely by the House alone leads to a 15 percentage point reduction in the epistemic competence of the US Congress (from just over 98 percent to just under 83 percent). Of course, direct democracy would do epistemically even better, but only by a small margin. It is strong bicameralism that carries the much greater epistemic cost (i.e. probability of reaching an incorrect decision). The Senate would only cease to be an epistemic bottleneck if it reached a collective competence close to one – which would require an extraordinarily strong Deliberation Effect, or extremely competent Senators. And say what one might about the competence or the superior deliberative capacities of the smaller upper chamber, it seems highly unlikely that they will be able to close a gap of that magnitude.Footnote 21
Parliaments with small pivotal parties
In the real world, representative assemblies are almost invariably organized along party lines. In strongly whipped parliaments, representatives almost always vote in line with the instructions of their party, rather than exercising their own independent judgment or following the instructions of their constituents.
In a parliament with strong parties of that sort, there would be only as many independent sources of judgment at the point of voting on legislation as there are parties. Of course there are more physical bodies in the lobby. But all those MPs are not independent sources of judgment: they all vote purely as their party dictates. Reducing the number of points of independent judgment in that way can have dire epistemic consequences, the CJT teaches us.
Let us suppose, however, that those MPs are not purely lobby fodder, just doing as they are told and not exercising any independent judgment anywhere at all. Let us suppose that they exercise independent judgment in the party room rather than on the floor. That is to say, rather than exercising independent judgment when voting on legislation, they instead exercise independent judgment within the party in trying to shape its policies in ways they think best.Footnote 22 For the purposes of our modeling below, we will suppose that party policy is chosen purely by a majority vote among the party's MPs. That is clearly contrary to fact: but again, that is a conservative assumption; if there are more independent influences on party policy than that (e.g. the entire party conference), so much the better epistemically.Footnote 23
Imagine a parliament that is strictly whipped, in which MPs always vote strictly according to their party's instructions. Assume that there are 603 MPs and that 302 votes are required to enact legislation. Further assume that MPs are individually p = 0.55 competent, both across the parliament as a whole and within each of the parties into which they are divided.
With those assumptions in place, we can calculate the epistemic consequences of dividing legislatures into parties. Suppose first that there were no parties at all in the parliament, as was the Federalist's original vision (Madison, Reference Madison1787a/1961: 57). On the assumptions sketched above, the probability that a majority vote in such a parliament would be correct is PN = 0.993. Introducing parties as such does not do much to reduce that. If the parliament were divided into two almost-equal parties (one with 302 MPs, the other with 301), the probability that a majority vote in parliament would be correct – which is just the probability that the majority among those 302 MPs in the government party room would be correct – is still 0.954. And if there were three equal parties (201 MPs each), then on the above assumptions the probability that a majority comprising any two of them would reach the correct decision would be 0.983.Footnote 24
It is not the dividing of parliament into well-whipped parties that causes epistemic damage, but rather the introduction of unequally sized parties. And even that, as such, is not the problem. Suppose the parliament were composed of two large parties (275 MPs each) and one small one (with 51 MPs). If any combination of parties could constitute a majority, then the probability of a correct decision emerging from that parliament is still 0.976. Epistemic problems arise only where the small party is pivotal, that is to say, has to be included in the winning majority.Footnote 25 On the assumptions described above, having to include such a small pivotal party in the majority reduces the probability that the parliament would make a correct decision to 0.762.
If pivotal, the small party forms an epistemic bottleneck. The calculations just presented enable us now to say just how much of a bottleneck it might be, in real-world sized political settings. On the assumptions above, having to include a small pivotal party of the sort described would reduce the epistemic competence of the parliament in question from over 97 percent to just over 76 percent – a drop of 21 percentage points. Of course, from an epistemic point of view, not dividing parliament into strong parties would be even better, but only by a couple of percentage points. It is small parties being pivotal that carries the much greater epistemic cost, in terms of increasing the probability of an incorrect decision.
Delegate vs. trustee-style representatives
Having calibrated the epistemic costs of familiar ‘bottlenecks’ – small upper chambers, small pivotal parties – in the institutional architecture of legislative assemblies, let us turn our attention back to the representatives themselves.
So far, we have been assuming that representatives act as ‘trustees’, each voting on the basis of his or her own independent judgment as to what is best. Alternatively, of course, representatives could act as ‘delegates’, each voting strictly according to instructions given by his or her electors. To what extent would representatives taking this different attitude toward their task solve the epistemic problem (i.e. the increased risk that comes from letting decisions be made by smaller representative assemblies rather than by direct vote of the electorate at large) as discussed in the sections ‘The Condorcet Jury Theorem’ and ‘The folly of the Selection Effect’ above?
Constituency-based representation with delegates
Making public decisions through a legislative assembly composed entirely of representatives who take a delegate-style attitude toward their task is different from direct democracy. The reason is that representatives represent particular constituencies. First let us reflect upon the epistemic consequences of bunching voters into constituencies, even if their representatives see their role purely as delegates.
To calculate the probability that delegate-style representatives acting on instruction in this way reach the correct collective decision, we need to apply formula (1) twice. First we use it to calculate the probability that the majority of constituents is correct in the instructions they give to their representative. Since the representatives always vote as instructed, we equate that with the probability that each representative's own vote will be correct. We then use formula (1) for a second time to calculate the probability that a majority of representatives with that level of individual competence will reach the correct collective decision.
Think, for example, about the 2005 British General Election. Each of the 646 Members of Parliament was elected by around 40,000 voters. So long as the individual competence of each voter is relevantly over one-half, with numbers like 40,000 the probability that the majority among them is correct is going to be very close to 1.Footnote 26 Representatives instructed by these large constituencies will reach correct decisions with near certainty. Thus, there is virtually no epistemic difference between delegate-style representative government and direct democracy, where electorates are of the size we typically see in real-world politics.
An epistemic advantage of trustee-style representatives
Is there any credible scenario by which the collective competence of a group of trustee-style representatives might overtake that of groups of delegate-style representatives? Looking at it one way, it seems unlikely. After all, trustees would have to be individually as competent as delegates for their collective competence literally to equal that of the same number of delegates under the same circumstances. And since the competence of each delegate following the instructions of large numbers of relevantly better-than-random constituents is very close to 1, the individual competence of trustees would have to be near 1 to match that.
But assemblies the size of the 2005 UK House of Commons could perform almost as well (be within one percentage point as accurate) as the electorate as a whole with individual competences of individual representatives being as low as p = 0.546 (see Appendix 2). The Deliberation Effect can plausibly close that sort of a gap.
Notice that the Deliberation Effect can work only for trustee-style representatives, not for delegate-style ones. However much delegate-style representatives discuss matters with one another in the assembly, at the end of the day delegates must by definition vote as they have been instructed by their constituents.Footnote 27 Trustee-style delegates are not so bound, and their votes can indeed be changed by deliberation within the chamber. That affords trustee-style representative government some real scope for ‘closing the epistemic gap’ between it and delegate-style representative government.
Assemblies with both delegates and trustees
So far, we have been treating representative assemblies as if they were composed either of all trustees or all delegates. A more realistic scenario is that, on any given issue, some representatives will behave as trustees and some as delegates.Footnote 28
In modeling that scenario, we take (for analytic convenience) the small-scale case of an assembly with 99 representatives, each of whom is elected by 101 constituents. We assume the individual competence of each representative, when acting as a trustee on the basis of his or her own judgment, is p = 0.550. We assume that the individual competence of each constituent is also p = 0.550, which (from formula 1) means that the collective competence of a majority among the 101 of them is PN = 0.844. By definition, a representative who acts as a delegate votes strictly in accordance with the majority of his or her constituents, so each delegate-representative's individual competence in this scenario is also, therefore, p = 0.844.
Figure 1 displays the way in which the collective competence of such an assembly varies, depending upon the number of representatives who behave as delegates.Footnote 29 Where there are no delegates and all trustees among the 99 members of the assembly, the collective competence of the assembly is 0.841. But as we see from Figure 1, it rises rapidly with just a few delegate-style representatives, closely approaching 1 by the time there are just 20 delegate-style representatives in the 99-member assembly.
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary-alt:20160626112427-10006-mediumThumb-S1755773911000245_fig1g.jpg?pub-status=live)
Figure 1 The epistemic competence of an assembly mixing delegates and trustees. The x-axis shows the number of delegates among 99 representatives, the y-axis group competence (trustees with p = 0.55, delegates are instructed by 101 voters with p = 0.55).
Thus, it does not take anything like half of the representatives behaving as delegates for the collective competence of an assembly to come very close to one. At first blush, this might come as a surprise. But upon reflection, it should not be so surprising. If the individual competence is relatively low, most results will end up around an equal split of opinions. If there is a small group that votes more reliably, that increases substantially the group's collective competence overall.Footnote 30
But what if the electorate is worse than random?
Throughout our discussion, we have been assuming that the individual competence of members of the electorate is relevantly greater than one-half.Footnote 31 If it is, then given the size of real-world electorates it is almost certain that the majority of them reach the correct decision. But if it is not – if voters are worse than random in making their political choices – then by the same CJT logic it is virtually certain that the majority of them will reach the wrong conclusion. In that case, trustee-style representatives exercising their own independent judgment – even if the probability of their being right is low – would be epistemically superior to the even greater likelihood of error by the majority among the electorate at large.Footnote 32
One final way of vindicating the Federalist's conjecture, then, would be to suppose that the competence of individual voters is p < 0.5. Then trustee-style representative government is always epistemically superior to direct democracy. But that way of vindicating the Federalist's conjecture would leave another large question unanswered, which is why we should let people with competence p < 0.5 choose their own representatives. By the CJT logic, we know that they would virtually always make the wrong choice.
Maybe this line of thought could vindicate the Federalist's conjecture if the electorate were better at choosing people than policies. Here are some reasons why they might be:
• Perhaps the electorate is not systematically better than random at making complex decisions, but is better than random at choosing trustees on the basis of their evaluations of their general character and competence. They might thereby end up electing trustee-style representatives who are much better than electors themselves at complex decisions.
• Direct democracy asks people to decide on many issues in which their personal stakes are very low (Downs, Reference Downs1957; Owen and Grofman, Reference Owen and Grofman1984). Under such conditions, biases and attempts to influence voters are more likely to succeed.
• Similarly, if voters think their competence is low, they might be tempted to ‘epistemically free-ride’ by just siding with the majority. Such a lack of independence may lead to cascades and hence bad outcomes (List and Pettit, Reference List and Pettit2004; Vermeule, Reference Vermeule2009: 46–47).
• It is possible that many people do not make up their own mind but rather follow opinion leaders. In the extreme case, people follow either Fox News or MSNBC, so that the station with more followers always wins, and n is effectively reduced to 1.
Of course, the last three of these forces will work to contaminate voters’ choices of representatives as well. Those last three forces will also work to contaminate the decisions of representatives themselves – who will after all be accountable to their electorates come the next election, and who are likely to act as ‘delegates’, at least in part, in view of that fact.
Beyond all that, however, notice that the Federalist's conjecture will be vindicated if and only if the electorate is better at choosing people than policies in a very precise way. Individual voters must be individually less competent than 0.5 at choosing policies, but more competent than 0.5 at choosing people. Perhaps sometimes the electorate's differential competence for choosing people and policies takes that very specific knife-edge form. But again, that seems unlikely to be the case with any generality, which is what would be required for the Federalist's conjecture to be vindicated via this route.
Conclusion
Epistemic models like the CJT provide guidance on how best to design institutions to reach correct conclusions, insofar as there are truths that political decision making is trying to track and insofar as political decision makers are genuinely trying to track them. Relying on fewer rather than more independent sources of opinion is generally epistemically disadvantageous. But those effects can be relatively easily overcome insofar as the smaller group is better able to deliberate together or insofar as the smaller group contains a sufficient proportion of people following the lead of a much larger number.
Acknowledgments
Earlier versions were presented at Australian National University, Trinity College Dublin, University of Maryland, Washington University in St Louis, and the American Political Science Association Annual Meetings in Toronto. We are grateful for comments, then and later, from Geoff Brennan, Jan Burgers, Randy Calvert, Dave Estlund, Clarissa Hayward, Christian List, Jenny Mansbridge, Tony McGann, David Miller, Cara Nine, Joe Oppenheimer, Andrew Rehfeld, Norman Schofield, Piotr Swistak, Larry Temkin, and this journal's editor and anonymous referees.
Appendix 1 Estimation of group competence with large n
We calculate results for PN(n, p) for large n by normal approximation:
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20160309110001774-0834:S1755773911000245_eqnU6.gif?pub-status=live)
where Φ is the cumulative distribution function of the normal distribution.
Appendix 2 Calibrating the Selection Effect in Real-world Representative Assemblies
In Table A2.1, we explore the collective epistemic competence of three actual assemblies:
Table A2.1 Estimated necessary competence of representatives to make their collective decision epistemically equal to that of the electorate or no more than 1% point worse
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary-alt:20160626112441-11212-mediumThumb-S1755773911000245_tab2.jpg?pub-status=live)
• First is the case most immediately on Hamilton's mind: the Electoral College for the first US presidential election in 1789. There were 69 members voting in the Electoral College that year.Footnote 33 There were 38,818 votes cast for electors that year.Footnote 34
• Second is the Electoral College in 2008. That year 538 members of the Electoral College were chosen on the basis of 131,237,603 popular votes.
• Third is the UK House of Commons in 2005. That year 27,110,727 popular votes were cast for 646 MPs.
For each of those representative assemblies, we calculate two statistics. The first, appearing as the upper entry for each assembly and competence level in Table A2.1, is how competent each individual in the representative assembly would have to be in order for the assembly's collective competence literally to match that of the electorate as a whole. That is the individual competence of each representative that would be required in order for the Federalist's conjecture literally to be true of those assemblies by virtue of the Selection Effect alone.
We calculate those isocompetence values using the Grofman–Dummkopf–Witkopf theorem (Grofman et al., Reference Grofman, Owen and Feld1983). Let there be two groups, the first of size n with individual competence p 1, the second with size n + y and competence p 2, such that p 1 > p 2. The Grofman–Dummkopf–Witkopf theorem states that the two groups are approximately equally competent if
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20160309110001774-0834:S1755773911000245_eqnU7.gif?pub-status=live)
Solving for p 1 yields
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20160309110001774-0834:S1755773911000245_eqnU8.gif?pub-status=live)
which we use to derive the isocompetence values stated in Table A2.1.
With the lower number for each assembly and competence level in Table A2.1, we report how individually competent each member of the representative assembly would have to be, in order for the assembly's collective competence to be within one percentage point of that of the electorate as a whole.
The results reported in the lower right corner of the cells in Table A2.1 are derived by using the inequality
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20160309110001774-0834:S1755773911000245_eqnU9.gif?pub-status=live)
with n 1 < n 2 and p 1 > p 2. We estimate PN(n 2, p 2) with normal approximation, substitute using (1), and use Mathematica© to solve for p 1.
Appendix 3 Multi-party parliaments
Let there be three parties A, B, and C with respective collective competences P A, P B, and P C. Assume that any combination of two parties constitutes a majority. To determine the competence of a parliament with strictly whipped parties P Parl, we calculate the probability of a majority for the correct alternative:
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20160309110001774-0834:S1755773911000245_eqnU10.gif?pub-status=live)
If, for some reason, party A has to be part of the winning majority, the competence of the parliament reduces to
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20160309110001774-0834:S1755773911000245_eqnU11.gif?pub-status=live)
Appendix 4 Group competence with delegate–trustee mix
Let there be an assembly with d delegates and t trustees, with overall size n = d + t. Delegates have competence pd and trustees competence pt. Let v = 〈v 1,v 2, …, vn〉 be a voting profile. The set of all logically possible voting profiles that instantiate a majority for the correct solution is denoted W. Let cd(v) be the number of delegates, and ct(v) be the number of trustees that vote for the correct solution in v. The group competence is
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20160309110001774-0834:S1755773911000245_eqnU12.gif?pub-status=live)