Condorcet proved that if each member of a group has a probability greater than one-half of making a correct decision on a binary question, the probability that the majority decision is correct increases with group size, approaching certainty. The theorem has a dark mirror: if individual reliability is below chance, larger groups converge on error with the same mathematical certainty. The theorem is therefore a double-edged sword — a justification for inclusive governance under specific conditions, and a diagnostic of how collective decision-making can systematically amplify mistake. Modern machine learning ensembles — random forests, boosting algorithms, voting classifiers — operate on the same mathematical structure. The theorem is not a metaphor for AI ensembles; it is the literal foundation beneath them.
There is a parallel reading that begins not with the mathematical elegance of independent judgments but with the material substrate that makes modern AI possible. The theorem assumes independence as a given condition, but the actual infrastructure of AI — from the semiconductor supply chain concentrated in Taiwan to the cloud computing oligopoly of AWS, Azure, and Google Cloud — creates systemic correlation at every level. When all large language models depend on NVIDIA's CUDA architecture, when all training data flows through the same web scraping pipelines, when all compute runs through the same three providers, the independence condition isn't merely violated; it becomes structurally impossible.
The theorem's application to AI governance reveals an even deeper problem. The entry correctly notes that broadening participation could improve collective reliability, but this assumes that 'adequate information' about AI development is accessible to diverse participants. In practice, the knowledge required to meaningfully evaluate AI systems — understanding transformer architectures, assessing dataset biases, evaluating safety protocols — is locked behind corporate NDAs, academic paywalls, and the sheer technical complexity that years of specialized training barely penetrates. The fishbowl isn't just a failure mode; it's the only mode available when the material conditions of AI development create insurmountable barriers to genuine independence. The theorem thus becomes not a guide for better governance but a mathematical proof of why current AI development, structured as it is, cannot achieve the conditions for reliable collective judgment. The convergence on error isn't a bug; it's the predictable outcome of a system where correlation is baked into the infrastructure itself.
The theorem has two critical conditions: individual reliability greater than chance, and independence of errors among participants. The first condition is what makes universal education a mathematical necessity rather than merely a social good — the reliability of democratic decisions depends on the reliability of the individual judgments composing them.
The independence condition is what makes diversity a mathematical requirement rather than merely a political value. A 2024 study applying the theorem to ensembles of large language models found that majority voting across multiple LLMs produced only marginal improvements — the models, despite apparent diversity, had been trained on overlapping data and shared architectures, so their errors correlated. When errors correlate, the theorem's guarantee collapses. The fishbowl becomes the governing failure mode.
Researchers have explicitly deployed the theorem in neural network ensembles for medical diagnosis — combining outputs of multiple deep learning models trained on radiograph images, using majority voting to achieve diagnostic accuracy exceeding any individual model. The theorem provides the mathematical guarantee, and the guarantee depends on precisely the conditions Condorcet specified.
The theorem's application to AI governance is direct. Decisions about AI development are currently made by a narrow group whose information, assumptions, and professional networks substantially overlap. Under the theorem's conditions, their collective reliability is constrained by the correlation of their errors. Broadening participation — genuinely diverse participation, not demographic tokenism — would, under the theorem, produce more reliable collective decisions, provided the new participants are adequately informed.
The theorem appeared in the Essai sur l'application de l'analyse à la probabilité des décisions rendues à la pluralité des voix (1785), a 500-page treatise applying probability calculus to collective judgment — a founding document of social choice theory, decision theory, and the statistical analysis of testimony.
The result was rediscovered and formalized by Duncan Black in the 1950s, generalized in machine learning by Robert Schapire's 1990 paper 'The Strength of Weak Learnability,' and is now standard material in computer science curricula — usually without mention of the eighteenth-century mathematician whose proof it is.
Individual reliability matters. Below chance, larger groups converge on error; above chance, they converge on truth.
Independence is essential. Correlated errors nullify the theorem's guarantee, regardless of group size.
Diversity is mathematical, not decorative. A homogeneous group produces a chorus, not a jury.
The theorem runs inside AI. Random forests, boosting, and ensemble voting are literal implementations of Condorcet's proof.
Critics have noted that the theorem assumes binary decisions and independent errors — conditions rarely met exactly in practice. Defenders respond that approximate satisfaction of the conditions produces approximate benefits, and that the theorem's diagnostic value lies less in its exact predictions than in its identification of the conditions under which collective judgment is reliable or fails.
The fundamental tension here isn't between optimism and pessimism but between mathematical idealization and material constraint. When we ask 'Can the Condorcet theorem guide AI governance?' the entry's framework is 100% correct — the mathematics are sound, the applications to ensemble methods proven, the diagnostic value clear. The theorem does identify precisely what conditions would enable reliable collective judgment. But shift the question to 'Will these conditions actually obtain?' and the contrarian view dominates at 80% — the infrastructure of AI development, from compute monopolies to knowledge barriers, systematically produces correlation rather than independence.
The synthesis emerges when we recognize that Condorcet's theorem functions best not as a prescription but as a diagnostic tool that reveals the gap between ideal and actual conditions. Where the entry sees the theorem running 'literally' inside AI systems, this is 70% true for the mathematical structure but only 30% true for the independence assumptions — modern ensembles often achieve pseudo-diversity through techniques like dropout and data augmentation rather than genuine independence. The medical diagnosis example succeeds precisely because radiograph interpretation allows for meaningfully different training approaches; the LLM voting fails because language models share too much substrate.
The proper frame for Condorcet's theorem in the AI age is as a measurement instrument for systemic correlation. Rather than asking whether we can broaden participation (we can) or whether infrastructure determines outcomes (it largely does), we should ask: what specific interventions reduce correlation in AI development? Open-source models trained on different architectures, regulatory requirements for algorithmic diversity, compute access that doesn't flow through three companies — these become not just nice-to-haves but mathematical necessities for achieving the theorem's conditions. The theorem thus transforms from a justification for inclusive governance into a precise diagnostic of where technical monoculture creates systematic error amplification.