And when it breaks down
Condorcet proved in 1785 that if each juror independently has probability p of being correct on a binary decision, the probability that the majority is correct grows rapidly with the number of jurors — provided p > 0.5. Below p = 0.5 the theorem reverses: more jurors means a worse collective decision. At exactly p = 0.5 the group is no better than a coin flip, no matter how many jurors you add.
When agents share the same LLM backbone their errors are correlated — they tend to be wrong about the same things. We model this with a correlation parameter ρ (rho). At ρ = 0 agents are fully independent (classic theorem). At ρ = 1 every agent gives the same answer — your committee of seven is really just one voice repeated. The effective number of independent agents is neff = n / (1 + (n−1)ρ). Even small correlations (ρ = 0.3) collapse a panel of 31 agents down to roughly 3 effective independent voters.
In practice agents are not equally competent. A specialist model might have p = 0.85 on its domain while a generalist sits at p = 0.60. Simple majority treats them equally. Weighted majority assigns each agent a weight proportional to its log-odds: wi = log(pi / (1−pi)), which is the information-theoretically optimal weighting for independent binary voters. The chart below runs 50,000 Monte Carlo simulations comparing the two strategies for 7 agents with competences spread between pmin and pmax.