The conventional wisdom is that correlation between people's estimates is bad for collective intelligence — correlated errors don't cancel out. But this framing is too simple. Correlation that results from calibrated social influence — where more accurate people have more influence — can improve both individual and crowd accuracy. What matters is not how correlated people are, but what process generates the correlation.
This project brings together two complementary lines of research. The first (Pilgrim & Becker, arXiv) shows that social communication reliably improves individual accuracy even when group accuracy gets worse — and derives a mathematical decomposition that explains why. The second (CI Abstract 2024) shows that even a perfectly accurate crowd can make the wrong decision by vote, because the crowd average and the plurality vote can be statistically "decoupled."
Citation: Pilgrim & Becker (2024). "Communication Reliably Improves Individual But Not Group Accuracy." arXiv:2407.00199.
After social influence, the change in mean squared error decomposes as:
| Study | N | P(Individual Improves) | P(Crowd Improves) | Key Takeaway |
|---|---|---|---|---|
| Gurcay et al. 2015 | — | computing… | computing… | Unbounded + bounded questions |
| Navajas et al. 2018 | — | computing… | computing… | Social influence with swarms |
| Jayles et al. 2021 | — | computing… | computing… | Sorted vs random social info |
| Becker et al. 2017 | — | computing… | computing… | Network centralization effects |
Table shows proportion of group-question trials where accuracy improved after social influence. Results are computed live from the R pipeline output.
Direct empirical comparison of crowd accuracy (group average) versus individual accuracy across 3 datasets with known true values (Gurcay et al. 2015, Navajas et al. 2018, Jayles et al. 2021), using mean absolute error (MAE) on round-1 estimates. Improvement ratio = individual MAE / crowd MAE; values above 1 indicate the crowd outperforms the average individual. Pooled geometric mean ratio = 2.06 (95% bootstrap CI: 1.77–2.43) across 178 questions — the crowd consistently beats the average individual by roughly a factor of two. (Kao 2018 and Mercier et al. 2022 are in the combined dataset but lack ground-truth values and are therefore excluded from this accuracy comparison.)
Citation: Pilgrim & Becker (2024). "Estimates are Not Decisions: How Wise Crowds Can Make Unwise Choices." Collective Intelligence 2024.
N people estimate M options. Option k has mean μk = k × S, standard deviation σ = 1. Each person votes for their highest estimate.
| Option 1 | Option 2 | Vote | |
|---|---|---|---|
| Voter 1 | 40 | 20 | 1 |
| Voter 2 | 100 | 70 | 1 |
| Voter 3 | 10 | 110 | 2 |
| Average | 50 | 67 | – |
Average correctly identifies Option 2 as best (67 > 50), but the vote goes 2–1 for Option 1.
| Study | Year | Design | N (approx) | Questions | Accuracy figs? |
|---|---|---|---|---|---|
| Gurcay, Mellers & Baron | 2015 | Two-round bounded/unbounded estimation | ~200 | 16 | Yes |
| Navajas et al. | 2018 | Swarm AI social influence | ~5000 | 6 | Yes |
| Jayles et al. | 2021 | Sorted vs random social information | ~1000 | 10 | Yes |
| Kao & Couzin | 2018 | Jar estimation, social info quantity | ~400 | 1 | No true value |
| Mercier et al. | 2022 | Discussion effects on estimation | ~300 | varies | No true value |
| Becker, Brackbill & Centola | 2017 | Network structure (PNAS) | ~1200 | 5 | Data unavailable |
The two papers together suggest that (1) social influence is not inherently harmful to collective intelligence — it can reliably improve individual accuracy — but (2) the aggregation method used to convert beliefs into decisions critically matters. Averaging dominates voting across a wide range of empirically realistic parameters.
Inspect the empirical datasets underlying the analysis.
Each point = one participant-question pair. Blue = round 1 estimate, orange = round 2 estimate. Dashed line = true value. Hover over points to see details.