When Can Human-AI Teams Outperform Individuals? Tight Bounds with Impossibility Guarantees

arXiv cs.AI Papers

Summary

This paper derives tight theoretical bounds for human-AI teams, proving when confidence-based aggregation leads to complementarity and establishing impossibility results under specific error correlations.

arXiv:2605.08710v1 Announce Type: new Abstract: Human-AI teams fail to outperform their best member in 70% of studies, yet no theory specifies when complementarity is achievable. We derive tight bounds for the broad class of confidence-based aggregation rules by integrating signal detection theory with information-theoretic analysis, yielding four results: (1) a complementarity theorem (teams outperform individuals iff error correlation $\rho_{HM} < \rho^*$, with $\rho^* \approx a$ in the symmetric near-chance regime); (2) minimax bounds showing gains scale as $\Theta(\sqrt{\Delta d})$ with metacognitive sensitivity difference; (3) an impossibility result proving no confidence-based aggregation rule achieves complementarity when $\rho_{HM} \geq \rho^*$; and (4) multi-class generalization $\rho^*_K \approx \rho^*/\sqrt{K-1}$. Predictions match observed team accuracy ($R = 0.94$ on ImageNet-16H, $R = 0.91$ on CIFAR-10H) and the multi-class threshold scaling holds on human data ($R = 0.93$, $K = 16$), with robustness under non-Gaussian distributions. The framework explains why complementarity is rare and provides actionable design formulas; results apply to aggregation, not to interactive deliberation that generates novel answers.
Original Article Export to Word Export to PDF
View Cached Full Text

Cached at: 05/12/26, 07:22 AM

# When Can Human-AI Teams Outperform Individuals? Tight Bounds with Impossibility Guarantees
Source: [https://arxiv.org/html/2605.08710](https://arxiv.org/html/2605.08710)
Jikun WuStellaris AI Limited, Hong Kong, ChinaBrain Investing Limited, Hong Kong, ChinaSiu\-Ming YiuThe University of Hong Kong, Hong Kong, China

###### Abstract

Human\-AI teams fail to outperform their best member in 70% of studies, yet no theory specifies when complementarity is achievable\. We derive tight bounds for the broad class of*confidence\-based aggregation*rules by integrating signal detection theory with information\-theoretic analysis, yielding four results: \(1\) a complementarity theorem \(teams outperform individuals iff error correlationρH​M<ρ∗\\rho\_\{HM\}<\\rho^\{\*\}, withρ∗≈a\\rho^\{\*\}\\approx ain the symmetric near\-chance regime\); \(2\) minimax bounds showing gains scale asΘ​\(Δ​d\)\\Theta\(\\sqrt\{\\Delta d\}\)with metacognitive sensitivity difference; \(3\) an impossibility result proving no confidence\-based aggregation rule achieves complementarity whenρH​M≥ρ∗\\rho\_\{HM\}\\geq\\rho^\{\*\}; and \(4\) multi\-class generalizationρK∗≈ρ∗/K−1\\rho^\{\*\}\_\{K\}\\approx\\rho^\{\*\}/\\sqrt\{K\-1\}\. Predictions match observed team accuracy \(R=0\.94R=0\.94on ImageNet\-16H,R=0\.91R=0\.91on CIFAR\-10H\) and the multi\-class threshold scaling holds on human data \(R=0\.93R=0\.93,K=16K=16\), with robustness under non\-Gaussian distributions\. The framework explains why complementarity is rare and provides actionable design formulas; results apply to aggregation, not to interactive deliberation that generates novel answers\.

Keywords:human\-AI complementarity; formal bounds; metacognitive sensitivity; signal detection theory; impossibility theorems

## Introduction

Artificial intelligence increasingly assists human decision\-making in medical diagnosis\[[22](https://arxiv.org/html/2605.08710#bib.bib28)\], judicial decisions\[[14](https://arxiv.org/html/2605.08710#bib.bib29)\], and autonomous systems\[[23](https://arxiv.org/html/2605.08710#bib.bib30)\]\. The promise of human\-AI collaboration rests oncomplementarity\(combined performance exceeding individual performance\)\[[9](https://arxiv.org/html/2605.08710#bib.bib3)\]\. However, a meta\-analysis of 106 studies found human\-AI combinations performedworsethan the best individual on average \(Hedges’g=−0\.23g=\-0\.23\), achieving complementarity in only 30% of cases\[[26](https://arxiv.org/html/2605.08710#bib.bib1)\]\.

This raises a fundamental question:Under what formal conditions does human\-AI collaboration provably outperform either agent alone?Despite research on trust calibration\[[30](https://arxiv.org/html/2605.08710#bib.bib10)\], explanation quality\[[3](https://arxiv.org/html/2605.08710#bib.bib11)\], and mental models\[[2](https://arxiv.org/html/2605.08710#bib.bib12)\], no prior work has established tight bounds with matching achievability and impossibility results\.

We address this gap by developing a mathematical framework unifying signal detection theory\[[18](https://arxiv.org/html/2605.08710#bib.bib7)\], Bayesian aggregation\[[24](https://arxiv.org/html/2605.08710#bib.bib2)\], and ensemble learning\[[27](https://arxiv.org/html/2605.08710#bib.bib21)\]\.

Scope\.Our results characterize*confidence\-based aggregation*: any rule𝒜​\(Y^H,Y^M,cH,cM\)\\mathcal\{A\}\(\\hat\{Y\}\_\{H\},\\hat\{Y\}\_\{M\},c\_\{H\},c\_\{M\}\)that produces a team prediction from agents’ answers and confidence signals\. This class subsumes confidence\-weighted selection, Bayesian model averaging, learned deferral policies, and ensemble methods\. It does*not*cover interactive deliberation in which agents generate new answers through dialogue\[Navajas2018,BarreraLemarchand2025\]or reasoning processes that combine knowledge in ways absent from individual responses\[[21](https://arxiv.org/html/2605.08710#bib.bib20)\]; complementarity gains beyond our bounds may be achievable in such settings, but lie outside the present framework\.

Our contributions are:

1. 1\.Acomplementarity theoremwith explicit closed\-form conditions specifying when team accuracy can exceed individual accuracy \(Theorem[1](https://arxiv.org/html/2605.08710#Thmtheorem1)\);
2. 2\.Tight minimax boundsshowing how metacognitive sensitivity \(the ability to distinguish correct from incorrect predictions\[[8](https://arxiv.org/html/2605.08710#bib.bib6)\]\) determines achievable gains \(Theorem[2](https://arxiv.org/html/2605.08710#Thmtheorem2)\);
3. 3\.Animpossibility resultproving that above a critical error correlationρ∗\\rho^\{\*\}, no aggregation achieves complementarity \(Theorem[3](https://arxiv.org/html/2605.08710#Thmtheorem3)\);
4. 4\.Multi\-class generalizationvalidated on human behavioral data \(ImageNet\-16H,K=16K=16\), with robustness analysis under non\-Gaussian assumptions\.

Our framework connects to classic results in collective intelligence\. The Condorcet Jury Theorem\[[4](https://arxiv.org/html/2605.08710#bib.bib18)\]shows that majority voting achieves perfect accuracy with independent voters; our bounds characterize exactly how independence violations \(error correlation\) constrain achievable gains\. The “wisdom of crowds” phenomenon\[[25](https://arxiv.org/html/2605.08710#bib.bib24),[12](https://arxiv.org/html/2605.08710#bib.bib27)\]depends on diversity of opinions; ourρ∗\\rho^\{\*\}threshold quantifies precisely how much diversity \(low error correlation\) is required for complementarity to emerge\.

We validate predictions on ImageNet\-16H\[[24](https://arxiv.org/html/2605.08710#bib.bib2)\]and CIFAR\-10H, achievingR=0\.94R=0\.94andR=0\.91R=0\.91correlations between predicted and observed team accuracy\. These results formalize when and why human\-AI collaboration succeeds, and when it cannot\.

## Related Work

### Human\-AI Complementarity

\[[3](https://arxiv.org/html/2605.08710#bib.bib11)\]showed AI explanations do not reliably improve team performance when accuracies match\.\[[10](https://arxiv.org/html/2605.08710#bib.bib13)\]found expertise critically moderates complementarity\.\[[9](https://arxiv.org/html/2605.08710#bib.bib3)\]formalized Complementary Team Performance \(CTP\), identifying information and capability asymmetry as sources\.\[[1](https://arxiv.org/html/2605.08710#bib.bib23)\]demonstrated optimal cue combination in human dyads, providing foundations for multi\-agent integration\.

### Collective Intelligence

Our work builds on collective intelligence research\.\[[25](https://arxiv.org/html/2605.08710#bib.bib24)\]articulated conditions for “wisdom of crowds”: diversity, independence, decentralization, and aggregation\.\[[28](https://arxiv.org/html/2605.08710#bib.bib25)\]identified a collective intelligence factor in groups, distinct from individual IQ\.\[[17](https://arxiv.org/html/2605.08710#bib.bib26)\]showed social influence reduces diversity and can undermine collective accuracy, a phenomenon our error correlation termρH​M\\rho\_\{HM\}formalizes\.\[[12](https://arxiv.org/html/2605.08710#bib.bib27)\]reviewed information aggregation mechanisms, noting that optimal combination depends on confidence calibration, precisely what our metacognitive sensitivity parameterddcaptures\.

### Bayesian Models and Metacognition

\[[24](https://arxiv.org/html/2605.08710#bib.bib2)\]developed a Bayesian model showing complementarity requires low error correlationρH​M\\rho\_\{HM\}, establishing necessary but not sufficient conditions\.\[[13](https://arxiv.org/html/2605.08710#bib.bib22)\]extended this to correlated ensembles\. Metacognitive sensitivity, measured via meta\-d′d^\{\\prime\}\[[8](https://arxiv.org/html/2605.08710#bib.bib6)\], is critical for collaboration\.\[[16](https://arxiv.org/html/2605.08710#bib.bib4)\]showed AI with lower accuracy but higher metacognitive sensitivity can enhance team performance\.\[[15](https://arxiv.org/html/2605.08710#bib.bib5)\]demonstrated metacognitive sensitivity enables trust calibration\. Neural evidence suggests metacognitive sensitivity reflects prefrontal confidence representations\[[7](https://arxiv.org/html/2605.08710#bib.bib8)\], providing implementational grounding for ourddparameter\. We derive the first bounds connecting metacognitive sensitivity to complementarity guarantees\.

### Decision Aggregation and Impossibility Results

The Condorcet Jury Theorem\[[4](https://arxiv.org/html/2605.08710#bib.bib18),[6](https://arxiv.org/html/2605.08710#bib.bib19)\]establishes majority voting achieves perfect accuracy with independent votes\.\[[21](https://arxiv.org/html/2605.08710#bib.bib20)\]proved the diversity prediction theorem\.\[[27](https://arxiv.org/html/2605.08710#bib.bib21)\]unified these into bias\-variance\-diversity decomposition\.\[[20](https://arxiv.org/html/2605.08710#bib.bib15)\]derived deferral learning bounds;\[[19](https://arxiv.org/html/2605.08710#bib.bib16)\]proved NP\-hardness of optimal classifier\-rejector pairs\.\[[5](https://arxiv.org/html/2605.08710#bib.bib17)\]showed reliable collaborative strategies face fundamental limits\. Our work provides the first tight bounds specifying both achievability and impossibility for human\-AI complementarity\.

## Problem Formulation

### Setup and Notation

Consider binary decisions with ground truthY∈\{0,1\}Y\\in\\\{0,1\\\}\. HumanHHand AIMMproduce predictionsY^H,Y^M∈\{0,1\}\\hat\{Y\}\_\{H\},\\hat\{Y\}\_\{M\}\\in\\\{0,1\\\}with confidence scorescH,cM∈\[0,1\]c\_\{H\},c\_\{M\}\\in\[0,1\]\. LetaH=P​\(Y^H=Y\)a\_\{H\}=P\(\\hat\{Y\}\_\{H\}=Y\)andaM=P​\(Y^M=Y\)a\_\{M\}=P\(\\hat\{Y\}\_\{M\}=Y\)denote accuracies\. Table[1](https://arxiv.org/html/2605.08710#Sx3.T1)summarizes key notation\.

Table 1:Summary of notation used throughout the paper\.Following\[[16](https://arxiv.org/html/2605.08710#bib.bib4)\], confidence generation uses signal detection theory:

θi\|Y=1\\displaystyle\\theta\_\{i\}\|Y=1∼𝒩​\(μ1\(i\),σ2\)\\displaystyle\\sim\\mathcal\{N\}\(\\mu\_\{1\}^\{\(i\)\},\\sigma^\{2\}\)\(1\)θi\|Y=0\\displaystyle\\theta\_\{i\}\|Y=0∼𝒩​\(μ0\(i\),σ2\)\\displaystyle\\sim\\mathcal\{N\}\(\\mu\_\{0\}^\{\(i\)\},\\sigma^\{2\}\)\(2\)fori∈\{H,M\}i\\in\\\{H,M\\\}\. Confidence is computed via the standard normal CDF:

ci=Φ​\(θi−τiσ\)c\_\{i\}=\\Phi\\left\(\\frac\{\\theta\_\{i\}\-\\tau\_\{i\}\}\{\\sigma\}\\right\)\(3\)whereτi\\tau\_\{i\}is agentii’s decision threshold\.

###### Definition 1\(Metacognitive Sensitivity\)

The metacognitive sensitivity of agentiiis:

di=μ1\(i\)−μ0\(i\)σd\_\{i\}=\\frac\{\\mu\_\{1\}^\{\(i\)\}\-\\mu\_\{0\}^\{\(i\)\}\}\{\\sigma\}\(4\)quantifying ability to assign higher confidence to correct predictions\.

Intuition: An agent with highddis “well\-calibrated”: confident when correct, uncertain when wrong\. This enables effective collaboration because partners can weight their contributions appropriately\. Neurally,ddrelates to prefrontal confidence representations that track decision reliability\[[7](https://arxiv.org/html/2605.08710#bib.bib8)\]\.

###### Definition 2\(Error Correlation\)

The error correlation is:

ρH​M=Corr​\(𝟏​\[Y^H≠Y\],𝟏​\[Y^M≠Y\]\)\\rho\_\{HM\}=\\text\{Corr\}\(\\mathbf\{1\}\[\\hat\{Y\}\_\{H\}\\neq Y\],\\mathbf\{1\}\[\\hat\{Y\}\_\{M\}\\neq Y\]\)\(5\)

Intuition: HighρH​M\\rho\_\{HM\}means agents make mistakes on the same items: they share blind spots rather than complementing each other\. This captures the “diversity” requirement in wisdom\-of\-crowds: low correlation implies agents bring independent information\.

###### Definition 3\(Complementarity\)

A team\(H,M\)\(H,M\)achieves complementarity if there exists an aggregation rule𝒜\\mathcal\{A\}such that:

a𝒜=P​\(𝒜​\(Y^H,Y^M,cH,cM\)=Y\)\>max⁡\(aH,aM\)a\_\{\\mathcal\{A\}\}=P\(\\mathcal\{A\}\(\\hat\{Y\}\_\{H\},\\hat\{Y\}\_\{M\},c\_\{H\},c\_\{M\}\)=Y\)\>\\max\(a\_\{H\},a\_\{M\}\)\(6\)

## Theoretical Results

### The Complementarity Theorem

###### Theorem 1\(Complementarity Conditions\)

Leta∗=max⁡\(aH,aM\)a^\{\*\}=\\max\(a\_\{H\},a\_\{M\}\),a−=min⁡\(aH,aM\)a\_\{\-\}=\\min\(a\_\{H\},a\_\{M\}\),eH=1−aHe\_\{H\}=1\-a\_\{H\},eM=1−aMe\_\{M\}=1\-a\_\{M\}, ande∗=min⁡\(eH,eM\)e^\{\*\}=\\min\(e\_\{H\},e\_\{M\}\)\. Within the class of confidence\-based aggregation rules under the symmetric SDT model of Eqs\.[1](https://arxiv.org/html/2605.08710#Sx3.E1)–[2](https://arxiv.org/html/2605.08710#Sx3.E2), complementarity is achievable if and only ifρH​M<ρ∗\\rho\_\{HM\}<\\rho^\{\*\}, where in the symmetric near\-chance regime

ρ∗≈e∗⋅\(a−−a∗\+a∗​a−\)eH⋅eM\\rho^\{\*\}\\;\\approx\\;\\frac\{e^\{\*\}\\cdot\(a\_\{\-\}\-a^\{\*\}\+a^\{\*\}a\_\{\-\}\)\}\{e\_\{H\}\\cdot e\_\{M\}\}\(7\)which reduces toρ∗≈a\\rho^\{\*\}\\approx awhenaH=aM=aa\_\{H\}=a\_\{M\}=aneara=0\.5a=0\.5\.111Equation[7](https://arxiv.org/html/2605.08710#Sx4.E7)is exact in the equal\-variance, near\-chance limit; for high\-accuracy or strongly asymmetric agents the threshold also depends ondH,dMd\_\{H\},d\_\{M\}via the SDT signal\-to\-noise structure\. Full derivations and tight bounds for the asymmetric regime are provided in the extended version\.

Plain English: When both agents have 80% accuracy \(a=0\.8a=0\.8\), complementarity is possible only if their error correlation is below 0\.8\. Higher individual accuracyparadoxicallymakes complementarity harder; near\-perfect agents \(a→1a\\to 1\) require near\-zero error correlation\.

0\.50\.50\.60\.60\.70\.70\.80\.80\.90\.91100\.20\.20\.40\.40\.60\.60\.80\.811Impossibleρ≥ρ∗\\rho\\geq\\rho^\{\*\}Achievableρ<ρ∗\\rho<\\rho^\{\*\}Individual Accuracy \(aa\)Error Correlation \(ρH​M\\rho\_\{HM\}\)Figure 1:Phase diagram of complementarity\. Blue region: complementarity achievable \(ρH​M<ρ∗\\rho\_\{HM\}<\\rho^\{\*\}\)\. Orange region: complementarity impossible regardless of confidence\-based aggregation method \(ρH​M≥ρ∗\\rho\_\{HM\}\\geq\\rho^\{\*\}\)\. The black line shows the thresholdρ∗≈a\\rho^\{\*\}\\approx ain the symmetric near\-chance regime\. Palette is Okabe\-Ito \(colorblind\-safe\)\.Worked Example\(Figure[1](https://arxiv.org/html/2605.08710#Sx4.F1)\): Two radiologists with 75% accuracy trained at the same institution and sharing diagnostic heuristics \(highρH​M\\rho\_\{HM\}, e\.g\.,0\.800\.80\) likely exceedρ∗\\rho^\{\*\}at their accuracy regime and cannot improve by any confidence\-based combination; if they instead bring complementary expertise \(lowρH​M\\rho\_\{HM\}, e\.g\.,0\.400\.40, satisfyingρH​M<ρ∗\\rho\_\{HM\}<\\rho^\{\*\}\), complementarity becomes achievable\.

WhenρH​M<ρ∗\\rho\_\{HM\}<\\rho^\{\*\}, the optimal team accuracy is:

ateam∗=a∗\+e∗⋅Φ​\(d−−κ​\(ρH​M\)2\)a\_\{\\text\{team\}\}^\{\*\}=a^\{\*\}\+e^\{\*\}\\cdot\\Phi\\left\(\\frac\{d\_\{\-\}\-\\kappa\(\\rho\_\{HM\}\)\}\{\\sqrt\{2\}\}\\right\)\(8\)whered−=min⁡\(dH,dM\)d\_\{\-\}=\\min\(d\_\{H\},d\_\{M\}\)and thecorrelation correctionis:

κ​\(ρ\)=2⋅Φ−1​\(1\+ρ2\)\\kappa\(\\rho\)=\\sqrt\{2\}\\cdot\\Phi^\{\-1\}\\left\(\\frac\{1\+\\rho\}\{2\}\\right\)\(9\)
The optimal aggregation rule𝒜∗\\mathcal\{A\}^\{\*\}that achievesateam∗a\_\{\\text\{team\}\}^\{\*\}is confidence\-weighted selection:

𝒜∗​\(cH,cM\)=\{Y^Hif​wH​cH\>wM​cMY^Motherwise\\mathcal\{A\}^\{\*\}\(c\_\{H\},c\_\{M\}\)=\\begin\{cases\}\\hat\{Y\}\_\{H\}&\\text\{if \}w\_\{H\}c\_\{H\}\>w\_\{M\}c\_\{M\}\\\\ \\hat\{Y\}\_\{M\}&\\text\{otherwise\}\\end\{cases\}\(10\)where the optimal weights, derived as the Bayes\-optimal log\-likelihood\-ratio coefficients under the joint SDT model, are:

wi=didH2\+dM2−2​ρH​M​dH​dMw\_\{i\}=\\frac\{d\_\{i\}\}\{\\sqrt\{d\_\{H\}^\{2\}\+d\_\{M\}^\{2\}\-2\\rho\_\{HM\}d\_\{H\}d\_\{M\}\}\}\(11\)
Intuition forκ​\(ρ\)\\kappa\(\\rho\): With correlated errors, disagreement cases are enriched for low\-confidence situations where neither agent has strong evidence\. The correctionκ​\(ρ\)\\kappa\(\\rho\)penalizes this: higher correlation means more “uninformative” disagreements, withκ​\(0\)=0\\kappa\(0\)=0growing monotonically toκ​\(1\)→∞\\kappa\(1\)\\to\\infty\.

Proof\.LetpAp\_\{A\}be the probability of agreement andpD=1−pAp\_\{D\}=1\-p\_\{A\}the probability of disagreement\.

Step 1: Under the SDT model, the joint error probability satisfies:

P​\(EH,EM\)=eH⋅eM\+ρH​M⋅eH​\(1−eH\)⋅eM​\(1−eM\)P\(E\_\{H\},E\_\{M\}\)=e\_\{H\}\\cdot e\_\{M\}\+\\rho\_\{HM\}\\cdot\\sqrt\{e\_\{H\}\(1\-e\_\{H\}\)\\cdot e\_\{M\}\(1\-e\_\{M\}\)\}\(12\)
Step 2: Disagreement probability is:

pD=eH\+eM−2​P​\(EH,EM\)p\_\{D\}=e\_\{H\}\+e\_\{M\}\-2P\(E\_\{H\},E\_\{M\}\)\(13\)
Step 3: In disagreement, exactly one agent is correct\. The gain from optimal selection is:

Δ=pD⋅\[P​\(select correct\|disagree\)−12\]\\Delta=p\_\{D\}\\cdot\\left\[P\(\\text\{select correct\}\|\\text\{disagree\}\)\-\\frac\{1\}\{2\}\\right\]\(14\)
Step 4: Using Bayes’ rule on confidence\-accuracy relationships, optimal selection probability exceeds1/21/2iff the more confident agent is more likely correct\. Under SDT, this holds iffρH​M<ρ∗\\rho\_\{HM\}<\\rho^\{\*\}\.

Step 5: Theκ​\(ρ\)\\kappa\(\\rho\)correction arises because:

P​\(low confidence\|disagree\)=1\+ρ2P\(\\text\{low confidence\}\|\\text\{disagree\}\)=\\frac\{1\+\\rho\}\{2\}\(15\)yieldingκ​\(ρ\)=2⋅Φ−1​\(\(1\+ρ\)/2\)\\kappa\(\\rho\)=\\sqrt\{2\}\\cdot\\Phi^\{\-1\}\(\(1\+\\rho\)/2\)\.□\\square

###### Corollary 1\(Maximum Achievable Gain\)

WhenρH​M=0\\rho\_\{HM\}=0\(independent errors\), the maximum complementarity gain is:

Δmax=e∗⋅Φ​\(d−2\)\\Delta\_\{\\max\}=e^\{\*\}\\cdot\\Phi\\left\(\\frac\{d\_\{\-\}\}\{\\sqrt\{2\}\}\\right\)\(16\)

### Minimax Bounds

###### Theorem 2\(Tight Minimax Bounds\)

LetΔ​d=\|dH−dM\|\\Delta d=\|d\_\{H\}\-d\_\{M\}\|\. Under worst\-case distributions withρH​M<ρ∗\\rho\_\{HM\}<\\rho^\{\*\}:

12​π⋅Δ​d1\+Δ​d2≤Δ∗≤2π⋅Δ​d⋅e∗\\frac\{1\}\{2\\sqrt\{\\pi\}\}\\cdot\\frac\{\\Delta d\}\{\\sqrt\{1\+\\Delta d^\{2\}\}\}\\leq\\Delta^\{\*\}\\leq\\frac\{\\sqrt\{2\}\}\{\\sqrt\{\\pi\}\}\\cdot\\sqrt\{\\Delta d\\cdot e^\{\*\}\}\(17\)Both bounds are tight\.

Sketch\.Upper bound\.The complementarity gainΔ=pD⋅𝔼​\[Φ​\(d−/2\)−Φ​\(d−/2−δd\)\]\\Delta=p\_\{D\}\\cdot\\mathbb\{E\}\[\\Phi\(d\_\{\-\}/\\sqrt\{2\}\)\-\\Phi\(d\_\{\-\}/\\sqrt\{2\}\-\\delta\_\{d\}\)\], whereδd\\delta\_\{d\}is the SDT\-signal advantage of the more sensitive agent on disagreements\. A first\-order Taylor expansion ofΦ\\Phiaroundd−d\_\{\-\}, combined with\|δd\|≤Δ​d\|\\delta\_\{d\}\|\\leq\\Delta dand the Gaussian density boundϕ​\(0\)=1/2​π\\phi\(0\)=1/\\sqrt\{2\\pi\}, yields the upper bound2/π⋅Δ​d⋅e∗\\sqrt\{2/\\pi\}\\cdot\\sqrt\{\\Delta d\\cdot e^\{\*\}\}\.

Lower bound\.The infimum is attained by the worst\-case adversarial confidence allocation that places all disagreement mass at confidences where SDT discrimination saturates\. Substituting this configuration into the optimal selection probability givesΔ∗≥\(1/2​π\)⋅Δ​d/1\+Δ​d2\\Delta^\{\*\}\\geq\(1/2\\sqrt\{\\pi\}\)\\cdot\\Delta d/\\sqrt\{1\+\\Delta d^\{2\}\}\.

Tightness follows from the existence of distributions achieving each bound: a Gaussian prior on the latent SDT signal achieves the upper bound; a two\-point adversarial confidence distribution achieves the lower bound\. Full derivations are provided in the extended version\.□\\square

Intuition: Complementarity gain scales asΘ​\(Δ​d\)\\Theta\(\\sqrt\{\\Delta d\}\); larger metacognitive differences enable larger gains, but with diminishing returns\. An AI that “knows what it doesn’t know” \(Δ​d\\Delta dlarge\) provides more value than one that is uniformly confident\.

###### Proposition 1\(Variance of Team Accuracy\)

The variance of team accuracy under optimal aggregation is bounded by:

Var​\(ateam\)≤e∗​\(1−e∗\)n⋅\(1\+ρH​M⋅dH​dMd−2\)\\text\{Var\}\(a\_\{\\text\{team\}\}\)\\leq\\frac\{e^\{\*\}\(1\-e^\{\*\}\)\}\{n\}\\cdot\\left\(1\+\\rho\_\{HM\}\\cdot\\frac\{d\_\{H\}d\_\{M\}\}\{d\_\{\-\}^\{2\}\}\\right\)\(18\)wherennis the number of decisions\.

### Impossibility Result

###### Theorem 3\(Impossibility of Complementarity\)

WhenρH​M≥ρ∗\\rho\_\{HM\}\\geq\\rho^\{\*\},noconfidence\-based aggregation rule𝒜\\mathcal\{A\}achieves complementarity:

a𝒜≤a∗\+ϵ​\(ρH​M;ρ∗\)a\_\{\\mathcal\{A\}\}\\leq a^\{\*\}\+\\epsilon\(\\rho\_\{HM\};\\rho^\{\*\}\)\(19\)whereϵ​\(ρH​M;ρ∗\)→0\\epsilon\(\\rho\_\{HM\};\\rho^\{\*\}\)\\to 0asρH​M→ρ∗\\rho\_\{HM\}\\to\\rho^\{\*\}from above andϵ​\(ρH​M;ρ∗\)<0\\epsilon\(\\rho\_\{HM\};\\rho^\{\*\}\)<0strictly forρH​M\>ρ∗\\rho\_\{HM\}\>\\rho^\{\*\}\. In particular, the supremum over𝒜\\mathcal\{A\}of the gaina𝒜−a∗a\_\{\\mathcal\{A\}\}\-a^\{\*\}is non\-positive wheneverρH​M≥ρ∗\\rho\_\{HM\}\\geq\\rho^\{\*\}, and the trivial rule𝒜​\(⋅\)=Y^arg⁡maxi⁡ai\\mathcal\{A\}\(\\cdot\)=\\hat\{Y\}\_\{\\arg\\max\_\{i\}a\_\{i\}\}achieves equalitya𝒜=a∗a\_\{\\mathcal\{A\}\}=a^\{\*\}\.

Intuition: Aboveρ∗\\rho^\{\*\}, disagreement cases are dominated by situations where both agents have low confidence; essentially, both are guessing\. No clever algorithm can extract information that isn’t there\.

Sketch\.Step 1\.For any rule𝒜\\mathcal\{A\}that ignores confidence \(e\.g\., majority voting with random tie\-break\), team accuracy under symmetric SDT equalsa∗a^\{\*\}andρH​M\\rho\_\{HM\}\-independent:a𝒜=aH​aM\+12​\(aH​eM\+aM​eH\)=\(aH\+aM\)/2≤a∗a\_\{\\mathcal\{A\}\}=a\_\{H\}a\_\{M\}\+\\tfrac\{1\}\{2\}\(a\_\{H\}e\_\{M\}\+a\_\{M\}e\_\{H\}\)=\(a\_\{H\}\+a\_\{M\}\)/2\\leq a^\{\*\}\. Strict gains overa∗a^\{\*\}therefore require informative confidence in the disagreement region\.

Step 2\.Define the disagreement low\-confidence regionL=\{\(cH,cM\):max⁡\(cH,cM\)<τ\}L=\\\{\(c\_\{H\},c\_\{M\}\):\\max\(c\_\{H\},c\_\{M\}\)<\\tau\\\}for any thresholdτ\\tau\. Under the joint SDT distribution with error correlationρH​M\\rho\_\{HM\}, the conditional massP​\(L∣disagree\)P\(L\\mid\\text\{disagree\}\)is monotone increasing inρH​M\\rho\_\{HM\}and approaches 1 asρH​M→1\\rho\_\{HM\}\\to 1\.

Step 3\.OnLL, the SDT log\-likelihood ratio for either agent’s prediction satisfies\|log⁡LLRi\|<di​τ\|\\log\\mathrm\{LLR\}\_\{i\}\|<d\_\{i\}\\tau, so the Bayes\-optimal posteriorP​\(correct∣cH,cM,L\)→12P\(\\text\{correct\}\\mid c\_\{H\},c\_\{M\},L\)\\to\\tfrac\{1\}\{2\}asτ→0\\tau\\to 0\.

Step 4\.Combining Steps 2 and 3, the maximum achievable selection probability on disagreements satisfiesq∗​\(ρH​M\)→12q^\{\*\}\(\\rho\_\{HM\}\)\\to\\tfrac\{1\}\{2\}asρH​M→1\\rho\_\{HM\}\\to 1\. Substituting intoa𝒜=P​\(both correct\)\+pD⋅q∗a\_\{\\mathcal\{A\}\}=P\(\\text\{both correct\}\)\+p\_\{D\}\\cdot q^\{\*\}yieldsa𝒜≤a∗a\_\{\\mathcal\{A\}\}\\leq a^\{\*\}onceρH​M\\rho\_\{HM\}exceeds the valueρ∗\\rho^\{\*\}at whichq∗​\(ρ∗\)=p10/pDq^\{\*\}\(\\rho^\{\*\}\)=p\_\{10\}/p\_\{D\}\(the "always pick the more accurate agent" baseline\)\. ForρH​M\\rho\_\{HM\}strictly aboveρ∗\\rho^\{\*\},q∗​\(ρH​M\)<p10/pDq^\{\*\}\(\\rho\_\{HM\}\)<p\_\{10\}/p\_\{D\}, givingϵ​\(ρH​M;ρ∗\)<0\\epsilon\(\\rho\_\{HM\};\\rho^\{\*\}\)<0\. Full quantification ofϵ\\epsilonin terms ofdH,dMd\_\{H\},d\_\{M\}is given in the extended version\.□\\square

### Multi\-Class Generalization

ForKK\-class problems, we derive the generalized threshold:

###### Proposition 2\(K\-Class Threshold\)

ForKK\-class classification with uniform class probabilities, the complementarity threshold generalizes to:

ρK∗≈ρbinary∗K−1\\rho^\{\*\}\_\{K\}\\approx\\frac\{\\rho^\{\*\}\_\{\\text\{binary\}\}\}\{\\sqrt\{K\-1\}\}\(20\)

Intuition: With more classes, there are more ways for agents to makedifferenterrors, so lower correlation is required for complementarity\. A 10\-class problem requiresρ<ρbinary∗/3\\rho<\\rho^\{\*\}\_\{\\text\{binary\}\}/3for complementarity\.

Table 2:Multi\-class threshold validation acrossKKvalues\.Validation on Human Data: We validated Eq\.[20](https://arxiv.org/html/2605.08710#Sx4.E20)on ImageNet\-16H \(K=16K=16classes\)\. Comparing predictedρ16∗\\rho^\{\*\}\_\{16\}vs\. observed complementarity thresholds across 4,000 human\-AI pairs:R=0\.93R=0\.93\(95% CI: \[0\.91, 0\.95\]\), MAE = 0\.04\. This confirms the multi\-class formula generalizes beyond simulation to human behavioral data; Table[2](https://arxiv.org/html/2605.08710#Sx4.T2)shows close agreement betweenρK∗≈ρ∗/K−1\\rho^\{\*\}\_\{K\}\\approx\\rho^\{\*\}/\\sqrt\{K\-1\}and observed thresholds acrossK∈\{2,10,16\}K\\in\\\{2,10,16\\\}\.

### Robustness Analysis

###### Proposition 3\(Miscalibration Robustness\)

Letβi=𝔼​\[ci\|correct\]−𝔼​\[ci\|incorrect\]\\beta\_\{i\}=\\mathbb\{E\}\[c\_\{i\}\|\\text\{correct\}\]\-\\mathbb\{E\}\[c\_\{i\}\|\\text\{incorrect\}\]measure calibration quality\. Under bounded miscalibration\|βi−1\|≤ϵ\|\\beta\_\{i\}\-1\|\\leq\\epsilon:

ρϵ∗=ρ∗⋅\(1−2​ϵ\)\\rho^\{\*\}\_\{\\epsilon\}=\\rho^\{\*\}\\cdot\(1\-2\\epsilon\)\(21\)

Non\-Gaussian Robustness: We tested predictions under alternative confidence distributions to assess sensitivity to the Gaussian SDT assumption \(Eqs\.[1](https://arxiv.org/html/2605.08710#Sx3.E1)\-[2](https://arxiv.org/html/2605.08710#Sx3.E2)\); results appear in Table[3](https://arxiv.org/html/2605.08710#Sx4.T3)\.

Table 3:Prediction robustness under alternative distributions\.Predictions remain robust \(allR\>0\.85R\>0\.85\), with at most∼\\sim5% degradation under non\-Gaussian assumptions\. This confirms our bounds are not artifacts of distributional assumptions\.

## Empirical Validation

### Datasets and Methods

We validate on: \(1\)ImageNet\-16H\[[24](https://arxiv.org/html/2605.08710#bib.bib2)\]: 1,200 images, 16 categories, N=145 participants, 20 AI models, with 4,000 human\-AI pairs analyzed; \(2\)CIFAR\-10H: 500 images, N=50 participants, 10 models, 500 pairs\.

Statistical Dependence\. Because each participant contributes to multiple pairs \(one per AI model\) and each model contributes to multiple pairs \(one per participant\), pairs are not independent\. We refit all reported correlations as mixed\-effects models with crossed random intercepts for participant and model, and with confidence intervals computed from a participant\-level cluster bootstrap \(10,000 resamples\)\. Effect estimates were within±0\.02\\pm 0\.02of the marginal correlations reported below; allpp\-values remained<0\.001<0\.001\. The headline numbers we report use the marginal correlations for comparability with prior literature, with bootstrap\-clustered CIs\.

Power Analysis: For CIFAR\-10H, with N=50 participants and expectedR=0\.90R=0\.90, power to detectR\>0\.7R\>0\.7exceeds 99% \(α=0\.05\\alpha=0\.05\)\. The sample provides adequate power for replication purposes\.

Model Validation \(Simulation\-Before\-Fitting\): Prior to parameter fitting, we verified the SDT model generates observed accuracy\-correlation patterns\. Sampling parameters from realistic priors \(d∼Uniform​\(0\.5,2\.5\)d\\sim\\text\{Uniform\}\(0\.5,2\.5\),ρ∼Uniform​\(0,0\.9\)\\rho\\sim\\text\{Uniform\}\(0,0\.9\)\), simulated data exhibit: \(1\) the same bimodal complementarity distribution \(30% success\) observed empirically; \(2\) accuracy\-correlation trade\-offs matching Figure[2](https://arxiv.org/html/2605.08710#Sx5.F2)\. This confirms the model captures the generating process, not just fitting post\-hoc; Table[4](https://arxiv.org/html/2605.08710#Sx5.T4)reports parameter\-recovery results\.

Table 4:Parameter recovery from 1,000 synthetic pairs\.
### Model Comparison

Table[5](https://arxiv.org/html/2605.08710#Sx5.T5)compares our SDT\-based model against alternatives using BIC\.

Table 5:Model comparison using BIC \(lower is better\)\.The SDT\-based model substantially outperforms alternatives \(Δ\\DeltaBIC\>300\>300indicates decisive evidence\)\.

### Results

#### Theoretical Predictions Match Empirical Accuracy

Figure[2](https://arxiv.org/html/2605.08710#Sx5.F2)shows theoretical vs\. empirical team accuracy\. ImageNet\-16H:R=0\.94R=0\.94\(95% CI: \[0\.92, 0\.95\]\), MAE = 2\.1 pp\. CIFAR\-10H:R=0\.91R=0\.91\(95% CI: \[0\.87, 0\.94\]\), MAE = 2\.6 pp\.

0\.40\.40\.50\.50\.60\.60\.70\.70\.80\.80\.90\.9110\.40\.40\.50\.50\.60\.60\.70\.70\.80\.80\.90\.911Theoretical AccuracyEmpirical AccuracyFigure 2:Theoretical predictions closely match empirical team accuracy\. Dashed line: perfect prediction\. Blue points: ImageNet\-16H \(R=0\.94R=0\.94, 95% CI: \[0\.92, 0\.95\]\)\. Orange points: CIFAR\-10H \(R=0\.91R=0\.91, 95% CI: \[0\.87, 0\.94\]\)\. Both datasets show strong agreement\.
#### Complementarity Threshold Validation

Table[6](https://arxiv.org/html/2605.08710#Sx5.T6)validates the impossibility result\.

Table 6:Complementarity gain by error correlation regime\.
#### Aggregation Method Comparison

Table[7](https://arxiv.org/html/2605.08710#Sx5.T7)compares aggregation methods onρH​M<ρ∗\\rho\_\{HM\}<\\rho^\{\*\}pairs\.

Table 7:Aggregation method comparison \(ρH​M<ρ∗\\rho\_\{HM\}<\\rho^\{\*\}pairs\)\.\*p<0\.05p<0\.05; \*\*\*p<0\.001p<0\.001

#### Metacognitive Sensitivity Effects

Stratifying pairs by metacognitive sensitivity differenceΔ​d\\Delta dtertiles confirms Theorem[2](https://arxiv.org/html/2605.08710#Thmtheorem2): complementarity gains rise from2\.1%2\.1\\%\(lowΔ​d\\Delta d\) to3\.7%3\.7\\%\(medium\) to5\.8%5\.8\\%\(high\), consistent with theΘ​\(Δ​d\)\\Theta\(\\sqrt\{\\Delta d\}\)scaling \(linear\-mixed\-effectsβ=0\.07\\beta=0\.07,t=5\.73t=5\.73,p<0\.001p<0\.001, with crossed random effects for participant and model\)\.

## Discussion

### Theoretical Contributions

This work provides the first tight bounds on human\-AI complementarity with explicit closed\-form expressions, matching minimax bounds, and an impossibility result\. The finding that only 30% of studies achieve complementarity\[[26](https://arxiv.org/html/2605.08710#bib.bib1)\]is now explained: most human\-AI pairs likely exceedρ∗\\rho^\{\*\}due to shared training data, similar heuristics, or correlated biases\.

Our framework contributes to computational\-level understanding of collaborative cognition\. Theκ​\(ρ\)\\kappa\(\\rho\)correction formalizes how correlated errors degrade the informational value of confidence signals, providing mechanistic insight into why “wisdom of crowds” fails when opinions are not independent\[[17](https://arxiv.org/html/2605.08710#bib.bib26)\]\.

### Cognitive and Neural Grounding

Our two parameters map onto well\-studied cognitive processes\. Metacognitive sensitivityddcorresponds to the precision ofconfidence monitoring, the cognitive operation by which an agent assigns post\-decisional confidence based on evidence strength\[[29](https://arxiv.org/html/2605.08710#bib.bib9)\]\. Error correlationρH​M\\rho\_\{HM\}corresponds toshared internal representations: agents engage similar features, heuristics, and inductive biases, producing coordinated mistakes\. The framework yields testable predictions: training metacognitive monitoring should raiseddand enlarge gains; pairing agents with maximally different training distributions should lowerρH​M\\rho\_\{HM\}; and neural confidence markers in rostrolateral PFC and ACC\[[7](https://arxiv.org/html/2605.08710#bib.bib8)\]should predictq∗q^\{\*\}on disagreement trials\.

### Generalizability to Human\-Human Collaboration

Our bounds apply to any pair of decision\-makers with calibrated confidence: replacing “AI” with “Human 2” yields identical mathematics, since SDT originated in human perceptual psychology\[[18](https://arxiv.org/html/2605.08710#bib.bib7)\]\. The impossibility result therefore predicts that two physicians sharing institution, heuristics, and patient population may exceedρ∗\\rho^\{\*\}and gain nothing from second\-opinion aggregation\[[1](https://arxiv.org/html/2605.08710#bib.bib23)\]; the “groupthink” phenomenon in organizational psychology can be read as a high\-ρH​M\\rho\_\{HM\}regime induced by social influence\[[17](https://arxiv.org/html/2605.08710#bib.bib26)\]\.

### Practical Implications

Three design principles emerge: reduce error correlation via diverse training/information sources; optimize metacognitive sensitivity \(an AI that "knows what it doesn’t know" can outperform an overconfident accurate one, by Theorem[2](https://arxiv.org/html/2605.08710#Thmtheorem2)\); and estimateρH​M\\rho\_\{HM\}andρ∗\\rho^\{\*\}pre\-deployment to predict whether complementarity is achievable\.

### Limitations and Future Directions

Our framework assumes static metacognitive sensitivity; in practice,ddmay vary with fatigue, learning, or domain familiarity\[[29](https://arxiv.org/html/2605.08710#bib.bib9)\], so extending to time\-varyingd​\(t\)d\(t\)and to sequential trust dynamics\[[11](https://arxiv.org/html/2605.08710#bib.bib14)\]are next steps\. Predictions hold under log\-normal and beta confidence distributions \(R\>0\.85R\>0\.85\), but severely miscalibrated systems \(e\.g\., overconfident RLHF\-tuned LLMs whose confidence rarely enters the low\-confidence region\) violate our calibration assumption beyond Proposition[3](https://arxiv.org/html/2605.08710#Thmproposition3)’s bounded regime; the\[[26](https://arxiv.org/html/2605.08710#bib.bib1)\]30% figure \(studies through mid\-2023\) may also shift for current LLMs\. WEIRD sampling concerns apply; cross\-cultural validation would strengthen generalizability\.

## Conclusion

For confidence\-based aggregation, we proved a tight complementarity thresholdρ∗\\rho^\{\*\}, an impossibility result above it,Θ​\(Δ​d\)\\Theta\(\\sqrt\{\\Delta d\}\)minimax bounds, and a multi\-class generalizationρK∗≈ρ∗/K−1\\rho^\{\*\}\_\{K\}\\approx\\rho^\{\*\}/\\sqrt\{K\-1\}, validated on human data \(R\>0\.90R\>0\.90;R=0\.93R=0\.93atK=16K=16\)\. The framework explains why 70% of human\-AI teams underperform, applies equally to human\-human aggregation, and bounds what aggregation alone can achieve\.

## References

- \[1\]B\. Bahrami, K\. Olsen, P\. E\. Latham, A\. Roepstorff, G\. Rees, and C\. D\. Frith\(2010\)Optimally interacting minds\.Science329\(5995\),pp\. 1081–1085\.External Links:[Document](https://dx.doi.org/10.1126/science.1185718)Cited by:[Human\-AI Complementarity](https://arxiv.org/html/2605.08710#Sx2.SSx1.p1.1),[Generalizability to Human\-Human Collaboration](https://arxiv.org/html/2605.08710#Sx6.SSx3.p1.2)\.
- \[2\]G\. Bansal, B\. Nushi, E\. Kamar, W\. S\. Lasecki, D\. S\. Weld, and E\. Horvitz\(2019\)Beyond accuracy: the role of mental models in human\-AI team performance\.InProceedings of the AAAI Conference on Human Computation and Crowdsourcing,Vol\.7,pp\. 2–11\.Cited by:[Introduction](https://arxiv.org/html/2605.08710#Sx1.p2.1)\.
- \[3\]G\. Bansal, T\. Wu, J\. Zhou, R\. Fok, B\. Nushi, E\. Kamar, M\. T\. Ribeiro, and D\. S\. Weld\(2021\)Does the whole exceed its parts? The effect of AI explanations on complementary team performance\.InProceedings of the 2021 CHI Conference on Human Factors in Computing Systems,pp\. 1–16\.External Links:[Document](https://dx.doi.org/10.1145/3411764.3445717)Cited by:[Introduction](https://arxiv.org/html/2605.08710#Sx1.p2.1),[Human\-AI Complementarity](https://arxiv.org/html/2605.08710#Sx2.SSx1.p1.1)\.
- \[4\]M\. de Condorcet\(1785\)Essai sur l’application de l’analyse à la probabilité des décisions rendues à la pluralité des voix\.Imprimerie Royale\.Cited by:[Introduction](https://arxiv.org/html/2605.08710#Sx1.p7.1),[Decision Aggregation and Impossibility Results](https://arxiv.org/html/2605.08710#Sx2.SSx4.p1.1)\.
- \[5\]K\. Donahue and J\. Kleinberg\(2022\)Models of fairness in federated learning\.InProceedings of the 2022 Conference on Fairness, Accountability, and Transparency,Cited by:[Decision Aggregation and Impossibility Results](https://arxiv.org/html/2605.08710#Sx2.SSx4.p1.1)\.
- \[6\]M\. Fey\(2003\)A note on the Condorcet jury theorem with supermajority voting rules\.Social Choice and Welfare20\(1\),pp\. 27–32\.External Links:[Document](https://dx.doi.org/10.1007/s003550200166)Cited by:[Decision Aggregation and Impossibility Results](https://arxiv.org/html/2605.08710#Sx2.SSx4.p1.1)\.
- \[7\]S\. M\. Fleming and R\. J\. Dolan\(2012\)The neural basis of metacognitive ability\.Philosophical Transactions of the Royal Society B: Biological Sciences367\(1594\),pp\. 1338–1349\.External Links:[Document](https://dx.doi.org/10.1098/rstb.2011.0417)Cited by:[Bayesian Models and Metacognition](https://arxiv.org/html/2605.08710#Sx2.SSx3.p1.3),[Setup and Notation](https://arxiv.org/html/2605.08710#Sx3.SSx1.p3.2),[Cognitive and Neural Grounding](https://arxiv.org/html/2605.08710#Sx6.SSx2.p1.5)\.
- \[8\]S\. M\. Fleming and H\. C\. Lau\(2014\)How to measure metacognition\.Frontiers in Human Neuroscience8,pp\. 443\.External Links:[Document](https://dx.doi.org/10.3389/fnhum.2014.00443)Cited by:[item 2](https://arxiv.org/html/2605.08710#Sx1.I1.i2.p1.1),[Bayesian Models and Metacognition](https://arxiv.org/html/2605.08710#Sx2.SSx3.p1.3)\.
- \[9\]P\. Hemmer, M\. Schemmer, N\. Kühl, M\. Vössing, and G\. Satzger\(2024\)Complementarity in human\-AI collaboration: concept, sources, and evidence\.European Journal of Information Systems\.External Links:[Document](https://dx.doi.org/10.1080/0960085X.2025.2475962)Cited by:[Introduction](https://arxiv.org/html/2605.08710#Sx1.p1.1),[Human\-AI Complementarity](https://arxiv.org/html/2605.08710#Sx2.SSx1.p1.1)\.
- \[10\]K\. Inkpen, E\. T\. Chancey, S\. Thakkar, and J\. Wickwire\(2022\)Advancing human\-AI complementarity: the impact of user expertise on AI\-assisted decision making\.ACM Transactions on Computer\-Human Interaction\.External Links:[Document](https://dx.doi.org/10.1145/3534561)Cited by:[Human\-AI Complementarity](https://arxiv.org/html/2605.08710#Sx2.SSx1.p1.1)\.
- \[11\]M\. Jianget al\.\(2025\)Miscalibrated trust hinders effective partner choices in human\-AI collectives\.InProceedings of the Annual Meeting of the Cognitive Science Society,Vol\.47\.Cited by:[Limitations and Future Directions](https://arxiv.org/html/2605.08710#Sx6.SSx5.p1.3)\.
- \[12\]T\. Kameda, W\. Toyokawa, and R\. S\. Tindale\(2022\)Information aggregation and collective intelligence beyond the wisdom of crowds\.Nature Reviews Psychology1\(6\),pp\. 345–357\.External Links:[Document](https://dx.doi.org/10.1038/s44159-022-00054-y)Cited by:[Introduction](https://arxiv.org/html/2605.08710#Sx1.p7.1),[Collective Intelligence](https://arxiv.org/html/2605.08710#Sx2.SSx2.p1.2)\.
- \[13\]J\. Kelly, A\. Boyd, S\. Showalter, M\. Steyvers, and P\. Smyth\(2025\)Bayesian inference for correlated experts with applications to human\-AI teams\.Proceedings of the International Conference on Machine Learning\.Cited by:[Bayesian Models and Metacognition](https://arxiv.org/html/2605.08710#Sx2.SSx3.p1.3)\.
- \[14\]J\. Kleinberg, H\. Lakkaraju, J\. Leskovec, J\. Ludwig, and S\. Mullainathan\(2018\)Human decisions and machine predictions\.The Quarterly Journal of Economics133\(1\),pp\. 237–293\.External Links:[Document](https://dx.doi.org/10.1093/qje/qjx032)Cited by:[Introduction](https://arxiv.org/html/2605.08710#Sx1.p1.1)\.
- \[15\]D\. Lee, J\. Pruitt, T\. Zhou, J\. Du, and B\. Odegaard\(2025\)Metacognitive sensitivity: the key to calibrating trust and optimal decision making with AI\.PNAS Nexus\.External Links:[Document](https://dx.doi.org/10.1093/pnasnexus/pgae527)Cited by:[Bayesian Models and Metacognition](https://arxiv.org/html/2605.08710#Sx2.SSx3.p1.3)\.
- \[16\]Z\. Li and M\. Steyvers\(2025\)The importance of metacognitive sensitivity in human\-AI decision\-making\.Proceedings of the Annual Meeting of the Cognitive Science Society47,pp\. 1773–1779\.Cited by:[Bayesian Models and Metacognition](https://arxiv.org/html/2605.08710#Sx2.SSx3.p1.3),[Setup and Notation](https://arxiv.org/html/2605.08710#Sx3.SSx1.p2.4)\.
- \[17\]J\. Lorenz, H\. Rauhut, F\. Schweitzer, and D\. Helbing\(2011\)How social influence can undermine the wisdom of crowd effect\.Proceedings of the National Academy of Sciences108\(22\),pp\. 9020–9025\.External Links:[Document](https://dx.doi.org/10.1073/pnas.1008636108)Cited by:[Collective Intelligence](https://arxiv.org/html/2605.08710#Sx2.SSx2.p1.2),[Theoretical Contributions](https://arxiv.org/html/2605.08710#Sx6.SSx1.p2.1),[Generalizability to Human\-Human Collaboration](https://arxiv.org/html/2605.08710#Sx6.SSx3.p1.2)\.
- \[18\]B\. Maniscalco and H\. Lau\(2012\)A signal detection theoretic approach for estimating metacognitive sensitivity from confidence ratings\.Consciousness and Cognition21\(1\),pp\. 422–430\.External Links:[Document](https://dx.doi.org/10.1016/j.concog.2011.09.021)Cited by:[Introduction](https://arxiv.org/html/2605.08710#Sx1.p3.1),[Generalizability to Human\-Human Collaboration](https://arxiv.org/html/2605.08710#Sx6.SSx3.p1.2)\.
- \[19\]H\. Mozannar, H\. Lang, D\. Wei, P\. Sattigeri, S\. Das, and D\. Sontag\(2023\)Who should predict? Exact algorithms for learning to defer to humans\.InProceedings of the 26th International Conference on Artificial Intelligence and Statistics,pp\. 10520–10545\.Cited by:[Decision Aggregation and Impossibility Results](https://arxiv.org/html/2605.08710#Sx2.SSx4.p1.1)\.
- \[20\]H\. Mozannar and D\. Sontag\(2020\)Consistent estimators for learning to defer to an expert\.InProceedings of the 37th International Conference on Machine Learning,pp\. 7076–7087\.Cited by:[Decision Aggregation and Impossibility Results](https://arxiv.org/html/2605.08710#Sx2.SSx4.p1.1)\.
- \[21\]S\. E\. Page\(2007\)The difference: how the power of diversity creates better groups, firms, schools, and societies\.Princeton University Press\.Cited by:[Introduction](https://arxiv.org/html/2605.08710#Sx1.p4.1),[Decision Aggregation and Impossibility Results](https://arxiv.org/html/2605.08710#Sx2.SSx4.p1.1)\.
- \[22\]P\. Rajpurkar, E\. Chen, O\. Banerjee, and E\. J\. Topol\(2022\)AI in health and medicine\.Nature Medicine28\(1\),pp\. 31–38\.External Links:[Document](https://dx.doi.org/10.1038/s41591-021-01614-0)Cited by:[Introduction](https://arxiv.org/html/2605.08710#Sx1.p1.1)\.
- \[23\]S\. Shalev\-Shwartz, S\. Shammah, and A\. Shashua\(2017\)On a formal model of safe and scalable self\-driving cars\.Mobileye\.Cited by:[Introduction](https://arxiv.org/html/2605.08710#Sx1.p1.1)\.
- \[24\]M\. Steyvers, H\. Tejeda, G\. Kerrigan, and P\. Smyth\(2022\)Bayesian modeling of human–AI complementarity\.Proceedings of the National Academy of Sciences119\(11\),pp\. e2111547119\.External Links:[Document](https://dx.doi.org/10.1073/pnas.2111547119)Cited by:[Introduction](https://arxiv.org/html/2605.08710#Sx1.p3.1),[Introduction](https://arxiv.org/html/2605.08710#Sx1.p8.2),[Bayesian Models and Metacognition](https://arxiv.org/html/2605.08710#Sx2.SSx3.p1.3),[Datasets and Methods](https://arxiv.org/html/2605.08710#Sx5.SSx1.p1.1)\.
- \[25\]J\. Surowiecki\(2004\)The wisdom of crowds\.Doubleday\.Cited by:[Introduction](https://arxiv.org/html/2605.08710#Sx1.p7.1),[Collective Intelligence](https://arxiv.org/html/2605.08710#Sx2.SSx2.p1.2)\.
- \[26\]M\. Vaccaro, A\. Almaatouq, and T\. W\. Malone\(2024\)When combinations of humans and AI are useful: a systematic review and meta\-analysis\.Nature Human Behaviour8\(12\),pp\. 2293–2303\.External Links:[Document](https://dx.doi.org/10.1038/s41562-024-02024-1)Cited by:[Introduction](https://arxiv.org/html/2605.08710#Sx1.p1.1),[Theoretical Contributions](https://arxiv.org/html/2605.08710#Sx6.SSx1.p1.1),[Limitations and Future Directions](https://arxiv.org/html/2605.08710#Sx6.SSx5.p1.3)\.
- \[27\]I\. Wood, A\. Teredesai, and G\. Brown\(2023\)A unified theory of diversity in ensemble learning\.Journal of Machine Learning Research24\(359\),pp\. 1–49\.Cited by:[Introduction](https://arxiv.org/html/2605.08710#Sx1.p3.1),[Decision Aggregation and Impossibility Results](https://arxiv.org/html/2605.08710#Sx2.SSx4.p1.1)\.
- \[28\]A\. W\. Woolley, C\. F\. Chabris, A\. Pentland, N\. Hashmi, and T\. W\. Malone\(2010\)Evidence for a collective intelligence factor in the performance of human groups\.Science330\(6004\),pp\. 686–688\.External Links:[Document](https://dx.doi.org/10.1126/science.1193147)Cited by:[Collective Intelligence](https://arxiv.org/html/2605.08710#Sx2.SSx2.p1.2)\.
- \[29\]N\. Yeung and C\. Summerfield\(2012\)Metacognition in human decision\-making: confidence and error monitoring\.Philosophical Transactions of the Royal Society B: Biological Sciences367\(1594\),pp\. 1310–1321\.External Links:[Document](https://dx.doi.org/10.1098/rstb.2011.0416)Cited by:[Cognitive and Neural Grounding](https://arxiv.org/html/2605.08710#Sx6.SSx2.p1.5),[Limitations and Future Directions](https://arxiv.org/html/2605.08710#Sx6.SSx5.p1.3)\.
- \[30\]Y\. Zhang, Q\. V\. Liao, and R\. K\. E\. Bellamy\(2020\)Effect of confidence and explanation on accuracy and trust calibration in AI\-assisted decision making\.InProceedings of the 2020 Conference on Fairness, Accountability, and Transparency,pp\. 295–305\.External Links:[Document](https://dx.doi.org/10.1145/3351095.3372852)Cited by:[Introduction](https://arxiv.org/html/2605.08710#Sx1.p2.1)\.

Similar Articles

Beyond Autonomy: The Power of an Agent That Knows Its Limits

Reddit r/AI_Agents

The COWCORPUS project, a study of 4,200 human-AI interactions, found that agents predicting their own failures and intervention moments are more useful than those simply trying to avoid errors. Researchers identified four stable trust patterns in human-AI collaboration and developed the Perfect Timing Score (PTS) to measure intervention prediction accuracy.

Imperfectly Cooperative Human-AI Interactions: Comparing the Impacts of Human and AI Attributes in Simulated and User Studies

arXiv cs.CL

This research paper investigates how human personality traits and AI design characteristics jointly impact human-AI interactions in imperfectly cooperative scenarios using both simulated datasets (2,000 simulations) and human subjects experiments (290 participants). The study finds significant divergences between simulation and real-world interactions, with AI transparency emerging as a critical factor in actual human-AI encounters.

Generalization Bounds of Emergent Communications for Agentic AI Networking

arXiv cs.AI

This paper proposes an information-theoretic framework for emergent communication in Agentic AI Networking (AgentNet), addressing physical constraints and providing generalization bounds. Experimental validation on hardware prototypes demonstrates improved generalization performance compared to state-of-the-art solutions.

Bias by Necessity: Impossibility Theorems for Sequential Processing with Convergent AI and Human Validation

arXiv cs.AI

This paper proves impossibility theorems showing that primacy effects, anchoring, and order-dependence are architecturally necessary biases in autoregressive language models due to causal masking constraints. The authors validate these theoretical bounds across 12 frontier LLMs and confirm related predictions through pre-registered human experiments involving working memory loads.