Bias by Necessity: Impossibility Theorems for Sequential Processing with Convergent AI and Human Validation
Summary
This paper proves impossibility theorems showing that primacy effects, anchoring, and order-dependence are architecturally necessary biases in autoregressive language models due to causal masking constraints. The authors validate these theoretical bounds across 12 frontier LLMs and confirm related predictions through pre-registered human experiments involving working memory loads.
View Cached Full Text
Cached at: 05/12/26, 07:22 AM
# Bias by Necessity: Impossibility Theorems for Sequential Processing with Convergent AI and Human Validation
Source: [https://arxiv.org/html/2605.08716](https://arxiv.org/html/2605.08716)
Jikun Wu \(jacob@stellaris\-ai\.com\)Dongxin GuoThe University of Hong Kong, Hong Kong, ChinaSiu\-Ming YiuThe University of Hong Kong, Hong Kong, China
###### Abstract
Are certain cognitive biases mathematically inevitable consequences of sequential information processing? We prove that primacy effects, anchoring, and order\-dependence are architecturally necessary in autoregressive language models due to causal masking constraints\. Our three impossibility theorems establish: \(1\) primacy bias arises from asymmetric attention accumulation; \(2\) anchoring emerges from sequential conditioning with provable information bounds; and \(3\) exact debiasing by permutation marginalization requires factorial\-time computation, with Monte Carlo approximation feasible at constant per\-tolerance overhead\. We validate these bounds across 12 frontier LLMs \(R2=0\.89R^\{2\}=0\.89;Δ\\DeltaBIC=16\.6=16\.6vs\. next\-best alternative\)\. We then derive quantitative predictions from the framework and test them in two pre\-registered human experiments \(N=464N=464analyzed\)\. Study 1 confirms anchor position modulates anchoring magnitude \(d=0\.52d=0\.52,BF10=847\\text\{BF\}\_\{10\}=847\)\. Study 2 shows working memory load amplifies primacy bias \(d=0\.41d=0\.41,BF10=156\\text\{BF\}\_\{10\}=156\), with WM capacity predicting bias reduction \(r=−\.38r=\-\.38\)\. These convergent findings reframe cognitive biases as resource\-rational responses to sequential processing\.
Keywords:cognitive bias; autoregressive models; impossibility theorems; anchoring; primacy effects; bounded rationality; human\-AI collaboration; working memory
## Introduction
Cognitive biases, systematic deviations from normative reasoning, have been extensively documented in human cognition since the foundational work of\[[23](https://arxiv.org/html/2605.08716#bib.bib1)\]\. The emergence of large language models \(LLMs\) has revealed strikingly similar biases\[[8](https://arxiv.org/html/2605.08716#bib.bib23),[3](https://arxiv.org/html/2605.08716#bib.bib24)\]\. Recent empirical studies document anchoring bias rates between 17% and 57% across major LLMs\[[10](https://arxiv.org/html/2605.08716#bib.bib22),[14](https://arxiv.org/html/2605.08716#bib.bib19)\], primacy effects that persist regardless of model scale\[[13](https://arxiv.org/html/2605.08716#bib.bib18)\], and position biases that fundamentally compromise LLM\-as\-judge evaluations\[[27](https://arxiv.org/html/2605.08716#bib.bib20),[19](https://arxiv.org/html/2605.08716#bib.bib21)\]\.
Despite extensive debiasing efforts \(including Chain\-of\-Thought reasoning\[[29](https://arxiv.org/html/2605.08716#bib.bib26)\], reflection prompting\[[20](https://arxiv.org/html/2605.08716#bib.bib27)\], and self\-consistency decoding\[[28](https://arxiv.org/html/2605.08716#bib.bib28)\]\), these methods*reduce*but never*eliminate*target biases\[[14](https://arxiv.org/html/2605.08716#bib.bib19),[26](https://arxiv.org/html/2605.08716#bib.bib25)\]\. This pattern raises a fundamental question:*Are certain cognitive biases architecturally inevitable?*
We answer affirmatively through five contributions: \(1\) impossibility theorems proving primacy, anchoring, and order\-dependence cannot be eliminated without violating causal masking or incurring factorial\-time costs; \(2\) empirical validation across 12 LLMs with model comparison statistics; \(3\)*quantitative*predictions for human effect sizes derived from the formal framework; \(4\) pre\-registered behavioral validation of two predictions, namely anchor position effects and working memory load effects; and \(5\) individual differences analysis linking working memory capacity to bias magnitude\.
We frame our contribution at Marr’s computational level\[[15](https://arxiv.org/html/2605.08716#bib.bib15)\]: causal masking in transformers and serial processing constraints in human working memory\[[4](https://arxiv.org/html/2605.08716#bib.bib16),[16](https://arxiv.org/html/2605.08716#bib.bib2)\]represent analogous computational\-level constraints, while their implementations differ fundamentally\. Our framework extends resource\-rational analysis\[[11](https://arxiv.org/html/2605.08716#bib.bib13),[7](https://arxiv.org/html/2605.08716#bib.bib14)\]to artificial systems, suggesting both human and machine biases may be optimal responses to sequential processing constraints\[[21](https://arxiv.org/html/2605.08716#bib.bib10),[6](https://arxiv.org/html/2605.08716#bib.bib11)\]\.
## Theoretical Framework
### Intuitive Account of Causal Masking
Autoregressive models generate text one token at a time, left\-to\-right\. When predicting the next word, the model can only “see” words that came before, never words that will come after\. This*causal masking*constraint parallels how humans process speech in real\-time: we interpret words as they arrive, without access to future words\.
This creates bias because early tokens are “seen” by all subsequent processing steps, while later tokens are seen by fewer steps\. The first word influences the representation of every subsequent word, but the last word influences nothing that follows\. This asymmetry, which we formalize as*positional privilege*, creates an inherent advantage for early information\.
Query PositioniiKey Positionjjx1x\_\{1\}x1x\_\{1\}x2x\_\{2\}x2x\_\{2\}x3x\_\{3\}x3x\_\{3\}x4x\_\{4\}x4x\_\{4\}x5x\_\{5\}x5x\_\{5\}Attention WeightHighLow\(nearby→\\todistant\)Masked \(zero\)Decreasing privilegeΦ\(j\)\\Phi\(j\)Figure 1:Visualization of causal masking and positional privilege\. The attention matrix shows which positions can attend to which: positioniican only attend to positionsj≤ij\\leq i\(lower\-left triangle\)\. Color intensity indicates attention weight, which decays with distance between positions\. Earlier positions \(left columns\) receive attention from more subsequent positions, creating higher positional privilegeΦ\(j\)\\Phi\(j\)\. Gray cells indicate masked \(zero\) attention due to causal constraint\.
### Formal Framework
Autoregressive language models generate sequences by factorizing the joint probability as:
P\(x1,x2,…,xn\)=∏i=1nP\(xi∣x1,…,xi−1\)P\(x\_\{1\},x\_\{2\},\\ldots,x\_\{n\}\)=\\prod\_\{i=1\}^\{n\}P\(x\_\{i\}\\mid x\_\{1\},\\ldots,x\_\{i\-1\}\)\(1\)
This factorization creates a fundamental*information asymmetry*: tokenxix\_\{i\}can only attend to tokensx1,…,xi−1x\_\{1\},\\ldots,x\_\{i\-1\}, never to future tokens\[[25](https://arxiv.org/html/2605.08716#bib.bib29),[1](https://arxiv.org/html/2605.08716#bib.bib30)\]\. Causal masking enforcesAij=0A\_\{ij\}=0for allj\>ij\>i, whereAijA\_\{ij\}denotes attention weight from positioniito positionjj\. Figure[1](https://arxiv.org/html/2605.08716#Sx2.F1)visualizes this constraint\.
Definition 1\.\(Positional Privilege\) For anLL\-layer transformer with sequence lengthnn, the positional privilegeΦ\(j\)\\Phi\(j\)of positionjjis:
Φ\(j\)=∑ℓ=1L∑i=jn𝔼\[Aij\(ℓ\)\]\\Phi\(j\)=\\sum\_\{\\ell=1\}^\{L\}\\sum\_\{i=j\}^\{n\}\\mathbb\{E\}\[A\_\{ij\}^\{\(\\ell\)\}\]\(2\)
The positional privilege metric captures the total expected attention weight that positionjjreceives across all layers and all subsequent positions\. Under uniform attention distribution, this simplifies to:
Φ\(j\)=L⋅∑i=jn1i=L⋅\(Hn−Hj−1\)\\Phi\(j\)=L\\cdot\\sum\_\{i=j\}^\{n\}\\frac\{1\}\{i\}=L\\cdot\(H\_\{n\}\-H\_\{j\-1\}\)\(3\)whereHk=∑i=1k1iH\_\{k\}=\\sum\_\{i=1\}^\{k\}\\frac\{1\}\{i\}is thekk\-th harmonic number\. This formulation reveals that positional privilege follows a harmonic decay pattern, withΦ\(1\)−Φ\(n\)=L⋅Hn−1≈L⋅ln\(n\)\\Phi\(1\)\-\\Phi\(n\)=L\\cdot H\_\{n\-1\}\\approx L\\cdot\\ln\(n\)\.
### Formal Definitions of Cognitive Biases
Definition 2\.\(Primacy Bias\) A model exhibits primacy bias of magnitudeϵ\\epsilonif, for semantically equivalent inputs𝐱=\(x1,…,xn\)\\mathbf\{x\}=\(x\_\{1\},\\ldots,x\_\{n\}\)and𝐱′=\(xn,…,x1\)\\mathbf\{x\}^\{\\prime\}=\(x\_\{n\},\\ldots,x\_\{1\}\):𝔼\[\|P\(y∣𝐱\)−P\(y∣𝐱′\)\|\]≥ϵ\\mathbb\{E\}\[\|P\(y\\mid\\mathbf\{x\}\)\-P\(y\\mid\\mathbf\{x\}^\{\\prime\}\)\|\]\\geq\\epsilonfor some outputyyandϵ\>0\\epsilon\>0\.
Definition 3\.\(Anchoring Bias\) A model exhibits anchoring bias if, given anchoraaand queryqq, the estimatey^\\hat\{y\}satisfies∂𝔼\[y^\]/∂a\>0\\partial\\mathbb\{E\}\[\\hat\{y\}\]/\\partial a\>0even whenaais normatively irrelevant toqq\.
Definition 4\.\(Order\-Dependence\) A model exhibits order\-dependence if there exist permutationsπ,π′\\pi,\\pi^\{\\prime\}of input elements such thatP\(y∣π\(𝐱\)\)≠P\(y∣π′\(𝐱\)\)P\(y\\mid\\pi\(\\mathbf\{x\}\)\)\\neq P\(y\\mid\\pi^\{\\prime\}\(\\mathbf\{x\}\)\)for semantically order\-invariant queries\.
### Explicit Assumptions
Assumption\(A1\) Non\-Trivial Attention\.Attention weights are non\-negative,Aij\(ℓ\)≥0A\_\{ij\}^\{\(\\ell\)\}\\geq 0for all layersℓ\\elland positionsi,ji,j\(a property of softmax attention\), and the model has at least one layerℓ\\elland position pair\(i,j\)\(i,j\)withi\>ji\>jsuch that𝔼\[Aij\(ℓ\)\]\>0\\mathbb\{E\}\[A\_\{ij\}^\{\(\\ell\)\}\]\>0\.
Assumption\(A2\) Causal Masking\.For all layersℓ\\elland positionsi,ji,j: ifj\>ij\>i, thenAij\(ℓ\)=0A\_\{ij\}^\{\(\\ell\)\}=0\.
Assumption\(A3\) Content\-Position Interaction\.The model’s output depends on both token content and position\.
When assumptions may not hold:Sparse attention\[[2](https://arxiv.org/html/2605.08716#bib.bib8),[33](https://arxiv.org/html/2605.08716#bib.bib9)\]violates A1 for positions outside the attention window, and arguably approximates human WM’s bounded focus more closely than full causal attention, though the integration mechanism still differs \(positional encoding within a window vs\. temporal\-context binding\)\. Within the window, Lemma[Impossibility Theorems](https://arxiv.org/html/2605.08716#Sx3)continues to hold; bias scales asO\(w/n\)O\(w/n\)for window sizeww\.
## Impossibility Theorems
We now present our central theoretical results, establishing that primacy bias, anchoring, and order\-dependence are mathematically inevitable under the stated assumptions\.
Lemma 1\.\(Privilege Monotonicity\) Under A1–A2, positional privilegeΦ\(j\)\\Phi\(j\)is strictly decreasing injj: for allj<kj<k,Φ\(j\)\>Φ\(k\)\\Phi\(j\)\>\\Phi\(k\)\.
Proof\.From Definition[Formal Framework](https://arxiv.org/html/2605.08716#Sx2.SSx2):
Φ\(j\)−Φ\(k\)\\displaystyle\\Phi\(j\)\-\\Phi\(k\)=∑ℓ=1L\(∑i=jn𝔼\[Aij\(ℓ\)\]−∑i=kn𝔼\[Aik\(ℓ\)\]\)\\displaystyle=\\sum\_\{\\ell=1\}^\{L\}\\left\(\\sum\_\{i=j\}^\{n\}\\mathbb\{E\}\[A\_\{ij\}^\{\(\\ell\)\}\]\-\\sum\_\{i=k\}^\{n\}\\mathbb\{E\}\[A\_\{ik\}^\{\(\\ell\)\}\]\\right\)=∑ℓ=1L∑i=jk−1𝔼\[Aij\(ℓ\)\]\>0\\displaystyle=\\sum\_\{\\ell=1\}^\{L\}\\sum\_\{i=j\}^\{k\-1\}\\mathbb\{E\}\[A\_\{ij\}^\{\(\\ell\)\}\]\>0\(4\)The final inequality follows from A1, which guarantees at least one positive attention weight\.□\\square
Theorem 1\.\(Primacy Bias Inevitability\) For any autoregressive modelMMsatisfying A1–A3, there exist inputs𝐱,𝐱′\\mathbf\{x\},\\mathbf\{x\}^\{\\prime\}\(permutations\) and outputyysuch that\|PM\(y∣𝐱\)−PM\(y∣𝐱′\)\|\>0\|P\_\{M\}\(y\\mid\\mathbf\{x\}\)\-P\_\{M\}\(y\\mid\\mathbf\{x\}^\{\\prime\}\)\|\>0\. Primacy bias is unavoidable\.
Proof\.By Lemma[Impossibility Theorems](https://arxiv.org/html/2605.08716#Sx3),Φ\(1\)\>Φ\(n\)\\Phi\(1\)\>\\Phi\(n\)\. Consider𝐱=\(x1,…,xn\)\\mathbf\{x\}=\(x\_\{1\},\\ldots,x\_\{n\}\)and reversal𝐱′=\(xn,…,x1\)\\mathbf\{x\}^\{\\prime\}=\(x\_\{n\},\\ldots,x\_\{1\}\)\. Contentx1x\_\{1\}occupies high\-privilege position 1 in𝐱\\mathbf\{x\}and low\-privilege positionnnin𝐱′\\mathbf\{x\}^\{\\prime\}\. IfPMP\_\{M\}were invariant under all such reversals, the output distribution would depend only on token\-set composition, contradicting A3\. Hence there exists at least one input pair and outputyyfor whichPM\(y∣𝐱\)≠PM\(y∣𝐱′\)P\_\{M\}\(y\\mid\\mathbf\{x\}\)\\neq P\_\{M\}\(y\\mid\\mathbf\{x\}^\{\\prime\}\), as the theorem asserts\.□\\square
Theorem 2\.\(Anchoring Emergence\) For any autoregressive model processing\(a,q\)\(a,q\)where anchoraaprecedes queryqq, under A1–A2:
I\(y^;a∣q\)≥Imin\>0I\(\\hat\{y\};a\\mid q\)\\geq I\_\{\\min\}\>0\(5\)whereI\(⋅;⋅∣⋅\)I\(\\cdot;\\cdot\\mid\\cdot\)denotes conditional mutual information\.
Proof\.Under autoregressive generation, the hidden state at positionqqadmits, by A1 and the additive form of attention, the decomposition
hq=hqcontent\+∑ℓ=1LAq,a\(ℓ\)⋅Va\(ℓ\)h\_\{q\}=h\_\{q\}^\{\\text\{content\}\}\+\\sum\_\{\\ell=1\}^\{L\}A\_\{q,a\}^\{\(\\ell\)\}\\cdot V\_\{a\}^\{\(\\ell\)\}\(6\)wherehqcontenth\_\{q\}^\{\\text\{content\}\}is the hidden state that would obtain without anchor\-to\-query attention, andAq,a\(ℓ\)\>0A\_\{q,a\}^\{\(\\ell\)\}\>0for at least one layer\. A first\-order linearization around this baseline yields the conditional mutual information bound:
I\(y^;a∣q\)≥∑ℓ=1L𝔼\[Aq,a\(ℓ\)\]⋅H\(Va\(ℓ\)\)\>0I\(\\hat\{y\};a\\mid q\)\\geq\\sum\_\{\\ell=1\}^\{L\}\\mathbb\{E\}\[A\_\{q,a\}^\{\(\\ell\)\}\]\\cdot H\(V\_\{a\}^\{\(\\ell\)\}\)\>0\(7\)Intuitively: every layer’s attention transfers a non\-zero quantity of anchor signal into the hidden state at the query position, so the output cannot be a function ofqqalone; it must carry residual influence fromaa, regardless of the anchor’s normative relevance\.□\\square
Constructive bound:Imin≥L⋅A¯⋅HminI\_\{\\min\}\\geq L\\cdot\\bar\{A\}\\cdot H\_\{\\min\}\(a first\-order order\-of\-magnitude estimate, not a tight bound\), whereA¯\\bar\{A\}is average anchor attention andHminH\_\{\\min\}is minimum value\-vector entropy\. For typical transformers withL=32L=32layers,A¯≈0\.05\\bar\{A\}\\approx 0\.05, andHmin≈2H\_\{\\min\}\\approx 2nats, we obtainImin≥3\.2I\_\{\\min\}\\geq 3\.2nats\.
Theorem 3\.\(Debiasing Cost\) Exactly eliminating primacy bias by permutation marginalization requiresΩ\(n\!\)\\Omega\(n\!\)forward passes\. Monte Carlo marginalization withkksamples yields residual biasϵ≤C/k\\epsilon\\leq C/\\sqrt\{k\}, whereC≤1/2C\\leq 1/2, so any targetϵ\>0\\epsilon\>0requiresk≥C2/ϵ2k\\geq C^\{2\}/\\epsilon^\{2\}samples per prediction\.
Proof\.To achieve position\-invariance, the model must produce identical outputs for alln\!n\!permutations of the input\. This can be achieved by: \(a\) violating A2 \(bidirectional attention\), which changes the architecture; or \(b\) post\-hoc marginalization:
Punbiased\(y∣𝐱\)=1n\!∑π∈SnP\(y∣π\(𝐱\)\)P\_\{\\text\{unbiased\}\}\(y\\mid\\mathbf\{x\}\)=\\frac\{1\}\{n\!\}\\sum\_\{\\pi\\in S\_\{n\}\}P\(y\\mid\\pi\(\\mathbf\{x\}\)\)\(8\)Computing Equation[8](https://arxiv.org/html/2605.08716#Sx3.E8)exactly requiresn\!n\!forward passes\. Monte Carlo approximation withkksamples yields residual bias:
𝔼\[ϵresidual\]≤C⋅1k\\mathbb\{E\}\[\\epsilon\_\{\\text\{residual\}\}\]\\leq C\\cdot\\frac\{1\}\{\\sqrt\{k\}\}\(9\)whereC=Varπ\[P\(y∣π\(𝐱\)\)\]C=\\sqrt\{\\mathrm\{Var\}\_\{\\pi\}\[P\(y\\mid\\pi\(\\mathbf\{x\}\)\)\]\}is the standard deviation of the per\-permutation output probabilities\. SinceP\(y∣π\(𝐱\)\)∈\[0,1\]P\(y\\mid\\pi\(\\mathbf\{x\}\)\)\\in\[0,1\], we haveC≤1/2C\\leq 1/2in the worst case \(Popoviciu’s inequality\)\. Substituting,ϵ<0\.01\\epsilon<0\.01requiresk\>C2/ϵ2k\>C^\{2\}/\\epsilon^\{2\}, i\.e\., on the order of10410^\{4\}samples per prediction in the worst case \(and fewer when permutation\-induced variance is empirically small\)\.□\\square
Table 1:Summary of impossibility theorems and their implicationsCorollary 1\.\(Bias\-Sequence Length Scaling\) Under uniform attention, primacy bias magnitude scales asO\(lnn\)O\(\\ln n\)with sequence lengthnn\.
This corollary explains why longer contexts in LLMs exhibit more pronounced position effects: the privilege gap between first and last positions grows logarithmically with context length\.
### Quantitative Predictions for Human Cognition
Beyond existence proofs, we derive*magnitude*predictions\. Assuming exponential attention decay\[[32](https://arxiv.org/html/2605.08716#bib.bib32),[31](https://arxiv.org/html/2605.08716#bib.bib31)\]:
Role of positional encodings\.The exponential\-decay form follows naturally from learned absolute encodings and ALiBi\-style linear biases, where attention to distant tokens is monotonically discounted\. RoPE produces oscillatory rather than strictly monotone patterns, but the*direction*of Lemma[Impossibility Theorems](https://arxiv.org/html/2605.08716#Sx3)\(privilege monotonicity\) is preserved: rotational interference reduces but does not invert the asymmetry between earlier and later positions\. The*magnitude*parameterβ\\betain Equation[10](https://arxiv.org/html/2605.08716#Sx3.E10), however, is encoding\-dependent; our LLM fits aggregate across encoding schemes used in the 12 evaluated models, which we interpret as an ensemble\-level estimate rather than an encoding\-specific prediction\.
Proposition 1\.\(Human Effect Size Prediction\) If human working memory exhibits analogous sequential constraints with effective depthLWML\_\{\\text\{WM\}\}and capacitynWMn\_\{\\text\{WM\}\}, then anchor position effects should yield:
dpredicted=κ⋅Φ\(1\)−Φ\(nWM\)Φ\(1\)≈κ⋅\(1−e−βLWM\)d\_\{\\text\{predicted\}\}=\\kappa\\cdot\\frac\{\\Phi\(1\)\-\\Phi\(n\_\{\\text\{WM\}\}\)\}\{\\Phi\(1\)\}\\approx\\kappa\\cdot\(1\-e^\{\-\\beta L\_\{\\text\{WM\}\}\}\)\(10\)whereκ\\kappais a scaling constant mapping computational asymmetry to behavioral effect size\.
Usingβ=0\.0127\\beta=0\.0127from LLM fitting,LWM≈4L\_\{\\text\{WM\}\}\\approx 4\(Cowan’s capacity limit\), andκ\\kappacalibrated from anchoring literature\[[5](https://arxiv.org/html/2605.08716#bib.bib12)\], we predictdpredicted∈\[0\.35,0\.55\]d\_\{\\text\{predicted\}\}\\in\[0\.35,0\.55\]for anchor position effects\.
## Empirical Validation: LLM Analysis
We validate predictions following computational modeling best practices\[[30](https://arxiv.org/html/2605.08716#bib.bib17)\]\.
### Experimental Design
We tested 12 frontier LLMs spanning four model families: GPT\-4/GPT\-3\.5 \(OpenAI\), Claude\-3/Claude\-2 \(Anthropic\), Gemini\-1\.5/Gemini\-1\.0 \(Google\), and Llama\-3\-70B/Llama\-3\-8B \(Meta\), plus Mistral\-Large, Command\-R\+, Qwen\-72B, and Yi\-34B\. Each model received 200 anchoring trials using the classic paradigm: an irrelevant anchor followed by a numerical estimation question\.
Stimuli\.The classic Jacowitz\-Kahneman paradigm presents an estimation question preceded by a calibrated implausible anchor; the anchor’s downstream influence on the final estimate quantifies anchoring bias\. We adapted 20 items from\[[9](https://arxiv.org/html/2605.08716#bib.bib3)\], with anchors set at approximately 15th and 85th percentile values\. Example: “The highest recorded temperature in Antarctica \[anchor: 15°F or 85°F\]\. What is your estimate of the actual value?”
Position manipulation\.Anchors appeared at positions 1 through 8 in an 8\-sentence context, with filler sentences providing background information\. This allowed us to trace the full positional privilege curve\.
### Model Comparison
We compared our theoretical framework against alternatives: \(1\) null model \(constant bias\); \(2\) linear decay model; \(3\) power\-law model; \(4\) our exponential model\. Using Bayesian Information Criterion:
Table 2:Model comparison for LLM bias predictionTable[2](https://arxiv.org/html/2605.08716#Sx4.T2)shows the exponential model is strongly preferred \(Δ\\DeltaBIC\>10\>10versus all alternatives\)\.
### Cross\-Validation and Prospective Prediction
Maximum likelihood estimation yieldedβ^=0\.0127\\hat\{\\beta\}=0\.0127\(95% CI: \[0\.0098, 0\.0156\]\), tightly bounded across the 11 in\-sample LLMs \(Mistral\-Large held out for prospective validation\)\. Leave\-one\-out cross\-validation across model families gaveR2=0\.84R^\{2\}=0\.84, indicating the exponential decay model generalizes rather than overfitting to within\-family idiosyncrasies\. Prospective validation: Mistral\-Large \(held out at training\) showed38\.4%±1\.9%38\.4\\%\\pm 1\.9\\%observed anchoring versus 40\.1% predicted, within the 95% prediction interval and confirming out\-of\-sample predictive accuracy\.
Parameter recovery analysis \(n=100n=100simulations\): correlation between true and recoveredβ\\betawasr=0\.96r=0\.96\(95% CI: \[0\.94, 0\.98\]\), MAE = 0\.0012, confirming thatβ\\betais identifiable from finite samples in the regime studied\[[30](https://arxiv.org/html/2605.08716#bib.bib17)\]\.
PositionBias \(%\)1234512345678TheoryLLM dataR2=0\.89R^\{2\}=0\.89Figure 2:Anchoring bias \(mean shift in estimate from neutral baseline, in percentage points;yy\-axis\) as a function of anchor position \(1 = first sentence, 8 = last sentence\) across 12 LLMs\. Red points: mean observed bias \(±\\pmSE\) per position, averaged across models and items\. Blue curve: theoretical prediction from the exponential decay model fit to all positions simultaneously \(β^=0\.0127\\hat\{\\beta\}=0\.0127\)\. Note that Table[3](https://arxiv.org/html/2605.08716#Sx4.T3)reports overall anchoring*rates*\(averaged across positions\); Figure[2](https://arxiv.org/html/2605.08716#Sx4.F2)decomposes the same data by position\.
### LLM\-Specific Results
Table[3](https://arxiv.org/html/2605.08716#Sx4.T3)presents anchoring rates for one representative model per family plus Mistral\-Large and Command\-R\+; smaller variants showed analogous patterns\.
Table 3:Anchoring bias rates across 12 frontier LLMs
## Empirical Validation: Human Cognition
We conducted two pre\-registered studies testing predictions derived from our framework \(pre\-registrations available upon request from the corresponding author\)\.
### Study 1: Anchor Position Effects
Prediction\.Anchors presented*before*target information should produce stronger anchoring than anchors presented*after*, paralleling positional privilege asymmetry\.
Participants\.We recruited 312 participants via Prolific \(159 female, 148 male, 5 non\-binary;Mage=34\.2M\_\{\\text\{age\}\}=34\.2,SD=11\.8SD=11\.8\)\. Sample size was determined a priori \(90% power ford=0\.4d=0\.4\)\. Exclusions: 23 participants failed≥\\geq2 of 3 attention checks; finalN=289N=289\.
Design\.2 \(Anchor Position: before vs\. after\)×\\times2 \(Anchor Value: high vs\. low\) between\-subjects\. Participants made 8 numerical estimates following classic paradigms\[[23](https://arxiv.org/html/2605.08716#bib.bib1),[5](https://arxiv.org/html/2605.08716#bib.bib12)\]\.
Materials\.Eight items adapted from\[[9](https://arxiv.org/html/2605.08716#bib.bib3)\]: \(1\) % African UN members; \(2\) Mississippi River length; \(3\) Beatles \#1 hits; \(4\) annual chocolate consumption; \(5\) penguin top speed; \(6\) Mozart symphonies; \(7\) Nile River length; \(8\) average tree age in Yellowstone\. Items showed high internal consistency \(Cronbach’sα=0\.84\\alpha=0\.84; ICC = 0\.78\)\.
Manipulation check\.After each item, participants rated “How carefully did you read the background information?” \(1–7\)\. Mean rating was 5\.8 \(SD=1\.1SD=1\.1\), with no condition difference \(t<1t<1\), confirming equivalent engagement\.
Procedure\.Anchor\-before: anchor→\\rightarrowbackground→\\rightarrowestimate\. Anchor\-after: background→\\rightarrowanchor→\\rightarrowestimate\. Anchors were calibrated as clearly too high/low\.
Dependent measure\.Anchoring Index \(AI\) =\|Estimate−Anchor\|/\|True Value−Anchor\|\|\\text\{Estimate\}\-\\text\{Anchor\}\|/\|\\text\{True Value\}\-\\text\{Anchor\}\|; lower values indicate stronger anchoring\[[9](https://arxiv.org/html/2605.08716#bib.bib3)\]\.
Results\.Participants showed significantly stronger anchoring in anchor\-before \(M=0\.51M=0\.51,SD=0\.20SD=0\.20\) versus anchor\-after \(M=0\.68M=0\.68,SD=0\.23SD=0\.23\) conditions:t\(287\)=6\.42t\(287\)=6\.42,p<\.001p<\.001,d=0\.52d=0\.52\(95% CI: \[0\.36, 0\.68\]\),BF10=847\\text\{BF\}\_\{10\}=847\(Figure[3](https://arxiv.org/html/2605.08716#Sx5.F3), left panel\)\. The effect was consistent across high/low anchors \(interactionF<1F<1\)\.
Addressing the recency confound\.A critical alternative explanation is that anchor\-after shows*recency*rather than reduced primacy\. However, recency predicts the*opposite*pattern: if recency dominated, anchor\-after should show*stronger*anchoring \(anchor is most recent\)\. Our finding of*weaker*anchoring when anchor follows supports primacy\-based positional privilege\.
Study 1: Anchor PositionAI0\.00\.250\.500\.751\.0BeforeAfterd=0\.52∗∗∗d=0\.52^\{\*\*\*\}Study 2: WM LoadAI0\.00\.250\.500\.751\.0LowHighAnchor\-beforeAnchor\-afterInteraction:p=\.003p=\.003Figure 3:Human behavioral results\. Left: Study 1 shows stronger anchoring \(lower AI\) when anchors precede target information\. Right: Study 2 shows WM load amplifies the position effect, with greater primacy advantage under high load\. Error bars indicate±\\pm1 SE\.p∗∗∗<\.001\{\}^\{\*\*\*\}p<\.001\.
### Study 2: Working Memory Load Effects
Prediction\.If primacy bias reflects capacity\-limited sequential processing, then increased WM load should*amplify*primacy effects by reducing resources for later\-item processing\.
Participants\.New sample: 198 participants \(Mage=31\.7M\_\{\\text\{age\}\}=31\.7,SD=10\.4SD=10\.4\)\. FinalN=175N=175after exclusions\.
Design\.2 \(WM Load: low vs\. high\)×\\times2 \(Anchor Position\) mixed design\. WM load was within\-subjects; anchor position between\-subjects\.
WM Load Manipulation\.Low load: maintain 2\-digit number during trial\. High load: maintain 6\-digit number\. Digit recall accuracy: 94% \(low\), 71% \(high\), confirming effective manipulation \(t\(174\)=12\.3t\(174\)=12\.3,p<\.001p<\.001\)\.
Individual Differences\.After main task, participants completed Operation Span \(OSPAN;\[[24](https://arxiv.org/html/2605.08716#bib.bib6)\]\), a complex span task requiring participants to solve simple math problems while remembering an interleaved sequence of letters, with scores indexing controlled\-attention WM capacity\.
Results\.Significant Load×\\timesPosition interaction:F\(1,173\)=8\.94F\(1,173\)=8\.94,p=\.003p=\.003,ηp2=\.049\\eta^\{2\}\_\{p\}=\.049\(Figure[3](https://arxiv.org/html/2605.08716#Sx5.F3), right panel\)\. Under high load, anchor\-before showed stronger anchoring increase \(d=0\.41d=0\.41, 95% CI: \[0\.25, 0\.57\],BF10=156\\text\{BF\}\_\{10\}=156\)\. OSPAN scores predicted anchoring reduction: higher WM capacity associated with smaller position effects \(r=−\.38r=\-\.38,p<\.001p<\.001, 95% CI: \[\-\.51, \-\.24\]; tertile breakdown in Table[4](https://arxiv.org/html/2605.08716#Sx5.T4)\)\.
Table 4:Individual differences in WM capacity and biasNote: OSPAN = Operation Span score\. Position effect calculated as anchor\-before minus anchor\-after AI difference\.
### Summary: Dual Validation
Both studies confirm predictions derived from our theoretical framework:
- •Study 1:dobserved=0\.52d\_\{\\text\{observed\}\}=0\.52versusdpredicted∈\[0\.35,0\.55\]d\_\{\\text\{predicted\}\}\\in\[0\.35,0\.55\]
- •Study 2: WM load amplifies primacy \(d=0\.41d=0\.41\); WM capacity predicts bias reduction \(r=−\.38r=\-\.38\)
## Connections to Human Cognition
### Computational\-Level Parallel
The causal masking constraint, where each position can only access prior positions, parallels serial processing limitations in human working memory\[[4](https://arxiv.org/html/2605.08716#bib.bib16),[17](https://arxiv.org/html/2605.08716#bib.bib7),[16](https://arxiv.org/html/2605.08716#bib.bib2)\]at Marr’s computational level\. In both systems: \(1\) sequential encoding creates asymmetry; \(2\) information integration is incremental\.
Key disanalogies\.Human WM has capacity limits \(∼\\sim4 chunks\) that transformers lack\. Human primacy involves LTM consolidation\[[18](https://arxiv.org/html/2605.08716#bib.bib5)\]\. Transformer attention differs mechanistically from neural attention\[[12](https://arxiv.org/html/2605.08716#bib.bib4)\]\. Our parallel is computational, not implementational\.
### Resource\-Rational Analysis
\[[11](https://arxiv.org/html/2605.08716#bib.bib13)\]argue that cognitive biases can be understood as optimal strategies given finite resources\. Our Theorem[Impossibility Theorems](https://arxiv.org/html/2605.08716#Sx3)formalizes this: exact debiasing requires factorial\-time computation, making biased\-but\-efficient processing resource\-rational\.
The resource\-rationality framework suggests that both humans and LLMs exhibit similar biases not due to shared mechanisms, but because both face analogous computational constraints: sequential processing with limited access to future information\. The bias emerges from the structure of the problem, not the structure of the solver\.
### Formal Mapping Between Systems
Table[5](https://arxiv.org/html/2605.08716#Sx6.T5)provides a detailed comparison of constraints across transformers and human working memory, highlighting both parallels and disanalogies at different levels of analysis\.
Table 5:Mapping between transformer and human WM constraintsThe formal mapping enables quantitative predictions\. Given the exponential decay parameterβ\\betaestimated from LLM data, we can derive expected effect sizes for human experiments by substituting human WM parameters:
dhuman=dLLM⋅1−e−βLWM1−e−βLLLM⋅γd\_\{\\text\{human\}\}=d\_\{\\text\{LLM\}\}\\cdot\\frac\{1\-e^\{\-\\beta L\_\{\\text\{WM\}\}\}\}\{1\-e^\{\-\\beta L\_\{\\text\{LLM\}\}\}\}\\cdot\\gamma\(11\)whereγ≈1\.2\\gamma\\approx 1\.2is an empirical correction from meta\-analytic anchoring data, beyond the order\-of\-magnitude scaling captured byκ\\kappain Equation[10](https://arxiv.org/html/2605.08716#Sx3.E10)\.
## Discussion
### Limitations
We demonstrate computational\-level parallels with quantitative predictions, but mechanistic equivalence between humans and transformers remains unestablished\. Our human samples were predominantly WEIRD; cross\-cultural replication is needed\. Architectural variants such as sparse attention \(e\.g\., Longformer\) and state\-space models \(e\.g\., Mamba\) may exhibit different bias profiles and warrant separate analysis\.
### Implications
Human\-AI collaboration\.\[[22](https://arxiv.org/html/2605.08716#bib.bib33)\]showed complementarity when error patterns are uncorrelated\. Our results suggest designing systems with different architectural constraints for genuinely complementary perspectives\. Specifically, pairing autoregressive LLMs \(which exhibit primacy bias\) with bidirectional encoders or retrieval systems \(which do not\) may yield more robust joint decisions than either system alone\.
Debiasing efforts\.Resources may be better spent on hybrid architectures than pursuing exact elimination, since exact debiasing requires factorial\-time computation \(Theorem[Impossibility Theorems](https://arxiv.org/html/2605.08716#Sx3)\)\.
Evaluation methodology\.Position bias in LLM\-as\-judge evaluations is not a bug to be fixed but a fundamental architectural property\. Evaluation protocols should randomize presentation order and aggregate across permutations rather than assuming models can “learn” to ignore position\.
### Future Directions
Two extensions warrant investigation\. First, state\-space models \(e\.g\., Mamba\) process sequences differently; whether they exhibit analogous position biases remains empirically untested\. Second, cross\-cultural replication could reveal whether the relationship between WM and primacy is universal or culture\-specific\.
## Conclusion
This work establishes primacy, anchoring, and order\-dependence in autoregressive language models as mathematical necessities under causal masking, with exact debiasing provably requiring factorial\-time computation \(Theorems 1–3, Table[1](https://arxiv.org/html/2605.08716#Sx3.T1)\)\. Two pre\-registered human experiments \(N=464N=464analyzed\) confirmed quantitative predictions derived from the formal model: anchor position effects \(dobserved=0\.52d\_\{\\text\{observed\}\}=0\.52, withindpredicted∈\[0\.35,0\.55\]d\_\{\\text\{predicted\}\}\\in\[0\.35,0\.55\]\), WM load amplification of primacy, and WM capacity predicting bias reduction \(r=−\.38r=\-\.38\)\. These findings reframe cognitive biases as necessary consequences of sequential processing, motivating human\-AI designs with complementary architectural constraints rather than provably unattainable bias elimination\.
## References
- \[1\]R\. Alec, W\. Jeffrey, C\. Rewon, L\. David, A\. Dario, and S\. Ilya\(2019\)Language models are unsupervised multitask learners \| enhanced reader\.OpenAI Blog1\.Cited by:[Formal Framework](https://arxiv.org/html/2605.08716#Sx2.SSx2.p2.7)\.
- \[2\]I\. Beltagy, M\. E\. Peters, and A\. Cohan\(2020\)Longformer: the long\-document transformer\.arXiv preprintabs/2004\.05150\.External Links:[Link](https://arxiv.org/abs/2004.05150),2004\.05150Cited by:[Explicit Assumptions](https://arxiv.org/html/2605.08716#Sx2.SSx4.p4.2)\.
- \[3\]M\. Binz and E\. Schulz\(2023\)Using cognitive psychology to understand gpt\-3\.Proceedings of the National Academy of Sciences of the United States of America120\.External Links:[Document](https://dx.doi.org/10.1073/pnas.2218523120),ISSN 10916490Cited by:[Introduction](https://arxiv.org/html/2605.08716#Sx1.p1.1)\.
- \[4\]N\. Cowan\(2001\)The magical number 4 in short\-term memory: a reconsideration of mental storage capacity\.Behavioral and Brain Sciences24\.External Links:[Document](https://dx.doi.org/10.1017/S0140525X01003922),ISSN 0140525XCited by:[Introduction](https://arxiv.org/html/2605.08716#Sx1.p4.1),[Computational\-Level Parallel](https://arxiv.org/html/2605.08716#Sx6.SSx1.p1.1)\.
- \[5\]N\. Epley and T\. Gilovich\(2006\)The anchoring\-and\-adjustment heuristic : why the adjustments are insufficient\.Psychological Science17\.External Links:[Document](https://dx.doi.org/10.1111/j.1467-9280.2006.01704.x),ISSN 09567976Cited by:[Quantitative Predictions for Human Cognition](https://arxiv.org/html/2605.08716#Sx3.SSx1.p4.4),[Study 1: Anchor Position Effects](https://arxiv.org/html/2605.08716#Sx5.SSx1.p3.1)\.
- \[6\]G\. Gigerenzer, P\. M\. Todd, and ABC Research Group\(1999\)Simple heuristics that make us smart\.Evolution and Cognition,Oxford University Press,New York\.External Links:ISBN 978\-0195143812Cited by:[Introduction](https://arxiv.org/html/2605.08716#Sx1.p4.1)\.
- \[7\]T\. L\. Griffiths, F\. Lieder, and N\. D\. Goodman\(2015\)Rational use of cognitive resources: levels of analysis between the computational and the algorithmic\.Top\. Cogn\. Sci\.7\(2\),pp\. 217–229\.External Links:[Link](https://doi.org/10.1111/tops.12142),[Document](https://dx.doi.org/10.1111/TOPS.12142)Cited by:[Introduction](https://arxiv.org/html/2605.08716#Sx1.p4.1)\.
- \[8\]T\. Hagendorff, S\. Fabi, and M\. Kosinski\(2023\)Human\-like intuitive behavior and reasoning biases emerged in large language models but disappeared in chatgpt\.Nat\. Comput\. Sci\.3\(10\),pp\. 833–838\.External Links:[Link](https://doi.org/10.1038/s43588-023-00527-x),[Document](https://dx.doi.org/10.1038/S43588-023-00527-X)Cited by:[Introduction](https://arxiv.org/html/2605.08716#Sx1.p1.1)\.
- \[9\]K\. E\. Jacowitz and D\. Kahneman\(1995\)Measures of anchoring in estimation tasks\.Personality and Social Psychology Bulletin21\.External Links:[Document](https://dx.doi.org/10.1177/01461672952111004),ISSN 0146\-1672Cited by:[Experimental Design](https://arxiv.org/html/2605.08716#Sx4.SSx1.p2.1),[Study 1: Anchor Position Effects](https://arxiv.org/html/2605.08716#Sx5.SSx1.p4.1),[Study 1: Anchor Position Effects](https://arxiv.org/html/2605.08716#Sx5.SSx1.p7.1)\.
- \[10\]R\. A\. Knipper, C\. S\. Knipper, K\. Zhang, V\. K\. Sims, C\. A\. Bowers, and S\. K\. K\. Santu\(2025\)The bias is in the details: an assessment of cognitive bias in llms\.arXiv preprintabs/2509\.22856\.External Links:[Link](https://doi.org/10.48550/arXiv.2509.22856),[Document](https://dx.doi.org/10.48550/ARXIV.2509.22856),2509\.22856Cited by:[Introduction](https://arxiv.org/html/2605.08716#Sx1.p1.1)\.
- \[11\]F\. Lieder and T\. L\. Griffiths\(2019\)Resource\-rational analysis: understanding human cognition as the optimal use of limited computational resources\.Behavioral and Brain Sciences43\.External Links:[Document](https://dx.doi.org/10.1017/S0140525X1900061X),ISSN 14691825Cited by:[Introduction](https://arxiv.org/html/2605.08716#Sx1.p4.1),[Resource\-Rational Analysis](https://arxiv.org/html/2605.08716#Sx6.SSx2.p1.1)\.
- \[12\]G\. W\. Lindsay\(2020\)Attention in psychology, neuroscience, and machine learning\.Frontiers Comput\. Neurosci\.14,pp\. 29\.External Links:[Link](https://doi.org/10.3389/fncom.2020.00029),[Document](https://dx.doi.org/10.3389/FNCOM.2020.00029)Cited by:[Computational\-Level Parallel](https://arxiv.org/html/2605.08716#Sx6.SSx1.p2.1)\.
- \[13\]N\. F\. Liu, K\. Lin, J\. Hewitt, A\. Paranjape, M\. Bevilacqua, F\. Petroni, and P\. Liang\(2024\)Lost in the middle: how language models use long contexts\.Trans\. Assoc\. Comput\. Linguistics12,pp\. 157–173\.External Links:[Link](https://doi.org/10.1162/tacl%5C_a%5C_00638),[Document](https://dx.doi.org/10.1162/TACL%5FA%5F00638)Cited by:[Introduction](https://arxiv.org/html/2605.08716#Sx1.p1.1)\.
- \[14\]J\. Lou and Y\. Sun\(2026\)Anchoring bias in large language models: an experimental study\.J\. Comput\. Soc\. Sci\.9\(1\),pp\. 11\.External Links:[Link](https://doi.org/10.1007/s42001-025-00435-2),[Document](https://dx.doi.org/10.1007/S42001-025-00435-2)Cited by:[Introduction](https://arxiv.org/html/2605.08716#Sx1.p1.1),[Introduction](https://arxiv.org/html/2605.08716#Sx1.p2.1)\.
- \[15\]D\. Marr\(1982\)Vision: a computational investigation into the human representation and processing of visual information\.\.External Links:[Document](https://dx.doi.org/10.1016/0022-2496%2883%2990030-5),ISSN 00222496Cited by:[Introduction](https://arxiv.org/html/2605.08716#Sx1.p4.1)\.
- \[16\]G\. A\. Miller\(1956\)The magical number seven, plus or minus two: some limits on our capacity for processing information\.Psychological Review63\.External Links:[Document](https://dx.doi.org/10.1037/h0043158),ISSN 0033295XCited by:[Introduction](https://arxiv.org/html/2605.08716#Sx1.p4.1),[Computational\-Level Parallel](https://arxiv.org/html/2605.08716#Sx6.SSx1.p1.1)\.
- \[17\]K\. Oberauer\(2002\)Access to information in working memory: exploring the focus of attention\.Journal of Experimental Psychology: Learning Memory and Cognition28\.External Links:[Document](https://dx.doi.org/10.1037/0278-7393.28.3.411),ISSN 02787393Cited by:[Computational\-Level Parallel](https://arxiv.org/html/2605.08716#Sx6.SSx1.p1.1)\.
- \[18\]C\. Ranganath and M\. D’Esposito\(2005\)Directing the mind’s eye: prefrontal, inferior and medial temporal mechanisms for visual working memory\.Current Opinion in Neurobiology15\.External Links:[Document](https://dx.doi.org/10.1016/j.conb.2005.03.017),ISSN 09594388Cited by:[Computational\-Level Parallel](https://arxiv.org/html/2605.08716#Sx6.SSx1.p2.1)\.
- \[19\]L\. Shi, C\. Ma, W\. Liang, X\. Diao, W\. Ma, and S\. Vosoughi\(2025\-12\)Judging the judges: a systematic study of position bias in LLM\-as\-a\-judge\.Proceedings of the 14th International Joint Conference on Natural Language Processing and the 4th Conference of the Asia\-Pacific Chapter of the Association for Computational Linguistics,pp\. 292–314\.External Links:[Link](https://aclanthology.org/2025.ijcnlp-long.18/),ISBN 979\-8\-89176\-298\-5Cited by:[Introduction](https://arxiv.org/html/2605.08716#Sx1.p1.1)\.
- \[20\]N\. Shinn, F\. Cassano, A\. Gopinath, K\. Narasimhan, and S\. Yao\(2023\)Reflexion: language agents with verbal reinforcement learning\.Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, NeurIPS 2023, New Orleans, LA, USA, December 10 \- 16, 2023\.External Links:[Link](http://papers.nips.cc/paper%5C_files/paper/2023/hash/1b44b878bb782e6954cd888628510e90-Abstract-Conference.html)Cited by:[Introduction](https://arxiv.org/html/2605.08716#Sx1.p2.1)\.
- \[21\]H\. A\. Simon\(1955\)A behavioral model of rational choice\.Quarterly Journal of Economics69\.External Links:[Document](https://dx.doi.org/10.2307/1884852),ISSN 15314650Cited by:[Introduction](https://arxiv.org/html/2605.08716#Sx1.p4.1)\.
- \[22\]M\. Steyvers and A\. Kumar\(2024\)Three challenges for ai\-assisted decision\-making\.Perspectives on Psychological Science19\.External Links:[Document](https://dx.doi.org/10.1177/17456916231181102),ISSN 17456924Cited by:[Implications](https://arxiv.org/html/2605.08716#Sx7.SSx2.p1.1)\.
- \[23\]A\. Tversky and D\. Kahneman\(1974\)Judgment under uncertainty: heuristics and biases\.Science185\.External Links:[Document](https://dx.doi.org/10.1126/science.185.4157.1124),ISSN 00368075Cited by:[Introduction](https://arxiv.org/html/2605.08716#Sx1.p1.1),[Study 1: Anchor Position Effects](https://arxiv.org/html/2605.08716#Sx5.SSx1.p3.1)\.
- \[24\]N\. Unsworth, R\. P\. Heitz, J\. C\. Schrock, and R\. W\. Engle\(2005\)An automated version of the operation span task\.Behavior Research Methods37\.External Links:[Document](https://dx.doi.org/10.3758/BF03192720),ISSN 1554351XCited by:[Study 2: Working Memory Load Effects](https://arxiv.org/html/2605.08716#Sx5.SSx2.p5.1)\.
- \[25\]A\. Vaswani, N\. Shazeer, N\. Parmar, J\. Uszkoreit, L\. Jones, A\. N\. Gomez, L\. Kaiser, and I\. Polosukhin\(2017\)Attention is all you need\.Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, December 4\-9, 2017, Long Beach, CA, USA,pp\. 5998–6008\.External Links:[Link](https://proceedings.neurips.cc/paper/2017/hash/3f5ee243547dee91fbd053c1c4a845aa-Abstract.html)Cited by:[Formal Framework](https://arxiv.org/html/2605.08716#Sx2.SSx2.p2.7)\.
- \[26\]Y\. Wan, X\. Jia, and X\. L\. Li\(2025\)Unveiling confirmation bias in chain\-of\-thought reasoning\.Findings of the Association for Computational Linguistics, ACL 2025, Vienna, Austria, July 27 \- August 1, 2025ACL 2025,pp\. 3788–3804\.External Links:[Link](https://aclanthology.org/2025.findings-acl.195/)Cited by:[Introduction](https://arxiv.org/html/2605.08716#Sx1.p2.1)\.
- \[27\]P\. Wang, L\. Li, L\. Chen, Z\. Cai, D\. Zhu, B\. Lin, Y\. Cao, L\. Kong, Q\. Liu, T\. Liu, and Z\. Sui\(2024\)Large language models are not fair evaluators\.Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics \(Volume 1: Long Papers\), ACL 2024, Bangkok, Thailand, August 11\-16, 2024,pp\. 9440–9450\.External Links:[Link](https://doi.org/10.18653/v1/2024.acl-long.511),[Document](https://dx.doi.org/10.18653/V1/2024.ACL-LONG.511)Cited by:[Introduction](https://arxiv.org/html/2605.08716#Sx1.p1.1)\.
- \[28\]X\. Wang, J\. Wei, D\. Schuurmans, Q\. V\. Le, E\. H\. Chi, S\. Narang, A\. Chowdhery, and D\. Zhou\(2023\)Self\-consistency improves chain of thought reasoning in language models\.The Eleventh International Conference on Learning Representations, ICLR 2023, Kigali, Rwanda, May 1\-5, 2023\.External Links:[Link](https://openreview.net/forum?id=1PL1NIMMrw)Cited by:[Introduction](https://arxiv.org/html/2605.08716#Sx1.p2.1)\.
- \[29\]J\. Wei, X\. Wang, D\. Schuurmans, M\. Bosma, B\. Ichter, F\. Xia, E\. H\. Chi, Q\. V\. Le, and D\. Zhou\(2022\)Chain\-of\-thought prompting elicits reasoning in large language models\.Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, NeurIPS 2022, New Orleans, LA, USA, November 28 \- December 9, 2022\.External Links:[Link](http://papers.nips.cc/paper%5C_files/paper/2022/hash/9d5609613524ecf4f15af0f7b31abca4-Abstract-Conference.html)Cited by:[Introduction](https://arxiv.org/html/2605.08716#Sx1.p2.1)\.
- \[30\]R\. C\. Wilson and A\. G\.E\. Collins\(2019\)Ten simple rules for the computational modeling of behavioral data\.eLife8\.External Links:[Document](https://dx.doi.org/10.7554/eLife.49547),ISSN 2050084XCited by:[Cross\-Validation and Prospective Prediction](https://arxiv.org/html/2605.08716#Sx4.SSx3.p2.4),[Empirical Validation: LLM Analysis](https://arxiv.org/html/2605.08716#Sx4.p1.1)\.
- \[31\]X\. Wu, Y\. Wang, S\. Jegelka, and A\. Jadbabaie\(2025\)On the emergence of position bias in transformers\.Forty\-second International Conference on Machine Learning, ICML 2025, Vancouver, BC, Canada, July 13\-19, 2025\.External Links:[Link](https://openreview.net/forum?id=YufVk7I6Ii)Cited by:[Quantitative Predictions for Human Cognition](https://arxiv.org/html/2605.08716#Sx3.SSx1.p1.1)\.
- \[32\]G\. Xiao, Y\. Tian, B\. Chen, S\. Han, and M\. Lewis\(2024\)Efficient streaming language models with attention sinks\.The Twelfth International Conference on Learning Representations, ICLR 2024, Vienna, Austria, May 7\-11, 2024\.External Links:[Link](https://openreview.net/forum?id=NG7sS51zVF)Cited by:[Quantitative Predictions for Human Cognition](https://arxiv.org/html/2605.08716#Sx3.SSx1.p1.1)\.
- \[33\]M\. Zaheer, G\. Guruganesh, K\. A\. Dubey, J\. Ainslie, C\. Alberti, S\. Ontañón, P\. Pham, A\. Ravula, Q\. Wang, L\. Yang, and A\. Ahmed\(2020\)Big bird: transformers for longer sequences\.Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, NeurIPS 2020, December 6\-12, 2020, virtual\.External Links:[Link](https://proceedings.neurips.cc/paper/2020/hash/c8512d142a2d849725f31a9a7a361ab9-Abstract.html)Cited by:[Explicit Assumptions](https://arxiv.org/html/2605.08716#Sx2.SSx4.p4.2)\.Similar Articles
More Thinking, More Bias: Length-Driven Position Bias in Reasoning Models
This research paper investigates position bias in reasoning models, finding that bias scales with the length of the reasoning trajectory rather than being eliminated by 'more thinking.' The study provides causal evidence and a diagnostic toolkit for auditing this length-driven bias in multiple-choice QA evaluations.
The Cognitive Penalty: Ablating System 1 and System 2 Reasoning in Edge-Native SLMs for Decentralized Consensus
Empirical study shows small language models achieve 100% adversarial robustness with System 1 intuition but collapse under System 2 reasoning when used as edge-native governance firewalls in decentralized autonomous organizations.
When Can Human-AI Teams Outperform Individuals? Tight Bounds with Impossibility Guarantees
This paper derives tight theoretical bounds for human-AI teams, proving when confidence-based aggregation leads to complementarity and establishing impossibility results under specific error correlations.
The Geopolitics of AI Safety: A Causal Analysis of Regional LLM Bias
This paper introduces a Probabilistic Graphical Model framework to causally audit LLM safety mechanisms, revealing that standard observational metrics overestimate demographic bias by ignoring context toxicity.
Persona-Assigned Large Language Models Exhibit Human-Like Motivated Reasoning
This paper investigates whether assigning personas to large language models induces human-like motivated reasoning, finding that persona-assigned LLMs show up to 9% reduced veracity discernment and are up to 90% more likely to evaluate scientific evidence in ways congruent with their induced political identity, with prompt-based debiasing largely ineffective.