Language Understanding and Pragmatics LLM Reasoning and Architecture

Do large language models make the same causal reasoning mistakes as humans?

Research on collider structures reveals whether LLMs share human biases in causal inference. This matters because if both fail identically, collaboration might reinforce rather than correct errors.

Note · 2026-02-22 · sourced from Reasoning Methods CoT ToT
Where exactly does language competence break down in LLMs? What kind of thing is an LLM really?

The collider structure C1 → E ← C2 (two independent causes with a shared effect) is a diagnostic test for normative causal reasoning. When you observe the effect E, observing one cause should lower your estimate of the other (explaining away). When E is absent, C1 and C2 should remain independent.

Humans systematically fail this test in characteristic ways:

The "Do LLMs Reason Causally Like Us?" paper (CLADDER dataset) finds that LLMs exhibit the same two biases in the same direction as humans. This is not the usual finding of LLM inferiority — it is a finding of human-like systematic error. LLMs are not categorically worse at causal reasoning; they err in the same direction.

This matters for several reasons. First, it undermines clean human-vs-LLM comparisons in causal reasoning tasks: if both fail in the same way, the relevant comparison shifts from "who is better" to "are the failure modes compatible." Second, it raises the question of mechanism: humans likely err due to the associative nature of pattern-matching; LLMs likely err for structurally related reasons (training on human text that exhibits the same biases). The shared error direction is evidence that Why do LLMs handle causal reasoning better than temporal reasoning? — the training data itself has these biases baked in.

Third, the finding has implications for high-stakes causal reasoning: medical diagnosis (collider structures appear in disease-symptom networks), legal reasoning (independent causes with shared outcomes), and policy analysis all involve collider-type structures. Human and LLM collaborators sharing the same biases may reinforce rather than correct each other's errors.


Source: Reasoning Methods CoT ToT

Related concepts in this collection

Concept map
13 direct connections · 115 in 2-hop network ·medium cluster

Click a node to walk · click center to open · click Open full network for a force-directed map

your link semantically near linked from elsewhere
Original note title

llms exhibit human-like causal biases — weak explaining away and markov violations in collider networks