LLM Reasoning and Architecture Language Understanding and Pragmatics

How do language models perform syllogistic reasoning internally?

Does formal symbolic reasoning exist as a distinct neural circuit in LLMs, or is it inevitably contaminated by world knowledge associations? Understanding the mechanism could reveal whether pure logical reasoning is separable from semantic inference.

Note · 2026-02-23 · sourced from MechInterp
What kind of thing is an LLM really? How should we allocate compute budget at inference time? How should researchers navigate LLM reasoning research?

Mechanistic analysis of syllogistic inference reveals a three-stage reasoning mechanism:

  1. Naive recitation — the model begins by reciting information from the first premise
  2. Middle-term suppression — duplicated middle-term information is suppressed (e.g., in "All A are B; All B are C," the shared term B is suppressed)
  3. Mediation — mover attention heads transfer information to derive the valid conclusion, connecting A to C through the suppressed B

This circuit is content-independent — it operates on symbolic variables, not on the specific content of premises. When tested on schemes instantiated with commonsense knowledge, the same mechanism is still necessary. But additional attention heads encoding contextualized world knowledge contaminate the formal circuit, creating belief bias: conclusions that align with real-world knowledge are easier to derive than those that don't.

The contamination scales with model size: larger models show more complex attention head contributions, suggesting increasing interference from world knowledge. This is precisely the opposite of what you might hope — scaling doesn't purify the reasoning circuit, it adds more contamination from richer world knowledge.

The circuit is sufficient and necessary for all unconditionally valid syllogistic schemes where the model achieves ≥60% accuracy. For schemes with lower accuracy, the circuit alone is insufficient — suggesting these harder schemes require additional mechanisms the model hasn't developed.

Cross-architecture compatibility: similar suppression mechanism patterns and information flow appear across GPT-2, Pythia, Llama, and Qwen families. The reasoning mechanism is architecturally general, not model-specific.

This provides mechanistic evidence for Do large language models reason symbolically or semantically?: the model has a formal reasoning circuit, but it is inherently contaminated by semantic associations. Pure formal reasoning and world knowledge are not cleanly separable — they share neural substrate.


Source: MechInterp

Related concepts in this collection

Concept map
14 direct connections · 146 in 2-hop network ·dense cluster

Click a node to walk · click center to open · click Open full network for a force-directed map

your link semantically near linked from elsewhere
Original note title

syllogistic reasoning circuits use a three-stage mechanism — recitation suppression mediation — contaminated by world knowledge bias