Why do neural networks fail at compositional generalization?
Exploring whether the binding problem from neuroscience explains neural networks' inability to systematically generalize. The binding problem has three aspects—segregation, representation, and composition—each creating distinct failure modes in how networks handle structured information.
Greff et al. (2012.05208) argue that the binding problem — well-studied in neuroscience and cognitive psychology — is the underlying cause of neural networks' failure to achieve human-level generalization. The binding problem has three aspects that each create distinct failure modes.
Segregation: forming meaningful entities from unstructured sensory inputs. Neural networks struggle to decompose inputs into discrete objects without architectural inductive biases (slot attention, object-centric representations). Without segregation, the network works with undifferentiated feature maps rather than structured entities.
Representation: maintaining separation of information at a representational level. Even when entities can be identified, distributed representations entangle them. A network may know that "red triangle" and "blue circle" are present, but fail to maintain the binding of red-to-triangle and blue-to-circle. This is the classic variable binding problem.
Composition: using entities to construct new inferences, predictions, and behaviors. Even with segregation and representation, composing entities into novel combinations (never seen during training) requires systematic reuse of learned structure. This is where distributional shift fragility appears — agents trained with RL are fragile under distributional shift and require substantially more training data than humans.
The deeper tension: connectionist representations are directly grounded in input data (unlike symbols, which require human interpretation for grounding — the symbol grounding problem). But this grounding advantage comes at the cost of compositional structure. Since Do large language models reason symbolically or semantically?, the binding problem may explain WHY semantic decoupling collapses reasoning: without compositional binding, removing semantic content removes the only glue holding the reasoning together.
Scaling can partially overcome the binding problem. The "Scaling can lead to compositional generalization" paper demonstrates that standard MLPs can compositionally generalize when data and model size are scaled sufficiently. The key theoretical result: MLPs can approximate compositional task families using only a linear number of neurons with respect to the number of task modules — compositionality does not inherently require exponential capacity. Empirically, when models successfully compositionally generalize, task constituents can be linearly decoded from hidden activations; this metric correlates with failures of image generation models to compose known concepts. This provides a partial counterpoint to the binding problem: while the fundamental challenge remains, scaling may create conditions where compositional representations emerge despite the lack of explicit binding mechanisms. The "Break It Down" paper provides structural evidence: models often implement solutions to subroutines via modular subnetworks, and pretraining encourages this structural compositionality. See Can neural networks learn compositional skills without symbolic mechanisms? and Do neural networks naturally break tasks into modular parts?.
Source: LLM Architecture; enriched from MechInterp
Related concepts in this collection
-
Do large language models reason symbolically or semantically?
Can LLMs follow explicit logical rules when those rules contradict their training knowledge? Testing whether reasoning operates independently of semantic associations reveals what computational mechanisms actually drive LLM multi-step inference.
semantic dependence may be a consequence of the binding problem
-
Do foundation models learn world models or task-specific shortcuts?
When transformer models predict sequences accurately, are they building genuine world models that capture underlying physics and logic? Or are they exploiting narrow patterns that fail under distribution shift?
heuristics bypass the binding problem by not requiring compositional structure
-
Do LLMs generalize moral reasoning by meaning or surface form?
When moral scenarios are reworded to reverse their meaning while keeping similar language, do LLMs recognize the semantic shift? This tests whether LLMs actually understand moral concepts or reproduce training distribution patterns.
surface similarity as a binding substitute
-
Can neural networks learn compositional skills without symbolic mechanisms?
Do neural networks need explicit symbolic architecture to compose learned concepts, or can scaling alone enable compositional generalization? This asks whether compositionality is an architectural feature or an emergent property of scale.
partial resolution: scaling creates conditions for compositional representations without explicit binding, but linear decodability may mask deeper structural issues (see FER tension)
-
Do neural networks naturally break tasks into modular parts?
Can standard neural networks decompose complex tasks into separate subroutines implemented in distinct subnetworks, or do they only memorize input-output patterns? Understanding whether compositionality emerges from gradient-based learning matters for interpretability and generalization.
structural evidence that pretraining encourages modular decomposition, partially addressing the composition aspect of the binding problem
-
Can explicit stack tracking improve how transformers learn recursive syntax?
Can adding an explicit stack tape to transformers help them track recursive structure more efficiently? This matters because standard transformers struggle with long-tail recursive patterns despite their size and data.
directly addresses the composition sub-problem with explicit recursive state tracking via stack tape, providing the constituent structure mechanism that standard attention lacks
-
Can recurrent hierarchies achieve reasoning that transformers cannot?
Can a dual-timescale recurrent architecture escape the computational limitations of standard transformers and solve complex reasoning tasks without explicit chain-of-thought? This explores whether architectural design, not scale, enables true algorithmic reasoning.
architectural evidence: HRM achieves near-perfect accuracy on Sudoku and maze tasks requiring compositional reasoning where standard transformers fail completely; hierarchical recurrence may provide the computational depth needed for the composition sub-problem without explicit symbolic binding
Click a node to walk · click center to open · click Open full network for a force-directed map
Original note title
the binding problem — segregation representation and composition — explains why neural networks fail at systematic generalization