Language Understanding and Pragmatics

Can large language models develop genuine world models without direct environmental contact?

Do LLMs extract meaningful world structures from human-generated text despite lacking direct sensory access to reality? This matters for understanding what kind of grounding and knowledge these systems actually possess.

Note · 2026-02-21 · sourced from Linguistics, NLP, NLU
What kind of thing is an LLM really? How should researchers navigate LLM reasoning research?

Current LLMs have not reached direct causal grounding — no unmediated contact with the physical world, modulo first multimodal approaches and robotics. But an indirect path is available.

Training data is produced by causally grounded beings: humans who interact with, perceive, and act in the world. The totality of text and language data is like a huge mirror of the world created by us. Modern LLMs are capable of extracting lawlike world structures and regularities from this data — forming representations that are structurally similar to parts of the world.

The argument from "Understanding AI" (Schneider 2024): LLM empirical successes would be "downright mysterious" without the assumption that these systems form grounded world models. The successes in world knowledge, physical reasoning, and factual recall point toward structured world representations, not just statistical fluency.

This is indirect causal grounding: functionally established through world model formation from causally grounded data, not through direct environmental interaction. It's grounding by proxy — the chain runs: world → human perception and action → human text → LLM training → LLM internal representation.

The limitation: the chain has gaps. LLMs cannot update world models through their own action and perception. They cannot verify claims against the world in real time. The models are frozen at training cutoff. But they are not worldless — the world is present in the representations, mediated.

This connects directly to Do language models actually use their encoded knowledge? — where even the encoded world knowledge may fail to influence outputs. Indirect causal grounding does not guarantee that world knowledge is actually used.


Source: Linguistics, NLP, NLU

Related concepts in this collection

Concept map
15 direct connections · 147 in 2-hop network ·dense cluster

Click a node to walk · click center to open · click Open full network for a force-directed map

your link semantically near linked from elsewhere
Original note title

llms develop world models that constitute indirect causal grounding despite lacking direct environmental contact