Language Understanding and Pragmatics

Does semantic grounding in language models come in degrees?

Rather than asking whether LLMs truly understand meaning, this explores whether grounding is actually a multi-dimensional spectrum. The question matters because it reframes the sterile understand/don't-understand debate into measurable, distinct capacities.

Note · 2026-02-21 · sourced from Linguistics, NLP, NLU
What kind of thing is an LLM really? How should researchers navigate LLM reasoning research?

The grounding debate — "do LLMs understand meaning?" — has been stuck in a binary framing that obscures more than it reveals. "Understanding AI" (Schneider 2024) proposes that semantic grounding is multi-criterial and gradual, distinguishing three fundamentally different types:

Functional grounding (system-internal): The capacity to integrate linguistic expressions into a coherent functional network of world knowledge and inference. LLMs have this strongly — their internal representations are structurally coherent, they form implicit world models, and their outputs show sophisticated semantic integration. This is purely about what happens inside the system boundary.

Social grounding (community-embedded): The capacity to participate in the "language games" of a linguistic community (Wittgenstein). Social grounding is constituted by the functional roles an agent plays through behavior in the world — "long-arm" functional roles. LLMs have this weakly. The more they are used as communicative partners and integrated into linguistic practices, the more they acquire it — Can LLMs acquire social grounding through linguistic integration?.

Causal grounding (world-connected): Direct contact with the physical world that language refers to. LLMs lack this directly. However, training data is produced by causally grounded humans, opening the door to Can large language models develop genuine world models without direct environmental contact? — an indirect causal grounding functionally established through world model formation.

The key insight: grounding is not a yes/no property but a graduated, multi-dimensional one. Arguments that "LLMs don't understand" often implicitly invoke causal grounding criteria; arguments that "LLMs do understand" typically invoke functional grounding criteria. Both are grounding, but different kinds.

This taxonomy connects to Do language models actually use their encoded knowledge? — which is a functional-to-causal gap: something can be encoded (functional) without influencing what is generated (and without causal world connection).

Peirce's semiotic framework provides a complementary vocabulary that maps precisely onto this tri-partite structure. Functional grounding corresponds to symbolic Thirdness (generalization, abstraction, pattern inference — the internal symbol system). Causal grounding corresponds to Secondness (direct encounter with brute fact, world resistance). Social grounding corresponds to socially-mediated Thirdness (shared, negotiated meaning within a community of interpreters). The Can AI systems achieve real alignment without world contact? insight follows from this mapping: a system operating in functional/Thirdness alone without Secondness and socially-mediated Thirdness cannot guarantee its goal representations track real-world values. Tool-use and RAG introduce what that paper calls "proto-indexicality" — beginning to address the Secondness/causal gap without fully closing it.


Source: Linguistics, NLP, NLU

Related concepts in this collection

Concept map
19 direct connections · 181 in 2-hop network ·dense cluster

Click a node to walk · click center to open · click Open full network for a force-directed map

your link semantically near linked from elsewhere
Original note title

llm semantic grounding is tri-partite functional grounding is strong social grounding is weak causal grounding is indirect