Language Understanding and Pragmatics Conversational AI Systems

Why do language models skip the calibration step?

Current LLMs assume shared understanding rather than building it through dialogue. This explores why that design choice persists and what breaks when it fails.

Note · 2026-02-21 · sourced from Linguistics, NLP, NLU
Where exactly does language competence break down in LLMs? How should researchers navigate LLM reasoning research?

Two fundamentally different communicative modes (from Chandu et al. 2021, formalized in Grounding 'Grounding' in NLP):

Static grounding: Common ground is presumed or pseudo-automatically established. The sequence is: human queries → agent retrieves from data → agent responds. The "common ground" here is the database itself, treated as universal truth. No negotiation occurs. The agent succeeds by linking the query to the right data.

Dynamic grounding: Common ground is built through interaction. The sequence involves clarification requests, acknowledgments, confirmations, and corrections — looping until mutual understanding is established. Only then does response delivery proceed. The interaction IS the grounding process.

Static grounding is the dominant mode in current LLM deployment. A user asks a question; the model generates an answer presuming shared understanding. This works when the query is unambiguous and the data retrieval is correct. It fails — silently — when the user's intent diverges from the model's interpretation, because there is no mechanism to detect or repair the divergence.

Dynamic grounding is what human dialogue depends on. Effective training domains (emotional support, conflict resolution, teaching) require dynamic grounding: the agent must detect when understanding has broken down and initiate repair. Since Does preference optimization damage conversational grounding in large language models?, the training that makes LLMs better at static grounding appears to make them worse at dynamic grounding.

Static grounding is the technical version of false punditry. The static/dynamic distinction names at the technical level what false punditry names at the social-media level: skipping the calibration step in which speakers verify shared understanding before treating claims as commonly accepted. Static grounding presumes the shared ground and proceeds; false punditry presumes the shared ground and proceeds. One is a design pattern in dialogue systems, the other is a genre pattern in AI-generated commentary, but the structural move is the same — the omission of the calibration step that would expose whether the ground is actually shared. Seen this way, false punditry is not a stylistic quirk of AI posts; it is static grounding transposed into public-facing genres where the absent calibration is not merely a conversational limitation but a legitimacy problem.

The distinction maps onto a structural asymmetry: static grounding is a retrieval problem; dynamic grounding is an intersubjectivity problem. LLMs are trained on the former.


Source: Linguistics, NLP, NLU

Related concepts in this collection

Concept map
22 direct connections · 179 in 2-hop network ·medium cluster

Click a node to walk · click center to open · click Open full network for a force-directed map

your link semantically near linked from elsewhere
Original note title

static grounding presumes common ground while dynamic grounding builds it through clarification and repair