Language Understanding and Pragmatics Conversational AI Systems

Why don't LLMs shorten messages like humans do?

Humans naturally develop shorter, efficient language during conversations. Do multimodal LLMs exhibit this same spontaneous adaptation, or do they lack this communicative behavior?

Note · 2026-02-22 · sourced from Conversation Topics Dialog
Where exactly does language competence break down in LLMs? What kind of thing is an LLM really? How should researchers navigate LLM reasoning research?

Humans spontaneously develop increasingly efficient language during interactions. A patient might start with "the medicine for my back pain in a small blue medicine bottle" and within a week say just "my back meds." This is lexical convention formation — a fundamental property of human communication documented extensively in reference game studies.

The ICCA framework evaluates whether multimodal LLMs exhibit this behavior. The results reveal an asymmetry: as listeners, models like GPT-4 show adaptation trends close to humans, improving accuracy as interactions progress. But as speakers, all models fail to spontaneously improve communication efficiency. Only with fairly heavy-handed instruction — explicitly telling the model to reduce message length and maintain lexical consistency — do GPT-4, Gemini, and Claude show partial adaptation.

Four prompting variants reveal the gradient:

The Word Novelty Rate (WNR) metric captures this precisely — counting word insertions and substitutions while ignoring deletions, since deletions reflect natural convention formation while additions indicate cognitive-load-increasing changes.

This finding is a concrete instantiation of a broader pattern. Since Why do language models fail at communicative optimization?, communicative efficiency through convention formation is precisely the optimization principle that training on form alone cannot produce. Convention formation requires modeling the listener's cognitive state and adjusting accordingly — a functional competence that next-token prediction does not select for.

A training-time solution has now been demonstrated. Since Can we teach LLMs to form linguistic conventions in context?, the convention formation gap is addressable through targeted post-training rather than architectural redesign. The approach — heuristically extracting coreference chains from TV scripts, constructing DPO preference pairs, and adding a [remention] planning token — produces general in-context convention formation behavior. The model spontaneously shortens references as interaction progresses, precisely the capability the ICCA framework found missing.


Source: Conversation Topics Dialog, Conversation Architecture Structure

Related concepts in this collection

Concept map
13 direct connections · 121 in 2-hop network ·dense cluster

Click a node to walk · click center to open · click Open full network for a force-directed map

your link semantically near linked from elsewhere
Original note title

multimodal LLMs do not spontaneously adapt their language for communication efficiency despite understanding their interlocutors increasingly efficient language