Conversational AI Systems Language Understanding and Pragmatics

Can we teach LLMs to form linguistic conventions in context?

Humans naturally shorten references as conversations progress, but LLMs don't adapt their language for efficiency even when they understand their partners do. Can training on coreference patterns teach this convention-forming behavior?

Note · 2026-02-22 · sourced from Conversation Architecture Structure
Where exactly does language competence break down in LLMs? Why do AI conversations reliably break down after multiple turns? How should researchers navigate LLM reasoning research?

Humans naturally form ad-hoc linguistic conventions during interaction — reducing "the medicine for my back pain in a small blue medicine bottle" to "my back meds" within a few exchanges. Since Why don't LLMs shorten messages like humans do?, this convention formation is absent from LLMs even when they understand the conventions as listeners.

The post-training solution uses three components:

1. Heuristic data extraction. Coreference resolution on 2,000 TV scripts identifies reference chains where a concept is initially mentioned with a full noun phrase and later re-mentioned with a more concise expression. Each chain can provide multiple demonstrations, showing the model that convention formation persists across the entire reference chain, not just the first re-mention.

2. Two types of preference pairs. Type 1 (convention demonstration): preferred = observed concise re-mention, dispreferred = verbose first-mention repeated as re-mention. This suppresses verbatim repetition and encourages adaptation. Type 2 (first-mention preservation): preferred = original full first mention, dispreferred = premature conventionalization. This prevents the model from pre-shortening before common ground exists. 11,106 Type 1 + 10,135 Type 2 pairs total.

3. Mention planning tokens. A special [remention] token precedes re-mentions, explicitly marking the distinction between initial and later references. This allows the model to separate its processing of first mentions (full description required) from re-mentions (convention formation appropriate). Additional preference pairs train the model to USE this token correctly.

The result is a general in-context behavior: post-trained models spontaneously form conventions as interactions progress, without task-specific fine-tuning. Evaluated on both a cognitively-motivated interaction benchmark and a document-grounded reference completion task.

This provides the training-time fix for what Why don't conversational AI systems mirror their users' word choices? identifies as a behavioral gap. The mechanism is elegant: rather than engineering convention formation rules, the training data reveals the pattern through naturally occurring coreference chains, and DPO optimization internalizes it.


Source: Conversation Architecture Structure

Related concepts in this collection

Concept map
14 direct connections · 131 in 2-hop network ·dense cluster

Click a node to walk · click center to open · click Open full network for a force-directed map

your link semantically near linked from elsewhere
Original note title

post-training for convention formation via DPO on coreference-identified preference pairs with mention planning tokens teaches LLMs to form ad-hoc conventions in context