Psychology and Social Cognition Language Understanding and Pragmatics

Can language models track how minds change during persuasion?

Do LLMs understand evolving mental states in persuasive dialogue, or do they only capture fixed attitudes? This explores whether models can update their reasoning as a person's beliefs shift across conversation turns.

Note · 2026-02-22 · sourced from Theory of Mind
How should researchers navigate LLM reasoning research? Why do LLMs excel at social norms yet fail at theory of mind?

PersuasiveToM evaluates LLM theory of mind through persuasive dialogue — a domain with asymmetric social status, evolving mental states, and strategic interaction. The core finding reveals an asymmetry in LLM ToM capability that structured benchmarks miss.

Static mental states: near-human. LLMs consistently identify the persuader's desire (their persuasion goal) throughout the dialogue. This is relatively fixed — the persuader wants the same thing from start to finish. Models perform competitively with humans on this.

Dynamic mental states: significantly worse than humans. The persuadee's desires shift — from initial refusal through hesitation to being persuaded. Tracking this evolution requires integrating cues from each utterance and updating a mental model of the persuadee's attitude. LLMs fail at this dynamic tracking. They also struggle to understand "the dynamics of mental states of the whole dialogue" — the overall trajectory rather than any single snapshot.

Distinct error patterns by role. Even when question types are identical (desire, belief, intention), LLMs exhibit different error patterns when reasoning about the persuader versus the persuadee. This suggests they are not applying a general mental-state-tracking mechanism but using different heuristics for different social roles.

CoT helps strategy prediction but not mental state reasoning. Chain-of-thought prompting enhances prediction of persuasion strategies but does not substantially improve reasoning about mental states themselves. This decoupling suggests that strategy prediction can be solved through surface patterns ("what usually comes next in a persuasion dialogue?") while genuine mental state tracking requires something CoT cannot provide.

The Belief-Desire-Intention (BDI) model structures what LLMs need but lack: the ability to reason about evolving desires (motivational states that shift in response to interaction), dynamically updating beliefs (attitudes toward the persuasion goal that change as dialogue progresses), and contextual intentions (persuasion strategies mapped to underlying goals). The static/dynamic split suggests LLMs can snapshot but not stream.


Source: Theory of Mind

Related concepts in this collection

Concept map
14 direct connections · 158 in 2-hop network ·dense cluster

Click a node to walk · click center to open · click Open full network for a force-directed map

your link semantically near linked from elsewhere
Original note title

llms track static mental states competitively with humans but fail at tracking dynamic mental state shifts in persuasive dialogue