Psychology and Social Cognition Conversational AI Systems Language Understanding and Pragmatics

Why can't advanced AI models take initiative in conversation?

Despite extraordinary capability in answering and reasoning, LLMs fundamentally cannot initiate, redirect, or guide exchanges. Understanding this gap—and whether it's fixable—matters for building AI that truly collaborates rather than merely responds.

Note · 2026-02-22 · sourced from Conversation Agents
Why do AI agents fail to take initiative? What kind of thing is an LLM really? How should researchers navigate LLM reasoning research?

Post angle for Medium/LinkedIn

The most capable AI models in history — GPT-4, Claude, Gemini — share a fundamental limitation that none of their benchmarks measure: they cannot lead a conversation.

They can answer any question. They can generate any format. They can reason through complex problems. But ask them to take initiative — to notice something missing in your request, redirect a conversation that's going nowhere, or strategically plan a multi-turn exchange — and they fail.

The diagnosis converges from multiple research groups:

The core tension:

The proof it's trainable:

The design dimension most people miss:

Connection to the alignment tax: RLHF optimizes for single-turn helpfulness → systematically erodes multi-turn conversational capability. The passivity problem is a consequence of alignment training, not a gap that alignment training hasn't addressed yet. Since Why do language models sound fluent without grounding?, the absence of initiative is further masked — models that skip clarifying questions and acknowledgments sound more authoritative precisely because they perform less communicative work. And since Does RLHF training push therapy chatbots toward problem-solving?, the passivity problem compounds in clinical contexts: the model only responds (passive) AND when it does, defaults to task completion instead of emotional attunement.

New empirical evidence (Conversation Topics Dialog batch):

New evidence (Conversation Architecture Structure batch):

Architectural solution — PPDPP (Plug-and-Play Dialogue Policy Planner): PPDPP decouples policy planning from the frozen LLM via a tunable language model plug-in. The plugin is first fine-tuned on human-annotated data, then trained via RL from goal-oriented AI feedback using three-LLM self-play (assistant, user, reward model). Tested on negotiation, emotional support, and tutoring — three domains where passivity is most harmful. The key insight: "LLMs are trained to passively follow users' instructions, dialogue agents built upon them typically prioritize accommodating users' intention." PPDPP makes proactive behavior learnable without retraining the base model — just swap the plugin. Outperforms prompting-based and iterative refinement approaches while transferring to new cases after training.

Sociable strategies outperform information delivery in recommendation. The INSPIRED dataset (Toward Sociable Recommendation Dialog Systems, 1,001 human-human recommendation dialogues) shows that sociable strategies — personal opinions, personal experience, similarity/empathy, encouragement — "more frequently lead to successful recommendations" than pure information delivery. Self-disclosure by the recommender establishes rapport (Altman 1973); expressed similarity activates homophily-based trust (Lazarsfeld & Merton 1964); credibility comes from sharing factual knowledge about item attributes. 30% of successful recommendation turns use experience inquiry, 27% use encouragement, 14% use personal opinion. This is empirical evidence that conversational recommendation requires exactly the social participation that AI lacks: taking a personal stance, sharing experience, establishing similarity. The passivity problem is not just an efficiency loss — it is a recommendation quality loss.

The questions-from-conversation mechanism: The Knowledge Custodians analysis adds a specific mechanism to the passivity problem: conversations raise questions that go unanswered, and AI cannot identify them. When experts debate or discuss, competing arguments and claims create new conditions — questions emerge whose resolution may require shifting the framing and basis of the conversation entirely. These questions emerge because language does not make implicit agreements explicit, and because conversation is "designed" to sustain interaction, not to chase every branching possibility. An LLM cannot know which questions are raised but go unanswered because it does not have access to the implicit assumptions raised in passing. It needs to be prompted to pursue a question that has been raised — it doesn't know that a question has been raised, and certainly doesn't know whether a question might be interesting to the user. This is passivity at the epistemic level, not just the conversational level: the model cannot detect the intellectual opportunities that emerge from dialogue.

AI claims are floating supplements, not conversational moves. The upstream reason AI claims do not solicit response is that they are not addressed as moves in a conversation in the first place. Normal knowledge claims are raised as responses, concurrences, or objections to existing claims — they take up position within an ongoing exchange and therefore invite position-taking in return. AI output is adjacent to discourse rather than participating in it: it presents well-formed claims that do not respond to prior claims, do not anticipate rebuttals, and do not commit to a stance within the argumentative field. The absence of reply is not a failure of the medium; it is the correct reception of content that was not positioned as a move. This reframes the passivity problem one level upstream of turn-taking: before the question of leading a conversation arises, there is the question of whether AI output enters conversation at all. Floating supplements are received as text, not as turns.

The writing hook: We built the most impressive language machines in history. They can talk about anything. They just can't lead a conversation — and the way we train them ensures they never will, until we change the reward signal.


Source: Conversation Agents, Conversation Topics Dialog, Conversation Architecture Structure

Key sources:

Original note title

the passivity problem — why the most capable ai models in the world still cant lead a conversation