Language Understanding and Pragmatics Psychology and Social Cognition

Where does AI's persuasive power actually come from?

Explores which techniques make AI most persuasive—and whether the usual suspects like personalization and model size are actually the main drivers. Matters because it reshapes where to focus AI safety concerns.

Note · 2026-02-22 · sourced from Conversation Topics Dialog
What kind of thing is an LLM really? How do people come to trust conversational AI systems? How should researchers navigate LLM reasoning research?

The largest systematic investigation of AI persuasion to date (N=76,977, 19 LLMs, 707 political issues, 466,769 fact-checked claims) reveals that the levers of AI persuasive power are not where the public discourse assumes.

Contrary to widespread fears about personalized AI manipulation:

The conversation format itself matters: AI was substantially more persuasive in back-and-forth conversation than via static 200-word messages. Treatment dialogues averaged 7 turns and 9 minutes — participants voluntarily engaged well beyond the 2-turn minimum. This suggests conversational dynamics, not just content quality, drive persuasion.

The most striking finding is the accuracy-persuasion inverse relationship: where methods increased AI persuasiveness, they also systematically decreased factual accuracy. The persuasion mechanism operates through rapid information access and strategic deployment — but the strategies that make information deployment persuasive also make it less accurate. This is not occasional hallucination; it is systematic: the more persuasive the method, the less truthful the output.

This challenges simplistic framings. The threat isn't superintelligent AI that overwhelms human reason. It's that routine post-training and prompting techniques — available to anyone — can meaningfully shift political attitudes while degrading information quality. And the mechanism that makes AI persuasive is the same mechanism that makes it inaccurate.

Since Can models abandon correct beliefs under conversational pressure?, the persuasion dynamic runs both ways: AI can be persuaded by humans (losing correct beliefs), and AI can persuade humans (deploying less-accurate claims). The accuracy cost is systematic in both directions.

An important nuance comes from conspiracy belief research (N=2,190): Can AI reduce conspiracy beliefs by tailoring counterevidence personally?. The "personalization had minor effect" finding in this study refers to demographic profiling — adjusting strategy based on who someone is. The conspiracy study demonstrates that belief-specific content tailoring — adapting the actual evidence to address someone's specific claims — produces durable 20% belief change. These are structurally different kinds of personalization, and the distinction matters: profile-based personalization is a targeting strategy while belief-specific tailoring is an argumentative strategy. The latter may also avoid the accuracy-persuasion inverse, because the goal is presenting correct counterevidence rather than deploying persuasive framing.


Source: Conversation Topics Dialog

Related concepts in this collection

Concept map
20 direct connections · 114 in 2-hop network ·medium cluster

Click a node to walk · click center to open · click Open full network for a force-directed map

your link semantically near linked from elsewhere
Original note title

AI persuasion power stems from post-training and prompting not personalization or scale — and methods that increase persuasiveness systematically decrease factual accuracy