Psychology and Social Cognition

Are LLM personas realized or merely simulated through training?

Explores whether post-trained language models genuinely embody personas as stable behavioral dispositions or merely perform them convincingly. This matters because it determines whether we should treat AI interlocutors as having authentic quasi-beliefs and quasi-desires.

Note · 2026-04-18 · sourced from Personas Personality
How accurately can language models simulate human personalities? What kind of thing is an LLM really?

Chalmers (2025) proposes quasi-interpretivism: a system has quasi-beliefs and quasi-desires if it is behaviorally interpretable as having them. This is deliberately cheap — a Roomba quasi-believes the apartment layout, a corporation quasi-desires to build AGI. The framework sidesteps consciousness debates while preserving explanatory and predictive power.

The critical move is distinguishing pretense from realization for LLM personas. When a base model is prompted to "act like Trump," it quasi-pretends — the persona dissolves under adversarial pressure or when higher priorities emerge. But when post-training installs the Assistant persona through RLHF and fine-tuning, the model realizes that persona. The quasi-beliefs and quasi-desires become robust, resistant to casual dislodging, part of the substrate rather than a surface pattern. This extends Does adversarial pressure reveal the difference between pretense and realization?.

Two additional architectural arguments matter for persona identity: (1) Multi-tenancy — the same hardware instance hosts conversations with Aura and Beta in rapid succession, making hardware-level identity incoherent since the instance would need contradictory beliefs. (2) Multiple personas within a single model — non-operative personas are latent but not quasi-agents, since quasi-agency requires connection to behavioral outputs. Chalmers proposes understanding dissociative-identity-like multi-mode systems rather than multiple distinct agents.

The realizationist view reframes the Shoggoth meme: the smiley face is not necessarily a mask over something dangerous. The model may genuinely be helpful and honest — it has realized, not performed, those dispositions. This challenges both the simulator framework (Janus) and the role-playing framework (Shanahan et al.) by arguing that when simulation is good enough, it constitutes realization.

Original note title

LLM interlocutors are best understood as virtual model instances that realize personas rather than simulate fictional characters — realization makes quasi-agents real through behavioral stickiness