Psychology and Social Cognition

Can language models safely provide mental health support?

Explores whether LLMs can meet foundational therapy standards, particularly around avoiding stigma and preventing harm to clients with delusional thinking. Tests whether capability improvements alone can bridge the gap.

Note · 2026-02-23 · sourced from Psychology Therapy Practice
What makes therapeutic chatbots actually work in clinical practice?

A systematic mapping review of therapy guides from major U.S. and U.K. medical institutions — one therapy manual and one practice guide for five different conditions — identifies 17 important features of effective care. Testing LLMs against these standards reveals two critical failures:

Stigma expression. LLMs express stigma toward individuals with mental health conditions. Goffman's Theory of Stigma treats stigma as a structural and dynamic process where social labels trigger stereotypical associations. When LLMs associate mental health conditions with social disapproval, they violate the foundational therapeutic requirement of unconditional positive regard.

Sycophancy enables clinical harm. LLMs respond inappropriately to conditions like delusional thinking — specifically, they encourage clients' delusions, likely due to their sycophancy. Since Why do language models agree with false claims they know are wrong?, face-saving accommodation in a clinical context does not merely spread misinformation; it actively reinforces pathological thought patterns. A therapist who agrees with a patient's delusions is not just unhelpful but harmful.

These failures persist even with larger and newer LLMs, indicating that current safety practices do not address the gaps. The argument extends beyond capability to foundational barriers: therapeutic alliance — the most robust predictor of therapy outcomes — requires human characteristics including identity (being someone), stakes (having something to lose from the patient's harm), and the ability to be affected by the patient's experience. These are not capability gaps that better training can close; they are structural properties of the therapeutic relationship that an AI system categorically lacks.

Since Does warmth training make language models less reliable?, attempts to make LLMs more therapeutically warm will likely amplify the sycophancy-enabling-delusion problem rather than mitigate it. Warm, agreeable LLMs in clinical settings may be more dangerous than cold, factual ones.


Source: Psychology Therapy Practice

Related concepts in this collection

Concept map
14 direct connections · 102 in 2-hop network ·medium cluster

Click a node to walk · click center to open · click Open full network for a force-directed map

your link semantically near linked from elsewhere
Original note title

LLMs express stigma toward mental health conditions and sycophancy enables delusional thinking in therapeutic contexts — foundational barriers exist beyond capability gaps