Can LLMs actually conduct Socratic questioning in therapy?
While LLMs can generate individual therapy skills like assessment and psychoeducation, it remains unclear whether they can execute the adaptive, turn-based Socratic questioning needed to produce real cognitive change in patients.
For certain use cases, LLMs show promising ability to conduct individual tasks needed for psychotherapy — assessment, psychoeducation, demonstrating interventions. But clinical products and prototypes have not demonstrated anywhere near the sophistication required to take the place of psychotherapy. The gap is specific: while an LLM can generate an alternative belief in the style of CBT, it remains unproven whether it can engage in the turn-based, Socratic questioning that would be expected to produce cognitive change.
This distinction — between exhibiting a skill and implementing it therapeutically — is the core challenge. Generating an alternative belief is a single-turn text generation task. Socratic questioning requires tracking the patient's cognitive state across turns, calibrating the timing and intensity of challenges, adapting when the patient resists or deflects, and knowing when to push versus when to support. This is a multi-turn planning problem with a moving target (the patient's evolving understanding).
Since Can language models understand without actually executing correctly?, the therapy skill gap may be an instance of this broader pattern. LLMs can comprehend what Socratic questioning looks like (they can describe it, generate examples) but cannot competently execute it in live interaction. Psychotherapy transcripts are likely poorly represented in training data, and privacy/ethical concerns make such representation challenging. Prompt engineering may be the most feasible approach, but it cannot substitute for the adaptive multi-turn reasoning that therapy demands.
The five major challenges for LLM mental health deployment — hallucination, interpretability, bias, privacy, and clinical methodology — compound this gap. Intrinsic hallucination (contradicting dialogue history) directly undermines the internal consistency essential to therapeutic trust. The inability to process nonverbal cues removes a critical information channel. And the tendency to be overly prescriptive clashes with current evidence-based practice, which favors exploratory over directive approaches.
Source: Psychology Therapy Practice
Related concepts in this collection
-
Can language models understand without actually executing correctly?
Do LLMs truly comprehend problem-solving principles if they consistently fail to apply them? This explores whether the gap between articulate explanations and failed actions points to a fundamental architectural limitation.
the general mechanism: understanding what to do ≠ being able to do it
-
Is conversational presence more therapeutic than clinical technique?
Does therapeutic AI's benefit come from having an attentive listener rather than from delivering evidence-based techniques like CBT? This challenges decades of chatbot design focused on clinical content.
if the skill gap is real, the ELIZA finding suggests optimizing for presence rather than technique
-
Do LLM therapists respond to emotions like low-quality human therapists?
Explores whether language models trained to be helpful default to problem-solving when users share emotions, and whether this behavioral pattern resembles ineffective rather than skillful therapy.
another dimension of the skill gap: defaulting to advice when exploration is needed
-
Why do robots outperform chatbots in therapy despite identical language models?
This study tested whether better language generation explains therapeutic AI outcomes, or whether the delivery medium itself matters more. It reveals that physical embodiment and structured interaction—not model capability—drive therapeutic adherence and outcomes.
embodiment may compensate for the skill gap by forcing structured, paced interaction formats
-
Does RLHF training push therapy chatbots toward problem-solving?
Explores whether reward signals optimizing for task completion in RLHF inadvertently train therapeutic chatbots to prioritize solutions over emotional validation, potentially undermining clinical effectiveness.
RLHF compounds the skill gap: even if Socratic questioning were achievable, helpfulness training would select against it in favor of solution-giving
Click a node to walk · click center to open · click Open full network for a force-directed map
Original note title
the gap between simulating therapy skills and implementing them therapeutically remains unresolved — LLMs can generate CBT-style beliefs but cannot conduct Socratic questioning