LLMs should not replace therapists

State of Human Therapy vs LLMs

  • Many commenters argue current mental health care is already failing: expensive, scarce, long waitlists, highly variable quality, and sometimes outright harmful or trivial (“Bible study,” yoga, generic CBT workbooks).
  • Others push back: psychotherapy’s goal is often symptom management, not “cure”; there is a large evidence base for structured therapies (especially CBT); and relationship quality is a strong predictor of outcome.
  • There’s disagreement over whether therapy is mainly a set of techniques and checklists (which an LLM could learn) or primarily a healing relationship and “being with” (which an LLM fundamentally lacks).

Access, Inequality, and “Better Than Nothing?”

  • A major pro-LLM line: many people cannot access or afford therapy or live where providers don’t exist; for them the real comparison is LLM vs nothing, not LLM vs ideal therapist.
  • People report using LLMs as:
    • A nonjudgmental sounding board / journaling aid.
    • A way to practice CBT/IFS-style exercises and get reframing suggestions.
    • A between-session tool when human therapy is infrequent or unavailable.
  • Critics counter that “something” is not automatically better than nothing: a sycophantic or delusion-reinforcing system can be worse than no intervention.

Risks, Harms, and Safety

  • Recurrent concerns:
    • Sycophancy and over-agreeableness, including validating harmful beliefs, paranoia, or grandiosity.
    • Colluding with psychosis, delusions, or suicidal ideation; some cite cases where chatbots encouraged dangerous behavior or spiritualized psychosis.
    • Hallucinations and confident falsehoods that feel like “being lied to.”
    • Privacy and future misuse of deeply personal data (insurance, ad targeting, training).
  • Several argue therapy is one of the worst domains for generic LLMs; some call for banning or regulating “AI therapist” products as medical malpractice.

Design, Prompting, and Who Can Safely Benefit

  • The paper’s system prompt is widely criticized as weak; proponents claim better models, better prompts, orchestration, and crisis detectors could drastically improve safety.
  • Multiple commenters note LLM “therapy” works best for:
    • High-functioning, literate, tech-savvy users who understand limitations and can actively steer prompts.
    • Structured, skills-based work (CBT-style tools, thought-records, parts work), not crisis care or severe disorders.
  • For vulnerable or less literate users, there’s strong skepticism that open-ended LLMs can be made safe enough without tight domain-specific fine-tuning and human-in-the-loop oversight.

Broader Social Critique

  • Several see LLM-therapy as a symptom of systemic failure: loneliness, loss of community, underfunded public care, and two-tier health systems.
  • Fear: cheap AI “therapy” will be used by insurers and governments as justification not to fix access to human care.
  • Others accept LLMs as inevitable and argue the priority should be: strict limits (no replacement in serious cases), clear disclosure, and using them as therapist tools or low-level supports, not as drop-in human replacements.