The LLMentalist Effect

Critique of the “LLMentalist / con-artist” analogy

  • Many commenters argue the psychic comparison is overstated: LLMs are continuously checked against ground truth (e.g., running code), while psychics cherry-pick safe claims.
  • Others accept a partial analogy: RLHF can reward answers that sound confident and emotionally satisfying even when unsupported, akin to cold reading.
  • Several think the article is internally inconsistent (both “it’s all an illusion” and “it stole all our work”) and light on technical detail; some see it as motivated by dislike of LLMs and note it’s dated (2023, pre-o1/o3).

What is “intelligence”?

  • Thread repeatedly notes that the article never pins this down.
  • Competing definitions:
    • Narrow/functional: ability to solve problems or use information effectively.
    • Richer: requires awareness, introspection, or “knowing with conscience.”
  • People point out you can’t resolve “are LLMs intelligent?” without first agreeing on a definition; otherwise the debate becomes circular.

Similarities and differences to human cognition

  • Some argue a large part of human cognition is statistical pattern-matching (especially grammar, conversational wandering), so LLMs plausibly mirror an aspect of mind.
  • Others emphasize missing facets: introspection, long-term memory, embodiment, consciousness, non-verbal reasoning, and the ability to notice and correct one’s own failures.
  • A minority worry about “religious” attitudes that insist human thinking must be fundamentally non-mechanical.

Capabilities, generalization, and benchmarks

  • Dispute over whether solving Olympiad/ARC-AGI/logic tasks shows real reasoning or just sophisticated pattern reuse/overfitting.
  • Some highlight LLM weaknesses on basic compositional tasks (like reliably counting letters) to argue limits of next-token prediction.
  • A custom NanoGPT sorting experiment is cited to counter “pure parroting,” sparking a technical subthread on what counts as genuine generalization.

Illusion, ELIZA effect, and world models

  • Several draw parallels to the ELIZA effect: humans over-attribute understanding to fluent text.
  • Others insist LLMs do build internal world models; the “illusion” is partly RLHF pushing them toward persuasive personas.
  • One framing: the chat persona is a fictional character users bond with, not the underlying system.

Usefulness and economics

  • Some say “do they think?” is secondary to “are they useful and cost-effective?”
  • Others see a gap between hype and current utility; examples like Copilot provoke disagreement over whether aggressive promotion reflects real value or a search for one.