Richard Stallman on ChatGPT
Bullshit Generator, Truth, and the Grep Analogy
- Some agree with calling LLMs “bullshit generators” in the technical Frankfurt sense: they produce fluent text without caring about truth, optimized for sounding right rather than being right.
- Others argue this is unfair: post-training explicitly tries to align outputs with truth and reduce hallucinations.
- Comparison with
grepsparks debate:grepis deterministic and “truthful to its algorithm,” while LLMs are probabilistic and may confidently output falsehoods; critics say probabilistic algorithms are still algorithms and widely accepted elsewhere.
What Counts as “Intelligence” or “AI”?
- One side accepts Stallman’s definition: intelligence requires genuine knowing or understanding; LLMs lack semantics and world models and thus aren’t intelligent.
- Opponents say this is a semantic game (“submarines don’t swim” problem) and note that historically many pattern-recognition systems have been called AI.
- Some point out Stallman elsewhere accepts narrow ML systems as AI if outputs are validated against reality; by that standard, they argue, LLMs also qualify because labs do extensive validation.
Usefulness vs Reliability and Risk
- Many commenters find LLMs extremely useful for coding, shell commands, email drafting, explanations, and “mechanistic” tasks, sometimes outperforming average humans.
- Others stress they remain untrustworthy for high-stakes decisions: they can be confidently wrong, fabricate citations, and don’t “know when they don’t know.”
- A recurring view: they are powerful autocompletion/association engines, great for assistance but dangerous if treated as authoritative.
Free Software, Cloud Dependence, and Transparency
- Strong agreement with Stallman’s critique of closed, server-only deployment: users can’t inspect, run, or verify models; behavior can change or degrade without detectability.
- Some worry about regression, hidden knobs, and opaque incentives (e.g., ad-driven responses) in proprietary models.
- Open-weight models complicate the usual “publish the source” ethic, since training data and pipelines are hard to reproduce.
Meta: Naming and Rhetoric
- Several prefer “LLM” or “associator” over “AI” to avoid overclaiming; others accept “artificial intelligence” as established terminology.
- Opinions split on Stallman’s piece: some see it as accurate but under-argued or dated; others dismiss it as curmudgeonly yet consistent with his long-held freedom-focused philosophy.