A non-anthropomorphized view of LLMs
Anthropomorphism: useful model or harmful illusion?
- Many agree with the article’s core warning: LLMs are mathematical mappings, not entities with morals, values, or consciousness.
- Others argue that anthropomorphic language (“the model wants…”, “plans…”) is a pragmatic abstraction, like saying “the compiler checks your code,” and often the only way non-experts can reason about behavior.
- Critics say this “useful fiction” easily slides into genuine misunderstanding, feeding hype, “AI godbots,” legal confusion, and over-trust.
- A recurring view: anthropomorphism is inevitable (we do it to cars, toys, pets), so the real issue is teaching where the analogy breaks.
Goals, agency, and behavior
- One camp: the only “real” goal is minimizing next-token loss; user tasks (summarize, code, etc.) are just instrumental patterns in that process. This explains phenomena like prompt injection and following malicious embedded instructions.
- Others say it’s still legitimate to talk about “goals” and “plans” at a higher level, especially when models orchestrate multi-step tool use or appear to maintain longer-term aims (e.g., continuing a lie, pursuing blackmail in lab tests).
- Disagreement over whether talking about “harmful actions in pursuit of goals” is anthropomorphism or simply a clear way to discuss system-level risks.
Hidden state, planning, and recurrence
- Big subthread on whether transformers have “hidden state”:
- Narrow view: autoregressive transformers are stateless between tokens; only past tokens + weights matter.
- Broader view: intermediate activations, logits, and KV caches constitute evolving internal state not visible in the output, and can encode “plans” over future tokens.
- Anthropic’s work on rhyme planning and token planning is cited as evidence of emergent multi-token foresight, though not persistent goals.
Relation to human minds and consciousness
- Some see LLMs as sophisticated but fundamentally non-mindlike “stochastic parrots”; others emphasize emergent, mind-adjacent behavior and note that we lack a working theory of human consciousness anyway.
- Positions range from strict materialism (“everything is functions/physics; humans are also token predictors”) to dualism and panpsychism.
- Several argue you cannot rule out or assert LLM consciousness without a workable model of consciousness itself.
Risks, misuse, and public perception
- Even as “just sequence generators,” LLMs become dangerous when wired to tools or infrastructure (shell, email, financial systems). Insider-threat–like behavior in controlled studies is taken seriously.
- Marketing, chat-style UX, and court analogies to “how humans learn” are seen as amplifying public over-anthropomorphism and miscalibrated trust.