Inductive or deductive? Rethinking the fundamental reasoning abilities of LLMs
Scope of “reasoning” in LLMs
- Strong disagreement over whether LLMs “reason” at all vs. being advanced statistical text predictors.
- Some argue they only map inputs to likely outputs from training data, with no goals, self-model, or awareness of questions/answers.
- Others say their behavior is best seen as approximating reasoning (or even as a kind of reasoning), just limited, brittle, and unlike human cognition.
- Several note that debates often reduce to differing definitions of “reason,” “intelligence,” and “consciousness.”
Deductive, inductive, abductive reasoning
- Multiple commenters note the paper’s focus on inductive vs. deductive is incomplete without abduction (inference to best explanation).
- One view: LLM behavior seems closer to abductive/Bayesian inference over token sequences than to clean symbolic deduction.
- Others say in practice the distinctions blur for LLMs, since they only ever see text, not real-world events.
Tokenization, “strawberry,” and failure modes
- The “How many ‘r’s in ‘strawberry’?” example is heavily discussed.
- Some see this as proof LLMs don’t understand letters/words, only tokens and distributions.
- Others argue the failure is a tokenization artifact; character-level models can count letters reliably, and prompts that change tokenization can fix it.
- Debate over whether such failures show “no reasoning” or simply current architectural/engineering limits.
Memorization vs. generalization
- Repeated concern: you cannot cleanly test reasoning without knowing what’s in the training set.
- Arithmetic in common bases, Caesar ciphers, and many “reasoning” benchmarks are likely in-distribution, so high scores may be memorization or pattern reuse.
- Some see base-dependent arithmetic performance as evidence of memorization rather than abstract rule learning.
- Others note that humans also rely heavily on learned patterns; the line between memorization and reasoning is fuzzy.
Consciousness and qualia
- Long subthread on whether LLMs are conscious or “aware” of anything, or merely manipulating symbols.
- Competing views: consciousness as graded world-modelling vs. requiring specific physical substrates (e.g., brain waves).
- No consensus; several stress that invoking qualia or “soul-like” properties does not help evaluate present systems.
Capabilities and current limits
- Commenters highlight LLM strengths in pattern mapping and language fluency but weaknesses in strict rule-following, robust math, logic puzzles, text-to-SQL, and ASCII art.
- Some argue these weaknesses show transformers are poor architectures for genuine reasoning (search + program execution); others see room for incremental improvement and hybrid systems.