LLMs understand nullability
What “understanding” means for LLMs
- Large part of the thread disputes whether LLMs can be said to “understand” anything at all.
- One camp: LLMs are just next-token predictors, like thermostats or photoreceptors; there is no mechanism for understanding or consciousness, so applying that word is misleading or wrong.
- Opposing camp: if a system consistently gives correct, context-sensitive answers, that’s functionally what we call “understanding” in humans; judging internal state is impossible for both brains and models, so insisting on a metaphysical distinction is empty semantics.
- Several comments note we lack precise, agreed scientific definitions of “understanding,” “intelligence,” and “consciousness,” making these discussions circular.
Brain vs LLM analogies
- Some argue the brain may itself be a kind of very large stochastic model; others respond that this analogy is too shallow, ignoring biology, embodiment, and non-linguistic cognition.
- Disagreement over whether future “true thinking” systems will look like scaled-up LLMs or require a fundamentally different architecture.
- Concern voiced that anthropomorphizing models (comparing them to humans) is dangerous, especially when used for high-stakes tasks like medical diagnosis.
Nullability, code, and the experiment itself
- Many find the visualization and probing of a “nullability direction” in embedding space very cool: subtracting averaged states reveals a linear axis corresponding to nullable vs non-nullable.
- There’s interest in composing this with other typing tools, especially focusing on interfaces/behaviors (duck typing) rather than concrete types.
- Some note that static type checkers already handle nullability well, so the value here is more about understanding how models internally represent code concepts, not adding new capabilities.
- One commenter links this work to similar findings of single “directions” for refusals/jailbreaking in safety research.
Reliability, evaluation, and limits
- Several people push for more rigorous reporting: showing probabilities over multiple runs rather than anecdotal “eventually it learns X,” given LLM output variance.
- Others emphasize that LLMs can reflect correct patterns for concepts like nullability because they’ve seen vast text/corpus coverage, not because they’ve executed programs.
- Critics argue models often fail at “simple but novel” code manipulations where a human programmer would generalize from semantics rather than surface patterns, suggesting a shallow form of competence.
Broader capability and hype
- Some see LLMs as a remarkable, surprising capability jump that already warrants the word “AI”; others view them as sophisticated autocomplete with overblown claims of understanding.
- There is shared fatigue over repeatedly re-litigating the same philosophical issues, with some proposing to avoid the verb “understand” entirely and instead talk in terms of “accuracy on tasks” and “capabilities over distributions of inputs.”