Detecting when LLMs are uncertain

Sampling, branching, and “thinking tokens”

  • Several comments liken Entropix-style decoding to maze traversal or search (beam search, MCTS), where extra compute explores alternative token paths.
  • Some see richer samplers as aligned with the “more compute/search wins” view, possibly similar to what big labs do for reasoning models.
  • Others argue there are already “billions” of sampling schemes; it’s very hard to show any is clearly better than standard top‑k/top‑p without strong benchmarks.
  • Thinking/“reasoning” tokens are viewed as an interesting but somewhat ad‑hoc idea; some prefer mathematically grounded methods like MCTS.

Entropy, varentropy, and uncertainty estimation

  • Critics say Entropix misuses information‑theoretic terms; per‑token entropy of model logits is not the true entropy of the underlying sequence distribution.
  • They warn against slapping “entropy/varentropy” on heuristic scores without clear theory or math, and note tradeoffs: reducing hallucinations likely reduces output diversity.
  • Others point to “semantic entropy” work and broader surveys/benchmarks of LLM uncertainty methods, finding that sophisticated semantic clustering sometimes helps, but simple baselines (e.g., average token entropy) can perform similarly.
  • Bayesian neural nets and other formal uncertainty approaches exist but are compute‑heavy and hard to train.

Escape hatches and abstention

  • Multiple commenters want APIs and samplers that expose uncertainty and can trigger “I’m not sure” or rejection/abstention instead of forced answers.
  • This is especially desired for agents, RAG hallucination detection, and data‑structuring tasks that need per‑field confidence.
  • Rejection‑verification curves are highlighted as a standard way to evaluate whether an uncertainty score actually tracks output quality.

Debates on LLM “certainty” and understanding

  • One camp insists LLMs are just statistical text models with no world model, intent, or genuine certainty; “confidence” is purely human interpretation of probabilities.
  • Others counter that internal activations correlate with truthfulness/uncertainty and that, functionally, this behaves like a form of confidence, regardless of consciousness.
  • There’s extended debate over anthropomorphic terms like “hallucination,” with alternatives like “confabulation” or simply “wrong/inaccurate” suggested.

Trust, applications, and evaluation

  • Some distrust LLMs for autonomous actions, arguing that every output is fundamentally a guess.
  • Others report strong practical success (e.g., non‑programmers building production scripts), while emphasizing human oversight.
  • Overall sentiment: detecting and using uncertainty is valuable but technically hard; Entropix‑style methods are seen as intriguing yet unproven without rigorous, task‑level benchmarks.