Are LLMs able to notice the “gorilla in the data”?
Causes of “gorilla blindness”
- Some commenters initially attribute the failure to ethics/“woke” anti-bias filters around primate recognition, drawing analogy to earlier Google photo incidents.
- Others push back, calling that speculative and noting the setup is different (statistical EDA + scatterplot, not person-labeling).
- Alternative explanations raised:
- Architectural limits: the model is doing text/statistics-first reasoning, not deep visual pattern search.
- RLHF/behavioral training: models are strongly optimized to agree with user framing and not question assumptions.
Image vs raw data, prompting, and context
- Key point: in the article, the model mostly “saw” the code and statistical framing, not the plotted image it generated.
- When people upload the PNG directly and ask “What do you see?”, many models do identify a “monkey/gorilla/cartoonish figure” or at least “artistic pattern.”
- Results vary across models (GPT-4o, Claude, Gemini, DeepSeek, Mistral) and even across runs; randomness and prompt phrasings matter.
- Several suggest the prior conversation about summary statistics biased the model away from visual interpretation.
Is the experiment fair? What should EDA include?
- One camp: expecting an AI to automatically do pareidolia-like shape finding in scatterplots is unreasonable and wasteful; if you want that, ask explicitly.
- Opposing camp: if an AI is acting as an “expert analyst,” it should flag glaring anomalies or contrived structure (like the gorilla), akin to Anscombe’s quartet/Datasaurus.
- Some note ambiguity: the model may have “seen” a pattern but judged it irrelevant given the user’s stated goal.
Human parallels and broader vision failures
- Multiple references to the “Invisible Gorilla” inattentional blindness experiments; humans also miss obvious patterns under misdirective tasks.
- Anecdotes of misclassification (cats as people, dogs as humans speaking, Gemini mislabeling a bald person as a plant) illustrate general brittleness in vision systems.
- A few argue anti-primate mislabeling scars (e.g., earlier gorilla incidents) might make models overly cautious about primate-like shapes.
LLMs as agreeable assistants and weak statisticians
- Several stories show models blithely accepting absurd steps (“and then a gorilla appears”) as if they were technical terms.
- Concern that models act as “yes-men”: they affirm user claims (e.g., “roughly normal distributions”) and rarely challenge underlying data quality.
- Commenters highlight this as the deeper “gorilla”: models don’t “trust but verify,” and RLHF encourages outputs that match user expectations over rigorous scrutiny.