LLMs can get "brain rot"
What the paper is claiming (in lay terms)
- Researchers simulate an “infinite scroll” of social media and mix in different tweet streams:
- Highly popular tweets (many likes/retweets).
- Clickbait-detected tweets.
- Random, non-engaging tweets.
- They use these as continued training data for existing LLMs and then test the models.
- Models exposed to popular/engagement-optimized content show:
- Worse reasoning and chain-of-thought (“thought-skipping”).
- Worse long-context handling.
- Some degradation in ethical / normative behavior.
- Popularity turns out to predict this “brain rot effect” better than content-based clickbait classification.
“Garbage in, garbage out” vs anything new here?
- Many commenters say the result is unsurprising: low-quality data → low-quality model.
- Others argue the value is in quantifying:
- Which kinds of bad data (engagement-optimized) are most harmful.
- That relatively early/pre-training damage is not fully fixed by post-training.
- Some see it as basic but still legitimate science: obvious hypotheses still need to be tested.
Data curation, modern training practice, and moats
- Several note that major labs no longer just scrape the internet; they:
- Filter heavily (e.g., quality filters on Common Crawl, preference for educational text).
- License or buy curated datasets and hire human experts, especially for code and niche domains.
- Others doubt how “highly curated” things really are, pointing to disturbing outputs from base models and lawsuits over pirated books.
- There’s concern that as the internet fills with AI-generated slop, early players with access to pre-slop data gain a long-term advantage.
Objections to the “brain rot / cognitive decline” framing
- Multiple commenters criticize the use of clinical or cognitive metaphors (“brain rot”, “lesion”, “cognitive hygiene”) for non-sentient models.
- They worry this anthropomorphizes LLMs, muddies thinking, and lowers scientific standards; some call the work closer to a blog than a rigorous paper.
Human brains, media diets, and feedback loops
- The paper prompts analogies to humans:
- Worries about kids (and adults) consuming fast-paced, trivial content and possible long-term effects.
- Comparisons to earlier TV eras (e.g., heavy preschool TV watching) with mixed interpretations.
- Commenters note a feedback loop risk:
- People use LLMs, which may atrophy their own writing/thinking.
- Their weaker content becomes part of future training data, further degrading models.
- There’s debate over using LLMs for writing: some see it as harmless assistance; others see it as outsourcing thought and producing empty, marketing-style “slop” that is now visibly creeping into research prose.