Will AI systems perform poorly due to AI-generated material in training data?
Watermarking and Detecting AI Output
- Some assume large labs watermark LLM outputs (statistical patterns, etc.) to later filter them from training sets; others think watermarking was largely abandoned as unreliable.
- Even if a vendor can track its own outputs, they cannot reliably filter outputs from competing models.
- Observed “anti‑GPT-ism” phrasing in system prompts (e.g., suppressing stock moralistic phrases) is taken as evidence that newer models’ training data is already contaminated with AI text.
Quality and Role of Synthetic Data
- Several commenters argue synthetic data is already central and beneficial:
- Llama 3’s post-training reportedly uses almost entirely synthetic answers from Llama 2.
- DeepSeek models and others are cited as heavily synthetic yet strong, contradicting simple “self‑training collapse” fears.
- Synthetic data is framed as an extension of training: used for classification, enhancement, and generating infinite math/programming problems, not just copying the web.
- Skeptics ask how synthetic data can exceed the quality of the original human data, especially in fuzzy domains without clear correctness checks.
Risk of Model Collapse and Error Accumulation
- One camp: repeated training on AI outputs leads to compounding “drift error,” where small hallucination rates amplify across generations until output becomes mostly wrong.
- The opposing camp: if selection/filters exist (human feedback, automated checks, tool use), retraining on model outputs can at worst preserve quality and often improves it.
- Some compare self-play in games (e.g., chess/Go) as evidence that self-generated data plus clear rewards can produce superhuman systems; critics counter that most real-world tasks lack such clean reward signals.
Human Data, Feedback Signals, and Privacy
- LLM chat logs are seen as a massive ongoing human data source, though many view the prompts/responses as low-quality or noisy.
- Weak behavior signals (rephrasing, follow-up prompts, “thumbs up/down,” scrolling) are considered valuable at scale, but skeptics doubt they can match rich, organically written content.
- There is concern about whether “opt-out from training” settings are genuine or dark patterns; enforcement ultimately depends on trust and legal penalties.
Reasoning vs Knowledge Base
- Some argue future progress will come from improved “core reasoning” and tool use, while encyclopedic knowledge from raw web text becomes less central and more polluted.
- Others question whether current chain-of-thought outputs demonstrate genuine reasoning or just plausible-looking text with unobserved jumps to the answer.
Broader Social and Cultural Feedback Loops
- Worry that humans are already being “trained on LLM garbage” (homework, coding, medical study aids) and will produce more derivative, low-quality text, further polluting training data.
- Counterpoint: human culture has always been self-referential; art and writing haven’t degraded just because humans learn from prior artifacts.
- Some foresee models learning to detect AI slop as a robustness feature; others fear a cultural “enshittification” equilibrium where both humans and AIs converge on bland, GPT-like language.