There are no new ideas in AI, only new datasets
Role of Data vs Methods
- Many argue recent AI gains mostly come from larger, cleaner, and more diverse datasets plus more compute, not fundamentally new algorithms.
- Others counter that architectural ideas (transformers, attention, long context, multimodal models, RL from human feedback, video world models) are substantive innovations, even if rooted in old math.
- Several note a cyclical pattern: new ideas → heavy scaling (data/compute) → diminishing returns → renewed search for new ideas.
Hardware, Infrastructure, and Timing
- Commenters stress that 1990s-era ideas only became practical due to:
- Massive GPU/parallel compute.
- Cloud-scale infrastructure and fast networks.
- The internet as a text-centric, labeled-data firehose.
- Some predict current architectures will hit a “compute coffin corner” where costs grow faster than quality, triggering an “AI winter”; others expect a plateau of productivity instead.
Generalization, Reasoning, and “Real” Intelligence
- RL/game examples: models can reach superhuman skill on one game yet fail to transfer to new levels or similar games, sometimes performing worse after prior training—seen as overfitting and weak abstraction.
- Debate over whether LLMs just memorize training data or truly generalize:
- One side: they are sophisticated pattern-matchers/compressors producing plausible outputs, not reasoning.
- Other side: they clearly solve novel tasks and manipulate unseen codebases, which implies some form of reasoning, even if alien to human cognition.
- Symbolic AI and meta-RL are cited as alternatives that sometimes show better true generalization.
Embodiment, Multimodality, and World Models
- One camp: language+vision already capture the core of human-level online intelligence; touch, smell, etc. mainly matter for robotics.
- Opposing camp: embodiment and rich sensorimotor experience are foundational for robust world models (space, causality, physical intuition); text/video alone are “shadows in a cave.”
- Robotics, simulation, and video world-models are seen as next frontiers, but current video models still lack object permanence and robust physics understanding.
Static Training vs Continuous Learning
- Current LLMs are static, not always-on learners; they don’t reliably update weights in real time or manage long-term memory/forgetting.
- Some see reinforcement learning, evolutionary methods, self-play, and agentic systems as paths toward continuous adaptation, but note they are compute-expensive and brittle.
Tool Use, Data Limits, and Future Directions
- A recurring wish: models that reliably delegate precise subproblems to traditional algorithms and tools without bespoke wiring for each task.
- Concern that web-scale high-quality text is close to exhausted; suggestions include biological, robotics, and video data, plus more synthetic/simulated experience.
- Several insist new architectures are being explored constantly, but most yield incremental gains; data quality/diversity often matter more in practice.