Ilya Sutskever, Yann LeCun and the End of “Just Add GPUs”
Article & source discussion
- Several commenters view the article as shallow or AI-generated “slop,” noting it largely paraphrases existing interviews.
- Others are fine with AI-written summaries for time-saving, but some prefer watching full interviews to judge nuance and intent.
- There is confusion/critique about grouping Sutskever and LeCun together, with the observation that Sutskever’s current stance has moved closer to long-standing critics of pure scaling.
Scaling vs. new paradigms
- Many argue the “just add GPUs” / scaling-hypothesis era is hitting limits: data scarcity, compute ceilings, and disappointing generalization despite great benchmark scores.
- Others insist scaling is still working—pointing to recent frontier models—and that progress remains “up and to the right,” especially when combined with better training tricks and tooling.
- A recurring theme: benchmarks and leaderboards overstate real capability; models look strong on exams but remain weak at robust reasoning and transfer.
Data, evaluation, and embodiment
- Proposed new data: real-world multimodal streams from robots, self-driving cars, surveillance, cloud storage troves, synthetic data, and video.
- Counterpoint: raw sensor data is mostly redundant “noise” without good evaluation functions; real-world reward signals (e.g., not crashing) are sparse and inefficient for learning complex behavior.
- Debate over whether next-state prediction in a physics-governed world can force good world models, or whether key architectural breakthroughs are still missing.
Compute, business models, and hype
- Questions about where the next 1000× FLOPs will come from; responses include more hardware, better efficiency, and massive energy buildout, not exotic megastructures.
- Frontier labs are seen as trapped: they know more research is needed but must keep selling growth and near-term ASI/AGI narratives to investors.
- Some argue big players have clear paths to profitability and won’t go bankrupt; others note current GPU spending and open-model competition could make margins thin.
Labor replacement and social impact
- One camp believes current LLMs plus “scaffolding” can automate a large share of white-collar tasks (QA, analysis, admin, parts of sales/support, etc.).
- Another camp sees this as detached from reality, emphasizing non-technical work complexity, risk aversion, brittleness of AI pipelines, and historical patterns where productivity gains don’t simply erase jobs.
Research culture and “scale is all you need”
- Several commenters from the “scaling is not enough” side express frustration: they feel sidelined for years while the community chased transformer scaling and benchmarks.
- They resent that prominent scaling advocates are now repositioned as thought leaders of the “age of research,” while those who argued for architectural diversity and deeper notions of generalization struggle for funding and publication.