LLMs have reached a point of diminishing returns
Perceived progress and diminishing returns
- Some argue LLMs are vastly better than two years ago and still improving, especially with multimodal capabilities coming online.
- Others see only mild gains for end users since early GPT‑4, with some reporting newer models feel “dumber” or more constrained in places.
- A news report (indirectly referenced) about a new frontier model with smaller gains than the GPT‑3 → GPT‑4 jump is cited as evidence of slowing progress; others say this just reflects known diminishing returns, not a hard “wall.”
- Several commenters expect a sigmoid‑shaped progress curve rather than endless exponential improvement and see current debates as overreacting to a predictable slowdown.
Usefulness and limitations in practice
- Many developers report significant productivity boosts: faster onboarding to new languages, easier boilerplate/glue code, help with “known unknowns,” and removal of some tedious tasks.
- Others find LLM code impressive but unreliable for complex integration; debugging AI‑generated code can erase much of the time savings and disrupt mental flow.
- There is concern that people learning with LLM “crutches” may skip foundational skills needed to judge correctness or idiomatic style.
- Outside coding, LLMs are seen as strong for learning assistance and rapid ideation, but not as a general panacea.
Scaling, architectures, and future directions
- One core disagreement: whether brute‑force transformer scaling alone can reach AGI versus needing new architectures (e.g., neuro‑symbolic hybrids, agents, RAG, tool use).
- Commenters note the shift toward system integration (RAG, function calling, agents) as pragmatic engineering, interpreted by some as evidence that pure scaling is insufficient, and by others as just another way to extract value while scaling continues.
- It’s emphasized that “diminishing returns” has always been part of scaling laws; loss keeps going down with more compute/data, just with smaller marginal gains.
Economic, legal, and societal implications
- Several expect that if scaling gains slow, companies must pivot from “data + GPUs” to application‑centric, services‑heavy business models.
- Some worry current gains mainly translate into workforce reduction and higher margins, not lower prices or better quality, and fear being left with higher unemployment and “lossy” AI outputs once subsidies end.
- There is interest in open‑source models trained only on clearly legal data, but skepticism that such models can match performance when most human knowledge is copyrighted.
- A side thread highlights non‑LLM AI risks: surveillance, predictive policing, autonomous weapons, and loss of privacy and political agency.
Assessment of the article itself
- Many see the piece as repetitive, self‑vindicating, and overly dramatic, pointing to earlier (pre‑GPT‑4) “wall” predictions as undermining credibility.
- Others value a skeptical counterweight to corporate AGI hype but criticize the lack of clear, quantified, falsifiable claims.
- A recurring desire is for more nuanced positions: LLMs may not scale to AGI alone, yet are already highly useful and likely to remain central tools.