The impact of competition and DeepSeek on Nvidia
Thread reception and meta
- Many readers praise the article as one of the clearest, most comprehensive breakdowns of the GPU/AI landscape, though some feel the title under‑sells the breadth of content.
- A few note technical quibbles (e.g., precision history, driver quality), but generally see it as informed and nuanced rather than typical finance-guy hot take.
Nvidia valuation and investment debate
- Broad agreement that Nvidia is “priced for perfection” and highly sensitive to any slowdown in growth, margin compression, or loss of share.
- Disagreement on overvaluation: some argue current P/E isn’t obviously excessive given growth; others say eventual commoditization and physical/power limits make current prices unsustainable.
- Comparisons with AMD: some see AMD as the better risk/reward; others warn simple P/E comparisons are naive and expectations matter more.
DeepSeek’s impact on demand and economics
- One camp: DeepSeek’s claimed ~45× training efficiency and much cheaper inference show massive over‑provisioning; future AI workloads may need far fewer top-end Nvidia GPUs, threatening margins.
- Opposing camp: cites Jevons paradox – cheaper, more efficient models expand use cases and total consumption; efficiency will increase, not reduce, aggregate AI compute demand, likely helping Nvidia/TSMC over time.
- Some stress that DeepSeek still uses Nvidia GPUs and that its main breakthrough is better algorithms and distillation, not non‑GPU hardware.
Competition and moats
- Nvidia’s moat is seen as multi‑layered: CUDA, mature tooling, Linux drivers (for compute), software stack, and high-speed interconnect/Mellanox.
- Counterpoints: higher-level frameworks and compilers (MLX, Triton, JAX, etc.) could erode CUDA lock‑in; cloud and hyperscaler custom silicon (TPUs, Apple, Huawei/China) may slowly chip away at Nvidia over years.
- AMD is viewed as real but lagging competition; ROCm and drivers draw mixed reviews, with anecdotes ranging from “hilariously bad” to “works great for desktop/gaming.”
Technical debates around DeepSeek
- Clarifications that mixed-precision and sub‑FP32 training have been used for years; DeepSeek pushes further (e.g., FP8 training, MoE routing, multi-token prediction, RL without labeled supervision).
- MoE discussion: generally agreed it saves per-token compute and bandwidth, not total VRAM (experts still loaded across GPUs; batching and routing matter).
- Some argue DeepSeek bundles many existing efficiency tricks (also seen in Llama) more aggressively rather than inventing something wholly new.
Infrastructure and physical limits
- Disagreement on whether electricity and cooling/water will be real constraints for Nvidia’s projected growth.
- Some argue current valuations implicitly assume AI datacenter power use can scale orders of magnitude, which skeptics doubt; others point to rapidly falling solar costs and geographic flexibility of training as mitigating factors.
Broader implications and sentiment
- Several see DeepSeek as accelerating commoditization of frontier models and compressing model-provider margins more than harming chipmakers.
- Others worry about Chinese strategic advantages, possible state backing, and potential propaganda/astroturf around DeepSeek’s narrative.
- A recurring theme is unease at AI’s pace and capital intensity, contrasted with excitement that efficiency gains might democratize model training and enable more players beyond mega‑caps.