Microsoft CTO says he wants to swap most AMD and Nvidia GPUs for homemade chips

Market power, pricing, and motives

  • Several see the announcement as an attempt to gain leverage over Nvidia’s pricing rather than a near-term technical shift.
  • Commenters note Nvidia’s “excess” profits and argue only hyperscalers threatening to go in-house can push prices down.
  • Some are cynical that this is also about sustaining the “AI growth” narrative for Wall Street in a broader tech-bubble pattern.

Vertical integration and big-tech playbook

  • Many compare this directly to Apple Silicon and Google TPUs, and to Amazon’s Graviton/Trainium: hyperscalers cutting out the middleman to save on per-wafer costs.
  • View that Microsoft is “late to the party,” but others say many hyperscaler silicon efforts actually began around 2018–2019, including inside Microsoft.
  • Discussion notes that these efforts are often based on ARM Neoverse cores plus custom accelerators, not fully custom CPU designs.

GPUs vs custom accelerators (training vs inference)

  • Broad agreement that inference (and much of training) is “embarrassingly parallel,” making custom ASICs and SoCs attractive.
  • Debate on whether an “inference-only” Nvidia chip is meaningfully distinct from a GPU; some cite TPUs, Groq, Tenstorrent, Etched as examples of more radical designs.
  • Several emphasize interconnects, memory bandwidth, and networking as the real bottlenecks and the hardest part to replicate, more than raw ALU performance.

Software ecosystem and CUDA moat

  • Strong consensus that Nvidia’s real advantage is CUDA and its mature tooling, not just hardware.
  • Some argue that for internal use you only need a small set of primitives (e.g., for transformers), so CUDA’s breadth matters less.
  • Others counter that developer inertia, ecosystem depth, and the scarcity of top-end engineers make it very costly to bet against CUDA at scale.

Microsoft’s credibility and impact on the ecosystem

  • Mixed views on Microsoft’s ability to execute: some point to prior in-house hardware (Catapult, Brainwave, Maia) and Azure systems; others call the company institutionally slow and see this as largely talk.
  • Concern that in-house chips across big tech could create hardware silos, limiting access for smaller players, though some hope it frees more Nvidia GPUs for consumers.