Local AI is driving the biggest change in laptops in decades

Memory, RAM Prices, and New Architectures

  • Many point out that exploding DRAM prices make “AI laptops with huge RAM” unrealistic in the near term; some expect 8 GB to re‑become the default.
  • Others argue DRAM cycles are historically feast/famine and high AI margins should eventually drive more capacity and lower prices, though current big-buyers (e.g., large AI labs) may distort competition.
  • Workstation laptops with 96–128 GB have existed for years; the move to 2‑slot, non‑upgradeable designs is seen as an artificial constraint.
  • Discussion of compute‑in‑flash, compute‑in‑DRAM, memristors and high‑bandwidth flash: seen as promising to host larger models cheaply, but with skepticism about latency, bandwidth figures, cost, and real‑world availability.

Critique of the Article and “AI PC” Branding

  • Multiple commenters call the article technically weak: misunderstanding TOPS, ignoring that required throughput can be computed, confusing millions vs billions of parameters, and underplaying existing open‑source benchmarks.
  • The article is criticized for ignoring the RAM price spike and for implying that most current hardware can’t run useful models.
  • “AI PC” and “Copilot+ PC” labels are widely seen as marketing; many current “AI” laptops mostly just have a cloud LLM endpoint plus an NPU that does little in practice.

Local vs Cloud AI: Capability, Economics, and Privacy

  • Enthusiasts report good experiences running mid‑sized models (e.g., 7–30B, GPT‑OSS 120B quantized) on Apple M‑series laptops with 24–128 GB, or on modest GPU desktops, for offline coding, CLI usage, and image generation.
  • Others argue that:
    • Truly frontier models (hundreds of GB) are far beyond typical consumer PCs for many years.
    • For most users, cheaper laptops + cloud subscriptions are more economical and higher quality.
  • “Good enough” is contested: some find current small models already practical; skeptics say average users will abandon them after a few visible mistakes compared to frontier cloud models.
  • Strong privacy arguments for local inference (personal data never leaving the device), but several believe most people will accept cloud trade‑offs.

GPUs, NPUs, and Specialized Accelerators

  • Debate over whether GPUs will be displaced by specialized AI chips:
    • One side expects distinct accelerators for LLMs vs diffusion.
    • Others say GPGPUs remain the best balance of power, flexibility, and cost.
  • Clarified that dense LLMs are extremely bandwidth‑bound: weights must effectively be read per token; HBM and low‑precision formats are key.
  • NPUs on consumer laptops are viewed as underpowered, fragmented, and poorly supported in software, mostly saving a bit of power for small on‑device tasks.

OS, Platforms, and Control

  • Apple silicon is repeatedly cited as currently the best laptop platform for local AI (unified memory, fast integrated GPU), though high‑RAM configs are expensive.
  • Critics note that many non‑Apple laptops marketed as “AI ready” are effectively just “can make HTTP requests to a cloud LLM.”
  • Concerns about Microsoft’s Copilot/Recall and pervasive telemetry drive some toward Linux, but gaming, creative tools (Adobe, video editing), and driver issues are significant barriers.
  • Some see aligned incentives: RAM‑hungry cloud AI competes with consumers for memory, nudging users toward being thin clients to datacenter models.

Overall Mood

  • The thread is sharply divided:
    • Optimists see local AI as already viable on high‑end consumer hardware and expect hardware to chase this use‑case.
    • Skeptics see “AI laptops” as mostly hype, with serious local AI remaining a niche akin to gaming rigs, while mainstream users rely on cheaper, more capable cloud models.