Apple M5 chip

LLM Performance, Memory Bandwidth & Capacity

  • Much debate centers on whether the base M5’s 153 GB/s unified memory bandwidth and 32 GB max RAM are enough for “proper” local LLM use.
  • Several comments explain that for large models every parameter effectively must be touched per token, making memory bandwidth and capacity the main bottlenecks; insufficient bandwidth caps tokens/sec regardless of compute.
  • Others argue the base M5 is fine for smaller or 4-bit quantized models and that Pro/Max/Ultra variants (projected ~600+ GB/s) plus 128–512 GB unified RAM will be the real LLM workhorses.
  • There’s disagreement over cost‑effectiveness versus high-end GPUs (e.g., 4090/DGX Spark): Apple wins on power and noise, loses on peak bandwidth.

Apple’s “AI” Branding and webAI Callout

  • Several note Apple previously avoided the generic “AI” label in favor of “machine learning” or “Apple Intelligence”, but the M5 press uses “AI” heavily.
  • The explicit mention of webAI as an example of on‑device LLM usage is seen as a mutually beneficial showcase of Apple’s local‑first strategy and a smaller partner that isn’t Meta/OpenAI/Chinese.

Hardware Outpaces Software

  • Strong consensus that Apple’s silicon and devices are outstanding, while macOS/iOS quality and UX have regressed.
  • Many describe Tahoe/iOS 26 as sluggish and glitchy, with laggy animations and battery drain; some tie this to an Electron private‑API bug, others to general bloat and “iOS‑ification”.
  • Long subthread rehashes whether Apple is fundamentally a hardware, software, or “systems/product” company, using revenue splits, historic quotes, and comparisons to Windows/Linux UX.

Gaming on Mac

  • Multiple users report great raw performance (e.g., Death Stranding, Cyberpunk, WoW) via native ports or GPTK/Wine, but overall gaming is “brittle”: anti‑cheat, Steam limitations, APIs, and OS instability across versions.
  • Explanations: small Mac gamer market, constant API/compat breaks (32‑bit removal, OpenGL deprecation, signing changes), Metal’s isolation vs DirectX/Vulkan, and Apple’s limited incentives (no cut on Steam).

Linux, Asahi & Openness

  • Strong desire for M‑series hardware with first‑class Linux support (or even an Apple‑sold Linux/Windows line).
  • Asahi is praised on M1/M2 but stalled on newer chips; missing sleep, TB/DP, video codecs, and full power management make it non‑mainstream.
  • Some say macOS still feels “dev‑hostile” versus Linux in openness, scripting, and window management despite its Unix base.

M5 Lineup, Specs & Marketing Claims

  • Confusion that only a base M5 exists so far, in a 14" MacBook Pro and iPad Pro, with no 16", Mac mini, or Pro/Max/Ultra yet; most expect higher‑end parts in the next cycle.
  • Bandwidth figure (1,224 Gbps = 153 GB/s) is seen as good for a base SoC but unimpressive versus earlier Max/Ultra parts and discrete GPUs.
  • Several question Apple’s “up to 3.5x” and “4x AI GPU” claims, noting that real‑world examples in the press releases mostly show ~1.2–2.3x improvements.

Neural Engine, “Neural Accelerators” & Software Stack

  • Discussion parses Apple’s multiple matmul paths: SIMD/AMX/SME on CPU, GPU tensor-style units, and the Neural Engine (ANE).
  • Some think the new “neural accelerators” are GPU tensor arrays; others highlight that ANE remains optimized for low‑power CNN‑style inference and is awkward to target (CoreML/ONNX only).
  • Developers complain that Apple’s ML stack (Metal, CoreML, MLX, MPS) is powerful but fragmented and less aligned with mainstream PyTorch/CUDA ecosystems.

AI Strategy, Energy Use & Climate Concerns

  • One camp sees Apple’s local‑first AI as a niche “persona/photo tricks” sideshow that’s dragging OS quality and wasting money; others argue on‑device private AI plus strong silicon is a long‑term differentiator.
  • Some refuse to use Apple Intelligence for climate reasons; others counter with claims that inference energy per query is small compared to overall personal footprint, though training remains energy‑intensive.
  • There’s skepticism that Apple can avoid a “Siri vs everyone else” repeat if its local models lag far behind cloud SOTA.