Claude Opus 4.6

Availability & Integrations

  • Users quickly noticed Opus 4.6 appearing in the Claude web app, Claude Code, Cursor, VS Code, and Copilot; some needed to update clients or restart.
  • Claude Code now exposes an “effort” slider (often defaulting to High) and supports Opus 4.6 as default; 1M context is API-only at launch, not on subscriptions or standard Claude Code sessions.

Agent Teams & Coding Workflows

  • Major novelty is multi‑agent “teams” in Claude Code; people see this as early “agent swarms,” powerful but extremely token‑hungry and easy to misuse (agents keep reusing each other instead of terminating).
  • Some compare built‑in teams to external MCP-based mail/coordination tools, noting built-in wins on friction but is session‑scoped; cross‑tool, persistent coordination remains an open niche.

Model Quality, Benchmarks & Comparisons

  • Benchmarks show strong gains on “agentic search” / terminal‑bench; SWE‑Bench Verified dips by 0.1%, widely dismissed as noise on a saturated Python/Django benchmark.
  • Several users report 4.6 doing deeper, more accurate code analysis and bug-finding than 4.5, especially on complex flows and large repos.
  • Others find it worse at instruction-following and more prone to “running wild” (changing code unasked, not pausing for confirmation).
  • Comparisons with GPT‑5.2/5.3 Codex are mixed: some prefer Codex for direction-following and speed, others find Opus more capable on reasoning-heavy or non-coding tasks.

Context Window, Pricing & Limits

  • Headline feature: 1M-token context for Opus, but only via API/usage billing; above 200k tokens, per‑token prices jump (roughly 2× input, 1.5× output).
  • Subscription users (Pro/Max/Teams/Enterprise) lack 1M context at launch and complain about strict Opus usage caps; many reserve Opus for planning and use cheaper models for execution.

Claude Code Architecture & Reliability

  • Long thread on Claude Code being a React/Ink TUI on Bun, with heavy RAM footprint vs Rust‑based competitors; some see this as “AI slop,” others say it’s a reasonable tradeoff for fast feature iteration.
  • Repo has ~6000 open issues; users report frequent non‑critical bugs, flicker, latency, memory leaks, and sandboxing concerns. Some see each release as buggier; others say this is normal for a fast‑growing, complex tool.
  • Anthropic’s uptime record is criticized; some prefer accessing Claude via third‑party inference providers.

Memory, Context Compaction & Prefill Removal

  • New features: automatic project memory in Claude Code and automatic “context compaction” for long sessions. Some welcome not having to hand-roll summarization; others fear opaque, accumulating “junk memory.”
  • Users want easy ways to disable or tightly control memory and cross-chat recall.
  • API “prefill” (forcing the first tokens of a reply) is removed on 4.6, likely for safety/jailbreak reasons; structured outputs and prompts are suggested as replacements, to the disappointment of heavy API users.

Inference Economics & Business Strategy

  • Lengthy debate on whether frontier labs lose money per token: some argue costs per token are clearly falling and API inference is now marginally profitable; others insist overall training+infra is still heavily subsidized.
  • Consensus: per‑token inference may be profitable; overall programs can still be deeply unprofitable given training cadence and capex.
  • Many assume current flat‑rate $20–$200/month plans are introductory and will eventually rise, especially as models become more capable.

Pelican & Other Informal “Benchmarks”

  • The long‑running “SVG pelican on a bicycle” test resurfaces; Opus 4.6 produces one of the best yet, though still anatomically and mechanically wrong (arms, bike frame, clouds, etc.).
  • Users wonder whether Anthropic has effectively overfit to this meme prompt; similar concerns raised about benchmark “benchmaxxing” generally.

User Experiences & Meta Reactions

  • Some report stunning qualitative jumps (e.g., deep literary analysis of a 900‑poem corpus, long technical research sessions, large codebase understanding); others feel only incremental gains over Opus 4.5.
  • A noticeable fraction of the thread is humor, sarcasm, and job‑loss anxiety; some lament HN “turning into Reddit,” while others say joking is a healthy reaction to rapid, unsettling progress.