Claude Opus 4.6
Availability & Integrations
- Users quickly noticed Opus 4.6 appearing in the Claude web app, Claude Code, Cursor, VS Code, and Copilot; some needed to update clients or restart.
- Claude Code now exposes an “effort” slider (often defaulting to High) and supports Opus 4.6 as default; 1M context is API-only at launch, not on subscriptions or standard Claude Code sessions.
Agent Teams & Coding Workflows
- Major novelty is multi‑agent “teams” in Claude Code; people see this as early “agent swarms,” powerful but extremely token‑hungry and easy to misuse (agents keep reusing each other instead of terminating).
- Some compare built‑in teams to external MCP-based mail/coordination tools, noting built-in wins on friction but is session‑scoped; cross‑tool, persistent coordination remains an open niche.
Model Quality, Benchmarks & Comparisons
- Benchmarks show strong gains on “agentic search” / terminal‑bench; SWE‑Bench Verified dips by 0.1%, widely dismissed as noise on a saturated Python/Django benchmark.
- Several users report 4.6 doing deeper, more accurate code analysis and bug-finding than 4.5, especially on complex flows and large repos.
- Others find it worse at instruction-following and more prone to “running wild” (changing code unasked, not pausing for confirmation).
- Comparisons with GPT‑5.2/5.3 Codex are mixed: some prefer Codex for direction-following and speed, others find Opus more capable on reasoning-heavy or non-coding tasks.
Context Window, Pricing & Limits
- Headline feature: 1M-token context for Opus, but only via API/usage billing; above 200k tokens, per‑token prices jump (roughly 2× input, 1.5× output).
- Subscription users (Pro/Max/Teams/Enterprise) lack 1M context at launch and complain about strict Opus usage caps; many reserve Opus for planning and use cheaper models for execution.
Claude Code Architecture & Reliability
- Long thread on Claude Code being a React/Ink TUI on Bun, with heavy RAM footprint vs Rust‑based competitors; some see this as “AI slop,” others say it’s a reasonable tradeoff for fast feature iteration.
- Repo has ~6000 open issues; users report frequent non‑critical bugs, flicker, latency, memory leaks, and sandboxing concerns. Some see each release as buggier; others say this is normal for a fast‑growing, complex tool.
- Anthropic’s uptime record is criticized; some prefer accessing Claude via third‑party inference providers.
Memory, Context Compaction & Prefill Removal
- New features: automatic project memory in Claude Code and automatic “context compaction” for long sessions. Some welcome not having to hand-roll summarization; others fear opaque, accumulating “junk memory.”
- Users want easy ways to disable or tightly control memory and cross-chat recall.
- API “prefill” (forcing the first tokens of a reply) is removed on 4.6, likely for safety/jailbreak reasons; structured outputs and prompts are suggested as replacements, to the disappointment of heavy API users.
Inference Economics & Business Strategy
- Lengthy debate on whether frontier labs lose money per token: some argue costs per token are clearly falling and API inference is now marginally profitable; others insist overall training+infra is still heavily subsidized.
- Consensus: per‑token inference may be profitable; overall programs can still be deeply unprofitable given training cadence and capex.
- Many assume current flat‑rate $20–$200/month plans are introductory and will eventually rise, especially as models become more capable.
Pelican & Other Informal “Benchmarks”
- The long‑running “SVG pelican on a bicycle” test resurfaces; Opus 4.6 produces one of the best yet, though still anatomically and mechanically wrong (arms, bike frame, clouds, etc.).
- Users wonder whether Anthropic has effectively overfit to this meme prompt; similar concerns raised about benchmark “benchmaxxing” generally.
User Experiences & Meta Reactions
- Some report stunning qualitative jumps (e.g., deep literary analysis of a 900‑poem corpus, long technical research sessions, large codebase understanding); others feel only incremental gains over Opus 4.5.
- A noticeable fraction of the thread is humor, sarcasm, and job‑loss anxiety; some lament HN “turning into Reddit,” while others say joking is a healthy reaction to rapid, unsettling progress.