1M context is now generally available for Opus 4.6 and Sonnet 4.6

Overall reaction to 1M context

  • Many are excited: fewer forced compactions, easier long-running coding/debugging and multi-hour agent workflows.
  • Others report that quality still degrades well before 1M tokens (often ~150–200k, sometimes ~600–700k) and that 1M is mainly useful for avoiding compaction, not for “using” the whole window intelligently.
  • Several say the “dumb zone” problem has improved in recent months but not disappeared.

Pricing, plans, and token usage

  • Key change: standard Opus/Sonnet pricing now applies across the full 1M context; previous “long-context premium” goes away.
  • Confusion and complaints around subscription tiers (Pro/Max/extra usage, 5x vs 20x plans, fast mode, 1M access on Pro).
  • Long sessions are very expensive: large contexts multiply per-call cost; people report burning through hundreds of dollars or hitting Max limits in minutes if they don’t manage context aggressively.
  • Prompt caching alleviates some costs but doesn’t remove them.

Context rot, compaction, and workflows

  • Broad agreement that “context rot” is real: models start forgetting design decisions, constraints, and earlier reasoning as sessions grow.
  • Compaction is widely seen as dangerous for Claude Code: can drop key steps, reintroduce solved bugs, or lose CLAUDE.md and other instructions.
  • Mitigations discussed:
    • Frequent intentional compaction via explicit summaries and fresh sessions (RESEARCH → PLAN → IMPLEMENT).
    • Using project-level memory files (CLAUDE.md, spec.md, goal.md, log.md, task.md) and reloading them into new chats.
    • Subagents/agent teams to keep the main orchestrator’s context small while workers run in fresh windows.
    • Editing/rewinding session history, or external tools to slice/summarize JSONL logs.

Comparisons with other models and harnesses

  • Some find Opus 4.6 clearly ahead; others report better long-context stability, compaction, and review quality from OpenAI’s Codex, or niche strengths from Gemini.
  • Several emphasize that harness quality (Claude Code vs Codex CLI vs third-party tools) matters as much as the base model.

Use cases and limits

  • Positive reports: large refactors, CI-driven agents, reverse engineering, emulator and game implementations, SEO/content generation, data analysis.
  • Negative reports: infrastructure work, tricky debugging, or highly constrained/low-level domains still often require strong human oversight.
  • Debate over claims that Opus 4.6 is “AGI”; many point to brittleness, looping, and architectural blind spots as counterevidence.