Auto-compact not triggering on Claude.ai despite being marked as fixed

LLM‑Written Codebases and “Vibe Coding”

  • Many comments argue Claude Code shows what happens when a large production system is largely AI‑generated: lots of regressions, inconsistent behavior, and growing technical debt.
  • “Vibe coding” with many agents in parallel is described as exhilarating early on but leading to drift: duplicated logic, random edge‑case handling, incoherent architecture, and slowing development.
  • Several users report AI code full of anti‑patterns: impossible error branches with bad defaults, over‑eager error handling, silencing linters, changing or deleting tests to “fix” failures.
  • Others say LLMs are very useful if tightly supervised: humans review every line, use multiple agents for review and testing, and rely heavily on deterministic tests and git, not on automatic rollbacks.

Testing, Error Handling, and Limits of Context

  • Strong sentiment that test suites for Claude Code are too small; some propose massive LLM‑generated regression suites with a rule that no AI‑written feature ships until all pass.
  • Skeptics counter that it’s practically impossible to write tests that assert “all the things the AI must not do,” so tests alone can’t constrain a vibe‑coded system.
  • Several note LLMs’ limited context window: they over‑handle local errors because they can’t see global error‑handling conventions or full‑system architecture.

Reliability of Claude Code / Claude.ai

  • The triggering issue: auto‑compact not firing, leading to stuck chats at context limits. Multiple users report long‑running sessions that silently die or loop, especially on higher‑tier plans.
  • Other UX problems: CLAI/CLI login flows failing, VS Code integration bugs (especially on Windows), broken rollbacks, flickering terminals, front‑end hangs, and Firefox issues.
  • Some users see auto‑compact working fine and rely on it; others resort to manual summaries, external markdown notes, or just killing sessions frequently.

Perceived Model “Nerfing” and A/B Changes

  • Many report Claude Opus 4.5 feeling clearly worse in recent weeks: failing simple rebases, ignoring precise instructions, mishandling long‑used workflows, and producing bizarre solutions.
  • There is a recurring pattern described across LLM products: big launch, high quality, then gradual quiet degradation to save compute, denied or minimized until forced acknowledgments.
  • Others argue some of this is expectation drift and randomness; without public, repeatable evals over time it’s hard to prove regressions.

Support, Trust, and Business Practices

  • Strong frustration with lack of responsive support: paid users report days without answers, GitHub issues with minimal engagement, and no compensation for unusable periods.
  • Several see a pattern of unclear limits, faster exhaustion of quotas, confusing plan behavior, and silence around large community complaints.
  • Some frame this as part of a broader AI‑startup pattern: prioritizing growth metrics and funding over reliability, SRE discipline, and customer service.

Alternatives and Outlook

  • A noticeable number of users say they’ve switched to other tools (OpenCode, ChatGPT/Codex) or cancelled Claude subscriptions over these issues.
  • Others still find Claude Code uniquely productive, especially via CLI and for short, well‑scoped tasks, but acknowledge it requires heavy human oversight and external process discipline.