Over-editing refers to a model modifying code beyond what is necessary

Over‑editing vs Minimal Changes

  • Many report LLM coding agents fixing a small bug by rewriting large chunks of code or multiple files, increasing cognitive complexity and “code churn.”
  • Users distinguish useful refactoring from arbitrary rewrites; most say agents frequently do the latter, including unnecessary comments and try/catch blocks.
  • Some compare this to junior dev behavior: large, “clean-up” PRs for tiny feature changes, violating the norm of low‑impact changes and small diffs.

Refactor‑as‑You‑Go Debate

  • One camp defends the “Boy Scout rule”: modest refactors while you’re in an area, especially when not refactoring would force hacks.
  • Another camp stresses Chesterton’s Fence: don’t touch code outside the current task unless you deeply understand it, and keep refactors separate from feature commits/PRs.
  • Several suggest compromise: refactor, but in follow‑up PRs, or cap refactor scope (e.g., ~10% of a change).

Autonomous vs Steerable Agents

  • Strong concern about fully autonomous agents: they touch many files, run scripts/tests/deployments, and can wipe databases or mishandle credentials.
  • Some find “babysitting” every command exhausting; others accept slower, supervised workflows as the only safe and comprehensible option.
  • Many argue for semi‑autonomous, steerable agents with granular tools, strong constraints, and explicit human design/architecture decisions.

Learning, Understanding, and Cognitive Load

  • Several worry that offloading too much work atrophies skills and leaves them not understanding their own systems.
  • Others report productivity doubling while still reviewing every line and treating the agent as a junior pair‑programmer or code reviewer.
  • Some find agent‑driven workflows too tedious or anxiety‑inducing and prefer manual coding except for small tasks.

Safety, Sandboxing, and Secrets

  • Recommended mitigations: no prod creds, strict environment separation (VMs, containers, separate pods), no unrestricted shells, tool‑wrapped commands, and read‑only or short‑lived tokens.
  • Some note agents can behave securely one day and carelessly the next, reinforcing the need for structural protections, not just instructions.

Prompting, Harnesses, and Alignment

  • Explicit prompts like “minimal change,” “no refactors,” or “don’t modify unrelated code” help but don’t fully solve over‑editing.
  • Users report better behavior with:
    • Lower “thinking”/complexity modes.
    • Project‑specific guidelines (AGENTS.md, skills).
    • Quality gates, diff regression checks, and multi‑agent review.
  • One comment highlights research: RL‑based fine‑tuning seems to generalize better than SFT for enforcing minimal editing style.