Parallel coding agents with tmux and Markdown specs
Perceived productivity vs. “where’s the software?”
- Some are skeptical: if multi-agent setups are so productive, they ask why we don’t see lots of clearly “great” AI-built software.
- Others argue it’s early: tools only got good recently, feedback cycles for software quality are long, and much of the output is internal tools or personal projects.
- Several note that AI mostly increases volume of “mundane” but useful software, not necessarily greatness; average quality may even drop as volume rises.
Reported use cases and concrete wins
- Many describe internal or personal tools: finance and chat apps, window compositors, status dashboards, automation scripts, filesystem/Ansible helpers, browser automation, CI tooling, ZFS backends, games, and reverse‑engineered docs for legacy systems.
- One user claims a ~20%+ reduction in PR time-to-merge via parallel review agents, but this is challenged as over-extrapolated and not business‑proven.
- Others mention multi-agent code review at large companies and production-grade products built on agentic coding.
Orchestration patterns: tmux, specs, worktrees
- Common pattern: one markdown spec per agent/pane; agents work in parallel against git worktrees or repo copies to avoid clashes.
- Some prefer 2–3 focused agents (backend/frontend/tests) over 6–8, citing merge conflicts and cognitive overhead.
- Others build higher-level “factory” abstractions with a supervisor agent that decomposes work, spawns workers, and manages worktrees/merges.
Context management and documentation
- Major challenge is context drift across sessions; solutions include:
- Per-agent spec docs plus an orchestration doc.
- Tools like agent-doc, Beads, or entity-centered “NERDs” documents.
- PROJECT.md / SPEC.md to record direction, key decisions, and avoid scope creep.
Costs, quotas, and optimization
- Parallel agents rapidly exhaust top-tier subscriptions; several hit weekly limits in 3–4 days.
- Strategies: mix cheaper models, “oracle” agents for code questions, cheap supervisors to detect spec gaps, and strong planning/checkpointing to reduce wasted “thinking” tokens.
Quality, validation, and safety
- Strong emphasis on tests, verification commands, and sometimes separate reviewer agents.
- Concern about agents bypassing deny lists; some invert the model and require “proof of safety” (explicit intents, path checks, diffs, tests) before any action.
- Long-term impact on maintainability and design quality is seen as unclear and unmeasured.