Swarm, a new agent framework by OpenAI
Scope, Licensing, and Intended Use
- Swarm is MIT-licensed and positioned as an experimental “sample framework” for multi-agent systems, explicitly not production-ready and not officially supported.
- The repo states PRs and issues will not be reviewed, which some see as logically separate from “not for production” and effectively discouraging collaboration.
- A linked cookbook example exists, but people note basic spelling/grammar errors in official OpenAI content and question whether OpenAI actually uses its own models for documentation.
Design, Code Quality, and Comparison to Other Frameworks
- Some find the code “poorly written” (no async, heavy
deepcopy, print debugging) and see it as a simple reference rather than a serious framework. - Others argue that’s acceptable given it is explicitly a sample/experimental library.
- Multiple alternatives are suggested: LangChain, LangGraph, Autogen, txtai, Langroid, Microsoft Semantic Kernel, crewAI, griptape, and others.
- One view is that what LangChain-style frameworks do is simple enough that many teams just roll their own instead of adopting yet another abstraction.
Multi-Agent Orchestration & Technical Challenges
- Several commenters claim Swarm offers nothing fundamentally new versus many existing agent frameworks.
- A recurring theme: the “hard part” is not routing/triage of prompts but:
- Handling long-running, large-compute inference with robust message-passing.
- Dealing with high-bandwidth, multimodal data between many agents.
- Designing and optimizing agent graphs and workflows rather than a single prompt.
- There’s debate over infrastructure choices (Temporal, Kafka, etc.), with some dismissing them as reinventions of older ideas (e.g., Erlang-style systems).
Production Use, Reliability, and Hype
- Multiple people question whether multi-agent systems are actually working at scale in production, citing slowness, cost, and unreliability.
- Others report real use cases:
- Internal batch agents for large-scale code generation and testing.
- Support-fraud analysis systems where cost and latency are secondary to accuracy and analyst assistance.
- Personal agents used daily for research and data analysis.
- A key problem raised is “divergence”: ensembles of agents drift from goals, requiring strong constraints and ground-truth checks.
- Some argue that rapidly improving large-context models and newer APIs may make complex agentic setups less necessary; others counter that evals on real tasks still show benefit from carefully designed workflows.
Naming, Trademark, and Ecosystem Drama
- The name “Swarm” clashes with:
- A 1990s multi-agent simulation toolkit.
- A separate, heavily-promoted “Swarms” agent framework whose author has been criticized elsewhere for low-quality or non-functional repos.
- There is an ongoing trademark complaint around “swarms”; several commenters think it is unlikely to succeed given long-standing generic use of the term.