Building SQLite with a small swarm
Test Coverage, Correctness, and “Did It Work?”
- Multiple commenters ask whether the implementation passed SQLite’s official test suite; it did not.
- The project’s tests against SQLite as an “oracle” are minimal (a few simple SELECTs), far from SQLite’s tens of thousands/millions of cases.
- Lack of rigorous testing makes claims like “implemented most SQLite operations” unreliable; even the author later acknowledges over‑trusting the model’s self‑report.
Code Quality vs SQLite
- Reviewers who inspected the code describe it as basic and incomplete: no concurrency, linear free-list search, TODOs for critical behaviors (e.g., freeing overflow pages), naive buffer cloning, and a very limited query planner.
- It’s seen as potentially “basically working” for simple embedded use, but nowhere close to SQLite’s robustness, performance, or engineering standards.
- SQLite’s huge, public test suite and additional proprietary TH3 tests are repeatedly cited as the benchmark for quality.
Rust, Memory Safety, and SQLite Security
- One thread suggests a Rust, unsafe‑free implementation might avoid memory corruption vulnerabilities, even if it “eats your data.”
- Others push back, arguing SQLite’s CVEs are often overblown and that the project’s own security statements can feel dismissive or arrogant.
- Debate arises over whether SQLite’s C + exhaustive testing can be strictly “less safe” than a young Rust reimplementation.
Value, Naming, and “Simulacra”
- Strong criticism of calling this “building SQLite” when it fails the test suite; several prefer framing it as “wrote an embedded database.”
- Some argue these projects are mostly demos or props—“simulacra” of complex systems—useful for hype, not production.
- Others see genuine value in proving agents can approximate complex architectures from tests, or in the idea of clean‑room reimplementations.
Agents, Orchestration, and Validation
- The author frames the project as an experiment in multi‑agent orchestration (six heterogeneous models) rather than a viable DB.
- Commenters highlight validation as the real bottleneck; more agents and parallelism mostly create coordination overhead and messy code.
- There’s skepticism that agents can “iron out bugs” without introducing others, even with test suites.
Meta: Novelty, Licensing, and Practical Use
- Several point out that re‑creating existing OSS with LLMs is essentially “laundering” public code and offers little novelty.
- Others respond that most real‑world software is pattern‑rehash anyway, so brute‑forcing similar systems can still be economically valuable.
- Some call for more ambitious or genuinely new targets (e.g., “Wine for macOS apps”) rather than weaker clones of existing tools.