AI has a deep understanding of how this code works
Context of the PR
- A large PR (~13–22k LOC) added DWARF debugging support to OCaml, mostly generated by LLMs.
- The submitter openly described prompting Claude/ChatGPT and having them also write the explanations, copyright analysis, and even markdown planning files.
- The work appears influenced by an existing DWARF implementation in a forked compiler, which was also pointed at the AI as reference material.
Maintainers’ Concerns and Project Process
- Core complaint: a massive, first-time PR with no prior proposal, design discussion, or buy‑in, in an area where others are already working carefully in smaller, reviewable steps.
- Maintainers emphasized:
- Too big for the small core team to safely review.
- Insufficient tests for the amount and centrality of code.
- Design issues (DWARF library tightly coupled into the compiler, long‑term tech debt).
- Several commenters stressed that such a PR would be unacceptable even if written entirely by a human.
AI-Generated Code: Quality, Accountability, and Review Burden
- Many maintainers report AI code is harder to review than human code: it looks polished, but signals of author competence and intent are missing.
- Accountability problem: there is no evolving contributor behind the code, just one‑off artifacts; each PR might be disconnected from the last.
- Reviewers reject the idea that their role is to deeply vet code that the submitter themselves doesn’t fully understand.
Copyright and Provenance Issues
- Multiple files in the PR named another developer as author; the submitter’s answer (“AI decided, I didn’t question it”) became emblematic of the entire episode.
- Commenters see this as a red flag about provenance and as evidence that LLMs can silently “adapt” or copy from nearby codebases.
- Some argue accepting code with unknown origins is legally risky and socially corrosive, even if licenses are technically compatible.
Open Source Culture, Spam, and Platform Choices
- Maintainers describe a growing wave of AI‑generated, “drive‑by” PRs from contributors seeking résumé material or attention.
- Brandolini’s law is invoked: it takes orders of magnitude more effort to refute AI slop than to produce it.
- Proposed responses:
- Stricter contribution guidelines, explicit AI policies, and pre‑discussion requirements.
- Rejecting AI PRs outright, or at least massive ones.
- Moving away from GitHub or adding friction (self‑hosted repos, email patches, requiring local accounts) to filter out low‑investment contributors.
- Encouraging AI enthusiasts to maintain their own forks or greenfield projects instead of offloading maintenance onto existing teams.
Views on “Good” Uses of AI
- Some accept LLMs as personal tools: generating one‑off features for private forks, experiments, or non‑critical code, provided the user owns and understands the result.
- Many draw a hard line at merging large AI‑generated features into mature, shared codebases without thorough human design, ownership, and review.
Reaction to Maintainers’ Conduct
- Commenters widely praise the OCaml maintainers’ patience, clarity, and emotional maturity in handling the situation.
- There is debate over whether such politeness scales, or whether harsher, more “Torvalds‑like” responses will become necessary as AI‑driven spam increases.