Who owns the code Claude Code wrote?
Status of copyright for AI‑generated code
- Many note the US Copyright Office’s position: works “predominantly” generated by AI without meaningful human authorship are not copyrightable. What counts as “meaningful” is unresolved.
- Some argue that prompting, reviewing, and editing AI output can be enough to create a new, copyrightable work; others respond that such code is at best a derivative work, not a fresh copyright.
- Image cases (e.g., Midjourney‑generated comics) are cited: human text got copyright, AI images did not. Several argue code will be treated similarly.
- Others stress that agency rulings and one circuit’s decisions are not nationwide Supreme Court precedent; law remains unsettled, especially on “how much” human input is sufficient.
Employer ownership, work‑for‑hire, and trade secrets
- Consensus: models aren’t legal persons and can’t own IP.
- Ownership today mostly flows from employment and enterprise contracts: the company that directs the work and pays for the tools typically owns whatever rights exist.
- If code is uncopyrightable, contracts can still treat it as confidential work product or trade secret, but that’s weaker: once leaked, anyone else can freely use it.
Training data, infringement, and license contamination
- Strong concern that LLMs are trained on copyrighted and copyleft code (GPL/LGPL, textbooks, GitHub), enabling “copyright washing” of OSS.
- Others argue AI “learns” like humans do, not simply copy‑pastes, though counterexamples of regurgitated code and comments are mentioned.
- Debate over whether provenance from GPL/LGPL/BSD code “travels” into outputs; no clear case law yet, but some assume courts will treat infringing outputs like any other derivative work.
Practical risk, enforcement, and M&A
- Some claim this is mostly academic: very few lawsuits so far, and enforcement (especially of GPL) is rare and expensive.
- Others say the concrete pressure will come from M&A and fundraising: acquirers already ask about AI usage and license contamination; inability to prove human authorship or clean licensing can jeopardize deals.
Ethical views and the commons
- One camp sees AI as accelerating enclosure and exploitation of creators; another sees it as undermining overbroad copyright and pushing more artifacts into the commons, closer to copyright’s original limited‑term bargain.
Impact on software practice and liability
- Developers report “vibe‑coded” codebases, weaker reviews, and erosion of shared understanding when AI writes and reviews most code.
- Others celebrate faster “lone‑wolf” development with agents as power tools.
- On liability, most argue nothing fundamental changes: organizations remain responsible for shipped code, regardless of whether a human or an AI wrote it.
Unclear / open questions
- Exact threshold for “meaningful human authorship.”
- Whether employees can freely publish uncopyrightable, AI‑generated work made at their job.
- How courts will handle provable LLM regurgitation of protected code at scale.