Claude wrote a full FreeBSD remote kernel RCE with root shell
Role of LLMs in finding vs exploiting bugs
- Commenters stress the exploit was written from an existing advisory, but the underlying bug itself was also originally found with help from an LLM, according to the FreeBSD security notice.
- Several see this as moving the line on what “only humans can do,” especially around kernel exploit development and ROP-style work.
- Others emphasize the human-in-the-loop nature: the shared prompt history shows lots of steering, nudging, and iteration rather than a single-shot “write full exploit” request.
Effectiveness and evidence of AI bug-finding
- Some claim LLMs are already “expert level” at finding vulnerabilities, citing talks, AI CTFs, and tools like Xbow.
- Examples mentioned: internal use at companies discovering dozens of CVEs, framework maintainers finding several, and browser vendors reportedly finding hundreds of issues with LLM help.
- Skeptics push back, asking for harder evidence (e.g., bug bounty stats) and noting that some public red-team writeups looked underwhelming.
- There’s disagreement about how to interpret earlier Anthropic red-team work; one side calls it basically a glorified search for unsafe functions, others point to concrete, acknowledged browser CVEs as counterevidence.
Offense vs defense and the CVE flood
- Many see cheaper, automated CVE discovery as a net positive: defenders and maintainers can find bugs that previously only well-funded attackers would.
- Others worry about a flood of low-impact or duplicate CVEs, and about fixing becoming the bottleneck since patches are often non-trivial and style-sensitive.
- There’s debate over whether LLMs will help more on defense (finding and fixing) or primarily empower attackers, turning this into an arms race.
Fuzzing, testing, and workflows
- Commenters describe using LLMs to:
- Design fuzzing strategies and harnesses.
- Analyze crash logs and iterate on tests.
- Reverse engineer binaries/firmware with tools like Ghidra, radare2, and dynamic instrumentation.
- Several advocate simulation-style, standalone tests with rich logs, specifically to feed AI systems that can generate remediation guidance.
Exploit scope and FreeBSD security posture
- The exploit’s attack surface is relatively narrow: a specific NFS+Kerberos setup; one exploitation path further assumes SSH access for key injection.
- Discussion notes that FreeBSD lacks some mitigations (e.g., KASLR and certain stack protections in this context), though there’s confusion about existing ASLR knobs vs true kernel ASLR.
- Some argue KASLR is limited but still part of a “defense-in-depth” onion.
Meta: hype, newsworthiness, and risk
- Some find it concerning or overhyped that this is still “news”; they see it as an expected capability of frontier LLMs.
- Others see the autonomy angle—agents chaining bugs into working exploits—as what truly worries enterprises and drives calls for governance and safety.
- A few view the thread as unpaid marketing for AI vendors; others reply that the capabilities are now visible enough to “believe your eyes.”