AI cracks superbug problem in two days that took scientists years
What the AI Actually Did
- The system (Google’s “AI co-scientist”) generated a ranked list of hypotheses for a specific microbiology problem.
- Its top hypothesis matched a conclusion the team had spent years developing; several other hypotheses “made sense,” and one is now being investigated as new.
- The AI did not prove anything experimentally; it suggested lines of inquiry that still require lab verification.
Novelty of the Hypothesis
- Many commenters argue the hypothesis was likely present in some form in prior literature or “future work” sections, or implicit in the team’s earlier publications.
- Follow‑up reporting referenced in the thread says the system was explicitly fed a 2023 paper by the same group; it may simply have ignored a limitation that the humans had assumed.
- Some see this as “when given a decade of work as premises, the computer produced the conclusion in hours,” not discovery from scratch.
Prompting, Suggestion, and “Clever Hans”
- Several note that the original prompt itself contained strong hints (e.g., mentioning the “tail”), making it easier for the model to converge on the desired answer.
- This is compared to Clever Hans or mentalism: the model appears smarter because the human encodes much of the solution in the question.
- Others frame it as “rubber‑ducking”: carefully explaining the problem to the AI helps the human see connections.
Media Hype and Misleading Framing
- The headline “cracks superbug problem in two days” is widely criticized as sensational and inaccurate.
- Commenters emphasize: AI proposed a plausible hypothesis based on existing work; it didn’t independently solve how to kill superbugs.
- Some call the article “PR for Google,” noting the timing with the product launch and the BBC’s lack of technical skepticism.
LLMs as Advanced Search / Synthesis Tools
- Many see this as LLMs shining at literature synthesis and hypothesis generation—“the next generation of search,” not magic reasoning.
- Value is in connecting scattered, published “bits” across papers and disciplines faster than humans can manually.
Data, Privacy, and Training Concerns
- Debate centers on whether the model might have used private data (Gmail, Drive, unpublished drafts).
- Google’s Workspace policy text is dissected; interpretations differ on what “without permission” and “outside of Workspace” really mean.
- There’s broader worry about opaque training sources and over‑attributing capabilities that may just be regurgitated or recombined prior work.
Impact on Science and Society
- Some wonder how such tools will affect the role of grad students and exploratory “messing around” in research.
- Others warn of societal risks in believing AI has capacities (true understanding, originality) it does not demonstrably possess.