Why Most Published Research Findings Are False (2005)
Scope and intent of the paper
- Many commenters see the paper as a clear, accessible synthesis of long‑standing critiques of p‑values, low power, and biased study designs.
- Others argue it overgeneralizes from medical simulations to “all research,” noting that fields differ markedly in methods and error profiles.
- Some follow‑up work (cited in the thread) suggests the original quantitative claims (e.g., proportion false) were overstated, but directionally right.
Replication crisis and field differences
- Strong consensus that replication problems are worst in medicine, psychology, social sciences, ecology/climatology, and some life sciences.
- Physical sciences, some areas of chemistry/biochemistry, and engineering are seen as more robust, partly because experiments can be repeated many times with high signal‑to‑noise.
- Computer science is split: some see widespread non‑reproducibility (e.g., sensitivity to random seeds, missing methods/code), others stress that CS results are more abstract and not meant as drop‑in industrial solutions.
Peer review, media, and public trust
- Peer review is described as a weak filter: more a “stamp” than verification or replication.
- Science journalism is heavily criticized for hyping single studies, omitting key details, and fostering the impression that each paper is a fact.
- This feeds public confusion (“science changes every week”) and politicized sloganizing (“trust the science”) despite fragile underlying evidence.
- Several argue decisions should rest on replicated results, meta‑analyses, and convergence of evidence, not single papers.
Non‑replication, “truth,” and usefulness
- Some emphasize that non‑replication often signals “failure to generalize” rather than outright falsity, especially in heterogeneous human/medical contexts.
- Others counter that, given most hypotheses are false a priori, repeated non‑replication should strongly increase confidence in the null.
- Key point: even nominally true but irreproducible results are a poor foundation for further work; science needs stable, usable building blocks.
Incentives, misconduct, and systemic issues
- “Publish or perish,” prestige chasing, and funding rules incentivize p‑hacking, overinterpretation, and omission of crucial methods.
- Fraud is viewed as rare but impactful; incompetence and sloppy practice are seen as far more common.
- Commenters note paper mills, retractions, and national/institutional pressures, but disagree on how pervasive outright fraud is.
Proposed reforms and alternatives
- Suggestions include: valuing replication, preregistration, better data management and sharing, publishing under pseudonyms, open code, meta‑science tools, and new publishing platforms.
- Some argue science is a noisy but convergent process (like gradient descent) that moves toward truth over generations; others worry it can get trapped in local minima due to politics, funding, and entrenched assumptions.