Why Most Published Research Findings Are False: Difference between revisions

From Wikipedia, the free encyclopedia
Content deleted Content added
m Bot: Fixing double redirect to John Ioannidis
Removed redirect to "John Ioannidis" and created a summary of the paper
Tags: Removed redirect Visual edit
Line 1: Line 1:
'''''Why Most Published Research Findings Are False'''''<ref>{{Cite journal|last=Ioannidis|first=John P. A.|date=2005|title=Why Most Published Research Findings Are False|url=https://www.ncbi.nlm.nih.gov/pmc/articles/PMC1182327/|journal=PLoS Medicine|volume=2|issue=8|pages=|doi=10.1371/journal.pmed.0020124|issn=1549-1277|pmc=1182327|pmid=16060722|via=}}</ref> is a 2005 research paper written by [[John Ioannidis]], a professor at the [[Stanford University School of Medicine|Stanford School of Medicine]], and published in [[PLOS Medicine]]. In the paper, Ioannidis argues that a large number, if not the majority, of published [[Scientific method|scientific research]] papers contain results that cannot be [[Reproducibility|replicated]].
#REDIRECT [[John Ioannidis]]

== Argument ==
Suppose that in a given scientific field there is a known baseline probability that a result is true, denoted by <math>\mathbb{P}(\text{True})</math>. When a study is conducted, the probability that a positive result is obtained is <math>\mathbb{P}(+)</math>. Given these two factors, we want to compute the [[conditional probability]] <math>\mathbb{P}(\text{True}|+)</math>, which is known as the [[Positive and negative predictive values#Positive predictive value|positive predictive value]] (PPV). Bayes' theorem allows us to compute the PPV as:<math display="block">\mathbb{P}(\text{True} | +) = {(1-\beta)\mathbb{P}(\text{True})\over{(1-\beta)\mathbb{P}(\text{True}) + \alpha\left[1-\mathbb{P}(\text{True})\right]}}</math>where <math>\alpha</math> is the [[type I error rate]] and <math>\beta</math> is the [[Type I and type II errors#Type II error|type II error rate]]; the [[Power (statistics)|statistical power]] is <math>1-\beta</math>. It is customary in most scientific research to desire <math>\alpha = 0.05</math> and <math>\beta = 0.2</math>. If we assume <math>\mathbb{P}(\text{True}) = 0.1</math> for a given scientific field. Then we may compute the PPV for different values of <math>\alpha</math> and <math>\beta</math>:
{| class="wikitable"
|
! colspan="9" |<math>\beta</math>
|-
!<math>\alpha</math>
!0.1
!0.2
!0.3
!0.4
!0.5
!0.6
!0.7
!0.8
!0.9
|-
!0.01
|0.91
|0.90
|0.89
|0.87
|0.85
|0.82
|0.77
|0.69
|0.53
|-
!0.02
|0.83
|0.82
|0.80
|0.77
|0.74
|0.69
|0.63
|0.53
|0.36
|-
!0.03
|0.77
|0.75
|0.72
|0.69
|0.65
|0.60
|0.53
|0.43
|0.27
|-
!0.04
|0.71
|0.69
|0.66
|0.63
|0.58
|0.53
|0.45
|0.36
|0.22
|-
!0.05
|0.67
|0.64
|0.61
|0.57
|0.53
|0.47
|0.40
|0.31
|0.18
|}
However, the simple formula for PPV derived from Bayes' theorem does not account for [[Bias (statistics)|bias]] in study design or reporting. In the presence of bias <math>u\in[0,1]</math>, the PPV is given by the more general expression:<math display="block">\mathbb{P}(\text{True}|+) = {\left[1-(1-u)\beta \right ]\mathbb{P}(\text{True})\over{\left[1-(1-u)\beta \right ]\mathbb{P}(\text{True}) + \left[(1-u)\alpha + u \right ]\left[1-\mathbb{P}(\text{True}) \right ] }}</math>The introduction of bias will tend to depress the PPV; in the extreme case when the bias of a study is maximized, <math>\mathbb{P}(\text{True}|+) = \mathbb{P}(\text{True})</math>. Even if a study meets the benchmark requirements for <math>\alpha</math> and <math>\beta</math>, and is free of bias, there is still a 36% probability that a paper reporting a positive result will be incorrect; if the base probability of a true result is lower, then this will push the PPV lower too. Furthermore, there is strong evidence that the average statistical power of a study in many scientific fields is well below the benchmark level of 0.8.<ref>{{Cite journal|last=Button|first=Katherine S.|last2=Ioannidis|first2=John P. A.|last3=Mokrysz|first3=Claire|last4=Nosek|first4=Brian A.|last5=Flint|first5=Jonathan|last6=Robinson|first6=Emma S. J.|last7=Munafò|first7=Marcus R.|date=2013|title=Power failure: why small sample size undermines the reliability of neuroscience|url=https://www.nature.com/articles/nrn3475|journal=Nature Reviews Neuroscience|language=en|volume=14|issue=5|pages=365–376|doi=10.1038/nrn3475|issn=1471-0048|via=}}</ref><ref>{{Cite journal|last=Szucs|first=Denes|last2=Ioannidis|first2=John P. A.|date=2017-03-02|title=Empirical assessment of published effect sizes and power in the recent cognitive neuroscience and psychology literature|url=https://journals.plos.org/plosbiology/article?id=10.1371/journal.pbio.2000797|journal=PLOS Biology|language=en|volume=15|issue=3|pages=e2000797|doi=10.1371/journal.pbio.2000797|issn=1545-7885|pmc=PMC5333800|pmid=28253258}}</ref><ref>{{Cite journal|last=Ioannidis|first=John P. A.|last2=Stanley|first2=T. D.|last3=Doucouliagos|first3=Hristos|date=2017|title=The Power of Bias in Economics Research|url=https://onlinelibrary.wiley.com/doi/abs/10.1111/ecoj.12461|journal=The Economic Journal|language=en|volume=127|issue=605|pages=F236–F265|doi=10.1111/ecoj.12461|issn=1468-0297}}</ref>

Given the realities of bias, low statistical power, and a small number of true hypotheses, Ioannidis concludes that the majority of studies in a variety of scientific fields are likely to report results that are false.

=== Corollaries ===
In addition to the main result, Ioannidis lists six corollaries for factors that can influence the reliability of published research:

# The smaller the studies conducted in a scientific field, the less likely the research findings are to be true.
# The smaller the [[Effect size|effect sizes]] in a scientific field, the less likely the research findings are to be true.
# The greater the number and the lesser the selection of [[Data dredging|tested relationships]] in a scientific field, the less likely the research findings are to be true.
# The greater the flexibility in [[Design of experiments|designs]], definitions, outcomes, and analytical modes in a scientific field, the less likely the research findings are to be true.
# The greater the [[Conflict of interest|financial and other interests]] and [[Groupthink|prejudices]] in a scientific field, the less likely the research findings are to be true.
# The hotter a scientific field (with more scientific teams involved), the less likely the research findings are to be true.

== Influence ==
Despite initial skepticism about the claims made in the paper, Ioannidis's argument has been accepted by a large number of researchers.<ref>{{Cite web|url=https://www.vox.com/2015/2/16/8034143/john-ioannidis-interview|title=John Ioannidis has dedicated his life to quantifying how science is broken|last=Belluz|first=Julia|date=2015-02-16|website=Vox|language=en|access-date=2020-03-28}}</ref> The growth of [[metascience]] and the recognition of a scientific [[replication crisis]] have bolstered the paper's credibility, and led to calls for methodological reforms in scientific research.<ref>{{Cite web|url=https://statmodeling.stat.columbia.edu/2018/02/18/low-power-replication-crisis-learned-since-2004-1984-1964/|title=Low power and the replication crisis: What have we learned since 2004 (or 1984, or 1964)? « Statistical Modeling, Causal Inference, and Social Science|website=statmodeling.stat.columbia.edu|language=en-US|access-date=2020-03-28}}</ref><ref>{{Cite journal|last=Wasserstein|first=Ronald L.|last2=Lazar|first2=Nicole A.|date=2016-04-02|title=The ASA Statement on p-Values: Context, Process, and Purpose|url=https://doi.org/10.1080/00031305.2016.1154108|journal=The American Statistician|volume=70|issue=2|pages=129–133|doi=10.1080/00031305.2016.1154108|issn=0003-1305}}</ref>

== See also ==

* [[Bayes' theorem]]
* [[Metascience]]
* [[Replication crisis]]
** [[Data dredging]]
** [[Publication bias]]
** [[Reproducibility Project]]

== References ==
<references />

== Further reading ==

* [https://www.stat.cmu.edu/~ryantibs/journalclub/ioannidis.pdf Summary and discussion of: “Why Most Published Research Findings Are False”]
* [https://marginalrevolution.com/marginalrevolution/2005/09/why_most_publis.html Why Most Published Research Findings are False]
* De Long, J. Bradford; Lang, Kevin (1992). [https://www.jstor.org/stable/2138833?seq=1 "Are all Economic Hypotheses False?"] ''Journal of Political Economy''. '''100''' (6): 1257-1272.

== External links ==

* "Why Most Published Research Findings are False" ([https://www.youtube.com/watch?v=wM0vXVclQZg Part I], [https://www.youtube.com/watch?v=NWjN67vqXOo Part II], [https://www.youtube.com/watch?v=_GAeTwLB24c Part III])
* [https://www.youtube.com/watch?v=GPYzY9I78CI John Ioannidis: "Reproducible Research: True or False?" | Talks at Google]

Revision as of 05:45, 28 March 2020

Why Most Published Research Findings Are False[1] is a 2005 research paper written by John Ioannidis, a professor at the Stanford School of Medicine, and published in PLOS Medicine. In the paper, Ioannidis argues that a large number, if not the majority, of published scientific research papers contain results that cannot be replicated.

Argument

Suppose that in a given scientific field there is a known baseline probability that a result is true, denoted by . When a study is conducted, the probability that a positive result is obtained is . Given these two factors, we want to compute the conditional probability , which is known as the positive predictive value (PPV). Bayes' theorem allows us to compute the PPV as:

where is the type I error rate and is the type II error rate; the statistical power is . It is customary in most scientific research to desire and . If we assume for a given scientific field. Then we may compute the PPV for different values of and :

0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9
0.01 0.91 0.90 0.89 0.87 0.85 0.82 0.77 0.69 0.53
0.02 0.83 0.82 0.80 0.77 0.74 0.69 0.63 0.53 0.36
0.03 0.77 0.75 0.72 0.69 0.65 0.60 0.53 0.43 0.27
0.04 0.71 0.69 0.66 0.63 0.58 0.53 0.45 0.36 0.22
0.05 0.67 0.64 0.61 0.57 0.53 0.47 0.40 0.31 0.18

However, the simple formula for PPV derived from Bayes' theorem does not account for bias in study design or reporting. In the presence of bias , the PPV is given by the more general expression:

The introduction of bias will tend to depress the PPV; in the extreme case when the bias of a study is maximized, . Even if a study meets the benchmark requirements for and , and is free of bias, there is still a 36% probability that a paper reporting a positive result will be incorrect; if the base probability of a true result is lower, then this will push the PPV lower too. Furthermore, there is strong evidence that the average statistical power of a study in many scientific fields is well below the benchmark level of 0.8.[2][3][4]

Given the realities of bias, low statistical power, and a small number of true hypotheses, Ioannidis concludes that the majority of studies in a variety of scientific fields are likely to report results that are false.

Corollaries

In addition to the main result, Ioannidis lists six corollaries for factors that can influence the reliability of published research:

  1. The smaller the studies conducted in a scientific field, the less likely the research findings are to be true.
  2. The smaller the effect sizes in a scientific field, the less likely the research findings are to be true.
  3. The greater the number and the lesser the selection of tested relationships in a scientific field, the less likely the research findings are to be true.
  4. The greater the flexibility in designs, definitions, outcomes, and analytical modes in a scientific field, the less likely the research findings are to be true.
  5. The greater the financial and other interests and prejudices in a scientific field, the less likely the research findings are to be true.
  6. The hotter a scientific field (with more scientific teams involved), the less likely the research findings are to be true.

Influence

Despite initial skepticism about the claims made in the paper, Ioannidis's argument has been accepted by a large number of researchers.[5] The growth of metascience and the recognition of a scientific replication crisis have bolstered the paper's credibility, and led to calls for methodological reforms in scientific research.[6][7]

See also

References

  1. ^ Ioannidis, John P. A. (2005). "Why Most Published Research Findings Are False". PLoS Medicine. 2 (8). doi:10.1371/journal.pmed.0020124. ISSN 1549-1277. PMC 1182327. PMID 16060722.{{cite journal}}: CS1 maint: unflagged free DOI (link)
  2. ^ Button, Katherine S.; Ioannidis, John P. A.; Mokrysz, Claire; Nosek, Brian A.; Flint, Jonathan; Robinson, Emma S. J.; Munafò, Marcus R. (2013). "Power failure: why small sample size undermines the reliability of neuroscience". Nature Reviews Neuroscience. 14 (5): 365–376. doi:10.1038/nrn3475. ISSN 1471-0048.
  3. ^ Szucs, Denes; Ioannidis, John P. A. (2017-03-02). "Empirical assessment of published effect sizes and power in the recent cognitive neuroscience and psychology literature". PLOS Biology. 15 (3): e2000797. doi:10.1371/journal.pbio.2000797. ISSN 1545-7885. PMC 5333800. PMID 28253258.{{cite journal}}: CS1 maint: PMC format (link) CS1 maint: unflagged free DOI (link)
  4. ^ Ioannidis, John P. A.; Stanley, T. D.; Doucouliagos, Hristos (2017). "The Power of Bias in Economics Research". The Economic Journal. 127 (605): F236–F265. doi:10.1111/ecoj.12461. ISSN 1468-0297.
  5. ^ Belluz, Julia (2015-02-16). "John Ioannidis has dedicated his life to quantifying how science is broken". Vox. Retrieved 2020-03-28.
  6. ^ "Low power and the replication crisis: What have we learned since 2004 (or 1984, or 1964)? «  Statistical Modeling, Causal Inference, and Social Science". statmodeling.stat.columbia.edu. Retrieved 2020-03-28.
  7. ^ Wasserstein, Ronald L.; Lazar, Nicole A. (2016-04-02). "The ASA Statement on p-Values: Context, Process, and Purpose". The American Statistician. 70 (2): 129–133. doi:10.1080/00031305.2016.1154108. ISSN 0003-1305.

Further reading

External links