Publication bias in psychology: a diagnosis based on the correlation between effect size and sample size

PLoS One. 2014 Sep 5;9(9):e105825. doi: 10.1371/journal.pone.0105825. eCollection 2014.

Abstract

Background: The p value obtained from a significance test provides no information about the magnitude or importance of the underlying phenomenon. Therefore, additional reporting of effect size is often recommended. Effect sizes are theoretically independent from sample size. Yet this may not hold true empirically: non-independence could indicate publication bias.

Methods: We investigate whether effect size is independent from sample size in psychological research. We randomly sampled 1,000 psychological articles from all areas of psychological research. We extracted p values, effect sizes, and sample sizes of all empirical papers, and calculated the correlation between effect size and sample size, and investigated the distribution of p values.

Results: We found a negative correlation of r = -.45 [95% CI: -.53; -.35] between effect size and sample size. In addition, we found an inordinately high number of p values just passing the boundary of significance. Additional data showed that neither implicit nor explicit power analysis could account for this pattern of findings.

Conclusion: The negative correlation between effect size and samples size, and the biased distribution of p values indicate pervasive publication bias in the entire field of psychology.

Publication types

  • Research Support, Non-U.S. Gov't

MeSH terms

  • Epidemiologic Research Design
  • Humans
  • Mental Disorders* / diagnosis
  • Mental Disorders* / epidemiology
  • Psychology*
  • Publication Bias*
  • Research*
  • Sample Size

Grants and funding

This research was supported by a DOC-fFORTE-fellowship of the Austrian Academy of Sciences to Astrid Fritz. The funder had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.