p-Curve and Effect Size Correcting for Publication Bias Using Only Significant Results

Loading...
Thumbnail Image
Penn collection
Operations, Information and Decisions Papers
Degree type
Discipline
Subject
publication bias
p-hacking
p-curve
Other Psychology
Theory and Philosophy
Funder
Grant number
License
Copyright date
Distributor
Related resources
Author
Simonsohn, Uri
Nelson, Leif D
Simmons, Joseph P
Contributor
Abstract

Journals tend to publish only statistically significant evidence, creating a scientific record that markedly overstates the size of effects. We provide a new tool that corrects for this bias without requiring access to nonsignificant results. It capitalizes on the fact that the distribution of significant p values, p-curve, is a function of the true underlying effect. Researchers armed only with sample sizes and test results of the published findings can correct for publication bias. We validate the technique with simulations and by reanalyzing data from the Many-Labs Replication project. We demonstrate that p-curve can arrive at conclusions opposite that of existing tools by reanalyzing the meta-analysis of the “choice overload” literature.

Advisor
Date Range for Data Collection (Start Date)
Date Range for Data Collection (End Date)
Digital Object Identifier
Series name and number
Publication date
2014-11-01
Journal title
Perspectives on Psychological Science
Volume number
Issue number
Publisher
Publisher DOI
Journal Issue
Comments
Recommended citation
Collection