Please use this identifier to cite or link to this item:
http://hdl.handle.net/1942/31157
Title: | p-Curve and p-Hacking in Observational Research | Authors: | BRUNS, Stephan Ioannidis, John P. A. |
Issue Date: | 2016 | Publisher: | PUBLIC LIBRARY SCIENCE | Source: | PLoS One, 11 (2) , p. e0149144 (Art N° e0149144) | Abstract: | The p-curve, the distribution of statistically significant p-values of published studies, has been used to make inferences on the proportion of true effects and on the presence of p-hacking in the published literature. We analyze the p-curve for observational research in the presence of p-hacking. We show by means of simulations that even with minimal omitted-variable bias (e.g., unaccounted confounding) p-curves based on true effects and p-curves based on null-effects with p-hacking cannot be reliably distinguished. We also demonstrate this problem using as practical example the evaluation of the effect of malaria prevalence on economic growth between 1960 and 1996. These findings call recent studies into question that use the p-curve to infer that most published research findings are based on true effects in the medical literature and in a wide range of disciplines. p-values in observational research may need to be empirically calibrated to be interpretable with respect to the commonly used significance threshold of 0.05. Violations of randomization in experimental studies may also result in situations where the use of p-curves is similarly unreliable. The p-curve, the distribution of statistically significant p-values of published studies, has been used to make inferences on the proportion of true effects and on the presence of p-hacking in the published literature. We analyze the p-curve for observational research in the presence of p-hacking. We show by means of simulations that even with minimal omitted-variable bias (e.g., unaccounted confounding) p-curves based on true effects and p-curves based on null-effects with p-hacking cannot be reliably distinguished. We also demonstrate this problem using as practical example the evaluation of the effect of malaria prevalence on economic growth between 1960 and 1996. These findings call recent studies into question that use the p-curve to infer that most published research findings are based on true effects in the medical literature and in a wide range of disciplines. p-values in observational research may need to be empirically calibrated to be interpretable with respect to the commonly used significance threshold of 0.05. Violations of randomization in experimental studies may also result in situations where the use of p-curves is similarly unreliable. |
Keywords: | Humans;Malaria;Prevalence;Sample Size;Vibration;Observational Studies as Topic;Research;Statistics as Topic | Document URI: | http://hdl.handle.net/1942/31157 | ISSN: | 1932-6203 | e-ISSN: | 1932-6203 | DOI: | 10.1371/journal.pone.0149144 | ISI #: | WOS:000371218400062 | Category: | A1 | Type: | Journal Contribution |
Appears in Collections: | Research publications |
Show full item record
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.