Abstract

A focus on novel, confirmatory, and statistically significant results leads to substantial bias in the scientific literature. One type of bias, known as "p-hacking," occurs when researchers collect or select data or statistical analyses until nonsignificant results become significant. Here, we use text-mining to demonstrate that p-hacking is widespread throughout science. We then illustrate how one can test for p-hacking when performing a meta-analysis and show that, while p-hacking is probably common, its effect seems to be weak relative to the real effect sizes being measured. This result suggests that p-hacking probably does not drastically alter scientific consensuses drawn from meta-analyses.

Download full-text PDF

Link Source
Download Source 1https://journals.plos.org/plosbiology/article?id=10.1371/journal.pbio.1002106Web Search
Download Source 2http://www.ncbi.nlm.nih.gov/pmc/articles/PMC4359000PMC
Download Source 3http://dx.doi.org/10.1371/journal.pbio.1002106DOI Listing

Publication Analysis

Top Keywords

p-hacking
5
extent consequences
4
consequences p-hacking
4
p-hacking science
4
science focus
4
focus novel
4
novel confirmatory
4
confirmatory statistically
4
statistically leads
4
leads substantial
4