2015
Estimating the reproducibility of psychological science
Abstract: 52% Yes, a signiicant crisis 3% No, there is no crisis 7% Don't know 38% Yes, a slight crisis 38% Yes, a slight crisis 1,576 RESEARCHERS SURVEYED M ore than 70% of researchers have tried and failed to reproduce another scientist's experiments, and more than half have failed to reproduce their own experiments. Those are some of the telling figures that emerged from Nature's survey of 1,576 researchers who took a brief online questionnaire on reproducibility in research. The data reveal sometimes-contradictory a…
View preprint versions
Search citation statements
Paper Sections
Select...
7,566
436
411
79
Citation Types
67
2,441
5
14
Year Published
1993
2026
Publication Types
Select...
5,968
1,145
709
493
Relationship
0
8,315
Authors
Journals
Cited by 7,979 publications
(2,532 citation statements)
References 43 publications
67
2,441
5
14
“…However, a look at Figure 5 reveals that with studies published with pre-registration, hence potentially preventing publication bias, the correlation is indeed smaller but still far from zero. This result is in accordance with the result of the Reproducibility Project in Psychology (Open Science Collaboration, 2015), where for replication studies, the standard error was a significant predictor ( z = 3.47, p < 0.001) for the observed effect size. The authors concluded that “[b]ecause publication bias was absent, this positive effect of standard error was likely caused by using power analysis for replication studies, i.e., generally larger replication samples were used for smaller true effects” (Open Science Collaboration, 2015, supplemental information).…”
Section: Resultssupporting
confidence: 90%
“…However, a look at Figure 5 reveals that with studies published with pre-registration, hence potentially preventing publication bias, the correlation is indeed smaller but still far from zero. This result is in accordance with the result of the Reproducibility Project in Psychology (Open Science Collaboration, 2015), where for replication studies, the standard error was a significant predictor ( z = 3.47, p < 0.001) for the observed effect size. The authors concluded that “[b]ecause publication bias was absent, this positive effect of standard error was likely caused by using power analysis for replication studies, i.e., generally larger replication samples were used for smaller true effects” (Open Science Collaboration, 2015, supplemental information).…”
Section: Resultssupporting
confidence: 90%
“…This result is in accordance with the result of the Reproducibility Project in Psychology (Open Science Collaboration, 2015), where for replication studies, the standard error was a significant predictor ( z = 3.47, p < 0.001) for the observed effect size. The authors concluded that “[b]ecause publication bias was absent, this positive effect of standard error was likely caused by using power analysis for replication studies, i.e., generally larger replication samples were used for smaller true effects” (Open Science Collaboration, 2015, supplemental information). This general correlation between sample size and effect size due to statistical power might also have led to a learning effect: in research areas with larger effects, scientists may have learned that small samples are enough while in research areas with smaller effects, they know that larger samples are needed.…”
Section: Resultssupporting
confidence: 90%
“…An exploratory linear regression with the Big Five traits and social status predicting green behavior showed only one unique main effect, for Openness, β = 0.48, SE = 0.18, t(343) = 2.65, p = 0.009. Openness having the strongest relationship with pro-environmental behavior is consistent with previous research [29,31] and lends support to the idea that Study 3 measured a behavior that seemed pro-environmental to participants.…”
Section: Resultssupporting
confidence: 87%
“…With these findings, this study documents challenges in achieving reproducibility of ecological studies on insect behavior. In comparison to other systematic replication studies [ 3 , 6 , 7 ], however, we observed higher reproducibility rates, suggesting that while reproducibility issues do exist in insect studies, they might be less pronounced than in other areas of science.…”
Section: Discussioncontrasting
confidence: 91%
