Evaluating the Replicability of Social Science Experiments in Nature and Science Between 2010 and 2015

By SRAI News posted 08-30-2018 12:00 AM

  

Excerpt from Evaluating the replicability of social science experiments in Nature and Science between 2010 and 2015," posted on Nature.com, August 27, 2018.


To what extent can we trust scientific findings? The answer to this question is of fundamental importance and the reproducibility of published studies has been questioned in many fields. Until recently, systematic evidence has been scarce. The Reproducibility Project: Psychology (RPP)12 put the question of scientific reproducibility at the forefront of scientific debate. The RPP replicated 100 original studies in psychology and found a significant effect in the same direction as the original studies for 36% of the 97 studies reporting ‘positive findings.’ The RPP was followed by the Experimental Economics Replication Project (EERP), which replicated 18 laboratory experiments in economics and found a significant effect in the same direction as the original studies for 61% of replications. Both the RPP and the EERP had high statistical power to detect the effect sizes observed in the original studies. However, the effect sizes of published studies may be inflated even for true-positive findings owing to publication or reporting biases. As a consequence, if replications were well powered to detect effect sizes smaller than those observed in the original studies, replication rates might be higher than those estimated in the RPP and the EERP.

We provide evidence about the replicability of experimental studies in the social sciences published in the two most prestigious general science journals, Nature and Science (the Social Sciences Replication Project (SSRP)). Articles published in these journals are considered exciting, innovative and important. We include all experimental studies published between 2010 and 2015 that (1) test for an experimental treatment effect between or within subjects, (2) test at least one clear hypothesis with a statistically significant finding, and (3) were performed on students or other accessible subject pools. Twenty-one studies were identified to meet these criteria. We used the following three criteria in descending order to determine which treatment effect to replicate within each of these 21 papers: (a) select the first study reporting a significant treatment effect for papers reporting more than one study, (b) from that study, select the statistically significant result identified in the original study as the most important result among all within- and between-subject treatment comparisons, and (c) if there was more than one equally central result, randomly select one of them for replication. The interpretation of which was the most central and important statistically significant result within a study in criteria (b) above was made by us and not by the original authors.

Continue to full article


#NewsintheField

Permalink