Skip to main content
UKPMC Funders Author Manuscripts logoLink to UKPMC Funders Author Manuscripts
. Author manuscript; available in PMC: 2016 Jul 25.
Published in final edited form as: Cortex. 2013 Oct 10;51:107–108. doi: 10.1016/j.cortex.2013.09.011

Incentivising reproducible research

Katherine S Button 1,2,*, Marcus R Munafò 2
PMCID: PMC4959568  EMSID: EMS69299  PMID: 24239011

Wagenmakers and Forstmann argue that journals (and journal editors) should actively solicit replication attempts to confirm important findings (Wagenmakers & Forstmann, 2013). They point out, correctly in our opinion, that current incentive structures in science discourage scientists from adopting practices that are advantageous for science as a whole, because they leave individual scientists disadvantaged relative to their peers. Replication is one example of this, although there are others, such as the general lack of enthusiasm for publishing null results. Much of this arises from the current “publish or perish” culture within science.

These issues are not new e concerns about low statistical power and publication bias predate our own (Button et al., 2013a) by at least 50 years (Cohen, 1962; Sterling, 1959). Critically, exhortations that scientists should strive to change their practices often appear to have little effect; previous studies of statistical power, for example, have not led to an increase in statistical power of studies (Sedlmeier & Gigerenzer, 1989). We are therefore sanguine about the prospects that our recent work will have real long-term impact, despite the interesting debate and commentary that it has generated (Ashton, 2013; Bacchetti, 2013; Button et al., 2013b; Quinlan, 2013). At the same time, much of how science is conducted has changed over the last 50 years (Hubel, 2009). An obvious example is the enormous number of statistical tests that it is now possible to perform in minutes, which previously would have been impossible when statistics were calculated by hand. Multiple testing of this kind, when not reported and corrected for, invalidates key assumptions of null hypothesis significance testing (Sterne & Davey Smith, 2001).

Concerns about low statistical power and publication bias were raised 50 years ago, and it is unlikely that changes in our ability to repeatedly interrogate our data will have improved matters. Despite this, most technological advances are of course to the general advantage of science. However, other changes in how science is conducted are structural and relate to the incentives to which scientists respond. While new approaches can be initiated from the bottom-up by individual scientists, this is likely to be slow and potentially hindered by the short-term consequences of, for example, publishing fewer papers. It is therefore dangerous to assume that science is ultimately self-correcting (Ioannidis, 2012), in which case there is a clear need for change to be initiated from the top down. Journals, funding agencies and other key stakeholders have a critical role to play by incentivising research activity likely to benefit the field as a whole. The recommendation by Wagenmakers and Forstmann is one example of this, while the Registered Reports format now available at Cortex is another. What both illustrate is the need for creative thinking and leadership on the part of those in a position to encourage change.

References

  1. Ashton JC. Experimental power comes from powerful theories: the real problem in null hypothesis testing. Nature Reviews Neuroscience. 2013;14(8):585. doi: 10.1038/nrn3475-c2. [DOI] [PubMed] [Google Scholar]
  2. Bacchetti P. Small sample size is not the real problem. Nature Reviews Neuroscience. 2013;14(8):585. doi: 10.1038/nrn3475-c3. [DOI] [PubMed] [Google Scholar]
  3. Button KS, Ioannidis JP, Mokrysz C, Nosek BA, Flint J, Robinson ES, et al. Power failure: why small sample size undermines the reliability of neuroscience. Nature Reviews Neuroscience. 2013a;14(5):365–376. doi: 10.1038/nrn3475. [DOI] [PubMed] [Google Scholar]
  4. Button KS, Ioannidis JP, Mokrysz C, Nosek BA, Flint J, Robinson ES, et al. Confidence and precision increase with high statistical power. Nature Reviews Neuroscience. 2013b;14(8):585. doi: 10.1038/nrn3475-c4. [DOI] [PubMed] [Google Scholar]
  5. Cohen J. The statistical power of abnormal-social psychological research: a review. Journal of Abnormal and Social Psychology. 1962;65:145–153. doi: 10.1037/h0045186. [DOI] [PubMed] [Google Scholar]
  6. Hubel DH. The way biomedical research is organized has dramatically changed over the past half-century: are the changes for the better? Neuron. 2009;64(2):161–163. doi: 10.1016/j.neuron.2009.09.022. [DOI] [PubMed] [Google Scholar]
  7. Ioannidis JPA. Why science is not necessarily selfcorrecting. Perspectives on Psychological Science. 2012;7(6):645–654. doi: 10.1177/1745691612464056. [DOI] [PubMed] [Google Scholar]
  8. Quinlan PT. Misuse of power: in defence of small-scale science. Nature Reviews Neuroscience. 2013;14(8):585. doi: 10.1038/nrn3475-c1. [DOI] [PubMed] [Google Scholar]
  9. Sedlmeier P, Gigerenzer G. Do studies of statistical power have an effect on the power of studies? Psychological Bulletin. 1989;105(2):309–316. [Google Scholar]
  10. Sterling TD. Publication decisions and their possible effects on inferences drawn from tests of significance e or vice versa. Journal of the American Statistical Association. 1959;54(285):30–34. [Google Scholar]
  11. Sterne JA, Davey Smith G. Sifting the evidence-what’s wrong with significance tests? British Medical Journal. 2001;322(7280):226–231. doi: 10.1136/bmj.322.7280.226. [DOI] [PMC free article] [PubMed] [Google Scholar]
  12. Wagenmakers EeJ, Forstmann BU. Rewarding high power replication research. Cortex. 2013 doi: 10.1016/j.cortex.2013.09.010. [DOI] [PubMed] [Google Scholar]

RESOURCES