Abstract
Serious concerns about research quality have catalysed a number of reform initiatives intended to improve transparency and reproducibility and thus facilitate self-correction, increase efficiency and enhance research credibility. Meta-research has evaluated the merits of some individual initiatives; however, this may not capture broader trends reflecting the cumulative contribution of these efforts. In this study, we manually examined a random sample of 250 articles in order to estimate the prevalence of a range of transparency and reproducibility-related indicators in the social sciences literature published between 2014 and 2017. Few articles indicated availability of materials (16/151, 11% [95% confidence interval, 7% to 16%]), protocols (0/156, 0% [0% to 1%]), raw data (11/156, 7% [2% to 13%]) or analysis scripts (2/156, 1% [0% to 3%]), and no studies were pre-registered (0/156, 0% [0% to 1%]). Some articles explicitly disclosed funding sources (or lack of; 74/236, 31% [25% to 37%]) and some declared no conflicts of interest (36/236, 15% [11% to 20%]). Replication studies were rare (2/156, 1% [0% to 3%]). Few studies were included in evidence synthesis via systematic review (17/151, 11% [7% to 16%]) or meta-analysis (2/151, 1% [0% to 3%]). Less than half the articles were publicly available (101/250, 40% [34% to 47%]). Minimal adoption of transparency and reproducibility-related research practices could be undermining the credibility and efficiency of social science research. The present study establishes a baseline that can be revisited in the future to assess progress.
Keywords: transparency, reproducibility, meta-research, social sciences, open science
1. Introduction
Transparency and reproducibility are core scientific principles and their pursuit can be important for improving efficiency [1], facilitating self-correction [2] and enhancing the credibility of the published literature [3]. Although journal articles are often considered the principal commodity of the scientific ecosystem [4], they usually only provide an incomplete summary of a research project. A rich array of research resources such as (ideally pre-registered) protocols (i.e. hypotheses, methods and analysis plans), materials, raw data and analysis scripts provide the most detailed documentation of a study, and having access to them can support independent verification, replication, evidence synthesis and further discovery [5]. For example, access to raw data enables assessment of computational reproducibility [6–10] and analytic robustness [11], more sophisticated individual participant-level meta-analyses [12], and creative merging of datasets or use of novel analytic techniques [13]. However, although researchers appear to endorse the values of transparency and reproducibility in principle [14,15], they are routinely neglected in practice [16–19].
Concerns about low transparency and poor reproducibility [20–24] have catalysed an array of reform initiatives from researchers, journals, publishers, funders and universities. In some cases, it has been possible to isolate these interventions and evaluate their impact through meta-research (research on research) [25,26]. For example, journal data sharing policies have been associated with an increased quantity of publicly available datasets [6,27,28]. However, such assessments may not capture broader trends reflecting the cumulative contribution of disparate initiatives and growing awareness among researchers about the importance of transparency and reproducibility [15]. It is, therefore, valuable to zoom out and take a bird's eye view of the published literature. Building on similar studies in the biomedical domain [17,18], we manually examined a random sample of 250 articles in order to estimate the prevalence of a range of transparency and reproducibility-related indicators in the social sciences literature published between 2014 and 2017.
2. Methods
The study protocol was pre-registered on 3 July 2018 and is available at https://osf.io/u5bk9/. An amended version of the protocol was registered on 14th August 2019 and is available at https://osf.io/j5zsh/. All deviations from the original protocol are explicitly acknowledged in the main text. All data exclusions and measures conducted during this study are reported.
2.1. Design
This was an observational study with a cross-sectional design. Measured variables are shown in table 1.
Table 1.
applicable study designs | |
---|---|
articles | |
accessibility and retrieval method (can the article be accessed, is there a public version or is paywall access required?) | all |
protocols | |
availability statement (is availability, or lack of, explicitly declared?) | empirical studiesa, commentaries and meta-analyses |
content (what aspects of the study are included in the protocol?) | |
materials | |
availability statement (is availability, or lack of, explicitly declared?) | empirical studiesa |
retrieval method (e.g. upon request or via online repository) | |
accessibility (can the materials be accessed?) | |
raw data | |
availability statement (is availability, or lack of, explicitly declared?) | empirical studiesa, commentaries and meta-analyses |
retrieval method (e.g. upon request or via online repository) | |
accessibility (can the data be accessed?) | |
content (has all relevant data been shared?) | |
documentation (are the data understandable?) | |
analysis scripts | |
availability statement (is availability, or lack of, explicitly declared?) | empirical studiesa, commentaries and meta-analyses |
retrieval method (e.g. upon request or via online repository) | |
accessibility (can the scripts be accessed?) | |
pre-registration | |
availability statement (is availability, or lack of, explicitly declared?) | empirical studiesa, commentaries and meta-analyses |
retrieval method (which registry was used?) | |
accessibility (can the pre-registration be accessed?) | |
content (what was pre-registered?) | |
funding | |
disclosure statement (are funding sources, or lack of, explicitly declared?) | all |
conflicts of interest | |
disclosure statement (are conflicts of interest, or lack of, explicitly declared?) | all |
replication | |
statement (does the article claim to report a replication?) | all |
citation history (has the article been cited by a study that claims to be a replication?) | empirical studiesa |
evidence synthesis | |
meta-analysis citation history (has the article been cited by, and included in the evidence-synthesis component of, a meta-analysis) | empirical studiesa |
systematic review citation history (has the article been cited by, and included in the evidence-synthesis component of, a systematic review) | empirical studiesa |
article characteristics | |
subject area, year of publication, study design, country of origin (based on corresponding author's affiliation), human/animal subjects, 2016 journal impact factor (according to Thomson Reuters Journal Citation Reports) | all |
a‘Empirical studies’ encompasses the following study design classifications: field studies, laboratory studies, surveys, case studies, multiple types, clinical trials and ‘other’ designs.
2.2. Sample
2.2.1. Sampling frame
We identified a random sample of 250 articles published in the social sciences between January 2014 and April 2017. Specifically, the sample was drawn from a database based on Scopus-indexed content that classifies academic documents according to one of 12 broad fields of science using a model of the disciplinary structure of the scientific literature [29]. The model grouped documents into 91 726 clusters using citation information and 14 342 of these clusters were assigned to the social sciences category. For the time period of interest (2014–2017), the 14 342 social science clusters contained 648 063 documents with an article type of ‘article’ or ‘review’ (as determined by Scopus). However, not all articles included in these social science clusters may be strongly related to social sciences. Therefore, to further ensure the face validity of articles included in the sample, we additionally limited the documents to those that had an All Science Journal Classification (ASJC) code related to the social sciences, specifically ‘Economics, Econometrics and Finance’ (ECON), ‘Psychology’ (PSYCH), ‘Business, Management and Accounting’ (BUS) and ‘Social Sciences’ (SOC). Note that this means the sample would not capture social science articles published in multidisciplinary journals (e.g. Nature, Science, PNAS, RSOS) and/or journals that belong mainly to other disciplines. The number of documents in the database broken by ASJC code was as follows: BUS, 105 447; ECON, 92 348; PSYCH, 75 353; SOC, 324 618.
Based on our judgement of what sample size would be sufficiently large enough to be informative but also realistically feasible for manual data extraction, we decided to examine 250 articles. Random sampling was performed by using a random number generator to shuffle the order of the articles and selecting the top N articles required. Following insightful comments from a peer-reviewer, we detected a (now corrected) error in the original sampling procedure. Full details of the error and the steps taken to correct it are available at https://osf.io/7anx6/. The final sample of 250 articles represents a random sample of the 485 460 English-language articles available in the database [29] that were classified into one of the social sciences clusters, also had an ASJC code specifically related to the social sciences (BUS, ECON, PSYCH or SOC), and were published between January 2014 and April 2017.
2.3. Procedure
A structured form (https://osf.io/v4f59/) was created based on previous investigations in the biomedical domain [17,18] and used to guide data extraction and coding for a range of indicators related to transparency and reproducibility. The exact indicators measured depended on study design (table 1). We pilot tested the data extraction procedures using 15 articles that were not included in the final sample.
Between 50 and 71 articles were randomly assigned to each of four investigators (JDW, MCK, TB and SC) who performed initial extraction and coding. All articles were second coded by a fifth investigator (TEH) and any coding differences were resolved through discussion. Journal impact factors and citations histories were obtained by a single investigator only. Data collection took place between 7 May 2018 and 10 October 2019.
We attempted to access each article by searching with the Open Access Button (https://openaccessbutton.org/), Google (https://www.google.com/), and two of five university libraries (Stanford University, Yale University, University of Utah, Aarhus University, Charité – Universitätsmedizin Berlin) depending on the institutional affiliation of the assigned coders.
3. Results
Values in square brackets are 95% confidence intervals based on the Sison-Glaz method for multinomial proportions [30].
3.1. Sample characteristics
Accessible articles originated from journals with a wide range of journal impact factors (median 1.30, range 0.04–16.79; based on 2016 journal impact factor). For 122 articles, no journal impact factor was available. Other sample characteristics are displayed in table 2.
Table 2.
n | |
---|---|
subject area | |
education | 34 |
geography, planning and development | 18 |
sociology and political science | 16 |
cultural studies | 14 |
economics and econometrics | 14 |
business and international management | 12 |
finance | 10 |
44 other social sciences subject-areasa (accounting for less than 10 per area) | 118 |
year of publication | |
2014 | 63 |
2015 | 84 |
2016 | 69 |
2017 | 20 |
study design | |
no empirical data | 80 |
field study | 73 |
survey | 39 |
multiple types | 22 |
case study | 9 |
laboratory study | 7 |
commentary with analysis | 3 |
meta-analysis | 2 |
other | 1 |
country of origin | |
USA | 76 |
UK | 25 |
Australia | 17 |
Germany | 13 |
47 other countriesb (accounting for less than 10 per country) | 105 |
subjects | |
human | 105 |
animal | 0 |
neither human nor animal subjects involved | 131 |
aFor all subject areas, see https://osf.io/7fm96/.
bFor all countries, see https://osf.io/kg7j5/.
3.2. Article availability
Among the 250 articles, 101 (40% [34% to 47%]) had a publicly available version (figure 1a) whereas 135 (54% [48% to 61%]) were only accessible to us through a paywall. Of those paywalled articles, 34 (25%) had received public funding (this may be an underestimate because 67% of the paywalled articles did not provide a funding statement); 14 (6% [0% to 12%]) additional articles were not available to our team, highlighting that even researchers with broad academic access privileges cannot reach portions of the scientific literature.
3.3. Materials and protocol availability
Research protocols (documentation of the research methodology and analysis plan) and original research materials (such as survey instruments, stimuli or software) provide a detailed account of how a study was performed. The vast majority of articles we examined did not contain a materials availability statement (figure 1b) and none contained a protocol availability statement (figure 1c). Of the 16 of 151 (11% [7% to 16%]) articles where materials were reportedly available, two links to journal-hosted electronic supplementary material were broken and 1 was only ‘available upon request’ from authors. The remaining 13 sets of materials were available in the article itself (n = 10), via journal-hosted electronic supplementary material (n = 2), or in another paper (n = 1).
3.4. Data availability
Raw data are the core evidence that underlies scientific claims. However, the vast majority of the 156 relevant articles we examined did not contain a data availability statement (figure 1d). Eighteen articles (12% [6% to 18%]) mentioned using an external data source,1 but did not provide detailed instructions on how to obtain the specific data that was used. A further 11 (7% [2% to 13%]) datasets were reportedly available; however, eight of these were in fact not accessible due to non-functioning links to personal/institutional web pages (n = 4), a non-functioning link to journal electronic supplementary material (n = 1), only being ‘available upon request’ from authors (n = 2; note that we did not attempt to contact them), or involving a file that could not be opened because commercial software was required (n = 1). Of the three accessible datasets (two available via electronic supplementary material and one available via a personal/institutional web page) two were both complete and clearly documented and one was neither complete nor clearly documented.
3.5. Analysis script availability
Analysis scripts provide detailed step-by-step descriptions of performed analyses, often in the form of computer code (e.g. R, Python or Matlab) or syntax (SPSS, Stata, SAS). Two of 156 (1% [0% to 3%]) articles reported that analysis scripts were available (figure 1e) via journal-hosted electronic supplementary material. One of these was not accessible due to a non-functioning link.
3.6. Pre-registration
Pre-registration refers to the archiving of a read-only, time-stamped study protocol in a public repository (such as the Open Science Framework, https://osf.io/) prior to study commencement [31]. None of the articles specified that the study was pre-registered (figure 1f), highlighting that this concept still has fledgling status in the social sciences.
3.7. Conflicts of interest and funding statements
Statements about potential conflicts of interest and funding sources enable researchers to disclose pertinent factors that could introduce bias: 162 of 236 (69% [63% to 75%]) articles did not include a funding statement (figure 1g), 200 of 236 (85% [81% to 89%]) articles did not include a conflicts of interest statement (figure 1h), and 36 of 236 (15% [11% to 20%]) articles reported that there were no conflicting interests and no articles reported that there were conflicting interests. Of the articles that reported their funding sources, 59 had received public funding.
3.8. Replication and evidence synthesis
Replication studies repeat the methods of previous studies in order to systematically gather evidence on a given research question. Evidence gathered across multiple relevant studies can be formally collated and synthesized through systematic reviews and meta-analyses. Only 2 of the 156 (1% [0% to 3%]) articles we examined explicitly self-identified as a replication study. No articles were cited by another article reporting to be a replication. Seventeen (11% [7% to 16%]) articles were included in evidence synthesis in a systematic review. Two (1% [0% to 3%]) had been included in evidence synthesis in a meta-analysis (an additional 1 was cited but excluded). Overall, articles associated with empirical data were infrequently cited (median = 1, min = 0, max = 31).
4. Discussion
Our empirical assessment of a random sample of articles published between 2014 and 2017 suggests a serious neglect of transparency and reproducibility in the social sciences. Most research resources, such as materials, protocols, raw data and analysis scripts, were not explicitly available, no studies were pre-registered, disclosure of funding sources and conflicts of interest was modest, and replication or evidence synthesis via meta-analysis or systematic review was rare.
Poor transparency can have very real costs. For example, the Reproducibility Project in Cancer Biology—a major effort to replicate 50 high-impact cancer biology papers—recently had to abandon 32 replication attempts partly because pertinent methodological information about the original studies was not available [32]. Conversely, transparency and sharing can have tangible benefits. For example, a paper entitled ‘Growth in a Time of Debt’, which influenced government austerity policies around the world, was found to contain a serious analytic error by a student examining formulae and data in the original spreadsheet [33]. Social science research often addresses topics that are highly pertinent to policy makers, and policy decisions based on flawed research can have substantial economic, social and individual costs. To ensure that a credible evidence base is available to policy makers, it is imperative that social science research is held to high-quality standards [22].
The finding that less than half of the articles in our sample were publicly available is broadly consistent with a large-scale automated analysis which suggested that 55% of the scientific literature published in 2015 was not open access [34]. This suggests that critical stakeholders are potentially being deprived of relevant scientific information. Six per cent of the articles in our sample could not even be retrieved through some of the most comprehensive university library catalogues.
The low availability of materials and protocols that we observed is consistent with transdisciplinary investigations indicating that most articles lack availability statements for these resources [16–18]. Not being able to view original research materials and protocols precludes comprehensive evaluation of a study [3]. Furthermore, inability to obtain materials and protocols can disrupt or even prevent efforts to conduct high-fidelity replication attempts which are vital for both verification and the systematic accumulation of knowledge [2]. We did not evaluate whether all relevant research materials had been made available or whether protocols were sufficiently described in the articles themselves.
Assessments in other scientific disciplines have also shown that raw data are typically not available [10,16–18] including for some of the most influential articles [35]. Requesting data directly from authors typically yields only a modestly successful response [10,35–37]. Even data that have been made available can be poorly organized and documented, and do not guarantee the analytic reproducibility of the reported outcomes [6–10,38]. The inability to access the raw data that underlines scientific claims seriously limits verification and reuse, potentially wasting resources by impeding discovery and compromising the supposedly self-correcting nature of the scientific endeavour [1–3].
Sharing of analysis scripts seems extremely rare in many scientific fields [6,7,18]. Analysis scripts are important because the verbal description of the original analyses provided in a published paper can be ambiguous, incomplete, or even incorrect, undermining efforts to independently establish computational reproducibility [6–9] and precluding comprehensive peer scrutiny [3].
We found no studies had been pre-registered. Because pre-registration occurs prior to study commencement, there will be some lag before any changes in the uptake of pre-registration are detectable in the published literature. Additionally, pre-registration may be less pertinent when analyses of pre-existing data are intended to be entirely exploratory and no pre-conceived protocol really exists. Various modes of study pre-registration are gaining attention in several research communities, but still capture only a small proportion of total research output [31]. At the time of writing, the Open Science Framework contains over 41 000 pre-registrations (https://osf.io/registries/discover). Pre-registration is intended to facilitate the demarcation of exploratory (unplanned) and confirmatory (planned) analyses, and potentially mitigates or enables detection of questionable research practices, such as opaque ‘hypothesizing after the results are known’ [39], selective reporting [19,40,41] or ‘p-hacking’ [31,42,43]. When studies are not pre-registered, these potential advantages are lost.
Disclosure statements related to potential conflicts of interest and funding are also often neglected in the biomedical sciences [17,18], but the situation seems worse in the social sciences. Articles that do not include these disclosure statements cannot be evaluated in full context. It is possible that many authors neglect to include these statements because they believe there is nothing relevant to disclose. Nevertheless, a lack of funding or potential conflicts is still important to explicitly acknowledge because the absence of a statement is ambiguous.
We assessed article citation histories in order to gauge how often they had been cited overall and cited by replication studies, meta-analyses or systematic reviews specifically. It should be noted that our sample pertained to recently published studies and it may take some time before studies that build upon the original articles are themselves published. Nevertheless, these findings are generally consistent with evidence suggesting that replication studies are rarely conducted or reported and individual studies only infrequently contribute to the cumulative body of research evidence through inclusion in systematic reviews and meta-analyses [17,18,44].
Our study has several important limitations. Firstly, no database has perfect coverage of all journals in every field, but Scopus coverage is extensive. Google Scholar may have broader coverage than Scopus, but it is less transparent than Scopus about what journals and sources of information are included. Secondly, we relied solely on published information. If we had contacted authors, it is likely that we would have been able to obtain additional information. However, requests to authors for resources like raw data typically have modest yield and the modal outcome is to not respond [35–37,45]. Thirdly, as with previous similar investigations [17,18], we did not deeply assess the quality of information and resources that were shared. Given the low levels of transparency, such quality checks would not have been especially informative. However, it is important to bear in mind that transparency alone is rarely sufficient. For example, just because data are shared does not necessarily imply that they are reusable in practice [6]. Fourthly, achieving transparency is not always straightforward when there are overriding legal, ethical or practical concerns [46]. It is possible that a lack of transparency (in particular, a lack of data sharing) is well justified in some cases. However, no such justifications were stated in the articles we examined. It has been argued that when complete transparency is not possible, the reasons should be explicitly declared in any relevant scientific publications [47]. Finally, although our sample can be used to estimate the prevalence of the measured indicators broadly in the social sciences, those estimates may not necessarily generalize to specific contexts, for example, specific subfields or articles published in specific journals. It is known, for example, that specific journal policies can be associated with increases in data and materials availability [6,27,28].
Minimal adoption of transparency and reproducibility-related research practices impoverishes the scientific ecosystem [1], disrupts self-correction activities [2] and could undermine public trust in science [3]. A cascade of reform initiatives have been launched in an effort to improve transparency [21,22], and close monitoring and evaluation of specific reform initiatives will help to identify efforts that are the most successful and highlight areas for improvement [25,26]. The present study establishes a baseline estimate for the prevalence of transparency and reproducibility-related indicators in the social sciences that can be revisited in the future to assess progress.
Supplementary Material
Acknowledgements
We thank Kevin Boyack for assistance obtaining the sample of articles. We are grateful to an anonymous peer-reviewer whose comments led us to identify an error (now corrected) in our initial sampling procedure.
Footnotes
This particular classification was created post hoc and not included in the pre-registered protocol.
Data accessibility
All data (https://osf.io/u9fw8/), materials (https://osf.io/z9qtr/) and analysis scripts (https://osf.io/sbrez/) related to this study are publicly available. To facilitate reproducibility this manuscript was written by interleaving regular prose and analysis code and is available in a Code Ocean container (https://doi.org/10.24433/CO.2749769.v4) which re-creates the software environment in which the original analyses were performed.
Authors' contributions
T.E.H., J.D.W. and J.P.A.I. designed the study. T.E.H., J.D.W., M.C.K., T.B. and S.C. conducted the data collection. T.E.H. performed the data analysis. T.E.H., J.D.W. and J.P.A.I. wrote the manuscript. All the authors gave their final approval for publication.
Competing interests
In the past 36 months, J.D.W. received research support outside of the scope of this project through the Collaboration for Research Integrity and Transparency from the Laura and John Arnold Foundation, and through the Center for Excellence in Regulatory Science and Innovation (CERSI) at Yale University and the Mayo Clinic (U01FD005938). All other authors declare no competing interests.
Funding
The authors received no specific funding for this work. The Meta-Research Innovation Center at Stanford (METRICS) is supported by a grant from the Laura and John Arnold Foundation. The Meta-Research Innovation Center Berlin (METRIC-B) is supported by a grant from the Einstein Foundation and Stiftung Charité.
References
- 1.Chalmers I, Glasziou P. 2009. Viewpoint avoidable waste in the production and reporting of research evidence. The Lancet 374, 86–89. ( 10.1016/s0140-6736(09)60329-9) [DOI] [PubMed] [Google Scholar]
- 2.Ioannidis JPA. 2012. Why science is not necessarily self-correcting. Perspect. Psychol. Sci. 7, 645–654. ( 10.1177/1745691612464056) [DOI] [PubMed] [Google Scholar]
- 3.Vazire S. 2017. Quality uncertainty erodes trust in science. Collabra: Psychol. 3, 1–5. ( 10.1525/collabra.74) [DOI] [Google Scholar]
- 4.Young NS, Ioannidis JPA, Al-Ubaydli O. 2008. Why current publication practices may distort science. PLoS Med. 5, e201 ( 10.1371/journal.pmed.0050201) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.Klein O, et al. 2018. A practical guide for transparency in psychological science. Collabra: Psychol. 4, 20 ( 10.1525/collabra.158) [DOI] [Google Scholar]
- 6.Hardwicke TE, et al. 2018. Data availability, reusability, and analytic reproducibility: evaluating the impact of a mandatory open data policy at the journal cognition. R. Soc. open sci. 5, 180448 ( 10.1098/rsos.180448) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.McCullough BD, McGeary KA, Harrison TD. 2008. Do economics journal archives promote replicable research? Can. J. Econ. 4, 1406–1420. ( 10.1111/j.1540-5982.2008.00509.x) [DOI] [Google Scholar]
- 8.Wood BDK, Müller R, Brown AN. 2018. Push button replication: is impact evaluation evidence for international development verifiable? PLoS ONE 13, e0209416 ( 10.1371/journal.pone.0209416) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.Chang AC, Li P. 2015. Is economics research replicable? Sixty published papers from thirteen journals say ‘Usually Not’. Econ. Discuss. Ser. 2015, 1–26. ( 10.17016/feds.2015.083) [DOI] [Google Scholar]
- 10.Stockemer D, Koehler S, Lentz T. 2018. Data access, transparency, and replication: new insights from the political behavior literature. PS: Political Sci. Politics 51, 799–803. ( 10.1017/s1049096518000926) [DOI] [Google Scholar]
- 11.Steegen S, Tuerlinckx F, Gelman A, Vanpaemel W. 2016. Increasing transparency through a multiverse analysis. Perspect. Psychol. Sci. 11, 702–712. ( 10.1177/1745691616658637) [DOI] [PubMed] [Google Scholar]
- 12.Tierney JF, Vale C, Riley R, Smith CT, Stewart L, Clarke M, Rovers M. 2015. Individual participant data (IPD) meta-analyses of randomised controlled trials: guidance on their use. PLoS Med. 12, e1001855 16 ( 10.1371/journal.pmed.1001855) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Voytek B. 2016. The virtuous cycle of a data ecosystem. PLoS Comput. Biol. 12, e1005037 6 ( 10.1371/journal.pcbi.1005037) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Anderson MS, Ronning EA, Devries R, Martinson BC. 2010. Extending the Mertonian norms: scientists' subscription to norms of research. J. High. Edu. 81, 366–393. ( 10.1353/jhe.0.0095) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Baker M. 2016. Is there a reproducibility crisis? A nature survey lifts the lid on how researchers view the crisis rocking science and what they think will help. Nature 533, 452–454. ( 10.1038/521274a) [DOI] [PubMed] [Google Scholar]
- 16.Alsheikh-Ali AA, Qureshi W, Al-Mallah MH, Ioannidis JPA. 2011. Public availability of published research data in high-impact journals. PLoS ONE 6, e24357 4 ( 10.1371/journal.pone.0024357) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Iqbal SA, Wallach JD, Khoury MJ, Schully SD, Ioannidis JPA. 2016. Reproducible research practices and transparency across the biomedical literature. PLoS Biol. 14, e1002333 13 ( 10.1371/journal.pbio.1002333) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Wallach JD, Boyack KW, Ioannidis JPA. 2018. Reproducible research practices, transparency, and open access data in the biomedical literature, 2015–2017. PLoS Biol. 16, e2006930 ( 10.1371/journal.pbio.2006930) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.Necker S. 2014. Scientific misbehavior in economics. Evol. Hum. Behav. 43, 1747–1759. ( 10.1016/j.respol.2014.05.002) [DOI] [Google Scholar]
- 20.Munaf‘o MR, et al. 2017. A manifesto for reproducible science. Nat. Human Behav. 1, 1–9. ( 10.1038/s41562-016-0021) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Nosek BA. et al. 2015. Promoting an open research culture. Science 348, 1422–1425. ( 10.1126/science.aab2374) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22.Miguel E, et al. 2014. Promoting transparency in social science research. Science 343, 30–31. ( 10.1126/science.1245317) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23.Camerer CF, et al. 2016. Evaluating replicability of laboratory experiments in economics. Science 351, 1433–1436. ( 10.1126/science.aaf0918) [DOI] [PubMed] [Google Scholar]
- 24.Camerer CF, et al. 2018. Evaluating the replicability of social science experiments in nature and science between 2010 and 2015. Nat. Human Behav. 2, 637–644. ( 10.1038/s41562-018-0399-z) [DOI] [PubMed] [Google Scholar]
- 25.Ioannidis JPA, Fanelli D, Dunne DD, Goodman SN. 2015. Meta-research: evaluation and improvement of research methods and practices. PLoS Biol. 13, e1002264 ( 10.1371/journal.pbio.1002264) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Hardwicke TE, Serghiou S, Janiaud P, Danchev V, Crüwell S, Goodman S, Ioannidis JPA. 2019. Calibrating the scientific ecosystem through meta-research. Annu. Rev. Stat. Appl. ( 10.31222/osf.io/krb58) [DOI] [Google Scholar]
- 27.Nuijten MB, Borghuis J, Veldkamp CLS, Dominguez-Alvarez L, Assen MALMV, Wicherts JM. 2017. Journal data sharing policies and statistical reporting inconsistencies in psychology. Collabra: Psychol. 3, 31 ( 10.1525/collabra.102) [DOI] [Google Scholar]
- 28.Kidwell MC, et al. 2016. Badges to acknowledge open practices: a simple, low-cost, effective method for increasing transparency. PLoS Biol. 14, e1002456 ( 10.1371/journal.pbio.1002456) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Klavans R, Boyack KW. 2017. Research portfolio analysis and topic prominence. J. Informetrics 11, 1158–1174. ( 10.1016/j.joi.2017.10.002) [DOI] [Google Scholar]
- 30.Sison CP, Glaz J. 1995. Simultaneous confidence intervals and sample size determination for multinomial proportions. J. Am. Stat. Assoc. 90, 366–369. ( 10.1080/01621459.1995.10476521) [DOI] [Google Scholar]
- 31.Nosek BA, Ebersole CR, DeHaven AC, Mellor DT. 2018. The preregistration revolution. Proc. Natl Acad. Sci. USA 115, 2600–2606. ( 10.1073/pnas.1708274114) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 32.Kaiser J. 2018. Plan to replicate 50 high-impact cancer papers shrinks to just 18. Science ( 10.1126/science.aau9619) [DOI] [Google Scholar]
- 33.Herndon T, Ash M, Pollin R. 2013. Does high public debt consistently stifle economic growth? A critique of Reinhart and Rogoff. Camb. J. Econ. 38, 257–279. ( 10.1093/cje/bet075) [DOI] [Google Scholar]
- 34.Piwowar H, Priem J, Larivi‘ere V, Alperin JP, Matthias L, Norlander B, Farley A, West J, Haustein S. 2018. The state of OA: a large-scale analysis of the prevalence and impact of open access articles. PeerJ 6, e4375 ( 10.7717/peerj.4375) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35.Hardwicke TE, Ioannidis JPA. 2018. Populating the Data Ark: an attempt to retrieve, preserve, and liberate data from the most highly-cited psychology and psychiatry articles. PLoS ONE 13, e0201856 ( 10.1371/journal.pone.0201856) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 36.Vanpaemel W, Vermorgen M, Deriemaecker L, Storms G. 2015. Are we wasting a good crisis? The availability of psychological research data after the storm. Collabra 1, 1–5. ( 10.1525/collabra.13) [DOI] [Google Scholar]
- 37.Vines TH, et al. 2014. The availability of research data declines rapidly with article age. Curr. Biol. 24, 94–97. ( 10.1016/j.cub.2013.11.014) [DOI] [PubMed] [Google Scholar]
- 38.Eubank N. 2016. Lessons from a decade of replications at the quarterly journal of political science. PS: Political Sci. Politics 49, 273–276. ( 10.1017/s1049096516000196) [DOI] [Google Scholar]
- 39.Kerr NL. 1998. HARKing: hypothesizing after the results are known. Pers. Soc. Psychol. Rev. 2, 196–217. ( 10.1207/s15327957pspr0203_4) [DOI] [PubMed] [Google Scholar]
- 40.Franco A, Malhotra N, Simonovits G. 2014. Publication bias in the social sciences: unlocking the file drawer. Science 345, 1502–1505. ( 10.1126/science.1255484) [DOI] [PubMed] [Google Scholar]
- 41.John LK, Loewenstein G, Prelec D. 2012. Measuring the prevalence of questionable research practices with incentives for truth telling. Psychol. Sci. 23, 524–532. ( 10.1177/0956797611430953) [DOI] [PubMed] [Google Scholar]
- 42.Wagenmakers E-J, Wetzels R, Borsboom D, van der HLJ, Kievit RA. 2012. An agenda for purely confirmatory research. Perspect. Psychol. Sci. 7, 632–638. ( 10.1177/1745691612463078) [DOI] [PubMed] [Google Scholar]
- 43.Olken BA. 2015. Promises and perils of pre-analysis plans. J. Econ. Perspect. 29, 61–80. ( 10.1257/jep.29.3.61) [DOI] [Google Scholar]
- 44.Makel MC, Plucker JA, Hegarty B. 2012. Replications in psychology research. Perspect. Psychol. Sci. 7, 537–542. ( 10.1177/1745691612460688) [DOI] [PubMed] [Google Scholar]
- 45.Krawczyk M, Reuben E. 2012. (Un)available upon request: field experiment on researchers' willingness to share supplementary materials. Account. Res. 19, 175–186. ( 10.1080/08989621.2012.678688) [DOI] [PubMed] [Google Scholar]
- 46.Meyer MN. 2017. Practical tips for ethical data sharing. Adv. Methods Pract. Psychol. Sci. 1, 131–144. ( 10.1177/2515245917747656) [DOI] [Google Scholar]
- 47.Morey RD, et al. 2016. The Peer Reviewers’ Openness Initiative: incentivizing open research practices through peer review. R. Soc. open sci. 3, 150547 ( 10.1098/rsos.150547) [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
Raw data are the core evidence that underlies scientific claims. However, the vast majority of the 156 relevant articles we examined did not contain a data availability statement (figure 1d). Eighteen articles (12% [6% to 18%]) mentioned using an external data source,1 but did not provide detailed instructions on how to obtain the specific data that was used. A further 11 (7% [2% to 13%]) datasets were reportedly available; however, eight of these were in fact not accessible due to non-functioning links to personal/institutional web pages (n = 4), a non-functioning link to journal electronic supplementary material (n = 1), only being ‘available upon request’ from authors (n = 2; note that we did not attempt to contact them), or involving a file that could not be opened because commercial software was required (n = 1). Of the three accessible datasets (two available via electronic supplementary material and one available via a personal/institutional web page) two were both complete and clearly documented and one was neither complete nor clearly documented.
All data (https://osf.io/u9fw8/), materials (https://osf.io/z9qtr/) and analysis scripts (https://osf.io/sbrez/) related to this study are publicly available. To facilitate reproducibility this manuscript was written by interleaving regular prose and analysis code and is available in a Code Ocean container (https://doi.org/10.24433/CO.2749769.v4) which re-creates the software environment in which the original analyses were performed.