Skip to main content
PLOS Biology logoLink to PLOS Biology
. 2024 Nov 5;22(11):e3002870. doi: 10.1371/journal.pbio.3002870

Biomedical researchers’ perspectives on the reproducibility of research

Kelly D Cobey 1,2,*, Sanam Ebrahimzadeh 3, Matthew J Page 4, Robert T Thibault 5,6, Phi-Yen Nguyen 4, Farah Abu-Dalfa 1,7, David Moher 2,3
Editor: Cilene Lino de Oliveira8
PMCID: PMC11537370  PMID: 39499707

Abstract

We conducted an international cross-sectional survey of biomedical researchers’ perspectives on the reproducibility of research. This study builds on a widely cited 2016 survey on reproducibility and provides a biomedical-specific and contemporary perspective on reproducibility. To sample the community, we randomly selected 400 journals indexed in MEDLINE, from which we extracted the author names and emails from all articles published between October 1, 2020 and October 1, 2021. We invited participants to complete an anonymous online survey which collected basic demographic information, perceptions about a reproducibility crisis, perceived causes of irreproducibility of research results, experience conducting reproducibility studies, and knowledge of funding and training for research on reproducibility. A total of 1,924 participants accessed our survey, of which 1,630 provided useable responses (response rate 7% of 23,234). Key findings include that 72% of participants agreed there was a reproducibility crisis in biomedicine, with 27% of participants indicating the crisis was “significant.” The leading perceived cause of irreproducibility was a “pressure to publish” with 62% of participants indicating it “always” or “very often” contributes. About half of the participants (54%) had run a replication of their own previously published study while slightly more (57%) had run a replication of another researcher’s study. Just 16% of participants indicated their institution had established procedures to enhance the reproducibility of biomedical research and 67% felt their institution valued new research over replication studies. Participants also reported few opportunities to obtain funding to attempt to reproduce a study and 83% perceived it would be harder to do so than to get funding to do a novel study. Our results may be used to guide training and interventions to improve research reproducibility and to monitor rates of reproducibility over time. The findings are also relevant to policy makers and academic leadership looking to create incentives and research cultures that support reproducibility and value research quality.


There is growing interest in the reproducibility of research and ways to enhance transparency. This study uses an international survey of more than 1900 biomedical researchers to reveal the perceived causes of irreproducibility of research results, experience of conducting replication studies, and knowledge of relevant funding and training.

Introduction

There is growing interest in both the reproducibility of research and ways to enhance research transparency [14]. Terminology around reproducibility varies [5]; here, we define reproducibility as re-doing a study using similar methods and obtaining findings consistent with the original study and as irreproducible when the findings are not consistent with the original study. This definition allows for variation in methods (e.g., conceptual and direct replications) between the original study and the reproducibility study as well as different definitions of how “consistent results” are defined (i.e., using p-value, observing results in the same direction, comparing effect sizes). Reproducibility of research is core to maintaining research trustworthiness and fostering translation and progressive discovery. Despite the seemingly critical role of reproducibility of research and growing discussions surrounding reproducibility, the reality is that most studies, including pivotal studies within several disciplines, have never been formally subjected to a reproducibility effort. For example, in education research, an analysis of publications in the field’s top 100 journals showed that just 0.13% publications (221 out of 164,589) described reproducibility projects [6]. In psychology, a study examining 250 articles published between 2014 and 2017 found that 5% described a reproducibility effort [7], while a similar study examining reproducibility in social sciences found that just 1% of articles sampled described a focused on reproducing previous results [8]. Across disciplines, our knowledge about the proportion of studies that are reproducible tends to be dominated by a small number of large-scale reproducibility projects. In psychology, for example, a study which estimated the replicability of 100 foundational studies in top journals in the field reported that only 36% had statistically significant results (one measure of reproducibility), compared to 97% of the original studies [9].

In 2016, Nature reported on a survey of more than 1,500 researchers about their perceptions of reproducibility. They found that 83% agreed there was a reproducibility crisis in science, with 52% indicating that they felt the crisis was “significant” [10]. Survey studies like this play a powerful role in elucidating the determinants of reproducibility. Such information is essential to identify gaps in factors including training, research support, and incentives to ensure reproducible research. Given the global nature of research, capturing global perspectives as was done in the Nature survey is crucial to obtaining broad understanding of issues across the research ecosystem.

In this study, we aim to build on the 2016 Nature survey on reproducibility by surveying researchers in the biomedical community specifically. There is immediate importance to ensuring biomedical research is reproducible: here, studies that were subsequently not reproducible have led to patient harms [11,12]. By capturing a diverse and global group of biomedical researchers’ perceptions of reproducibility within the field we hope to better understand how to ensure reproducibility in biomedicine. While our work is inspired by the 2016 Nature survey, this is not a direct effort to reproduce that study: we sample a different community and use a sampling approach that differs from the original study which limits the ability for direct comparison. Our specific objectives were to: (1) explore biomedical researchers’ perceptions of reproducibility and their perceptions of causes of irreproducibility; and (2) describe biomedical researchers’ experiences conducting and publishing reproducibility projects. The study is descriptive, so we had no formal hypotheses. Understanding researcher perceptions is a key starting point to drive culture change and to create training interventions to drive improvement.

Methods

Open science statement

This study received ethics approval from the Ottawa Health Sciences Research Ethics Board (20210856-01H). This study protocol was posted a priori, and data and materials have been made available: https://osf.io/3ksvz/ [13].

Study design

We conducted an online cross-sectional closed survey of researchers who published a paper in a journal indexed in MEDLINE (RRID:SCR_002185). The survey was anonymous.

Sampling framework

We downloaded the MEDLINE database of journals. From this list of approximately 30,000 journals, we selected a random sample of 400 journals using the RAND() function in Excel (RRID:SCR_016137). We then extracted author names and emails from all articles published in those journals between October 1, 2020 and October 1, 2021. We included all authors whose names and emails were available and all article types/study designs. For full details on our semi-automated approach to extracting author emails, please see our search strategy in S1 File.

Participant recruitment

The survey was sent only to those researchers identified via our sampling procedure (i.e., a closed survey). Potential participants received an email containing a recruitment script which detailed the purpose of the study and invited them to review our informed consent form and complete our anonymous online survey. Participation in the survey served as implied consent; we did not require signed consent to maintain anonymity. To send emails to the sample of authors, we used mail merge feature in Microsoft 365. This tool allows for the personalization of emails without having to individually customize and send each out. In the case of non-response, we sent 3 reminder emails to potential participants at weekly intervals after the initial invitation. We closed the survey 4 weeks after the initial invitation. We did not provide any specific incentive to complete the survey.

Survey

The full survey is available in S2 File. The survey was administered using SurveyMonkey and could be completed in about 10 min. The survey contained a total of 19 questions starting with 4 demographic questions about the participants, including their gender, research role, research area, and country of residence. Participants were then asked to complete questions about their perceptions of reproducibility in biomedicine, questions about their experience with reproducibility, and questions about perceptions of barriers and facilitators to conducting reproducibility projects. The survey questions were presented sequentially and using adaptive formatting to present only certain items based on the participant’s response to previous questions. Most questions were multiple choice, with 2 questions asking participants to expand on their responses using a free-text box. The survey was purpose-built for the study by the research team, building directly off the previously published Nature reproducibility survey [10]. We included several of the previous study’s questions directly in this study, modified some slightly, and made some more specific to the biomedical research setting. We also introduced some novel questions on reproducibility. The survey was pilot tested by 3 researchers (not on the team) to ensure clarity and acceptability of format and we edited the survey to address their feedback regarding the clarity of survey questions and issues with illogical survey flow. Participants were able to skip any question.

Data management and analysis

Data were exported from SurveyMonkey and analyzed using SPSS 28 (RRID:SCR_002865). We report descriptive statistics including count and percentages for all quantitative items. For the qualitative items, we conducted a thematic content analysis. To do so, 2 researchers individually read all text-based responses and assigned a code to summarize the content of the text. This inductive coding approach involves creating codes based on the data itself, rather than from a pre-established coding framework. Codes were refined iteratively upon exposure to each text-based response read. After discussion to reach consensus on the codes used, we then grouped the agreed codes into themes for reporting in tables. Coders were members of the project team who were not blinded to the study aims.

Results

Protocol amendments

In our original protocol, we said we would take a random sample of 1,000 journals from MEDLINE and extract information from the first 20 authors. This approach required extensive manual extraction, so we opted to restrict our random sample to 400 journals and semi-automate extraction of author information for an entire year’s worth of publications. Our revised method meant that we obtained and used listed emails from all authors on an identified paper (i.e., we were not restricted to corresponding authors).

Demographics

A total of 24,614 emails were sent, but bounce backs were received from 1380, meaning 23,234 emails were sent successfully to potential participants. A total of 1,924 participants accessed our survey, of whom 1,630 participants provided completed responses (response rate 7%; this frequency is slightly lower than the estimated 1,800 responses reported in our protocol). Most participants were Faculty Members/Primary Investigators (N = 1,151, 72%) and more than half of participants were male (N = 943, 59%). Respondents were from more than 80 countries, with the USA (N = 450, 28%) having the highest representation. About half of participants reported working in clinical research (N = 819, 50%). Further demographic details by role, gender, country, and research area are provided in Table 1.

Table 1. Participant demographics.

Item Response options N %
Researcher role Graduate student 88 6
Postdoctoral fellow 129 8
Faculty member/PI 1,151 72
Research support staff 54 3
Scientist in industry 28 2
Scientist in third sector 27 2
Government scientist 54 3
Other 73 5
Missing data 26 -
Gender Female 643 40
Male 943 59
Non-binary 3 0.2
Prefer to self-describe 1 0.1
Prefer not to say 13 1
Missing data 27 -
Country of Employment (Top 3) USA 450 28
Canada 128 8
UK 105 7
Missing data 32 -
Research Area Clinical research 819 50
Preclinical research–in vivo 191 12
Preclinical research–in vitro 163 10
Health systems research 147 9
Methods research 81 5
Other, please specify 227 14
Missing data 2 -

Perceptions of reproducibility

When asked whether there was a reproducibility crisis in biomedicine most researchers agreed (N = 1,168, 72%), with 27% (N = 438) indicating the crisis was significant and 45% (N = 703) indicating a slight crisis (we note that a “slight crisis” is a bit of an oxymoron, but retained the wording from the original Nature survey for comparison purposes); see Fig 1 and S3 File for breakdown by discipline. Compared to the previously published Nature study (N = 819, 52%), fewer participants in our study felt there was a “significant reproducibility crisis” (N = 438, 27%). This difference was even larger when we restricted to the Nature study participants who indicated they worked in medicine (N = 203, 60%).

Fig 1. Participant perceptions of a reproducibility crisis.

Fig 1

Data is presented overall for all participants in the current study and is broken down by research focus area in medicine. Results are presented in context to the overall Nature study findings and specifically to participants from this study indicating they worked in medicine. The underlying data for this figure can be found at https://osf.io/dbh2a.

Participants were then asked what percentage of papers in each of biomedical research overall, clinical biomedical research, in vivo biomedical research, and in vitro biomedical research they thought were reproducible, see Fig 2. Only 5% (N = 77) thought more than 80% of biomedical research was reproducible. See S3 File for complete results. We provide a breakdown of responses between genders, between researchers in different biomedical research areas, and by career rank in S4 File.

Fig 2. Participants perceptions of the proportion of papers they think are reproducible in biomedicine overall and by biomedical research area.

Fig 2

The underlying data for this figure can be found at https://osf.io/dbh2a.

Determinants of irreproducibility

When presented with various potential causes of irreproducibility, more than half of participants responded that each presented factor contributes to irreproducibility. The top characteristic participants noted as “always contributing” to irreproducibility was pressure to publish (N = 300, 19%). Factors deemed least likely to contribute to irreproducibility were fraud (N = 320, 20%) and bad luck (N = 568, 36%). See Table 2 for complete results.

Table 2. Participant perceptions of the causes of irreproducibility.

  N(%) 
Always contributes Very often Contributes Sometimes Contributes Does not Contribute Unsure Missing data
Selective reporting of the published literature 131 (8) 638 (40) 714 (45) 43 (3) 73 (5) 31
Selective publication of entire studies 182 (11) 698 (44) 577 (36) 71 (4) 71 (4) 31
Pressure to publish 300 (19) 693 (43) 473 (30) 75 (5) 57 (4) 32
Low statistical power 185 (12) 706 (44) 579 (36) 76 (5) 48 (3) 36
Poor statistical analysis 197 (12) 615 (38) 649 (41) 99 (6) 44 (3) 26
Not enough internal replication (E.g., by the original lab/authors) 132 (8) 539 (34) 697 (44) 93 (6) 142 (9) 27
Insufficient study oversight 86 (5) 376 (24) 799 (50) 194 (12) 143 (9) 32
Lack of training in reproducibility 153 (10) 522 (33) 622 (39) 168 (11) 135 (8) 30
Failure to make materials openly available 141 (9) 449 (28) 722 (45) 191 (12) 99 (6) 28
Failure to make original study data openly available 137 (9) 476 (30) 685 (43) 205 (13) 94 (6) 33
Poor study design 208 (13) 584 (36) 678 (42) 96 (6) 38 (2) 26
Fraud 185 (12) 120 (8) 624 (40) 320 (20) 330 (21) 51
Poor quality peer review 140 (9) 437 (27) 755 (47) 192 (13) 72 (5) 34
Problems in the design of replication studies 103 (6) 406 (25) 809 (51) 162 (10) 123 (8) 27
Technical expertise required for replication 96 (6) 429 (27) 743 (46) 190 (12) 144 (9) 28
Variability of standard reagents 82 (5) 288 (18) 617 (39) 229 (14) 380 (24) 34
Bad luck 23 (1) 70 (4) 461 (29) 568 (36) 466 (29) 42

A total of 97 (6%) participants provided a written response to elaborate on what they perceived were causes of irreproducibility. Responses were coded into 16 unique codes and then thematically grouped into 7 categories: ethics, research methods, statistical issues, incentives, issues with journal and peer review, lack of resources, and other. For definitions and illustrative examples of the codes, see Table 3.

Table 3. Thematic analysis of perceived causes of irreproducibility.

Themes Codes N (97) % Example
Ethics Conflicts of interest 3 3 “Conflicts of interest, commercial interests, corporate interests”
Fraud 2 2 “Sure sometimes there is fraud or poor study design or execution…”
Research methods Complex research design or methods 2 2 “specialized or cutting edge techniques not adopted or fully appreciated by enough other labs”
Heterogeneity in biology/environment 25 26 “heterogeneity of included subjects”
Lack of standard methods 6 6 “lack of precise outcome measures for clinical studies”
Poor study design or planning 8 8 “Poor design of original studies with increased Type 1 error due to multiple comparisons/endpoints”
Statistical issues Discretion in statistical analysis 5 5 “Investigators conducting their own analyses”
Overreliance on statistics 3 3 “over interpretation of statistics—0.05 p value without thinking enough about methods behind it and meaning”
sample size/power issues 4 4 “Small effects of biomedical phenomena”
Incentives Lack of value for reproducibility studies 4 4 “no incentive to reproduce studies”
Preference for novelty 10 10 “…They are so fixated on novelty they absolutely discourage replication and/or the publishing alternative findings.”
Pressure to publish 2 2 “I think it is the pressure to publish..”
Researcher attitudes 2 2 “Preconceptions of investigators and reviewers..”
Issues with journals and peer review Issues with journals and peer review 22 23 “Journals not accepting replication studies.”
Lack of resources Lack of resources 14 14 “Almost no funding for replication studies”
Other Other 6 6 “Desire to have a convincing story, results”

Experiences with reproducibility

Participants were asked about their experience conducting reproducibility projects. Nearly a quarter of participants indicated that they had previously tried to replicate one of their own published studies and failed to do so (N = 373, 23%), whereas 31% (N = 501) indicated all such replications they had conducted had yielded the same result as the original, and slightly less than half of participants indicated that they had never tried to replicate any of their own published work (N = 734, 46%). Among the 874 participants who indicated they had tried to replicate one of their own studies, when asked to consider their most recent replication effort, 313 (36%) indicated they had published the results.

Almost half of the participants indicated that they had previously tried to replicate a published study conducted by another team and failed to do so (N = 724, 47%), whereas 10% (N = 156) indicated all replications that they had attempted were successful, while 43% (N = 666) indicated they had never tried to replicate someone else’s published research. Among those who had published their replication study, when asked to consider their most recent replication effort, 29% (N = 224) indicated it took about the same amount of time to publish as a typical non-replication paper. A quarter (N = 189, 25%) of participants who had attempted to replicate others’ research indicated they had no plans to publish their replication study. Eighty-five percent of participants (N = 1,316) indicated they had never been contacted by another researcher who was unable to reproduce a finding they previously published. See Table 4 for complete results.

Table 4. Participant experiences with reproducibility.

Item Response options N %
Have you ever tried to replicate a published study you previously conducted and failed? Yes 373 23
No- all replications I have completed of my own research have been successful 501 31
No – I have never tried to replicate my own research 734 46
Missing data 22 -
Did you publish your replication study results of your study? Yes – but it took longer to publish than other papers you’ve published that were not replications 139 17
Yes – and it took about the same amount of time to publish as other papers you’ve published that were not replications 152 19
Yes – but it was quicker to publish than other papers you’ve published that were not replications 22 3
No – I have submitted but not yet had the work accepted 29 4
No – I have not yet submitted, but intend to do so 84 10
No – I don’t intend to attempt to publish this study
No – Journals don’t appear interested in publishing replications
205

117
25

14
Other 75 91
Missing data 51 -
Have you ever tried to replicate a published study conducted by another team of authors and failed? Yes 724 47
No- all replications I have completed have been successful 156 10

No – I have never tried to replicate someone else’s published research

Missing
666

84
43

-
Did you publish your replication study results of the other researchers’ study? Yes – but it took longer to publish than other papers you’ve published that were not replications 139 18

Yes – and it took about the same amount of time to publish as other papers you’ve published that were not replications
224 29

Yes – but it was quicker to publish than other papers you’ve published that were not replications
14 2
No – I have submitted but not yet had the work accepted 21 3
No – I have not yet submitted, but intend to do so 79 10
No – I don’t intend to attempt to publish this study 189 25
Don’t appear interested in publishing replications 56 7
Other 43 6
Missing data 12 -
Have you ever been contacted by another researcher who was unable to reproduce a finding you published? Yes 164 11

No

I can’t remember
1316
53
85
3

Unsure
16 1

Missing data
81 -

A total of 724 participants responded to the item about why they replicated their own research, of which 675 (93%) provided relevant text-based responses. Responses were coded into 17 unique codes and then grouped into 7 themes: training, ensuring reproducibility, additional research, addressing concerns, joining/setting up a new lab, for publication or due to peer review, and other. For illustrative examples of the codes, see Table 5.

Table 5. Thematic analysis of reasons why participants replicated their own study.

Theme Code N % Example
Training Education purposes 15 2 “Used it for teaching purposes”
New students/staff replicate former results 27 4 “To teach new batch of PhD students and ask them replicate senior students experiments”
Ensuring reproducibility Could not reproduce a finding so studied why 15 2 “Students reported difficulty replicating original methods”
Findings were challenged 9 1 “Others have reported opposite findings so we wanted to verify our results.”
Interesting 8 1 “I was curious.”
Replication is part of research approach norms 22 3 “As a clinical researcher, we repeat some evaluations over time to detect modified trends.”
To confirm/validate the finding 309 46 “To check if finding were comparable over time”
Public/community value 6 1 “This is very important for public acceptability.”
Additional research To use as controls 33 5 “As controls in a follow-up study.”
Novel research 17 3 “Test new ideas”
Extension of study 232 34 “Replication and extension.”
Address potential concerns with original study To address limitations of the original study 30 4 “Confirmation with larger sample size after pilot proof of concept study”
To address heterogeneity of biology/environment 35 5 “Search for minor variations in population and results”
Improving quality 11 2 “Improve the research”
Joining/setting up a new lab Joining/setting up a new lab 8 1 “Started independent lab”
For publication or due to peer review For publication or due to peer review 11 2 “Journals request it”
Other Other 30 4 “Scientific ethics”

A total of 748 participants responded to the item about why they replicated someone else’s research, of whom 700 (94%) provided relevant text-based response data. Responses were coded into 19 unique codes and then grouped into 7 themes: trustworthiness, extending and improving research, application to new setting, new research, interest, training, and other. For illustrative examples of the codes, see Table 6.

Table 6. Thematic analysis of reasons why participants replicated someone else’s study.

Theme Definition Codes N (874) %
Trustworthiness Expressing doubt and seeking to verify the trustworthiness of the original results To verify the results 158 18
Wary of the original result 59 7
Obtained conflicting results 52 6
To follow-up/challenge findings 12 1
Extending and improving research Intending to extend and improve the original study To extend current research 129 15
Ability to extend study with improved methods 48 6
Interested in the same question 26 3
To provide a more current replication 16 2
Already running the same study 11 1
Collaborating with original team 10 1
Application to new setting Intending to apply the original study in a broader or new setting Replication in a new setting/population 116 13
To determine generalizability 8 1
New research Utilizing the original study and/or studies method in new projects Wanted to use the new studies method 73 8
Reproduced research for new projects 30 3
To use as baseline or control data 23 3
Interest Replicating the original study out of personal interest and/or curiosity Out of interest/curiosity 34 4
Training Replicating the original study for the purpose of understanding and gaining new knowledge To understand the methods better 17 2
For educational purposes/to gain new knowledge 16 2
Other Other Other 36 4

Support for initiatives to enhance reproducibility

Few participants reported that their research institution has established procedures to enhance reproducibility of biomedical research (N = 254, 16%), and almost half reported that their institutions did not provide training on how to enhance the reproducibility of research (N = 731, 48%) with an additional 503 (33%) reporting they were unsure of whether such training existed. We asked participants to provide information and links to relevant reproducibility training at their institution, which resulted in information or (functioning) links to 24 unique training resources (see 5). Among these 24 sources, just 9 (38%) clearly described specific openly available (e.g., no paywall) training related to reproducibility. Most researchers responded that they perceived their institution would value them doing new biomedical research studies more than replication studies (N = 1031, 67%). The majority also indicated that it would be harder to find funding to conduct a replication study than a new study (N = 1,258, 83%), with just 7% (N = 112) indicating they were aware of funders providing specific calls for conducting reproducibility related research. For full results, see Table 7.

Table 7. Participants perceived support for reproducibility.

Item Response options N %
Does your research institution have established procedures to enhance reproducibility of biomedical research? Yes 254 16
No 655 42
I do not know 637 41
Missing data 84 -
My institution would value me doing new biomedical studies more than me doing replication studies. True 1,031 67
False 147 10
I do not know 351 23
Missing data 101 -
In my biomedical research setting, it would be harder to find funding to conduct a replication study than it would be to find funding for a new study. True 1,258 83
False 72 5
Unsure 194 13
Missing data 106 -
Are you aware of funders providing specific calls for conducting reproducibility related research? Yes 112 7
No 1,417 93
Missing data 101 -
Does your research institution provide training on how to enhance the reproducibility of research? Yes, and I have taken it 184 12
Yes, but I have not taken it 102 7
No 731 48
Unsure 503 33
Missing data 110 -

We asked participants to indicate how much they agreed with the statement “I feel I am more concerned about reproducibility than the average researcher in my field and at my career stage” as a way to indirectly address potential bias in self-selection to complete the survey. Participants responded on a 5-point scale with endpoints, strongly disagree (1) and strongly agree (5). Participants reported a mean response of 3.2 (N = 1,402, SD = 0.89) that corresponds to the mid-point of the scale “neither agree nor disagree.”

Discussion

We report the results of an international survey examining perceptions of reproducibility. Almost 3 quarters of participants reported that they felt there was a reproducibility crisis in biomedicine. The concern appears to apply to biomedicine overall, but also specifically to clinical research, in vivo research, and in vitro research (11% or fewer participants indicated that they think more than 80% of papers in each category were reproducible). Researchers agreed that a variety of factors contribute to irreproducibility; however, the chief characteristic that most participants indicated “always contributes” to irreproducible research was a pressure to publish. Concerns about how the current system of academic rewards stresses quantity over quality have been expressed for decades [1416]—a sentiment supported by this study’s data, which suggests that researchers’ performance is negatively impacted, in terms of producing reproducible research, by what the academic system incentivizes.

More than half of participants reported having tried to replicate their own work previously, with almost a quarter indicating that when they did so they failed, and many indicating that they do not intend to publish their findings. Similar findings were reported when asked about whether participants had tried to replicate another researcher’s study, with 57% indicating they had done so, and 47% indicating the replication failed. The majority of participants had not been contacted by another researcher who was unable to reproduce their findings, which suggests that teams of researchers attempting to reproduce studies do not typically communicate despite the potential value for this contact to enhance reproducibility [9,17].

We asked several items about researchers’ perceptions of their institution’s support for reproducibility (Table 7) and the findings collectively suggest gaps in incentives and support to pursue reproducibility projects. For example, with just 16% of respondents reporting awareness that their institution has established procedures to enhance reproducibility, and 67% of researchers perceiving their institution values novelty over replication, our results suggest that overall, researchers perceive that institutions are not doing their part to effectively support and reward research reproducibility [18]. The growth of “Reproducibility Networks,” national peer-led consortiums aiming to promote reproducibility and understand factors related to irreproducibility, are a promising opportunity to rectify this situation [19]. For example, the UK Reproducibility Network (UKRN) provides an opportunity for institutions to formally commit to reproducibility by joining the network and requires a formal role within the senior management team of each member institution to support the delivery of network activities with the institution. The UKRN boasts a range of events and shared resources to support reproducibility [20]. The structure of reproducibility networks allows for harmonization but also flexibility in approach. Obtaining a sustained funding mechanism will be critical to their growth and ongoing success in terms of their long-term impact.

This study built off of an earlier study of more than 1,500 researchers surveyed by Nature about reproducibility [10]. The current study differs in several important ways. Firstly, the focus is exclusively on biomedicine, since to our knowledge no large scale and representative survey of biomedical researchers has been conducted to date. Indeed, just 203 (13%) of the 1,576 researchers who completed the original study indicated “medicine” as their main area of interest. Secondly, we randomly sampled researchers from publication lists, meaning we are able to report a response rate. This was not possible in the Nature survey, which was emailed to Nature readers and advertised on affiliated websites and social media outlets, meaning that the number of individuals encountering the survey is unknown. While it is possible that among those invited to take part there is bias among participants who choose to complete the survey, our approach has been chosen to help minimize surveying those explicitly active in reproducibility projects or related initiatives. Indeed, the finding that overall participants report not to differ from their belief of their peers regarding their level of concern about reproducibility provides some assurance that our sampling strategy was effective. We also find there is not much difference between different groups’ responses (S4 File).

We acknowledge several limitations in our approach. Firstly, our study survey was purpose-built, and while drawing on the Nature study, was not designed according to a particular theoretical framework. This type of approach lends itself well to the exploratory and descriptive goals of the current study; however, it makes it harder to test hypotheses or measure changes over time. A validated scale to assess perceptions of reproducibility and its cause would be of value to the community for future research. We did not look at correlations between responses to items on our survey, as doing so would be a post hoc change from our protocol. Future research using a validated scale may be better positioned to consider potential relationships between responses a priori (i.e., how one’s perception of a reproducibility crisis relates to the perceived proportion of research that is reproducible). Future surveys would also benefit from a more extensive approach to piloting for clarity and readability. Another potential limitation is that the coders conducting the thematic analysis were not blind to the study’s objective (i.e., were members of the research team). While the study is descriptive it remains possible that knowledge of the study aim and expertise in the reproducibility literature biased the coding approach. Moreover, only a small number of participants (6%) provided text-based descriptions to nuance their survey responses. It is perhaps not surprising given that the survey offered no incentive to take part, but it raises the possibility that those that took the time to respond to text-based items may differ from those who did not. Indeed, the top 3 most represented countries in our sample overall were the US, Canada, and the UK. While we had over 80 countries represented, our responses tended to be Western-centric which may present another bias, even in these countries tend to be among the most productive producers of research. Future research delving more specifically into recruiting from a specific country, or set of countries, could help improve our understanding of how national contexts may lead to differences in perceptions and practices related to reproducibility. Finally, our sampling approach allowed us to obtain any listed emails on a given study article. This means that more than one of the authors of a given article in our random sample may have responded to the survey. While our random sampling approach provided a strong basis for sampling, this nuance presents the possibility for some bias as authors of different studies may vary more considerably and we have no way of knowing how survey participants were represented within or between given papers. While it may be tempting to conduct comparison testing of our results compared to the earlier Nature replication survey, we are not positioned to do so. Our study did not aim to replicate the original Nature paper and was registered as a descriptive study. Given that we sampled exclusively biomedical researchers, and that our recruitment approach for sampling differed significantly, it would be very difficult to disentangle whether differences in results reflect sampling bias in the original study, temporal changes in the research ecosystem over time, different perceptions in biomedicine compared to research more broadly, or a combination of these factors. We know of no relevant framework that would have been appropriate to implement in order to test specific hypotheses within the biomedical community. Further, null hypothesis testing is not appropriate for determining beliefs and when hypothesis testing is used you need a formal sample size calculation [21,22].

Concerns about reproducibility are being widely recognized within research but also more broadly in the media [23]. The COVID-19 pandemic has shifted our thinking about research transparency [24] and highlighted issues with fraud in research, poor-quality reporting, and poor-quality study design [25,26], all factors that can contribute to irreproducible research. As stakeholders work to introduce training and interventions to enhance reproducibility, it will be critical to monitor the effectiveness of these interventions. This international survey provides a contemporary cross-section of the biomedical community. While our survey approach and direction of findings are consistent with the previous Nature study, ongoing monitoring of perceptions of reproducibility in the community is critical to gauge shifts over time. Indeed, conducting this same survey again in the future would allow for a temporal comparison on how perceptions and behaviors shift over time. This suggestion aligns with the approach of the EU Commission who in their report “Assessing the reproducibility of research results in EU Framework Programmes for Research” [27], describe an approach to understand and monitor the progress of reproducibility over time. In the work, researchers were also asked about their perceptions of reproducibility and 60% agreed that there is a reproducibility crisis, but also overwhelmingly agreed (82%) that reproducibility is “important” or “very important” in their discipline. Like our findings, participants indicated several key challenges to achieve reproducibility, with issues around cultural factors being paramount. Our finding that a pressure to publish was most likely to be rated as “always contributing” to irreproducibility is consistent with the conclusions of this report. Collectively, the outcomes of this body of evidence highlight perceived causes and constraints to producing reproducible research which should be prioritized within the biomedical community and targeted with interventions and supports to create improvements over time. An important consideration of future survey or reproducibility monitoring will be to be explicit about the language and definitions of reproducibility used—terminology in this space is complex and the same words can reflect different concepts between researchers within and between research areas. Clear definitions for reproducibility will help to ensure that differences observed in survey responses over time, or between groups, reflect more than simply differences in how surveys are interpreted.

Supporting information

S1 File. Search strategy used to obtain a random sample of journals to extract author contact information from to identify possible participants.

(DOCX)

pbio.3002870.s001.docx (17KB, docx)
S2 File. Survey administered to participants.

The survey was administered online and used survey logic to present relevant items.

(DOCX)

pbio.3002870.s002.docx (20.4KB, docx)
S3 File. Supplementary tables presenting a comparison of perceptions of reproducibility between our findings and the original Nature paper by Baker, presented overall and by discipline.

In addition, we present participant perceptions of reproducibility in different research areas.

(DOCX)

pbio.3002870.s003.docx (18.5KB, docx)
S4 File. Additional analyses of survey responses by gender and academic role.

(DOCX)

pbio.3002870.s004.docx (19.7KB, docx)
S5 File. List of training available on reproducibility mentioned by survey participants.

(DOCX)

pbio.3002870.s005.docx (19.4KB, docx)

Data Availability

All data and study materials are available on the Open Science Framework (https://osf.io/3ksvz/) or in the supplementary materials.

Funding Statement

The author(s) received no specific funding for this work.

References

  • 1.Buck S. Solving reproducibility. Science. 2015;348:6242. doi: 10.1126/science.aac8041 [DOI] [PubMed] [Google Scholar]
  • 2.Munafò MR, Nosek BA, Bishop DVM, et al. PERSPECTIVE A manifesto for reproducible science. Nat Hum Behav. 2017;1(January):1–9. doi: 10.1038/s41562-016-0021 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Collins FS, Tabak LA. NIH plans to enhance reproducibility. Nature. 2012;505(7485):612–613. doi: 10.1038/505612a [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Begley CG, Ioannidis JPA. Reproducibility in science: Improving the standard for basic and preclinical research. Circ Res. 2015;116(1):116–126. doi: 10.1161/CIRCRESAHA.114.303819 [DOI] [PubMed] [Google Scholar]
  • 5.Goodman SN, Fanelli D, Ioannidis JPA. What does research reproducibility mean? 2016;8(341). doi: 10.1126/scitranslmed.aaf5027 [DOI] [PubMed] [Google Scholar]
  • 6.Makel MC, Plucker JA. Facts Are More Important Than Novelty: Replication in the Education Sciences. Educ Res. 2014;43:304–316. doi: 10.3102/0013189X14545513 [DOI] [Google Scholar]
  • 7.Hardwicke TE, Thibault RT, Kosie JE, Wallach JD, Kidwell MC, Ioannidis JPA. Estimating the Prevalence of Transparency and Reproducibility-Related Research Practices in Psychology (2014–2017). Perspect Psychol Sci. Published online March 8, 2021:1745691620979806. doi: 10.1177/1745691620979806 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Hardwicke TE, Wallach JD, Kidwell MC, Bendixen T, Crüwell S, Ioannidis JPA. An empirical assessment of transparency and reproducibility-related research practices in the social sciences (2014–2017). R Soc Open Sci. 2020;7(2):190806. doi: 10.1098/rsos.190806 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Chang AC, Li P. Is Economics Research Replicable? Sixty Published Papers from Thirteen Journals Say”Usually Not”. 2015. [Google Scholar]
  • 10.Baker M. Is there a reproducibility crisis? Nature. 2016;533:452–454. doi: 10.1038/533452a [DOI] [PubMed] [Google Scholar]
  • 11.Le Noury J, Nardo JM, Healy D, et al. Restoring Study 329: Eficacy and harms of paroxetine and imipramine in treatment of major depression in adolescence. BMJ. 2015:351. doi: 10.1136/bmj.h4320 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Ending Medical Reversal | Johns Hopkins University Press Books. Accessed December 7, 2021. Available from: https://jhupbooks.press.jhu.edu/title/ending-medical-reversal. [Google Scholar]
  • 13.OSF | Reproducibility in biomedicine: A cross-sectional survey. Accessed June 27, 2023. Available from: https://osf.io/3ksvz/. [Google Scholar]
  • 14.Fanelli D. Do pressures to publish increase scientists’ bias? An empirical support from US states data. PLoS ONE. 2010;5(4). doi: 10.1371/journal.pone.0010271 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Garfield E. What is the primordial reference for the phrase “publish or perish”? Forensic Sci. 1996;10(12):11. [Google Scholar]
  • 16.Home. DORA. August 1, 2023. Accessed March 1, 2024. Available from: https://sfdora.org/.
  • 17.Nosek BA, Errington TM. What is replication? PLoS Biol. 2020;18(3):e3000691. doi: 10.1371/journal.pbio.3000691 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Begley CG, Buchan AM, Dirnagl U. Robust research: Institutions must do their part for reproducibility. Nature. 2015;525(7567):25–27. doi: 10.1038/525025a [DOI] [PubMed] [Google Scholar]
  • 19.From grassroots to global: A blueprint for building a reproducibility network | PLOS Biology. Accessed May 25, 2023. Available from: https://journals.plos.org/plosbiology/article?id=10.1371/journal.pbio.3001461. doi: 10.1371/journal.pbio.3001461 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.UK Reproducibility Network. Accessed June 27, 2024. Available from: https://www.ukrn.org/.
  • 21.Frick RW. The appropriate use of null hypothesis testing. Psychol Methods. 1996;1(4):379–390. doi: 10.1037/1082-989X.1.4.379 [DOI] [Google Scholar]
  • 22.Hernán MA, Greenland S. Why Stating Hypotheses in Grant Applications Is Unnecessary. JAMA. 2024;331(4):285–286. doi: 10.1001/jama.2023.27163 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Most scientists “can’t replicate studies by their peers.” BBC News. Available from: https://www.bbc.com/news/science-environment-39054778. February 22, 2017. Accessed May 25, 2023. [Google Scholar]
  • 24.Sharing Research Data and Findings Relevant to the Novel Coronavirus (COVID-19) Outbreak. Wellcome Trust. Available from: https://wellcome.ac.uk/coronavirus-covid-19/open-data. [Google Scholar]
  • 25.Abritis A, Marcus A, Oransky I. An “alarming” and “exceptionally high” rate of COVID-19 retractions? Account Res. 2021;28(1):58–59. doi: 10.1080/08989621.2020.1793675 [DOI] [PubMed] [Google Scholar]
  • 26.Glasziou PP, Sanders S, Hoffmann T. Waste in covid-19 research. BMJ. 369. doi: 10.1136/bmj.m1847 [DOI] [PubMed] [Google Scholar]
  • 27.Athena RC, Directorate-General for Research and Innovation (European Commission), Know-Center, PPMI. Assessing the Reproducibility of Research Results in EU Framework Programmes for Research: Final Report. Publications Office of the European Union; 2022. Accessed June 27, 2024. Available from: https://data.europa.eu/doi/10.2777/186782. [Google Scholar]

Decision Letter 0

Roland G Roberts

19 Dec 2023

Dear Kelly,

Thank you for submitting your manuscript entitled "Biomedical researchers’ perspectives on the reproducibility of research: a cross-sectional international survey" for consideration as a Meta-Research Article by PLOS Biology.

This revised version of your manuscript has now been evaluated by the PLOS Biology editorial staff and I'm writing to let you know that we would like to send your submission out for external peer review. Note that, given the upcoming holidays, it is likely that we will not invite reviewers until January.

However, before we can send your manuscript to reviewers, we need you to complete your submission by providing the metadata that is required for full assessment. To this end, please login to Editorial Manager where you will find the paper in the 'Submissions Needing Revisions' folder on your homepage. Please click 'Revise Submission' from the Action Links and complete all additional questions in the submission questionnaire.

Once your full submission is complete, your paper will undergo a series of checks in preparation for peer review. After your manuscript has passed the checks it will be sent out for review. To provide the metadata for your submission, please Login to Editorial Manager (https://www.editorialmanager.com/pbiology) within two working days, i.e. by Dec 21 2023 11:59PM.

If your manuscript has been previously peer-reviewed at another journal, PLOS Biology is willing to work with those reviews in order to avoid re-starting the process. Submission of the previous reviews is entirely optional and our ability to use them effectively will depend on the willingness of the previous journal to confirm the content of the reports and share the reviewer identities. Please note that we reserve the right to invite additional reviewers if we consider that additional/independent reviewers are needed, although we aim to avoid this as far as possible. In our experience, working with previous reviews does save time.

If you would like us to consider previous reviewer reports, please edit your cover letter to let us know and include the name of the journal where the work was previously considered and the manuscript ID it was given. In addition, please upload a response to the reviews as a 'Prior Peer Review' file type, which should include the reports in full and a point-by-point reply detailing how you have or plan to address the reviewers' concerns.

During the process of completing your manuscript submission, you will be invited to opt-in to posting your pre-review manuscript as a bioRxiv preprint. Visit http://journals.plos.org/plosbiology/s/preprints for full details. If you consent to posting your current manuscript as a preprint, please upload a single Preprint PDF.

Feel free to email us at plosbiology@plos.org if you have any queries relating to your submission.

Kind regards,

Roli

Roland Roberts, PhD

Senior Editor

PLOS Biology

rroberts@plos.org

Decision Letter 1

Roland G Roberts

21 Mar 2024

Dear Kelly,

Thank you for your patience while your manuscript "Biomedical researchers’ perspectives on the reproducibility of research: a cross-sectional international survey" was peer-reviewed at PLOS Biology. It has now been evaluated by the PLOS Biology editors, an Academic Editor with relevant expertise (this is a different Academic Editor from the one who saw the initial submission), and by three independent reviewers.

You'll see that reviewer #1 says that this is well-written and highly relevant, but is not fully convinced by the novelty with respect to the Baker study, or the design of the work. S/he finds it confirmatory and descriptive (lacking in hypotheses), wonders whether the design reduces the risk of bias, and thinks that the Discussion fails to move us forward. Although in balance negative, their review seems helpful and constructive. Reviewer #2 says that he could not find the supplementary materials (he’s right; these are missing), suggests some presentational improvements (better Table structure, heatmaps), asks for some additional analysis (correlation analyses), wants you to declare some limitations, asks about the distinction between replicability and reproducibility, and asks you to provide code. Reviewer #3 is the most explicitly positive, but wants much better presentation (“graphs, figures, colour” – there are currently no Figures, just Tables), and wants breakdown by country and career stage (she thinks it’s rather “western-centric”). Like rev #2, she thinks the piloting was rather limited and wants you to supply the code.

IMPORTANT: We discussed these comments with the Academic Editor, who agreed that we should invite a revision, saying "Although the study was built on Baker (2016), data are original because time went by and many new discussions and actions happened in the meanwhile. I found no problem with the qualitative approach when it is suitable to answer the research question, as is the case in this paper. However, I agree with reviewers that limitations should be discussed. Sure, data visualization needs improvement, maybe, combining table with plots."

IMPORTANT: I should add, from the point of view of the journal, that for our broad readership, to enhance the appeal and traction of the paper, improved dataviz is ESSENTIAL, and we strongly support reviewer #3's helpful advice on this point. Also all supplementary files must be supplied, and underlying data and code must be made available, in line with our data and code availability policies.

In light of the reviews, which you will find at the end of this email, we would like to invite you to revise the work to thoroughly address the reviewers' reports.

Given the extent of revision needed, we cannot make a decision about publication until we have seen the revised manuscript and your response to the reviewers' comments. Your revised manuscript is likely to be sent for further evaluation by all or a subset of the reviewers.

We expect to receive your revised manuscript within 3 months. Please email us (plosbiology@plos.org) if you have any questions or concerns, or would like to request an extension.

At this stage, your manuscript remains formally under active consideration at our journal; please notify us by email if you do not intend to submit a revision so that we may withdraw it.

**IMPORTANT - SUBMITTING YOUR REVISION**

Your revisions should address the specific points made by each reviewer. Please submit the following files along with your revised manuscript:

1. A 'Response to Reviewers' file - this should detail your responses to the editorial requests, present a point-by-point response to all of the reviewers' comments, and indicate the changes made to the manuscript.

*NOTE: In your point-by-point response to the reviewers, please provide the full context of each review. Do not selectively quote paragraphs or sentences to reply to. The entire set of reviewer comments should be present in full and each specific point should be responded to individually, point by point.

You should also cite any additional relevant literature that has been published since the original submission and mention any additional citations in your response.

2. In addition to a clean copy of the manuscript, please also upload a 'track-changes' version of your manuscript that specifies the edits made. This should be uploaded as a "Revised Article with Changes Highlighted" file type.

*Re-submission Checklist*

When you are ready to resubmit your revised manuscript, please refer to this re-submission checklist: https://plos.io/Biology_Checklist

To submit a revised version of your manuscript, please go to https://www.editorialmanager.com/pbiology/ and log in as an Author. Click the link labelled 'Submissions Needing Revision' where you will find your submission record.

Please make sure to read the following important policies and guidelines while preparing your revision:

*Published Peer Review*

Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out. Please see here for more details:

https://blogs.plos.org/plos/2019/05/plos-journals-now-open-for-published-peer-review/

*PLOS Data Policy*

Please note that as a condition of publication PLOS' data policy (http://journals.plos.org/plosbiology/s/data-availability) requires that you make available all data used to draw the conclusions arrived at in your manuscript. If you have not already done so, you must include any data used in your manuscript either in appropriate repositories, within the body of the manuscript, or as supporting information (N.B. this includes any numerical values that were used to generate graphs, histograms etc.). For an example see here: http://www.plosbiology.org/article/info%3Adoi%2F10.1371%2Fjournal.pbio.1001908#s5

*Protocols deposition*

To enhance the reproducibility of your results, we recommend that if applicable you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols

Thank you again for your submission to our journal. We hope that our editorial process has been constructive thus far, and we welcome your feedback at any time. Please don't hesitate to contact us if you have any questions or comments.

Sincerely,

Roli

Roland Roberts, PhD

Senior Editor

PLOS Biology

rroberts@plos.org

------------------------------------

REVIEWERS' COMMENTS:

Reviewer #1:

The meta-research article of Cobey and colleagues presents a cross-sectional survey of biomedical researchers' perspectives on the reproducibility of research. The study builds on the widely cited Baker-survey published in Nature in 2016, but provides a more biomedical-centered perspective. In total, 1630 researchers successfully participated in this study. Briefly, 72% of participants agreed that there is a reproducibility crisis in biomedicine, with the leading cause of poor reproducibility being the 'pressure to publish' (which is interesting!). Only 16% of the participants indicated that their institution had established procedures to enhance the reproducibility of biomedical research; and 83% expected it to be harder to get funding for a replication study than for a novel one. Finally, the authors emphasize the relevance of the findings for policy makers and academic leadership to create incentives and research cultures that support better reproducibility and value research quality.

Overall, this is a short, but well-written manuscript on a highly relevant topic that presents the results of a survey on (ir)reproducibility in a clear and easy-to-read way. As stated by the authors themselves, the survey directly builds on the famous Nature-survey published in summer 2016, but differs in terms of the following two points: (1) the focus is exclusively on biomedical research (the Nature survey included scientists from various different disciplines across life sciences), (2) researchers were sampled from publication lists, allowing to report a response rate (this was not the case for the Nature survey). The latter point appears to be important to reduce any bias in the participants, as not only those researchers respond to the survey that are already actively involved in reproducibility projects. Whereas I agree on the importance of the latter point, I am not convinced by the overall novelty and design of the presented work. More specifically, I would like to raise the following major points:

(1) First of all, I am not entirely convinced by the novelty of the findings or the value of the outcome that goes beyond the Baker study. The fascinating aspect of the Baker study - at least from my perspective - was that so many researchers agreed on the existence of a crisis at a time (ir)reproducibility only started to become a hot topic in science. Furthermore, it was striking that the phenomenon was reported throughout the life sciences, arguing for a general rather than a discipline-specific problem. And finally, the Baker study provided a number of interesting causes for poor reproducibility that were discussed a lot and experimentally explored over the following years (thereby already creating the incentives for change that are discussed in the present paper). The findings of the present survey more or less confirm what has been reported previously and do not go beyond the already known problems. The focus on biomedical research is interesting, but not "groundbreaking". In case of publication, I would therefore like to see more reasons for the originality of this study as well as for its expected impact on the research community.

(2) The study is purely descriptive and the authors do not present any expectations or hypotheses. Although I agree that understanding researcher perceptions is key to driving a culture change and creating training interventions, one could also argue that the Baker study already met the criteria of an explorative study, allowing for generating some more concrete hypotheses for follow-up studies. For example, one could have looked at specific hypotheses surrounding the biomedical research community. What would you expect to be specific for this research field? Is it different from other disciplines within the life sciences? If so, why?

(3) It is not entirely clear to me, whether the survey was designed in such a way that it allows for reducing bias and identifying non-valid answers. I am not an expert in questionnaires' techniques, but I would expect to see some questions included that ask for the same thing, but with different words. From my perspective, such a validation step would be necessary to be able to extract only valid answers (and exclude those that were given simply to give any answer). The description of the survey in the Methods section appears to be rather limited in this respect. So, I would strongly recommend to provide more information about this.

(4) The discussion is rather short and does not go beyond the already known arguments. I therefore think that the manuscript could greatly benefit from a more in-depth analysis of the presented observations. In particular, it would be interesting to elaborate a bit more on the point that the perceived main reason was the pressure to publish. What is the expected impact of these findings and what kind of follow-up steps would be necessary to improve the situation? Is this specific for the biomedical research community?

Besides these major concerns, I do have a number of minor comments to the manuscript:

* The definition of reproducibility given in the introduction is a rather vague one. What does "consistent with the original study" mean? Is it dependent on the p-value (i.e. is a finding significant or not?) or should the observed effect simply go in a similar direction? Furthermore, what is meant by "similar methods"? How much variation is allowed? Is reproducibility here conceptualized as "generalizability"? Please be more specific, as this is the basis for the whole manuscript.

* At the end of the first section of the introduction the authors use the term "replicability"? Is it defined in a different way? Or is it used as synonym for reproducibility?

* The whole introduction is rather short, focusing on the Baker study and previous efforts to reproduce studies in different research fields. Whereas these are important aspects of the reproducibility topic, the issue is certainly much more complex and touches about various different aspects that could be mentioned here. For example, what are the discussed causes of poor reproducibility and what has been done to improve the situation? In particular, the topic of publication bias is an important one when it comes to systematic replication efforts (when trying to replicate published studies, you have to deal with the underlying publication bias that might be a reason for poor reproducibility in itself). I would therefore strongly recommend to expand the introduction a bit and include some more aspects that currently play a role in the debate.

* The methods are only superficially reported. For example, the section about the study design should also give more information about the different phases of the study (planning, piloting, surveying). Likewise, the section about the survey should provide more information about the technique used. How many questions were included? How was the survey organized? Were the questions presented in always the same or a randomized order? Did the authors include any kind of validation steps (see above)? Why did the authors modify some questions of the Nature study and others not?

* The survey was piloted by three researchers. Although I appreciate these piloting steps, I doubt the representativeness and generalizability of this step, as the involvement of just three scientists appears to be rather limited. Why only three? And what was changed according to the feedback of these researchers?

* The assignment of codes as reported in the data analysis section is not clear to me. What exactly was done here? What are potential consequences of the codes not being blinded?

* At the beginning of the results section it is mentioned that the authors obtained and used listed emails from all authors on an identified paper. What does it mean exactly? Did they include more than just one author per study? If yes, this could cause kind of pseudoreplication, as the authors of one study are more dependent on each other than the authors of different studies. Publications with a huge number of authors could therefore bias the outcome towards a specific direction. Please comment on that.

* The results are presented on the basis of pure numbers, completed by a huge number of tables. It would be much more readable to present some simplified figures or illustrations, as it was also done in the Baker study.

* The reference list is not formatted consistently. It is furthermore condensed to a surprisingly low number of references that could be complemented by several additional papers (related to the point that both the introduction and the discussion could cover additional topics).

* The statements concerning the link to the Baker-study are confusing. Sometimes it is explicitly stated that the aim was not to compare the data and a few sections later it is said that this approach allowed for comparing the results to the original study. Please be more concise on the aims and the actual relation to the Nature survey.

Reviewer #2:

[identified himself as Timothy M. Errington]

This manuscript describes a survey investigating biomedical researchers' views about reproducibility, including perceptions of the causes of irreproducibility, experiences with reproducibility, and perceived support for reproducibility from their institutions. The study builds off the 2016 Nature survey on this topic. Below are suggestions to help clarify details of the study and to improve accessibility to a broader readership.

1. There appears to be missing information. I could not find S1 (search strategy), S2 (full survey), S3 (breakdown of responses), and S4 (training resources)? I don't see it in the downloaded submission file or on the OSF project linked (though I did find this file that is maybe S2: https://osf.io/a3spd)? Related, the discussion refers to Table 7 as 'institutions influence on reproducibility of research' yet Table 7 is 'Thematic analysis of reasons why participants replicated their own study'. Finally, the authors share information about responses to a statement of "I feel I am more concerned about reproducibility than the average researcher in my field and at my career stage", yet I don't see anything in Table 9 about this question and responses?

2. The paper would benefit from presenting the tables in more easily digestible ways. For example, Table 3 presented two different ways is a bit confusing. I'd recommend picking one table to present instead of both (e.g., present numbers with percentages in parathesis for any given cell like Table 4). Another example, is for Table 4 it would be helpful if it was easy to digest. Maybe a heatmap?

3. Have the authors looked at any correlations of responses? For example, responses for if there is a 'crisis' (slight or significant) is at 72%, yet, when asked to give their perception of what proportion are reproducible responses are at the ~50% mark. A correlation analysis to understand the relationship between these two questions might be a valuable addition. Relatedly, what do the authors make of many responses being centered around the 'sometimes contributes' category in Table 4? Did the authors look to see if any respondents gave similar responses across the board on those questions? Or was there sufficient variation?

4. A potential limitation is the drop-off rate in the survey. It would great if the authors presented this information (graphically works well, but something to indicate where drop-off occurred). For example, there were 1630 respondents to at least one question, yet it's clear there was drop off since "A total of 724 participants responded to the item about why they replicated their own research".

5. A limitation section would be great - for example, the 'insufficient oversight' question of causes is limited in insight (and potential bias) by the sample being so heavily faculty/PI based (i.e., those responsible for oversight).

6. I see in table 6, the work 'replicate' and 'replication' are used? Similarly, in the discussion the authors switch feely between reproducibility and replicability. But the authors indicated they took a broad stance in the definition of reproducibility, which I took as inclusive of words like replicate and replication? What impact might this have on the responses given? For example, is the question "Have you ever tried to replicate a published study conducted by another team of authors and failed?". Respondents might have answered this as failed to get reproducible results or failed to be able to 'repeat the study exactly' as defined in the survey? This is different than question 8 in the survey (on OSF) where the failed is defined as "re-ran an experiment but got different results from the original study). Same for asking participants about whether another researcher contacted them who was 'unable to reproduce their findings', which based on definitions shared in the survey and manuscript would give a different response than if the authors asked 'unable to replicate their findings'? The terminology is indeed complex (as the authors acknowledge in the introduction - and thus there might be influence in the responses despite the authors best attempt at preventing it).

7. The discussion mentions 'reproducibility networks'. I agree on the need to mention, this, but just mentioning this and providing a reference I think is insufficient - it would benefit readers who have no idea what this is to give a little more context.

8. This EC study on reproducibility might be valuable to include in the discussion: https://op.europa.eu/en/publication-detail/-/publication/36fa41a8-dbd5-11ec-a534-01aa75ed71a1/language-en

9. Was the pilot testing with researchers include any of the authors of this manuscript? Same with the researchers who coded text response, were they any of the authors? It would be good to declare this either way.

10. Can you share your scripts (e.g., exported files form SPSS)? It would be nice to share these for others to review and reuse.

11. Can you share your anonymized data (or did you not get consent for that?).

Reviewer #3:

[identifies herself as Katherine S. Button]

This was an interesting and informative survey of biomedical researchers. The methods are generally sound and the findings are interesting. My only suggestions are that the results could be presented in a more interesting and informative manner using better data visualisation (graphs, figures, colour), and that the findings could be broken down by key categories (such as country, and career stage) to gain greater insights into international variation .

Specific points below (page numbers would have helped):

Methods:

1) Piloting by three researchers seems a bit limited - a broader range including people outside of the research group might have been more rigorous.

2) Analysis in SPSS - does this mean that the analysis code isn't available? Perhaps comment on the availability of meta-data and syntax/analysis code in the manuscript alongside data availability in the analysis section. Make it available if it currently is not.

3) The thematic analysis method seems a bit thin. Consider citing some of the key thematic methods papers that informed your approach (e.g., Braun and Clarke) there are slightly varying approaches.

Results:

Minor points:

1) Determinants of irreproducibility: 97 (give %). Very small number - perhaps discuss as a limitation and it would be good to know who these people were.

2) Provide all countries of employment in Table 1 to show how 'international' this survey is. The top few make it look very western-centric.

Major points:

1) I think the presentation of data in tables is very dry. My major suggestion for improving your paper is better data visualisation. Graphs, figures and colour would help. Think about people using your findings in presentations. These tables will look very dense on PowerPoint slides.

2) Similarly, I think there is more analysis that can be done to look at variability across nations and career stages. I would suggest nice graphs/charts in the supplementary of the key findings broken down by nationality and career stage. Here in the UK for example we've had some movement re. institutional involvement and initiatives around reproducibility e.g., institutional members of UKRN. It would be interesting to see whether that comes across in the responses around intuitional support, for example. Also the reproducibility 'movement' has been readily taken up by early career researchers with grassroots initiatives led by PhD students and post-docs gaining international reach (e.g., Science RIOT club, ReproducibiliTea). So I'd be interested to see how these responses differ by career stage. Providing more fine-grained analysis along with the analysis code will help this survey to have its greatest impact. As above, helpful for people planning to use your analysis in presentations, discussions with their institutions ect.

Check typos in penultimate paragraph.

Discussion:

1) Depending on what you find in the additional analysis there might be key things to pull out in the discussion.

Decision Letter 2

Roland G Roberts

22 Aug 2024

Dear Kelly,

Thank you for your patience while we considered your revised manuscript "Biomedical researchers’ perspectives on the reproducibility of research: a cross-sectional international survey" for publication as a Meta-Research Article at PLOS Biology. This revised version of your manuscript has been evaluated by the PLOS Biology editors, the Academic Editor and two of the original reviewers.

Based on the reviews, we are likely to accept this manuscript for publication, provided you satisfactorily address the remaining points raised by the reviewers, and the following data and other policy-related requests.

IMPORTANT - Please attend to the following:

a) Please truncate your Title to simply ""Biomedical researchers’ perspectives on the reproducibility of research"

b) Please address the concerns raised by reviewer #2. Like him, it was my impression that the OSF deposition is incomplete, and may not be sufficient to reproduce the study.

c) I note that you declare that you received no specific funding. Can you just confirm that this is correct?

d) When you mention your ethics approval in the MS (“This study received ethics approval from the Ottawa Health Sciences Research Ethics Board”), please can you also include the protocol approval number. From the files in OSF, it looks like this is #20210856-01H...

e) Please address my Data Policy requests below; specifically, we need you to supply the numerical values underlying Figs 1 and 2, either as a supplementary data file or as a permanent DOI’d deposition. I note that you already have an associated OSF deposition, but the values underlying these Figs look like they may already be presented in the Supplementary Tables...? Please ensure that everything is made available.

f) Please cite the location of the data clearly in all relevant Figure legends, e.g. “The data underlying this Figure can be found in Table S1” or “The data underlying this Figure can be found in https://osf.io/3ksvz/"

g) Please make any custom code available, either as a supplementary file or as part of your data deposition.

As you address these items, please take this last chance to review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the cover letter that accompanies your revised manuscript.

We expect to receive your revised manuscript within two weeks.

To submit your revision, please go to https://www.editorialmanager.com/pbiology/ and log in as an Author. Click the link labelled 'Submissions Needing Revision' to find your submission record. Your revised submission must include the following:

- a cover letter that should detail your responses to any editorial requests, if applicable, and whether changes have been made to the reference list

- a Response to Reviewers file that provides a detailed response to the reviewers' comments (if applicable, if not applicable please do not delete your existing 'Response to Reviewers' file.)

- a track-changes file indicating any changes that you have made to the manuscript.

NOTE: If Supporting Information files are included with your article, note that these are not copyedited and will be published as they are submitted. Please ensure that these files are legible and of high quality (at least 300 dpi) in an easily accessible file format. For this reason, please be aware that any references listed in an SI file will not be indexed. For more information, see our Supporting Information guidelines:

https://journals.plos.org/plosbiology/s/supporting-information

*Published Peer Review History*

Please note that you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out. Please see here for more details:

https://blogs.plos.org/plos/2019/05/plos-journals-now-open-for-published-peer-review/

*Press*

Should you, your institution's press office or the journal office choose to press release your paper, please ensure you have opted out of Early Article Posting on the submission form. We ask that you notify us as soon as possible if you or your institution is planning to press release the article.

*Protocols deposition*

To enhance the reproducibility of your results, we recommend that if applicable you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols

Please do not hesitate to contact me should you have any questions.

Sincerely,

Roli

Roland Roberts, PhD

Senior Editor

rroberts@plos.org

PLOS Biology

------------------------------------------------------------------------

DATA POLICY:

You may be aware of the PLOS Data Policy, which requires that all data be made available without restriction: http://journals.plos.org/plosbiology/s/data-availability. For more information, please also see this editorial: http://dx.doi.org/10.1371/journal.pbio.1001797

Note that we do not require all raw data. Rather, we ask that all individual quantitative observations that underlie the data summarized in the figures and results of your paper be made available in one of the following forms:

1) Supplementary files (e.g., excel). Please ensure that all data files are uploaded as 'Supporting Information' and are invariably referred to (in the manuscript, figure legends, and the Description field when uploading your files) using the following format verbatim: S1 Data, S2 Data, etc. Multiple panels of a single or even several figures can be included as multiple sheets in one excel file that is saved using exactly the following convention: S1_Data.xlsx (using an underscore).

2) Deposition in a publicly available repository. Please also provide the accession code or a reviewer link so that we may view your data before publication.

Regardless of the method selected, please ensure that you provide the individual numerical values that underlie the summary data displayed in the following figure panels as they are essential for readers to assess your analysis and to reproduce it: Figs 1 and 2. NOTE: the numerical data provided should include all replicates AND the way in which the plotted mean and errors were derived (it should not present only the mean/average values).

IMPORTANT: Please also ensure that figure legends in your manuscript include information on where the underlying data can be found, and ensure your supplemental data file/s has a legend.

Please ensure that your Data Statement in the submission system accurately describes where your data can be found.

------------------------------------------------------------------------

CODE POLICY

Per journal policy, if you have generated any custom code during the course of this investigation, please make it available without restrictions. Please ensure that the code is sufficiently well documented and reusable, and that your Data Statement in the Editorial Manager submission system accurately describes where your code can be found.

Please note that we cannot accept sole deposition of code in GitHub, as this could be changed after publication. However, you can archive this version of your publicly available GitHub code to Zenodo. Once you do this, it will generate a DOI number, which you will need to provide in the Data Accessibility Statement (you are welcome to also provide the GitHub access information). See the process for doing this here: https://docs.github.com/en/repositories/archiving-a-github-repository/referencing-and-citing-content

------------------------------------------------------------------------

DATA NOT SHOWN?

- Please note that per journal policy, we do not allow the mention of "data not shown", "personal communication", "manuscript in preparation" or other references to data that is not publicly available or contained within this manuscript. Please either remove mention of these data or provide figures presenting the results and the data underlying the figure(s).

------------------------------------------------------------------------

REVIEWERS' COMMENTS:

Reviewer #1:

The authors made a good job in revising the manuscript. In particular, the discussion has improved a lot, covering a few more (novel) points and aspects. Thank you.

Reviewer #2:

[identifies himself as Timothy M. Errington]

Thank you for revising the manuscript. It is much improved from the previous version. The authors addressed my previous comments, however two minor aspects still remain.

1. While I appreciate the authors position on not wanting to do any correlations of the responses, it would benefit if the rationale given in the response to my comment was included in the manuscript somewhere. Not the response regarding a potential perception of the authors 'fishing' (which personally I would not think was happening since it would clearly labeled as exploratory and post-hoc), but more about the concern regarding being mindful of responses not being mutually exclusive with each other. This aspect could be added to the limitations section, especially since this would suggest caution for future researchers who might want to do this to look at between group differences.

2. I appreciate that the authors added their code and data to OSF, however I still can not view them. Maybe they are in a component that is still private?

Decision Letter 3

Roland G Roberts

1 Oct 2024

Dear Kelly,

Thank you for the submission of your revised Meta-Research Article "Biomedical researchers’ perspectives on the reproducibility of research" for publication in PLOS Biology. On behalf of my colleagues and the Academic Editor, Cilene Lino de Oliveira, I'm pleased to say that we can in principle accept your manuscript for publication, provided you address any remaining formatting and reporting issues. These will be detailed in an email you should receive within 2-3 business days from our colleagues in the journal operations team; no action is required from you until then. Please note that we will not be able to formally accept your manuscript and schedule it for publication until you have completed any requested changes.

Please take a minute to log into Editorial Manager at http://www.editorialmanager.com/pbiology/, click the "Update My Information" link at the top of the page, and update your user information to ensure an efficient production process.

PRESS: We frequently collaborate with press offices. If your institution or institutions have a press office, please notify them about your upcoming paper at this point, to enable them to help maximise its impact. If the press office is planning to promote your findings, we would be grateful if they could coordinate with biologypress@plos.org. If you have previously opted in to the early version process, we ask that you notify us immediately of any press plans so that we may opt out on your behalf.

We also ask that you take this opportunity to read our Embargo Policy regarding the discussion, promotion and media coverage of work that is yet to be published by PLOS. As your manuscript is not yet published, it is bound by the conditions of our Embargo Policy. Please be aware that this policy is in place both to ensure that any press coverage of your article is fully substantiated and to provide a direct link between such coverage and the published work. For full details of our Embargo Policy, please visit http://www.plos.org/about/media-inquiries/embargo-policy/.

Thank you again for choosing PLOS Biology for publication and supporting Open Access publishing. We look forward to publishing your study. 

Sincerely, 

Roli

Roland G Roberts, PhD, PhD

Senior Editor

PLOS Biology

rroberts@plos.org

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 File. Search strategy used to obtain a random sample of journals to extract author contact information from to identify possible participants.

    (DOCX)

    pbio.3002870.s001.docx (17KB, docx)
    S2 File. Survey administered to participants.

    The survey was administered online and used survey logic to present relevant items.

    (DOCX)

    pbio.3002870.s002.docx (20.4KB, docx)
    S3 File. Supplementary tables presenting a comparison of perceptions of reproducibility between our findings and the original Nature paper by Baker, presented overall and by discipline.

    In addition, we present participant perceptions of reproducibility in different research areas.

    (DOCX)

    pbio.3002870.s003.docx (18.5KB, docx)
    S4 File. Additional analyses of survey responses by gender and academic role.

    (DOCX)

    pbio.3002870.s004.docx (19.7KB, docx)
    S5 File. List of training available on reproducibility mentioned by survey participants.

    (DOCX)

    pbio.3002870.s005.docx (19.4KB, docx)
    Attachment

    Submitted filename: Revision 2 Plos Bio Reproducibility Survey -2024-6-24.docx

    pbio.3002870.s006.docx (49.3KB, docx)
    Attachment

    Submitted filename: Revision Cobey et al PLOS Bio- 2024-9-11.docx

    pbio.3002870.s007.docx (17.4KB, docx)

    Data Availability Statement

    All data and study materials are available on the Open Science Framework (https://osf.io/3ksvz/) or in the supplementary materials.


    Articles from PLOS Biology are provided here courtesy of PLOS

    RESOURCES