Skip to main content
Sage Choice logoLink to Sage Choice
. 2021 May 26;16(4):450–460. doi: 10.1177/15562646211018916

Towards a Research Agenda for Promoting Responsible Research Practices

Joeri K Tijdink 1,2,, Serge PJM Horbach 3,4, Michèle B Nuijten 5, Gareth O’Neill 6,7
PMCID: PMC8458678  PMID: 34037490

Abstract

This opinion piece aims to inform future research funding programs on responsible research practices (RRP) based on three specific objectives: (1) to give a sketch of the current international discussion on responsible research practices (RRPs); (2) to give an overview of current initiatives and already obtained results regarding RRP; and (3) to give an overview of potential future needs for research on RRP. In this opinion piece, we have used seven iterative methodological steps (including literature review, ranking, and sorting exercises) to create the proposed research agenda. We identified six main themes that we believe need attention in future research: (1) responsible evaluation of research and researchers, (2) the influence of open science and transparency on RRP, (3) research on responsible mentoring, supervision, and role modeling, (4) the effect of education and training on RRP, (5) checking for reproducibility, and (6) responsible and fair peer review. These themes have in common that they address aspects of research that are mostly on the level of the scientific system, more than on the level of the individual researcher. Some current initiatives are already gathering substantial empirical evidence to start filling these gaps. We believe that with sufficient support from all relevant stakeholders, more progress can be made.

Keywords: responsible research practices, research integrity, research agenda, research ethics, authorship, conflicts of interest/bias, educational research, questionable research practices

Introduction

While academic research is generally recognized as an institute of crucial importance to the well-being of our contemporary society, concerns over its functioning are growing. The current coronavirus disease 2019 (COVID-19) pandemic especially highlights the need for more open and collaborative research practices. Stakeholders from within and outside the academic community are increasingly worried about issues regarding research funding, the scientific publication system, research evaluation practices, and academic career trajectories. Fueled by several highly visible cases of scientific misconduct—both before the COVID-19 outbreak and during the current pandemic (Callaway, 2011; Mehra et al., 2020)—there is a growing awareness that the results of scientific research should be reliable, that research practices should be responsible, and that the workflows and results of research need to be transparent. While media coverage has mainly centered on large fraud cases, the research community has acknowledged that fundamental causes of current issues are embedded in more systemic aspects of evaluating, rewarding, and disseminating research (Biagioli et al., 2019; Bramstedt, 2020; Fleming et al., 2020). Experts of research integrity are convinced that, on an aggregate level, factors such as detrimental research practices, lack of supervision and mentoring, the system of research with hypercompetition, unidimensional assessment criteria, an individualistic research culture, and publication pressure are more threatening to the reliability and validity of research than (arguably rare) cases of misconduct (Begley & Ioannidis, 2015; Fanelli et al., 2015; Martinson et al., 2010). This has led to a shift in focus towards research on responsible research practices (RRP). The increased attention for this subject is reflected by the large number of international initiatives from various stakeholders to foster RRP (e.g., The Research on Research Institute, the Meta-Research Innovation Center at Stanford [METRICS]). These initiatives have mainly emerged from the biomedical sciences and psychology, and there is now an increasing need for initiatives tailored to other research disciplines and cultures (Haven et al., 2019b; Horbach & Halffman, 2019).

Research on RRP has shown several trends. Initial interest in this field was mainly in defining research integrity and gaining more insight in the different forms of research misbehaviors, leading to classifications of major and minor forms of misbehaviors (falsification, fabrication, and plagiarism and questionable research practices [QRPs], respectively; [Bouter et al., 2016; Steneck, 2006]). In addition, ample attention has been paid to estimating the frequency of transgressions and analyzing individual cases (Fanelli, 2009; John et al., 2012; van der Heyden et al., 2009). Repeatedly, boundaries between research integrity and research ethics have been questioned in this debate with other concepts such as responsible research and innovation prominently coshaping discussion on research integrity. This has led us to adhere to a broad interpretation of “research integrity” including aspects that some may classify as research ethics. The scope of our concepts and paper will be further clarified in the methods section, which is added as Supplemental file Appendix I.

Over the past years, research on RRPs and research integrity in all disciplinary domains is increasingly redirecting its focus towards root causes and possible solutions. This has resulted in a shift from microlevel analyses, mainly concerned with the individual and his/her characteristics, towards current discussions acknowledging the important role of broader cultural, organizational, and systemic factors, including the research climate, organizational settings, and incentive structures (Haven et al., 2019b; Martinson et al., 2010). As an example, this shift is represented in the new Dutch Code of Conduct for Research Integrity, dedicating a full section to institutional responsibilities (KNAW, 2018).

In addition, openness and transparency have been increasingly identified as drivers of RRP. A growing number of initiatives are calling for and facilitating increased transparency in research, voiced in terms like “FAIR Data,” “Open Access,” and “Open Science.” Initiatives include proposals as diverse as calls for openly sharing data, facilitating reproducibility and replication studies (including the humanities), publishing in Open Access journal articles, and using Open Peer Review formats (Nosek et al., 2015). Lastly, suspicion about published findings that are not replicable, mainly within psychology and the biomedical sciences, has directed research focus of RRP researchers towards ways of enhancing replicability. This includes studying publication bias, methodological flexibility, and transparency (Munafò et al., 2017; Wicherts, 2017).

While the number of initiatives aiming to foster RRP is growing and diversifying, the evidence base for which initiatives are actually successful is often lacking. In addition, initiatives are commonly restricted to specific niches or academic disciplines. In particular, most initiatives and studies originate from the biomedical and social sciences, leaving blind spots in other research disciplines, cultures and methods, and risking the tendency to overgeneralize both challenges of research integrity as well as their potential solutions. Therefore, more research is still needed to guide the implementation of RRP across the full range of academic disciplines, cultures, and settings.

This paper aims to shed light on the international debate on RRP addressing three key issues:

  1. A sketch of the current international discussion on RRP.

  2. An overview of current initiatives and already obtained results regarding RRP.

  3. An overview of potential future needs for research on RRP.

Our opinion piece will hence address these three points. Its ultimate aim is to serve as a starting point for future research agendas and more broadly for the international research field on research integrity and RRP. However, the current document is slightly focused on the Dutch context. This is partly due to the fact that the current debate on responsible research in the Netherlands was sparked by the high-profile fraud case of Diederik Stapel. Stapel was a professor of social psychology who fabricated data in over 50 published articles. The case itself triggered a plethora of Dutch initiatives to foster RRP (Abma, 2013). This also included initiatives from funding agencies such as the European Union in their H2020 program, the Office of Research Integrity in the United States, the funder ZonMw in the Netherlands, and the Welcome Trust in the United Kingdom, establishing multiple funding programs (e.g., [Replication Studies, 2019; Research on Research, 2020]). In these programs, several researchers are currently working on and assessing initiatives that foster RRP. Although we have tried to incorporate an international perspective, we found several initiatives from the Netherlands. Since the Dutch health funding organization was one of the first funding organizations that started funding on research integrity and numerous initiatives have been initiated in the Netherlands, the Netherlands can be considered one of the frontrunners in the area of initiatives and research fostering RRP, we believe this focus constitutes an appropriate lens to inventories state-of-the-art developments in the field. The progressive atmosphere in the Netherlands regarding research and initiatives on RRP is among others exemplified in several grant schemes by national funders related to this topic, as well as the early establishment of a national committee on research integrity. In our endeavors to establish a research agenda, we first summarize the current state of research on RRP and provide a conceptual overview of initiatives to foster RRP. Second, we identify major gaps of knowledge in the field of research on RRP that need to be addressed. This builds to a research agenda in which we classify the most pressing themes that need further study. With this knowledge, we aim to inform policymakers and funders on current issues for RRP. The paper thereby complements recent initiatives that have inventarised research on specific aspects related to RRP, such as peer review and research integrity (Aubert Bonn & Pinxten, 2019; Tennant & Ross-Hellauer, 2020).

Current Initiatives Promoting RRP

The scientific community has already started many initiatives to promote and address RRP. These initiatives cover all aspects of the scientific “empirical process” and are in differing stages of development and acceptance by researchers. In this section, we categorize the main types of initiatives and sketch the current state-of-the-art of RRP. A list of concrete examples of initiatives can be found in Supplemental file Appendix II, whereby we acknowledge that the list is not exhaustive and may overlook other initiatives.

RRP effectively encompasses the entire scientific ecosystem from overarching frameworks to the implementation system to the research process (Mejlgaard et al., 2020). In Figure 1, we have framed RRP as an overarching theme consisting of these three different levels: scientific frameworks; the scientific system; the empirical cycle. We are aware that we focus on research adhering to a research paradigm following the empirical cycle and thus put less emphasis on other methodologies. Other ways to structure our data could have been chosen, for instance a framework based on Latour and Woolgar's credibility cycle. Our choice for structuring the data along the empirical cycle is not so much because we believe it is the best way to describe research in general, but because we believe it is most appropriate for our current task as most initiatives are tailored to a research paradigm fitting this cycle (Latour & Woolgar, 1979). We acknowledge that some disciplines may not be fully covered by this paradigm. In fact, we will address this as one of the major shortcomings of previous research on RRP. Nevertheless, we believe that the empirical cycle is a useful framework to create and present an inventory of current initiatives that aim to foster RRP.

Figure 1.

Figure 1.

A schematic representation of the elements that interact in responsible research practices (RRP). The three elements all consist of initiatives that we have collected in Supplemental file Appendix I. The three rings in this figure represent the different levels at which we can look at RRP: scientific frameworks, the scientific system, and the empirical cycle. Figure 2 elaborates on the initiatives to promote RRP in the empirical cycle.

Scientific Frameworks

The overarching first level of scientific frameworks refers to three broad value frameworks under which all RRP are expected to be carried out and judged in the current socio-political climate: open science; research integrity; societal impact. These three frameworks are not to be seen as distinct but rather overlap and feed into each other. In fact, by making research more open and transparent, open science is expected to be an enabler for research integrity and societal impact (Moher et al., 2020).

The practice of an open science has existed for centuries in Europe in the form of open letters and select publications in scholarly outlets. A modern interpretation of open science has been proposed by the European Commission (2016) as the opening up of research workflows and outcomes via digital technology. This approach applies to the entire research process: designs and methodologies (open methodologies); data that is findable, accessible, interoperable, and reusable (www.go-fair.org) and open (open data); software code (open source); initial versions of publications and peer review (open peer review); and access to final versions of publications (open access). This approach also includes opening up study materials (open education) and involving citizens in research (citizen science). The European Union has taken a leading role in open science with Plan S (www.coalition-s.org).

The concept of research integrity is crucial to the scientific endeavor and refers to research practices that follow five key principles of honesty, scrupulousness, transparency, independence, and responsibility (see the European ALLEA code and Netherlands Code of Conduct for Research Integrity [European Code of Conduct for Research Integrity. Revised edition, 2017; KNAW, 2018]). Research misconduct arises from researchers intentionally or unintentionally not following these principles. High-profile cases of intentional misconduct often lead to public outcry and calls for changes in the way we do science. Examples of research integrity initiatives in the Netherlands are the Netherlands Research Integrity Network (www.nrin.nl) and the EU-funded Embassy of Good Science (www.embassy.science).

Similarly, there has been considerable public indignation about the (lack of) societal impact of publicly funded research. There is a growing call for research activities and outcomes to not only involve citizens but also be relevant for and benefit society. This can take many forms: opening up and communicating about the outcomes of research; direct application and commercialization of research technologies; directly involving the public in the design, collection, and analysis of research; or directly letting the public decide future research funding topics. The Netherlands aims to facilitate this by the participatory National Research Agenda (www.wetenschapsagenda.nl).

The Scientific System

The secondary level of the scientific system focuses on the steering and facilitation of RRP by governments, funding agencies, and academic institutions under six key areas: research policies; research practices; training researchers; evaluating research(ers); rewarding researchers; funding research(ers). Each of these areas may focus on specific aspects of the scientific framework as well as individual steps in the empirical cycle. We are convinced that one of the goals of the scientific system for RRP is to ensure that researchers learn about RRP and conduct their research in a “responsible” manner.

The policies for RRP relate to the legislation affecting research, funding programs, guidelines and codes of conduct, and general principles for how researchers are to be trained, evaluated, and rewarded. These policies crucially determine the initiatives that are intended to foster RRP. Any identified gaps in initiatives should lead to new policies for stimulating initiatives to fill such gaps. This position paper aims to identify any gaps for developing new policies. Research performing organizations (incl. universities) strongly determine which policies and governance structures are shaping the research environment individual researchers work in. This responsibility is oftentimes overlooked. This research environment is influencing the behavior of researchers (Haven et al., 2021) and is even perceived as an important issue that may trigger research misconduct (Davies, 2019). We would like to emphasize the crucial role and responsibility research performing organizations have to foster research integrity. The European SOPs4RI project has already started collecting tools to help institutions foster RRP (Mejlgaard et al., 2020). Aside from formal policies, actual research practices form the most effective initiatives to foster responsible research. While stakeholders agree that mere changes in policies, guidelines, and codes of conduct will not suffice to establish more responsible research, much attention has been given to these formal academic structures. Nonetheless, actual research practices, as well as informal rules, academic cultures, and interpersonal relations have been the target of recent initiatives and studies on RRP.

The raising of awareness and training of researchers in RRP is crucial for RRP to be carried out successfully and be widely adopted by the scientific community. A recent survey by the European Commission (O'Carroll et al., 2017, See Appendix 3 for the survey results which show that researchers have little knowledge of Open Science), however, has shown that the majority of researchers in Europe do not fully understand and are not being supported to implement the principles of open science. Initiatives for supporting and training researches in RRP are currently grassroots and locally organized, such as the Superb Supervision (https://amsterdamresearchclimate.nl/superb-supervision) course or the ReproducibiliTea journal clubs (www.reproducibilitea.org).

The evaluation and rewarding of research and researchers are perhaps the most important factors in promoting RRP. The current evaluation system is focused on judging research, and by extension researchers, according to the number of publications that appear in high impact factor and branded journals, instead of judging the quality of the research itself. Furthermore, researchers are to a much lesser extent (if at all) evaluated on the basis of other research-related activities such as teaching, supervision, unsuccessful grant applications, research designs, peer reviews, the publishing of datasets, and public engagement. These “perverse incentives” (Bouter, 2015) are exacerbated by the high publication pressure and strong competition over research funding and job opportunities that researchers experience (Martinson et al., 2009; Tijdink et al., 2014; van Dalen & Henkens, 2012). Examples of initiatives that aim at changing this incentive structure are the Declaration on Research Assessment (www.sfdora.org) and the Hong Kong Manifesto (Moher et al., 2020).

Finally, the funding system drives academic research and is also closely related to the careers of researchers. The funding system has the potential to directly influence the behavior of researchers via funding mandates and funding award criteria. One example is the mandating by funders of researchers to publish in Open Access journals and openly license their publications. Another example is the refusal by funders to use journal impact factors to evaluate researchers: this reduces the pressure on researchers to publish only in high-impact and branded journals. Funding therefore directly determines the research agenda of researchers and can be more aware of their crucial role in fostering RRP.

The Empirical Cycle

Most of the initiatives to promote RRP are aimed at the empirical cycle itself. Figure 2 shows the many types of initiatives at this level, including how they link to the different steps in the empirical cycle and how they relate to each other. In many fields and types of research, the empirical cycle often roughly follows the same steps. From a theory (step 1), a hypothesis is formulated (step 2). To test this hypothesis, a study is designed (step 3) and data are collected (step 4). Based on the analysis of the data, a conclusion is drawn (step 5) and the research is disseminated (step 6). Finally, during and after dissemination, the published literature can be corrected (step 7). These steps are loosely based on the empirical cycle as discussed by De Groot (1961). We realize that not every single scientific study will fit this representation (e.g., exploratory research where the data generates the hypothesis or conceptual research fields that do not employ data). However, most of the initiatives to promote RRP at the empirical level seem to fit within this framework. The fact that most initiatives fit this framework is also indicative of a potential lack of diversity in those initiatives, arguably not covering research methods and traditions working from different frameworks. In the paragraphs below, we briefly outline the initiatives to promote RRP at each of these steps in the empirical cycle. Furthermore, for each step and each practice we have collected a list of initiatives that fall under these practices. The full list is available in Supplemental file Appendix II.

Figure 2.

Figure 2.

Schematic overview of the current initiatives to promote responsible research practices at the level of the empirical cycle. The initiatives (in the ellipses) are linked to different elements of the empirical cycle (rounded squares) and to each other with the grey connector lines. Each underlined term links to a list of concrete initiatives in Supplemental file Appendix I.

In the first step of the empirical cycle, we emphasize the need for theory. To build theory or to assess which findings are sufficiently robust to build upon, researchers usually aim to compile results from existing studies. One method to synthesize results is via meta-analysis: a statistical summary of different studies with (more or less) the same research question.1 A problem with meta-analyses is that they are not immune to problems such as publication bias, which means that they can lead to biased results.

In the next step of the empirical cycle, a hypothesis is formulated. At this step, one of the most notable types of initiatives to increase RRP is preregistration (Kupferschmidt, 2018). There are different ways to preregister a study, but generally, researchers publish their hypotheses and research plan online, before conducting the study. Preregistration can have several goals. First, it allows for a clear distinction between confirmatory and exploratory analyses (Wagenmakers et al., 2012). Second, it could prevent exploiting flexibility in methods and data analyses to obtain the desired result, because the plans were registered beforehand (initiatives such as Compare [www.comparetrials.org] and EU Trials Tracker [eu.trialstracker.net] compare registrations with the accompanying published papers). Third, preregistration could decrease the effects of publication bias. The practice of preregistration has been standard in biomedicine for years, but has only recently gained popularity in the social sciences. One relatively new form of preregistration is a “Registered Report” (Chambers et al., 2015). Here, the preregistration is submitted to a journal and peer reviewed. Once the preregistration is approved, the authors can get an “in principle acceptance,” meaning that if they follow their preregistration, their paper will be published regardless of the outcome. Over 260 journals now accept this format.

The next phase in the empirical cycle is the study design phase. Two notable types of initiatives that promote RRP at this step are multilab collaborations and replication studies. Multilab collaborations are coordinated efforts to run the same study in different labs, sometimes across the world (see the Psychological Science Accelerator [psysciacc.org] and the Many Babies project [https://psyarxiv.com/27b43/]). This strategy increases statistical power and allows investigating the generalizability of the findings. It also requires detailed study protocols, which increase reliability. A second initiative to improve RRP at the design step is a stronger focus on replication studies. Replications affect the design phase, but they also link to many other steps in the empirical cycle; from theory to analysis. An example of a Dutch initiative to promote replication studies is the replication grant of NWO (Replication Studies, 2019). A special case of a replication study is a “registered replication report”: a multilab collaboration that performs a preregistered replication study. Another example of fostering responsible collaboration is the Montreal statement on research integrity in cross-boundary research collaborations (Lancet, 2013). This statement highlights the importance of equitable institutional cross-boundary collaborations.

Another clear point in the empirical cycle where RRP can be stimulated is when data has been collected. One of the most advocated strategies is to share data. Openly sharing data allows for reanalysis to detect and correct mistakes, to check robustness, and to answer new research questions. Several stakeholders (journals, funders, institutions) now require Open Data, and there is an increasing number of infrastructural solutions to facilitate data sharing. One crucial prerequisite for opening data, as noted earlier, is that the data is made FAIR (www.go-fair.org). Data that has been made FAIR can, but may not necessarily, be opened afterwards.

In the analysis step of the empirical cycle, RRP are promoted via initiatives related to statistical innovations and reproducible workflows. Statistical innovations include rethinking thresholds for statistical significance, advanced analyses, and promoting statistical frameworks (e.g., frequentist vs. Bayesian statistics and effect size estimation). An example is the development of the free statistical software such as JASP (jasp.stats.org) and R (www.r-project.org). Most of these innovations focus on increasing the reliability and robustness of statistical conclusions. Other initiatives focus on increasing the reproducibility of workflows. A study is reproducible2 if a reanalysis of the data, following the reported procedures, leads to the same results. Reproducibility requires that data should be available and that procedures should be clear. Reproducibility can be greatly improved if researchers manage their data according to the FAIR data principles. Researchers can share their data via online data repositories, such as the Dataverse Project, Figshare, or Dryad. Furthermore, platforms such as the Open Science Framework (www.osf.io) can serve as a data repository. Another initiative that highlights the importance of reproducibility is the introduction of “Verification Reports” (Chambers, 2020) in the journal Cortex. The sole purpose of this new type of publication is to assess the reproducibility of previously published papers.

To guarantee reproducibility of systematic reviews and meta-analyses, more and more journals require authors to follow specific reporting guidelines (see also policy), such as the CONSORT statement for randomized controlled trials, and PRISMA or MARS for meta-analyses and/or systematic reviews. These guidelines serve as checklists for authors to make sure that all important information is included in their article. They result in more standardized reporting, which makes it easier for researchers to find the information they need, and to verify results and conclusions.

The next step of the empirical cycle is dissemination. In most cases, this entails publishing a study in a peer-reviewed journal. Many initiatives promoting RRP appear to focus on this step. Consider, for instance, initiatives that shift the focus to prepublication peer reviews, by promoting preprints, published online in preprint archives (e.g., Biorxiv.org). Yet other initiatives aim to revise the publication system completely via innovative online platforms (e.g., the megajournal F1000; www.F1000research.com).

The final step in the empirical cycle is the correction of research reporting. Correction can take place before/during dissemination through peer review. Correction can also take place after dissemination through replication (for initiatives linked to replication, see above), and errata or retractions (see, e.g., the popular blog Retraction Watch [www.retractionwatch.com] that tracks scientific retractions and covers cases of misconduct).

Themes That Warrant Future Research

Before we start our discussion on themes that warrant future research, we would like to stress that we consider responsible research as the product of applying the overarching scientific frameworks of open science, research integrity, and societal impact across the scientific system and in the empirical cycle. This is a concern for the entire research enterprise, and for all scientific domains, even though it may manifest itself differently in diverse research disciplines. Currently, these differences are poorly understood. The traditional focus on several disciplines is reflected in the list of abovementioned initiatives, stemming mainly from the biomedical and some of the social sciences, leaving the natural sciences, engineering, humanities, and the other social sciences understudied. Some studies have already indicated how integrity challenges emerge differently across research disciplines (Haven et al., 2019a, 2019b; Horbach & Halffman, 2019), but future studies should amplify the current initiatives towards other disciplinary domains and put more efforts into studying and resolving these differences.

While the above list of current initiatives regarding RRP consists of laudable efforts, and sketches a hopeful picture, much remains to be elucidated. Many of the initiatives described are relatively understudied and there remains a paucity of evidence on their effectiveness and potential impact. Few controlled studies have compared the differences between new initiatives with existing frameworks, as well as the potential hurdles and consequences of implementing the initiatives in different research settings.3 This could be a theme that warrants future research in itself.

In this section, we will outline some of the most prominent knowledge gaps that might be effectively addressed in future funding programs addressing RRP.

Our themes emerged from a systematic approach. First, we created an overview of current initiatives concerning RRP across different scientific disciplines. We clustered these initiatives in a thematic map to identify any gaps. Based on this overview, we created a set of underrepresented themes in the current research on RRP and condensed this set into a list of six major themes that we believe lack a robust evidence base and that can be effectively addressed in future funding programs. We asked 16 experts in the field of meta-research and/or RRP for feedback on the identified themes and their rationale. Below, you will find a summary of the six most important themes. The full list of themes can be found in Supplemental file Appendix III. More information on the methodology section can be found in the Supplemental file Appendix I.

First, we believe the responsible evaluation of research and researchers is a crucially understudied issue. The current evaluation criteria are thought to create a perverse incentive structure, are unidimensionally focused on the “bean counting” of publications in high-impact and branded journals, and may nudge researchers unconsciously into QRP with a focus on publishing as many articles as possible, instead of getting it right. How these incentive structures actually influence research practices is still unknown and how the research evaluation criteria should be reformed needs further research. The urgency for this research was also voiced during the 6th World Conference on Research Integrity (WCRI2019) in Hong Kong in June 2019. Research on this theme may also shed light on the obstacles to address. Finding out the reason for the lack of uptake could help form better policies and incentives. Furthermore, we recommend formal consideration by funders of ways to offset risks to early-career researchers in engaging in open research practices and ensuring proper reward and incentive structures are in place (e.g., grants and fellowship schemes that take into account commitment to open practices).

Second, we believe that the influence of open science and transparency on RRP requires further study. The open science movement is quickly adopting various initiatives that help to create a more open and transparent science (sharing data, preregistering studies, and openness in peer review and in publishing). While researchers may agree that open science can benefit RRP, there is much misunderstanding on what “open science” actually means. More research is thus needed to address the possible benefits and disadvantages of this trend in the research community. We are also still unable to determine if preregistration and the sharing of data are reliable determinants that reduce publication and outcome reporting bias. A shift to open science will furthermore require a coherent operationalization of the many open practices across the empirical cycle so that researchers can work effectively and efficiently in an open manner. The relationship between open science and innovation, especially in collaborations between academic institutions and industry where openness can hamper innovation, is lastly understudied and needs further scrutiny. A shift to open science and increased transparency should also involve reducing the bias of solely publishing positive research: negative and null results are just as important for advancing science.

Third, we identified a lack of knowledge on research on responsible mentoring, supervision, and role modeling. Mentors and supervisors play a key role in establishing a responsible research climate for early-career researchers. As such, they have an enormous influence on the next generation of researchers and thus on the (future) practice of responsible research. Furthermore, most of the education in crucial phases of an academic career happens through socialization processes, largely influenced by mentors, supervisors, and role models. However, this pivotal role is not always fully acknowledged and it is seldom thoroughly reviewed in academic research settings. The role of bad supervision or even harassment can furthermore have detrimental effects on research and researchers. We believe that additional research may shed light on this, which will eventually assist in mentors and role models establishing a research culture that fosters RRP.

Fourth, the effect of education and training on RRP needs further study. The role of education and training is commonly proposed as one of the main interventions to foster RRP. However, the research on its effectiveness is inconclusive and restricted (Marusic et al., 2016). Besides, hardly any educational programs are organized with a focus on responsible research in senior researchers. Empirical research should assess what type of education and training is successful and describe what potentially or definitively can make a difference.

Fifth, future research programs should focus on increasing the reproducibility of their workflow and analyses. We need to investigate how we can facilitate and incentivize sharing data, analysis scripts, protocols, and other relevant materials that are necessary to retrace the steps that the original researchers took to reach their conclusion. More research could expose possible factors that limit researchers to reflect on the reproducibility of their results and make people aware that reproducibility, at least in many research disciplines and cultures, is one of the pillars of RRP.

Finally, responsible and fair peer review is currently understudied and does not receive sufficient attention in research funding programs. Peer review is often considered to be one of the cornerstones of academic research. Currently, it is still largely a black box at the risk of conflicts of interest, unfair procedures, serendipity, and inconsistency. This leaves questions such as: How can we improve peer review? What role can transparency play? Do we know enough about peer review to propose novel strategies and interventions that can make peer review more reliable? Can responsible peer review reduce publication bias? What can automated software do to alleviate the enormous pressure on the peer review system? More research can detect potential flaws and can search for novel techniques that help us to improve the peer review process, ultimately making it more efficient and trustworthy.

For all these themes we believe that it is crucial to acknowledge the epistemic, methodological, gender, and cultural diversity in research in order to comprehensively tackle issues of QRP. In this light, we think that different paradigms should be open and transparent and include alternative perspectives to look further, to expand thinking and to initiate collaborations with other paradigms. Relating to researchers with diverse foci in their solution approach to breaches of research integrity, for example, those focusing on methodological fixes to enhance reproducibility and replicability, and others focusing on open science initiatives (Murphy et al., 2020), our research agenda emphasizes that those approaches should be harmonized and should move away from the one-size-fits-all solutions. To move away from one-size-fits all approaches to solving these issues, we need to better understand which QRP are prevalent in the different academic disciplines and which RRP are most effective in these disciplines. For the themes mentioned above, this means for example that training, mentoring, and supervision should be tailored to the specific needs of a community, allowing different approaches across different disciplines and academic ranks.

Conclusion

In conclusion, we aimed to systematize and categorize the initiatives that foster and study RRP. We have mapped these initiatives and subsequently identified gaps of knowledge and underrepresented themes in RRP that we believe require further exploration. This can serve as a starting point for research agendas on RRP.

We have identified six main themes for further attention: responsible evaluation of research and researchers; the influence of open science and transparency on RRP; research on responsible mentoring, supervision, and role modeling; the effect of education and training on RRP; checking for reproducibility; and finally, responsible and fair peer review.

The themes that we find underrepresented are broad areas of research that focus on the level of the scientific system, rather than the microlevel of individual researchers’ research practices and the associated QRPs. Some current initiatives are already gathering empirical evidence to start filling these gaps. However, we do feel that future funding programs should take these gaps into account in order to bring the field of research on research integrity a step further.

We lastly recognize the complexity of the scientific ecosystem as well as diverging interests of stakeholders within that ecosystem. Some stakeholders have nonscientific, for example, professional, financial, or commercial incentives for maintaining the status quo and will not necessarily be open to change. This includes academic publishers that generate their revenue from subscription and paid open access models, university libraries that aim to secure the best access and publishing deals, and senior academics who owe their positions and status to the current recognition and rewards structure. The road to improving RRP should ultimately involve alignment across all key RRP stakeholders, which requires a sensitivity to stakeholders’ background and interests and requires a unique and theme-specific approach across our emerging themes that warrant further study.

In conclusion, we feel that there is an urgent need for more research on research and research practices. While the focus in our empirical material has been slightly biased towards the Dutch context, we feel it sufficiently represents the global state of affairs. Specifically, studies of RRP and how to foster them require additional support in a global context. Both the current academic debate on these issues, as well as wider societal concerns, are indicative of this urgency. We moreover believe that current initiatives have made important steps in the right direction already, and with sufficient support from various stakeholders, more progress can be made. We hence call upon all relevant stakeholders to actively engage in efforts to further support studies on RRP and initiatives to foster them.

Supplemental Material

sj-docx-1-jre-10.1177_15562646211018916 - Supplemental material for Towards a Research Agenda for Promoting Responsible Research Practices

Supplemental material, sj-docx-1-jre-10.1177_15562646211018916 for Towards a Research Agenda for Promoting Responsible Research Practices by Joeri K. Tijdink, Serge P.J.M. Horbach, Michèle B. Nuijten and Gareth O’Neill in Journal of Empirical Research on Human Research Ethics

Acknowledgments

The authors thank the experts Fenneke Blom, Lex Bouter, Chris Chambers, Roel Freriks, Sonja Jerak-Zuiderent, Jean Philippe de Jong, Padmini Khedoe, Daniël Lakens, Ana Marusic, Lucille Mattijssen, Frank Miedema, David Moher, Peter Novitzky, Sarah de Rijcke, Jeroen de Ridder, Simine Vazire, and Eric-Jan Wagenmakers who have provided us with valuable feedback that has significantly improved this manuscript.

Author Biographies

Joeri K. Tijdink is an assistant professor at the Ethics, Law and Humanities department at Amsterdam UMC, location VUmc and at the philosophy department at VU University in Amsterdam. He obtained his PhD in 2016 with the thesis entitled Publish & Perish; research on research and researchers. Currently he is involved in several research projects that investigate research integrity, publication pressure, open science, RI-governance, research culture, and responsible research practices and is the author of the self help guide: Scientist on the Sofa; how to survive at the university. He also works as a clinical psychiatrist.

Serge P.J.M. Horbach works as a postdoc at the Danish Centre for Studies in Research and Research Policy, Aarhus University, and the Centre for Science and Technology Studies, Leiden University. His research focuses on research integrity and the scientific publication system, working in a sociology of science and science and technology studies tradition. Currently he studies how erroneous or fraudulent research may enter the scientific literature and how editorial and peer review practices may be organized to prevent this. Wider research interests include scientific (e)valuation practices, open science, and the role of organizations in fostering research integrity.

Michèle B. Nuijten is an assistant professor in Methodology and Statistics at Tilburg University, where she is part of the Meta-Research Center. Michèle obtained her Bachelor's (2011) and Master's (2012) in Psychological Methods at the University of Amsterdam. In her PhD thesis (2018, Tilburg University), she focused on meta-scientific studies of problems and solutions in psychological science. Among other things, she codeveloped the tool “statcheck”: a spellchecker for statistics. Currently, her work focuses on reproducibility and replicability in psychology.

Gareth O’Neill is a doctoral candidate in linguistics at Leiden University and has represented early-career researchers extensively in the Netherlands and in Europe. He is former president of the European Council of Doctoral Candidates and Junior Researchers (Eurodoc) and has been active as contributor to the Dutch National Plan for Open Science (NPOS), expert on Open Science for the European Commission, and advisory board member for the FOSTER Plus project. He is currently an ambassador for Plan S for cOALition S and adviser for the European Open Science Cloud (EOSC).

1.

Researchers with a qualitative focus typically produce a systematic review involving the collection, critical appraisal, and synthesis of all relevant studies.

2.

Please note that we use the term reproducibility to refer to analytical reproducibility, which is not the same as replicability. Replication and replicability refer to a situation in which new data is collected to assess whether the original results can be replicated in a new sample.

3.

However, we do note the rise of “meta-scientific” studies and even entire meta-research groups (e.g., the Meta-Research Center (www.metaresearch.nl) at Tilburg University and METRICS at Stanford University; metrics.stanford.edu) that study exactly these questions.

Footnotes

Declaration of Conflicting Interests: This paper was the result of a request by the Netherlands Organization of Health Research and Development (ZonMw) to shed light on the international debate on RRP addressing three key issues: (1) a sketch of the current international discussion on RRP, (2) an overview of current initiatives and already obtained results regarding RRP, and (c) an overview of potential future needs for research on RRP. All authors have received funding to write this report. The final report for ZonMw can be found here: https://www.zonmw.nl/fileadmin/zonmw/documenten/Fundamenteel/BVO/PositionPaper_Promoting_Responsible_Research_Practices__2019_.pdf.

Funding: The authors disclosed receipt of the following financial support for the research, authorship, and/or publication of this article: This work was supported by The Netherlands Organisation for Health Research and Development ZonMw (grant number not applicable).

ORCID iD: Joeri K. Tijdink https://orcid.org/0000-0002-1826-2274

Supplemental Material: Supplemental material for this article is available online.

References

  1. Abma R. (2013). De publicatiefabriek: Over de betekenis van de affaire stapel. Uitgeverij Vantilt. [Google Scholar]
  2. Aubert Bonn N., Pinxten W. (2019). A decade of empirical research on research integrity: What have we (not) looked at? Journal of Empirical Research on Human Research Ethics, 14(4), 338-352. 10.1177/1556264619858534 [DOI] [PubMed] [Google Scholar]
  3. Begley C. G., Ioannidis J. P. A. (2015). Reproducibility in science. Circulation Research, 116(1), 116-126. 10.1161/CIRCRESAHA.114.303819 [DOI] [PubMed] [Google Scholar]
  4. Biagioli M., Kenney M., Martin B. R., Walsh J. P. (2019). Academic misconduct, misrepresentation and gaming: A reassessment. Research Policy, 48(2), 401-413. 10.1016/j.respol.2018.10.025 [DOI] [Google Scholar]
  5. Bouter L. M. (2015). Commentary: Perverse incentives or rotten apples? Accountability in Research, 22(3), 148-161. 10.1080/08989621.2014.950253 [DOI] [PubMed] [Google Scholar]
  6. Bouter L. M., Tijdink J., Axelsen N., Martinson B. C., ter Riet G. (2016). Ranking major and minor research misbehaviors: Results from a survey among participants of four world conferences on research integrity. Research Integrity and Peer Review, 1(1), 17. 10.1186/s41073-016-0024-5 [DOI] [PMC free article] [PubMed] [Google Scholar]
  7. Bramstedt K. A. (2020). Luxembourg’s approach to research integrity during the COVID-19 pandemic. Accountability in Research, 27(6), 396-400. 10.1080/08989621.2020.1778473 [DOI] [PubMed] [Google Scholar]
  8. Callaway E. (2011). Report finds massive fraud at Dutch universities. Nature, 479(7371), 15-15. 10.1038/479015a [DOI] [PubMed] [Google Scholar]
  9. Chambers C. D., Dienes Z., McIntosh R. D., Rotshtein P., Willmes K. (2015). Registered reports: Realigning incentives in scientific publishing. Cortex, 66(May), A1-A2. 10.1016/j.cortex.2015.03.022 [DOI] [PubMed] [Google Scholar]
  10. Chambers C. D. (2020). Verification reports: A new article type at Cortex. Cortex, 129(Aug), A1-a3. 10.1016/j.cortex.2020.04.020 [DOI] [PubMed] [Google Scholar]
  11. Davies S. R. (2019). An ethics of the system: Talking to scientists about research integrity. Science and Engineering Ethics, 25(4), 1235-1253. 10.1007/s11948-018-0064-y [DOI] [PubMed] [Google Scholar]
  12. De Groot A. D. (1961). Methodologie. Grondslagen van onderzoek en denken in de gedragswetenschappen. 's Gravenhage: Mouton. https://lib.ugent.be/nl/catalog/rug01:000464674 [Google Scholar]
  13. European code of conduct for research integrity. Revised edition. (2017). Berlin: All European Academies. [Google Scholar]
  14. Fanelli D. (2009). How many scientists fabricate and falsify research? A systematic review and meta-analysis of survey data. PLOS ONE, 4(5), e5738. 10.1371/journal.pone.0005738 [DOI] [PMC free article] [PubMed] [Google Scholar]
  15. Fanelli D., Costas R., Larivière V. (2015). Misconduct policies, academic culture and career stage, not gender or pressures to publish, affect scientific integrity. PLOS ONE, 10(6), e0127556. 10.1371/journal.pone.0127556 [DOI] [PMC free article] [PubMed] [Google Scholar]
  16. Fleming T. R., Labriola D., Wittes J. (2020). Conducting clinical research during the COVID-19 pandemic: Protecting scientific integrity. JAMA, 324(1), 33-34. 10.1001/jama.2020.9286 [DOI] [PubMed] [Google Scholar]
  17. Haven T. L., Bouter L. M., Smulders Y. M., Tijdink J. K. (2019a). Perceived publication pressure in Amsterdam: Survey of all disciplinary fields and academic ranks. PLOS ONE, 14(6), e0217931. 10.1371/journal.pone.0217931 [DOI] [PMC free article] [PubMed] [Google Scholar]
  18. Haven T. L., Tijdink J. K., Martinson B. C., Bouter L. M. (2019b). Perceptions of research integrity climate differ between academic ranks and disciplinary fields: Results from a survey among academic researchers in Amsterdam. PLOS ONE, 14(1), e0210599. 10.1371/journal.pone.0210599 [DOI] [PMC free article] [PubMed] [Google Scholar]
  19. Haven T. L., Tijdink J. K., Martinson B. C., Bouter L. M., Oort F. J. (2021). Explaining variance in perceived research misbehavior – results from a survey among academic researchers in Amsterdam. Res Integr Peer Rev, 6, 7. 10.1186/s41073-021-00110-w [DOI] [PMC free article] [PubMed] [Google Scholar]
  20. Horbach S. P. J. M., Halffman W. (2019). The extent and causes of academic text recycling or ‘self-plagiarism’. Research Policy, 48(2), 492-502. 10.1016/j.respol.2017.09.004 [DOI] [Google Scholar]
  21. John L. K., Loewenstein G., Prelec D. (2012). Measuring the prevalence of questionable research practices with incentives for truth telling. Psychological Science, 23(5), 524-532. 10.1177/0956797611430953 [DOI] [PubMed] [Google Scholar]
  22. KNAW, N., NWO, TO2-federatie, Vereniging Hogescholen, VSNU. (2018). Nederlandse gedragscode wetenschappelijke integriteit. DANS. 10.17026/dans-2cj-nvwu [DOI]
  23. Kupferschmidt K. (2018). More and more scientists are preregistering their studies. Should you? Science Magazine. 10.1126/science.aav4786 [DOI] [Google Scholar]
  24. Lancet T. (2013). Integrity in research collaborations: The Montreal statement. The Lancet, 382(9901), 1310. 10.1016/S0140-6736(13)62126-1 [DOI] [PubMed] [Google Scholar]
  25. Latour B., Woolgar S. (1979). Laboratory life: The construction of scientific facts. Princeton University Press. [Google Scholar]
  26. Martinson B. C., Crain A. L., Anderson M. S., De Vries R. (2009). Institutions’ expectations for researchers’ self-funding, federal grant holding, and private industry involvement: Manifold drivers of self-interest and researcher behavior. Academic Medicine, 84(11), 1491-1499. https://journals.lww.com/academicmedicine/Fulltext/2009/11000/Institutions__Expectations_for_Researchers_.19.aspx [DOI] [PMC free article] [PubMed] [Google Scholar]
  27. Martinson B. C., Crain A. L., De Vries R., Anderson M. S. (2010). The importance of organizational justice in ensuring research integrity. Journal of Empirical Research. on Human Research Ethics, 5(3), 67-83. 10.1525/jer.2010.5.3.67 [DOI] [PMC free article] [PubMed] [Google Scholar]
  28. Marusic A., Wager E., Utrobicic A., Rothstein H. R., Sambunjak D. (2016). Interventions to prevent misconduct and promote integrity in research and publication. The Cochrane Database of Systematic Reviews, 4(4), Mr000038. 10.1002/14651858.MR000038.pub2 [DOI] [PMC free article] [PubMed] [Google Scholar]
  29. Mehra M. R., Ruschitzka F., Patel A. N. (2020). Retraction—hydroxychloroquine or chloroquine with or without a macrolide for treatment of COVID-19: A multinational registry analysis. The Lancet, 395(10240), 1820. 10.1016/S0140-6736(20)31324-6 [DOI] [PMC free article] [PubMed] [Google Scholar]
  30. Mejlgaard N., Bouter L. M., Gaskell G., Kavouras P., Allum N., Bendtsen A. K., Charitidis C. A., Claesen N., Dierickx K., Domaradzka A., Reyes Elizondo A., Foeger N., Hiney M., Kaltenbrunner W., Labib K., Marušić A., Sørensen M. P., Ravn T., Veltri G. A. (2020). Research integrity: Nine ways to move from talk to walk. Nature, 586(7829), 358-360. 10.1038/d41586-020-02847-8 [DOI] [PubMed] [Google Scholar]
  31. Moher D., Bouter L., Kleinert S., Glasziou P., Sham M. H., Barbour V., … Dirnagl U. (2020). The Hong Kong principles for assessing researchers: Fostering research integrity. PLOS Biology, 18(7), e3000737. 10.1371/journal.pbio.3000737 [DOI] [PMC free article] [PubMed] [Google Scholar]
  32. Munafò M. R., Nosek B. A., Bishop D. V. M., Button K. S., Chambers C. D., Percie du Sert N., … Ioannidis J. P. A. (2017). A manifesto for reproducible science. Nature Human Behaviour, 1(1), 0021. 10.1038/s41562-016-0021 [DOI] [PMC free article] [PubMed] [Google Scholar]
  33. Murphy M. C., Mejia A. F., Mejia J., Yan X., Cheryan S., Dasgupta N., … Pestilli F. (2020). Open science, communal culture, and women’s participation in the movement to improve science. Proceedings of the National Academy of Sciences, 117(39), 24154. 10.1073/pnas.1921320117 [DOI] [PMC free article] [PubMed] [Google Scholar]
  34. Nosek B. A., Alter G., Banks G. C., Borsboom D., Bowman S. D., Breckler S. J., … Yarkoni T. (2015). Promoting an open research culture. Science, 348(6242), 1422-1425. 10.1126/science.aab2374 [DOI] [PMC free article] [PubMed] [Google Scholar]
  35. O'Carroll C., et al. (2017). Providing Researchers with the Skills and Competencies they Need to Practise Open Science. Luxembourg: Publications Office of the European Union. 10.2777/121253, https://cdn1.euraxess.org/sites/default/files/policy_library/ec-rtd_os_skills_report_final_complete_2207_1.pdf [DOI] [Google Scholar]
  36. European Commission. (2016). Open Innovation, Open Science, Open to the World. A Vision for Europe. Luxembourg: Publication office of the European Union. 10.2777/061652, http://publications.europa.eu/resource/cellar/3213b335-1cbc-11e6-ba9a-01aa75ed71a1.0001.02/DOC_2 [DOI]
  37. Replication Studies. (2019). https://www.nwo.nl/en/funding/our-funding-instruments/sgw/replication-studies/replication-studies.html
  38. Research on Research. (2020). https://wellcome.org/what-we-do/our-work/research-research
  39. Steneck N. H. (2006). Fostering integrity in research: Definitions, current knowledge, and future directions. Science and Engineering Ethics, 12(1), 53-74. 10.1007/PL00022268 [DOI] [PubMed] [Google Scholar]
  40. Tennant J. P., Ross-Hellauer T. (2020). The limitations to our understanding of peer review. Research Integrity and Peer Review, 5(1), 6. 10.1186/s41073-020-00092-1 [DOI] [PMC free article] [PubMed] [Google Scholar]
  41. Tijdink J. K., Verbeke R., Smulders Y. M. (2014). Publication pressure and scientific misconduct in medical scientists. Journal of Empirical Research on Human Research Ethics, 9(5), 64-71. 10.1177/1556264614552421 [DOI] [PubMed] [Google Scholar]
  42. van Dalen H. P., Henkens K. (2012). Intended and unintended consequences of a publish-or-perish culture: A worldwide survey. Journal of the American Society for Information Science and Technology, 63(7), 1282-1293. 10.1002/asi.22636 [DOI] [Google Scholar]
  43. van der Heyden M. A. G., van de Ven T., Opthof T. (2009). Fraud and misconduct in science: The stem cell seduction: Implications for the peer-review process. Netherlands Heart Journal: Monthly Journal of the Netherlands Society of Cardiology and the Netherlands Heart Foundation, 17(1), 25-29. 10.1007/BF03086211 [DOI] [PMC free article] [PubMed] [Google Scholar]
  44. Wagenmakers E. J., Wetzels R., Borsboom D., van der Maas H. L., Kievit R. A. (2012). An agenda for purely confirmatory research. Perspectives. on Psychological Science, 7(6), 632-638. 10.1177/1745691612463078 [DOI] [PubMed] [Google Scholar]
  45. Wicherts J. M. (2017). The weak spots in contemporary science (and how to fix them). Animals (Basel), 7(12), 90. 10.3390/ani7120090 [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

sj-docx-1-jre-10.1177_15562646211018916 - Supplemental material for Towards a Research Agenda for Promoting Responsible Research Practices

Supplemental material, sj-docx-1-jre-10.1177_15562646211018916 for Towards a Research Agenda for Promoting Responsible Research Practices by Joeri K. Tijdink, Serge P.J.M. Horbach, Michèle B. Nuijten and Gareth O’Neill in Journal of Empirical Research on Human Research Ethics


Articles from Journal of Empirical Research on Human Research Ethics are provided here courtesy of SAGE Publications

RESOURCES