Skip to main content
Wellcome Open Research logoLink to Wellcome Open Research
. 2023 Jan 27;7:82. Originally published 2022 Mar 9. [Version 2] doi: 10.12688/wellcomeopenres.17715.2

An overview of innovations in the external peer review of journal manuscripts.

Helen Buckley Woods 1,a, Johanna Brumberg 2, Wolfgang Kaltenbrunner 3, Stephen Pinfield 1, Ludo Waltman 3
PMCID: PMC9984734  PMID: 36879926

Version Changes

Revised. Amendments from Version 1

We have updated the title and abstract to more accurately reflect the scope of the article. We have also added a limitations section which states the focus of the review, namely on published literature and on innovations in review of journal manuscripts, and on innovations in the external peer review process, not including innovations in the editorial process. In the methods section, we have provided a citation to give a definition of the different review types included in the article. We have removed the term ‘meta-summary’ and defined the article as an overview (review of reviews) and provided a citation to define this term in the methods section. We have revisited each included study and included information on quality assessment where it was found in the included articles. We have provided a new, simplified, and accurate flow chart. We have refined our definitions of ‘peer review’ and ‘innovation’ to be more precise and better aligned with their usage in the article.

Abstract

Background: There are currently numerous innovations in peer review and quality assurance in scholarly publishing. The Research on Research Institute conducted a programme of co-produced projects investigating these innovations. This literature review was part of one such project ‘Experiments in peer review’ which created an inventory and framework of peer review innovations. The aim of this literature review was to aid the development of the inventory by identifying innovations in the external peer review of journal manuscripts reported in the scholarly literature and by providing a summary of the different approaches. This did not include interventions in editorial processes.

Methods: This review of reviews is based on data identified from Web of Science and Scopus limited from 2010 to 2021. A total of 291 records were screened, with six review articles chosen for the focus of the literature review. Items were selected that described approaches to innovating peer review or illustrated examples.  

Results: The overview of innovations are drawn from six review articles. The innovations are divided into three high-level categories: approaches to peer review, reviewer focussed initiatives and technology to support peer review with sub-categories of results presented in tabular form and summarised. A summary of all innovations found is also presented.

Conclusions: From a simple synthesis of the review authors’ conclusions, three key messages are presented: observations on current practice; authors’ views on the implications of innovations in peer review; and calls for action in peer review research and practice.

Keywords: peer review, scholarly publishing, innovation, meta-summary, review of reviews

Introduction

The Research on Research Institute (RoRI) conducted a number of co-produced projects with academic publishers and scholarly communication service providers. These projects investigated current experiments and innovations in quality assurance and peer review in scholarly publishing. This literature review is part of one such project entitled ‘Experiments in Peer Review’.

The ‘Experiments in Peer Review’ project aimed:

  • to identify, analyse, and evaluate current innovations in peer review and other forms of quality control/assurance of research outputs

  • to assess their potential impacts on scholarly communication in particular and the research environment in general.

The first phase of this project was to create an inventory and framework of experiments in peer review carried out by publishers and other scholarly communication organisations. The inventory is based on a widely distributed survey of scholarly publishers designed to retrieve information on current innovations at grass roots level ( Kaltenbrunner et al., 2022). The purpose of this literature review is to provide context and aid the development of the inventory. To do so, we identify publications reporting innovations or experiments with peer review in scholarly publishing and we create a summary of the different types of initiative as identified in the literature.

Definitions, framing and limitations

In this review the definition of peer review reflects the use of the term in the literature selected. That is an inclusive and broad interpretation of the phrase to include many aspects of evaluation and quality assessment. So, whilst a ‘peer’ is commonly understood as an individual researcher with significant expertise and interest in a given field, ‘peer review’ includes the actions of other stakeholders or their agents within the system such as copyeditors and formatters, artificial intelligence (AI) software, members of the public, patients, advocates and lobbyists. Peer review, used in this broad sense, also encapsulates activities designed to ensure research integrity such as plagiarism checks and monitoring compliance with data management policies. Similarly, peer review also includes informal responses, questions and comments posted on social media, pre-print servers, e-journals or other places online in response to a given research output. These types of informal responses were found in a study examining disciplinary knowledge production ( Woods, 2018), where examples of researchers’ peer review practice were identified. The results of group interviews with researchers in applied fields identified that several other types of ‘peer review’ which did not involve scientific experts, but members of the public, those from other professions, advocacy and lobbying groups were also common occurrences. Over 20 years ago Barnett (2000) spoke about knowledge becoming a commodity tested by consumer reaction. Particularly in applied fields, this is coming to fruition, with this type of research having a greater number of consumers, invoking greater and different types of reaction and review ( Hoepner, 2019).

The scope of this review is restricted to peer review in scholarly publishing, although similar observations about the benefits and limitations of peer review are found in publications concerned with other parts of the research system. One example is the onerous nature of peer review, which is a factor also associated with peer review in research funding ( Bendiscioli & Garfinkel, 2021), and research assessment ( Wilsdon, 2015), in addition to scholarly publishing ( Smith, 2006). It is also worth noting that this literature review is concerned with what people are doing, and what the innovations entail rather than why they are doing it. The motivations giving rise to innovation, for example to reduce bias or achieve greater efficiency are peripheral to the purpose of this study.

This literature review describes innovations. In a similar treatment to the interpretation of the term ‘peer review’ we adopt a broad interpretation of the term ‘innovation’. The definition proposed by Rogers (2003) is fitting for this review: ‘an idea, practice, or object that is perceived as new by an individual or other unit of adoption’ (p. 26). That is, the status of something as an innovation does not rest on the date of its inception, it could be in existence for 4 hours or 40 years, what makes it an innovation is if the practice or idea is new to those who are recommending or suggesting its use in their particular context. Within the academic publishing industry, what may be considered an innovation in one organisation, such as reviewers and authors being blind to each other’s identity, would no longer be considered an innovation in another. Broadly speaking, an innovation can be implemented in different ways within an organisation: first, by intervening in usual practice; second, by intervening on a smaller scale, in one area of work, to test something out before implementing it more widely; or third, in setting up a separate innovative project or initiative, outside existing processes. In this review all types of innovations are included, and the review is agnostic to breadth of implementation. All types of innovations are captured that were reported in the included studies, this includes potential models of peer review which are untried.

Like any scholarly work, this literature review is limited by its scope and data sample. It is an overview of approaches to external peer review of journal manuscripts. It is based on searches of academic bibliographic databases and does not include evidence identified through grey literature searches. The source documents are six literature reviews and the article provides an overview of the peer review approaches stated in these reviews. Reviewing these overview articles enabled generic categories to be created to encompass broad types of intervention, such as training or other support for peer reviewers. In line with our aims, not every individual example or implementation of these broad types of initiative was found in the included papers. Also, given the data sources, some approaches to peer review may have been omitted, such as volunteering to review.

In addition, the review focuses on the peer review process involving external reviewers, rather than the editorial side of the review process, hence examples of innovations of this nature are not included.

This review of reviews does not contain a quality assessment of the included review papers. Where a quality assessment of primary studies was performed within the included reviews, we have indicated this in the descriptions of the included studies.

The next section will describe the methodology used in this literature review and how we will present the results. This is followed by the results themselves in which we classify different types of innovations. We conclude with some discursive reflections on the current situation.

Methodology

Overall approach

This review was undertaken to set the context and inform an empirical study ( Kaltenbrunner et al., 2022). To complete this task, it was not necessary to identify every publication discussing peer review, rather to capture as many different forms of peer review discussed as possible. With this in mind we did not limit the search to a particular publication type and our search results included several review articles. On closer inspection of these articles, it was clear that they covered all the peer review innovations identified from screening and coding the results of the literature search. This is with the possible exception of modelling or scientometric studies examining aspects of the peer review system ( Ortega, 2017; Ragone et al., 2013) or proposing a framework for best practice for academic publishing ( Waters et al., 2020) or audit of publishing processes ( Crewe, 2020). However, these papers were slightly out of scope for the remit of the empirical work. Therefore, the decision was made to present the data through the organising structure of six recent literature reviews on the topic, as an overview ‘review of reviews’ ( Booth et al., 2021). That is to say, the synthesised findings of each review are presented and combined to create an overview of peer review types, rather than describing disaggregated findings from the primary studies included within each review. This was a pragmatic decision and was not intended to detract from primary studies such as Walker & Rocha da Silva (2015) which presents a thorough and useful overview of peer review types, which is cited by Tennant et al. (2017); Horbach & Halffman (2018); Tennant (2018); and Barroga (2020) in this overview. The articles included in this review comprise numerous review types, namely, systematic review, meta-analysis, narrative review, state of the art review, and narrative summary. See Sutton et al. (2019) for definitions of these review types.

Of high importance to this review and RoRI’s work with scholarly publishers is the Reimagine Review registry set up by Accelerating Science and Publication in biology (ASAPbio). The projects included in this registry provide live examples of the types of peer review innovations summarised in this review, such as post-publication review and pre-print review. More details of the registry are provided in Box 1.

Box 1. ASAPbio’s Reimagine Review.

Box 1. ASAPbio’s Reimagine Review

Reimagine Review is a registry of peer review experiments. As of January 6, 2023, it includes 62 registered projects.

The registry is presented as a searchable database. The user is able to filter the records in various ways including type of output, who initiates the review, whether the reviews are stand alone or linked to a specific publication, the level of transparency or ‘openness’, whether a decision is made at the end of the review (to publish or not), discipline, format of reviewing (such as comments or scores) and some characteristics of the process (for example, if professional editors are used, if comments are moderated). Alternatively, the top page enables authors to choose by output: pre-print, articles already accepted for publication by a journal, privately shared manuscripts and finally ‘other outputs’ such as protocols, data sets etc. This enables authors to choose the most appropriate service that fits with their needs. The types of innovation featured in the Reimagine Review inventory ( ASAPbio, 2021) such as post-publication review have been included in this review.

Data collection

A literature search was conducted in Web of Science and Scopus to identify relevant papers using synonyms for ‘peer review’ and ‘innovation’, records were screened to exclude studies that did not describe a type of innovation in peer review, or gave an example of a specific innovation. Further records were retrieved using citation searches of the remaining relevant records. Following these search and screening iterations, 68 records were initially included. The full texts of these records were retrieved and screened and six review articles were chosen for the focus of the literature review. Searches were conducted in January 2021, limited from 2010 and no study filters were applied. An example search strategy is presented below.

Example search strategy

Web of Science via Clarivate. TI=("peer reviewing" OR "peer reviewer" OR "peer review ") AND TI=(experiment* or pilot* or improvement* or innovation* or solution* or initiative* or intervention*)

Timespan: 2010–2021. Indexes: SCI-EXPANDED, SSCI, A&HCI, CPCI-S, CPCI-SSH, BKCI-S, BKCI-SSH, ESCI.

Study selection / coding

Search results were initially downloaded to EndNote (X9.3.2) to facilitate de-duplication after which study selection was completed in the Rayyan software Ouzzani et al. (2016). This allowed easy viewing of decision making by the project team. Initial categories of innovations were developed to include overview articles, types of peer review, reviewer focussed initiatives, technological initiatives and specific uses of peer review (such as use of language or plagiarism). This exercise of developing topic categories aided the organisation of material in the review. As previously stated, several review articles were identified in this process and on closer inspection of these articles, it was clear that they covered all the peer review innovations identified from screening and coding the results of the literature search. The review therefore focussed on six literature reviews in a review of reviews format. Please note an earlier version of this article can be found on SocArXiv (doi: 10.31235/osf.io/qaksd).

Results

Presentation of results

The results are presented using narrative and tabular formats, followed by a summary of the review and conclusions. The results section begins with a description of included studies, followed by a detailed description of innovations in peer review types using three high level categories: approaches to peer review, reviewer focussed initiatives, and technology to support peer review. This includes definitions of each type of innovation extracted from the included studies. A summary table of each innovation and where these have been reported is provided at the end of the review. See Figure 1 for a PRISMA diagram giving details of the search process.

Figure 1. Flow chart based on PRISMA diagram ( Moher et al., 2009).

Figure 1.

Details of included studies

The review includes six overview studies which range from a systematic review of randomised controlled trials to a succinct summary of review innovations. All the studies were valuable in capturing the different types of current innovation in peer review. Each study is described below, presented in chronological order. This is followed by a narrative summary of each type of innovation, followed by a summary table, giving definitions of each innovation and citations to the respective studies where they are found.

Bruce et al. (2016) is a systematic review and meta-analysis published in BMC Medicine which aims to evaluate the effectiveness of innovations to improve the quality of peer review for publications in biomedical science. It includes 22 randomised controlled trials.

The innovations that were evaluated were: reviewer training, addition of a statistical peer reviewer, open peer review (where reviewers’ identity is known), blinded peer review (where reviewers’ or authors’ identity is not known), and innovations to increase the speed of peer review. The unit of randomisation were either peer reviewers or manuscripts depending upon the innovation being assessed, for example comparing the effect of adding a statistical reviewer against the usual process of peer review in a given set of manuscripts. For details of study characteristics see Bruce et al. (2016). The review was based on a search of CENTRAL, MEDLINE (PubMed), Embase, Cochrane Database of Systematic Reviews, and WHO ICTRP databases. To gauge the quality of the included studies, the authors assessed the risk of bias within each RCT using the Cochrane Collaboration’s risk of bias tool ( Higgins et al., 2011). Several scales were used within the included studies, such as final manuscript quality ( Goodman et al., 1994) and quality of the peer review report ( Black et al., 1998) to assess the different outcome measures: final manuscript quality, quality of the peer review report, rejection rate, time spent on peer review, and time spent on the peer review process. The authors found, based on these outcome measures, that compared with standard peer review, reviewer training was not successful in improving the quality of the peer review report and use of checklists by peer reviewers to check the quality of a manuscript did not improve the quality of the final article. However, the addition of a specialised statistical reviewer did improve the quality of the final article and open peer review was also successful in improving the quality of review reports. It did not affect the time reviewers spent on their report. Open peer review also decreased the number of papers rejected. Finally, blinded peer review did not affect the quality of review reports or rejection rates. The authors conclude that there is a lack of evidence on the effectiveness of various peer review procedures, especially given its central role in science.

Tennant et al. (2017) is a narrative review published in F1000 Research. Its process is notable as the review has 33 authors. They are experts in scholarly publishing, and it is these expertise which formed the basis of the review. The authors identified papers through searching databases such as Google Scholar, Scopus, and Library & Information Science Abstracts (LISA). Additionally, there was a lengthy peer review process which was published with the review including reviewers reports and authors responses, a key part of the F1000 model (and itself an innovation in peer review). The authors review the history of peer review, its myriad shortcomings and details the potential solutions that have been developed to date to address these challenges. The pros and cons of newer innovations in peer review such as portable, de-coupled and collaborative peer review are then discussed as well as an examination of different levels of anonymity. They then go on to focus on a number of social web platforms, such as Reddit and expand on the benefits and limitations of each platform considering three criteria: quality control and moderation, certification, and incentive structures. Alongside the review of new innovations, the authors are clear to signal that there are particular benefits of peer review and that it has deep and far-reaching cultural significance within research practices which should not be underestimated. A hybrid model is suggested combining aspects of different platforms. The authors stress that any such innovation cannot succeed without engagement from researchers, but this is in tension with the structure of researcher incentives in the research system. The review concludes with two main points, one to decouple peer review from journal publishing in order to return to what the authors suggest would be a community-led process. Secondly, there is very little evidence to support the uptake of different methods of peer review, so research to measure the effectiveness of these different approaches in achieving different goals of peer review is essential. Three key initiatives are referred to as leaders in this respect: the PEERE initiative, the Research Integrity and Peer Review journal, and the International Congress on Peer Review and Scientific Publication.

Burley (2017) is a narrative summary published in Information Services and Use which reviews new approaches to peer review in scholarly publishing. The author is affiliated to the BMC Group at the Springer Nature Publishing Company. The article does not refer to scholarly literature, but refers to new innovations and ways of working in practice. The article begins by stating the benefits of peer review by key stakeholders. The author then goes on to summarise newer practices such as the increase in double-blind peer review in scientific disciplines. She then discusses open peer review, post-publication peer review and transparent peer review. Finally, initiatives aimed at increased efficiency such as cascading peer review and sectional or partial peer review are discussed. The article concludes by highlighting the increased focus of rewarding and training reviewers by scholarly publishers and learned societies and the overall improvement in reviewer recognition and efficiency, as well as the increase in experiments in transparent peer review.

Horbach & Halffman (2018) is a narrative review published in Research Integrity and Peer Review. It aims to describe current forms of peer review and their implementation and also consider the role and expectations of peer review. The authors present an historical account of peer review and describe methods of peer review to date, including recent technological advances. Four dimensions are identified in peer review innovations: ‘the selection conditions [such as the timing of the review], the identity and access among actors involved, the level of specialisation in the review process, and the extent to which technological tools have been introduced’ (p. 9). The authors then present a typology of peer review characteristics ordered by these four dimensions. This is followed by a discussion on the role of the academic publishing system and expectations of peer review. The authors underline the large diversity of review processes currently used. They also suggest four key expectations for peer review: ‘assuring quality and accuracy of research, establishing a hierarchy of published work, providing fair and equal opportunities to all actors and assuring a fraud-free research record’ (p. 12). The article concludes by highlighting the lack of empirical evidence to test the efficacy of peer review methods and the tensions that exist between what peer review can deliver and what is expected of it, for example its ability to identify fraudulent research or methodological errors. The authors suggest there is a new, additional perspective, in how research knowledge is perceived, fuelled by statistical reviews, post-publication reviews and other innovations, from a library of knowledge, to a set of scientific facts. They suggest that this perception of research as 100% accurate knowledge fuels retractions and rewriting of documents to create seemingly perfect accounts.

Tennant (2018) is a state-of-the-art review published in FEMS Microbiology Letters. Explicit methods are not stated. There are 87 references listed in the document. Commensurate with the aims of a state-of-the-art review ( Grant & Booth, 2009), this article centres on the current status of peer review and its role in scholarly publishing in a digital age. The author states a conceptual difference between peer review as an idea, or ‘a singular ideologue’ (p. 2) and a practice. Open peer review is cited as a return to the original purpose of peer review to be collegial, constructive, to improve arguments and gaps in logic. The author states that peer review now has an additional gatekeeping function, and it is also used by commercial organisations as a selling point. He goes on to summarise the benefits and drawbacks of new models of peer review considering what the job of peer review is and how these functions can be achieved in the future, making better use of the technology we now have. Success would be an open participative model of peer review that is a genuine alternative rather than an add-on to the status quo. However, the author also states that it is difficult to separate the value and prestige that comes from publishing in journals, and this drives particular behaviours and limits uptake of new models. He refers to the ‘penguin effect’ ( Choi, 1997) where the level of perceived risk is greater than the motivation to change. This effect is compounded by the fact that moving away from current practices is often not in sync with the behaviours required for researchers’ job security and career progression. The author concludes by advocating a new framework based on current technological and communication norms by revisiting the core purposes of peer review, links to incentives for researchers and a clear consideration of how all stakeholders fit into any new system.

Barroga (2020) is a narrative review published in the Journal of Korean Medical Science giving an overview of innovations in peer review. The review is based on a search of MEDLINE, Embase and Scopus databases and uses the peer review innovations as stated in Tennant et al. (2017) to organise the material. The author begins by dividing peer review into two types: ‘open peer review’ and ‘traditional peer review’. He compares the features of these two review types, using various criteria such as openness, bias, time and so on. He then takes eleven different innovations stated in Tennant et al. (2017) and compares them against the same features. After comparing a number of web platforms/models of peer review (also from Tennant et al., 2017) he briefly discusses delays to peer review and the issue of anonymity as manifested in ‘blinding’ of reviewers. Finally, the discussion turns to reviewer incentives and training. The author concludes that the increase of innovations has been rapid and there is a lack of evidence as to how effective newer methods of review are in identifying research malpractice. He also suggests that review quality may be compromised where financial incentives are given. He advocates for an honest appraisal of stakeholder’s contribution to the process with reviewer training, core competencies for reviewers and engagement by the research community on this issue being paramount.

See Table 1 for a summary of the general characteristics of the studies included.

Table 1. General characteristics of included studies.

Author / Year / Journal Type of review No of included studies (size
of review)
Bruce et al. / 2016 / BMC Medicine Systematic review
and meta-analysis
22 randomised controlled trials
Tennant et al. / 2017 / F1000 Research Narrative review 300
Burley / 2017 / Information Services and Use Narrative summary N/A
Horbach & Halffman / 2018 / Research Integrity
and Peer Review
Narrative review 119
Tennant / 2018 / FEMS Microbiology Letters State of the art review 87
Barroga / 2020 / Journal of Korean Medical Science Narrative review 48

Description of innovations in peer review

The innovations presented below are divided into three high-level categories: approaches to peer review, reviewer focussed initiatives and technology to support peer review. This section begins with narrative summaries of the different approaches to peer review identified, organised into the following subcategories: open / masked peer review; pre / post publication review; collaboration and decoupling; and focussed and specialised review. The descriptions will be followed by a summary table for each subcategory.

Approaches to peer review.

Open/masked peer review

All the studies in our review mention open peer review (OPR). The data reveals that it is not a clearly defined concept. However, understandings of OPR centre around the (i) identities of the reviewers, editors and authors being known to each other in various combinations or made public and (ii) reviewer reports and authors responses to comments being made public. Burley (2017) makes a distinction between the reviewer reports being signed, which she terms OPR, or not, which she terms transparent peer review. In his state-of-the-art review Tennant (2018) reflects a wider definition of OPR citing Ross-Hellauer (2017), who goes beyond transparency of identity to include other aspects of peer review, including open final version commenting, open pre-review manuscripts and open platforms. Finally, Tennant et al. (2017) refer to a survey of peer review stakeholders (OpenAIRE) which found 122 different definitions to be in use.

Pre/post publication review

The key feature of peer review innovations in this subcategory is the timing of the review. The authors’ definitions reveal a mixture of informal and formal peer review, open and confidential modes, expert and lay commentary. Barroga (2020) (after Tennant et al., 2017) distinguishes between pre and post publication review and commenting. The key difference between review and commenting is who is responding to the publication. In pre and post publication review, this is field experts. In pre and post publication commenting, this amounts to comments or feedback by any interested party, irrespective of their academic or disciplinary credentials. Burley (2017) also describes post-publication review as taking place after publication, in an open manner. In an interesting use of the term ‘post-publication’ Horbach & Halffman (2018) describe post-publication peer review on pre-print servers. This clearly problematises the established use of the word ‘publish’ to mean publication in a journal or monograph. If something has been posted on a pre-print server, then it is published, albeit in a self-published mode, incurring only initial checks for eligibility to be posted on the particular pre-print service.

One other review innovation in this subcategory is the use of registered reports or similar approaches. This is where a research design is evaluated before the research has begun. It typically applies to quantitative empirical research that follows a fixed a priori design. Once a study is designed, the protocol is reviewed, before any data is collected. The value of this method is to reduce questionable research practices, where researchers deviate from their original intention and methodology and indulge in malpractice such as p-hacking and cherry-picking results to create more eye-catching conclusions. Registered reports are championed by the Center for Open Science amongst others.

Collaboration and decoupling

The approaches to peer review in this subcategory reflect a loosening of established roles and have been organised within a summary table ( Table 2) to illustrate this, with the more marked changes presented last. The types of peer review move from increased collaboration and interaction between stakeholders (collaborative review) to reassignment of roles in organisational innovations (decoupled post-publication review). Collaborative review ( Barroga, 2020) is the process where reviewers, editors and other contributors pool their comments to offer one set of consolidated recommendations for authors to address. Horbach & Halffman (2018) present a similar process which they name ‘discussion during review’. In a step further, this same process takes place online so other people can follow the process and add their own comments. Barroga (2020) and Tennant (2018) suggest the additional participants are limited to ‘other interested scientists’ but it is unclear how this can be enforced given the public platform.

Table 2. Approaches to peer review.

    1. Open/masked peer review
Label Author Description from paper
Open peer review Barroga (2020) ‘Open peer review discloses the names of the editors and reviewers handling the paper to the authors.’
Open peer review Bruce et al. (2016) ‘“Open” peer review process, whereby peer reviewers are informed that their name would be revealed to the authors, other
peer reviewers, and/or the public.’
Open peer review Burley (2017) ‘Open peer review with signed reports that are available with the published article, where a response
by the author may be included.’
Transparent peer
review
Burley (2017) ‘Transparent peer review by which the unsigned review reports are made available alongside the published article, and a
response by the author may be included.’
Open review Horbach & Halffman (2018) ‘We use the term ‘open review’ merely to indicate that the identity of the authors and reviewers are mutually known to each
other.’
Open peer review Tennant (2018) ‘It has been diagnosed to refer to seven key aspects of peer review: open identities, open reports, open participation, open
interaction, open pre-review
manuscripts, open final-version commenting and open platforms (or ‘decoupled review’) ( Ross-Hellauer, 2017)
Open peer review Tennant et al. (2017) ‘No agreed definition.’ Cites OpenAIRE survey which found 122 different definitions in use ( Ross-Hellauer, 2017)
Blinded/masked peer
review
Bruce et al. (2016) ‘... reviewers are blinded to author names and
affiliation. Author names and/or potentially
identifying credentials are removed from manuscripts sent for peer review so as to remove or minimize peer reviewer biases
that arise from knowledge of and assumptions about author identities.’
    2. Pre/post publication review
Label Author Description from paper
Pre-peer review
commenting
Barroga (2020);
Tennant et al. (2017)
‘Pre-peer review commenting involves the informal commenting or discussion on a publicly available pre-publication
manuscript draft.’
Pre-publication peer
review
Barroga (2020);
Tennant et al. (2017)
‘Pre-publication peer review consists of a formal and editorially invited evaluation of research by selected experts in the relevant
field.’
Post-publication peer
review
Barroga (2020);
Tennant et al. (2017)
‘Post-publication peer review comprises a formal and optionally invited evaluation of research by selected experts in the
relevant field after publication.’
Post-publication
commenting
Barroga (2020);
Tennant et al. (2017)
‘Post-publication commenting consists of an informal discussion of published research independent of any formal peer review.’
Post publication peer
review
Horbach & Halffman (2018) ‘In these archives [pre-print servers], manuscripts usually go through a minor evaluation to check whether they meet minimal
standards of academic writing. Subsequently, the actual review is done by community members who comment on the
manuscript... Authors can then improve the manuscript and upload new versions to the archive.’
Post publication peer
review
Burley (2017) ‘... takes place only after publication and is usually fully open.’
Registered
reports
Horbach & Halffman (2018) ‘In this form of peer review, which is still restricted mainly to medical fields and psychology, manuscripts are usually reviewed
in two stages. The initial and most important review stage takes place after the study has been designed, but prior to data
collection.’
    3. Collaboration and decoupling
Label Author Description from paper
Collaborative review Barroga (2020);
Tennant et al. (2017)
‘Collaborative review involves manuscript assessment wherein referees, editors, and external readers provide interactive
comments leading to a consensus decision and a single set of revisions.’
Interactive peer review Barroga (2020);
Tennant et al. (2017)
‘... the reviewers interact online with the authors and other interested scientists for a more open and collaborative review.’
Discussion during
review
Horbach & Halffman 2018 ‘Some journals … have attempted to improve editorial decision making by introducing interactive stages in the
review process, during which reviewers and editors can share or discuss their reports and opinions on a manuscript before
communicating a final decision to the author’
Cascading peer review Barroga (2020); Tennant et al. (2017) [Using this method] … ’rejections are avoided by redirecting peer-reviewed but rejected papers to a more suitable publication
venue. The consortia enable papers with the referee reports to move easily between publishers, reducing time and expense
of repeated evaluation. Some pass on the peer reviews with the rejected papers. Occasionally, reviews from other journals
accompanying manuscripts rejected are used for other journals.’
Cascading peer review Horbach & Halffman (2018) ‘[This model…is now widely used, especially by larger publishing houses. The system aims to avoid final rejection of a
manuscript after peer review by redirecting critically reviewed manuscripts to potentially more suitable journals.’
Cascading peer review Burley (2017) ‘[This method]…is based on the principle of reviewing a submitted manuscript only once (if possible) and offering to the
author(s) a suitable publication venue in a tiered structure.
Peer review as a
separate service
Burley (2017) ‘More experiments are underway [such as] making peer review independent of the journal and providing it as a service’
Recommendation
services
Barroga (2020); Tennant et al. (2017) ‘Recommendation services review involves post-publication evaluation and recommendation of significant articles, often
through a peer-nominated consortium.’
Portable review Barroga (2020); Tennant et al. (2017) ‘Portable review means the authors pay a company...for a standard single-blind review that they can submit with the paper to
collaborating journals.’
Independent peer
review
Barroga (2020); Tennant et al. (2017) ‘In independent peer review … a number of companies provide pre-submission peer review for a fee...Thus, reports from
commercial reviewer platforms are used to assist in peer review. This involvement of commercial refereeing bodies allows
the dissociation of review from the journal publishing the article, thereby facilitating a faster review (e.g., Rubriq, Peerage of
Science, Axios Review) or the detection of integrity issues (e.g., Research Square). Some companies use an online “scorecard”
to determine strengths and weaknesses of a paper. For Peerage of Science, the fee is paid by the journal which publishes the
offering.’
Decoupled post-
publication review
Barroga (2020); Tennant et al. (2017) ‘Decoupled post-
publication review consists of adding notes directly to the highlighted sections of the work. These added notes
can be kept private or made public.’
Review by third parties Horbach & Halffman (2018) ‘In addition to the systems providing pre-publication review, other independent platforms have emerged... in which any reader
can comment on any published manuscript. These systems constitute examples
of post-publication review independent of journals and publishers. These new trends have increasingly widened the definition
of a peer, so that the term now refers … to anyone who feels capable of understanding and evaluating a given piece of
research.’
    4. Focussed and specialised review
Label Author Description from paper
Soundness only review Horbach & Halffman (2018) ‘A major development in [peer review models] … came with the launch of the open access journal PLoS ONE, by the
Public Library of Science (PLoS), in 2006. In this journal’s review process and business model, reviewers are asked to base
their recommendation for acceptance or rejection purely on the soundness and validity of the research, comprising the
methodology, soundness of results and reporting.’
Results free review Burley (2017) ‘Results-free review means that in a first step the paper is evaluated only for its rationale and method, not the results. If the
former is deemed suitable for publication, then this is offered in principle. In a second step, the results are reviewed too.
Publication may only be rejected if the results deviate unjustifiably from the stated aims and methods.’
Specialised review Horbach & Halffman (2018) ‘Over the past decades, new actors have joined the review process, thereby compelling peer review itself to
become more specialised. This applies to its content, for example introducing specialised statistical reviewers, as well as to the
process ...’
Specialised review Horbach & Halffman (2018) ‘... plagiarism detection software tool to assist in peer review, the CrossCheck system being the most common ...’

‘Automatic analysis that checks for the correct use of statistics in manuscripts using AI.’
Specialised review Horbach & Halffman (2018) ‘The assistance of software in detecting image manipulation, which is considered an increasing form of fraud in various
research areas... it has already become possible to check for bad reporting, … data fabrication and image manipulation …
usually done by … the editorial team or journal’s staff.’
Specialised review Bruce et al. (2016) ‘Addition of peer reviewers for specific tasks or
with specific expertise such as adding a statistical peer reviewer, whose main task is to detect the misuse of methods or
misreporting of statistical analyses.’

Moving away from increased interaction as a focus of innovation and the slight modification of traditional procedures, the next type of innovation in this category is cascading or transferring peer review. This innovation was found three times in this review: Barroga (2020) (after Tennant et al., 2017), Horbach & Halffman (2018), and Burley (2017). It is the process whereby an article that has already been peer reviewed and rejected by one journal is given the opportunity to be considered by another journal within the same publishing company. Barroga (2020) (after Tennant et al., 2017) also suggests the process on a larger scale where publishers band together in consortia, enabling papers to move between journals owned by different publishers.

The final set of innovations in this subcategory reflect the deregulation of academic publishing as several new businesses emerge onto the market to provide peer-review services. This is reported by Burley (2017). Four variations on this theme emerge from Barroga (2020) (after Tennant et al., 2017): recommendation services, portable peer review, independent peer review, and de-coupled post-publication review. A key aspect of all of these is that they are journal agnostic, that is, the process of peer review is not directly linked to a particular journal’s decision-making process in relation to the article. Recommendation is about promotion of particular articles that have been reviewed post-publication through a respected consortium of researchers such as F1000Prime, now Faculty Opinions ( Thorburn, 2020). As defined by Barroga (2020) (after Tennant et al., (2017) portable peer review involves paying for an article to be reviewed and receiving the reports to submit to a publisher alongside the article. Independent peer review is again a commercial review company providing a service for an author, with the difference being that some publishers foot the bill for the review when a paper is subsequently published in their journal. De-coupled post-publication review is described by Barroga (2020) (after Tennant et al., 2017) as articles being annotated online and notes added in the margins in either a private or a public mode. This fits with the broader sense of the term decoupling used in scholarly communication to mean the overall relaxation between peer review and dissemination ( Priem & Hemminger, 2012). Finally, Horbach & Halfmann (2018) highlight the open nature of some peer review services in their category ‘review by third parties’. This open nature enables anyone who feels they can comment on a piece of research the opportunity to do so, which they reflect ‘increasingly widen[s] the definition of a peer’.

Focussed and specialised review

This category captures types of peer review that focus on one aspect or section of a publication. Soundness only peer review ( Horbach & Halffman, 2018) refers to a method of reviewing in which only the rigour of the research (as opposed to its novelty or significance) is considered in making a decision on acceptance. It is akin to the critical appraisal method often adopted in reviewing health research for example in using a Critical Appraisal Skills Programme checklist ( CASP, 2018). The aim is to allow all results to be published which meet a particular quality threshold, not only the most interesting or novel results. ‘Results free peer review’ ( Burley, 2017) refers to a method of screening papers in a two-stage review process. This involves evaluating the rationale for the study and the methods. In the case of a positive evaluation, the paper is approved for publication in principle subject to a further full review that also includes the results. Horbach & Halffman (2018) and Bruce et al. (2016) both report instances of specialised review where a paper is reviewed with a focus on one aspect. This includes plagiarism detection, use of statistics and use of images. This work is done by various actors including researchers and editorial or journal staff, or by utilising specialist tools such as CrossCheck, a publisher initiative using the iThenticate text comparison software ( Feinstein, 2008). AI tools may also be used for this kind of work. Table 2 summarises the different types of peer review found in the literature.

Peer reviewer focussed initiatives

Reviewer incentives

This subcategory describes various incentives that are offered to induce researchers to act as peer reviewers. The incentives are manifest in direct and indirect rewards for peer review.

Direct rewards have been designed to reward peer review on top of the traditional indirect rewards. Direct rewards come in a range of forms, such as linking peer review to ORCID records, which Barroga (2020) (after Tennant et al., 2017) terms ‘crediting’, or making peer review activity visible in Publons, a platform ‘dedicated to publicly recognising reviewers’ ( Burley, 2017). Financial reward (which can be private, or publicly acknowledged) is referred to by Barroga (2020) (after Tennant et al., 2017), in the form of free access to articles, waivers of article processing charges, and fees for providing pre-publication review. Barroga (2020) states some difficulties with introducing payment for reviews, for example, commodifying peer review being in tension with academic culture. Tennant (2018) touches on the implicit / explicit and private / public nature of reward for peer review, commenting on the limitations of the scope of public reward due to the private nature of most peer review.

Indirect rewards are about being a good academic citizen, taking part in peer review as a usual part of academic work. These rewards are well established. As examples of indirect rewards of a private nature, Barroga (2020) mentions being up to date with one’s field and having the opportunity to influence the direction of the field. Other indirect rewards are of a public nature, such as being invited to be on an editorial board. Regardless of their private or public nature, indirect rewards do not bring an immediate benefit but instead help to promote one’s reputation, gain experience and contribute to the wider research system.

Reviewer support

Innovations to support reviewers include standards, training and tools for reviewers. Barroga (2020) cites informal training that researchers do for themselves such as reading instructions for authors, or asking colleagues for support. Also, training that is not designed specifically for reviewers but helps in a lateral way in undertaking the role, such as keeping up to date about advances in open access. In addition, he cites formal training courses set up by the Publons Academy. Bruce et al. (2016) also refers to reviewer training and mentoring programmes for peer reviewers to help them evaluate manuscripts appropriately. Core competencies for peer reviewers based on their responsibilities to readers, authors and editors are another form of reviewer support. These are based on the recommendations of particular associations such as the Council of Science Editors (CSE) and the Committee on Publication Ethics (COPE) ( Barroga, 2020). Finally, Bruce et al. (2016) reports on the use of checklists to aid peer review, such as reporting guidelines for different study types. Table 3 summarises the different types of reviewer focussed initiatives found in the literature.

Table 3. Peer reviewer focussed initiatives.

    1. Reviewer incentives
Label Author Description from paper
Non-financial Barroga (2020) ’Nonfinancial incentives may come in the forms of frequent reviewer invitations, being up-to-date with research developments,
opportunities to influence science, increased acumen in reviewing, free journal access or subscription, access to databases/research
platforms and digital libraries, acknowledgment in journal websites, publicized reviews, letter of thanks, certificates of excellence, and
editorial board appointment.’
Crediting Barroga (2020) ‘Crediting incentives may be given by formally recognising the reviewing work and linking peer review activity to ORCID records using
DOIs.’
Financial Barroga (2020) ‘Financial incentives can be received through the Rubriq system by providing pre-publication reviews or from compensation derived
from the article processing charge. Although cash incentives can hasten reviews, many journals cannot realistically afford it. Cash
incentives may also affect the quality of review, transform the review process into business, or damage the moral sentiments of
researchers. Other forms of financial incentives include waiver of publication charges and free access to paid articles.’
Reviewer credit Tennant (2018) ‘How to provide and receive appropriate credit for peer review is an ongoing debate … There is … currently
a great potential scope of providing more detailed information about peer review quality, in a manner that is further tied to researcher
reputation and certification. The main barrier that remains here is the fact that peer review is still largely a closed and
secretive process, which inhibits the distribution of any form of credit.’
Rewarding peer
review
Burley (2017) ‘... recognizing and rewarding peer reviewers has become a priority for scholarly societies, publishers, and service providers. For
example, societies publish lists of the most prolific and helpful reviewers; publishers give public credit and provide additional rewards;
and service providers enable the collection of data on reviews and reviewers to enhance reviewer visibility and rewards. Further still,
Publons is a start-up dedicated to publicly recognizing reviewers for their contribution, enabling reviewers to track and showcase their
activities.
     2. Reviewer support
Label Author Description from paper
Guidelines and
training
Barroga (2020) ‘Training is achieved when reviewing author instructions from journals, receiving guidance from academic peers, or continuing
education on digitization and open access ... training and orientation through the Publons Academy can be received to further develop
skills in reviewing.’
Core competencies Barroga (2020) ‘Core competencies among peer reviewers are based on the recommendations of associations concerned with the integrity of peer
review. These associations include the Council of Science Editors (CSE), World Association of Medical Editors (WAME), International
Committee of Medical Journal Editors (ICMJE), and Committee on Publication Ethics (COPE). The core competencies commonly
recommended by these associations may be categorized as reviewer's responsibilities to the authors, editors, and readers.’
Training and
mentoring
Bruce et al. (2016) ‘Training, which included training or mentoring programs for peer reviewers to provide instructional support for appropriately
evaluating important
components of manuscript submissions. These interventions directly target the ability of peer
reviewers to appropriately evaluate the quality of the manuscripts.’
Checklists Bruce et al. (2016) ‘Peer reviewers’ use of a checklist, such as reporting guideline checklists, to evaluate the quality of the manuscript.’

Technology to support peer review

AI support for peer review, research discovery tools and publishing platforms amongst other technologies feature amongst the innovations described so far in this review. However, Barroga (2020) (after Tennant et al., 2017) goes beyond the use of software tools and discusses possible future models of peer review based on particular types of technology. This section reports on these suggested future models. This is followed by Table 4 which summarises current use of technology and future models.

Table 4. Technology to support peer review.

    1. Current uses
Type Author Examples / Description from paper
Platforms/Servers/OA journals Barroga (2020) BMJ Open, Sage Open, PLOS ONE.
PR services Barroga (2020) Peerage of Science
Applications and Tools Barroga (2020) and
Burley (2017)
CrossCheck
Publons
AI-assisted peer review Barroga (2020) ‘Used for recognizing images, recommending content, detecting fraud,
evaluating teaching and assessment, or detecting plagiarism; requires human
final judgement’
    2. Potential models
Model Author Examples / description from paper
Reddit model Barroga (2020);
Tennant et al. (2017)
‘Platform for comments and original or linked content’
Stack exchange model ibid ‘Network of websites of question and answer sites’
Amazon model ibid ‘Model for posting reviews of published materials’
GitHub model ibid ‘Open-source distributed version control system with features transferable to
peer-review system’
Hypothesis model ibid ‘Web annotation tool for interactive education and collection of peer
perspectives’
Wikipedia model ibid ‘Collaborative authoring and review system’
Blockchain model ibid ‘Technology for possibly creating tokenized peer review system’
Hybrid peer review platform ibid ‘Consists of harmonization, certification, and incentivization’

Barroga (2020) reviews the potential models of peer review put forward by Tennant et al. (2017) and assesses them against six features of open access publishing: openness, anonymity, accountability, bias, time and incentive. All the proposed models are open, in that review reports are public, but the identity of authors and reviewers remains unknown. On the factors of bias (whether editorial decisions are made public) and anonymity (whether the identity of editors and reviewers are revealed to authors) no assessment is made due to the models being hypothetical. The Reddit, Stack Exchange, and Hypothesis models are rated as offering greater author - reviewer accountability due to more transparent interactions between these stakeholders. Greater efficiency may be found in the GitHub and Wikipedia models with review time shortened or delays minimised. Reviewer incentives are found embedded within the Stack Exchange, block chain and hybrid peer review models.

Discussion and summary

As review articles, the studies in our review draw conclusions based on several items of primary evidence. By bringing together these conclusions in a simple synthesis, it is possible to reveal some key messages, given that any similar conclusions drawn in the various review articles have the combined weight of all the primary evidence reviewed. The conclusions of the review articles have been integrated below, to highlight observations on current practice, perspectives on the implications of new ways of working and calls for action. The strongest conclusions, based only on frequency, are the need for more research to determine the effectiveness of new models of peer review, and the need for a full reflection on the peer review system, including all stakeholders.

Observations on current practice

There are a number of observations summing up the current state of affairs in peer review: the increase in innovation has been rapid ( Barroga, 2020), there has been an overall improvement in reviewer recognition and efficiency of peer review processes, and an increase in initiatives trialing transparent peer review ( Burley, 2017). Technology is not being used to its full potential in the peer review system ( Tennant, 2018). Moving to reviewer focussed innovations, two of the reviews covered highlight the use of reviewer training, Burley (2017) commenting that scholarly publishers and learned societies are increasing their focus on training (and rewarding) reviewers, and Barroga (2020) suggesting that reviewer training and core competencies are important to consider as part of a broader reflection on the peer review system as a whole.

Perspectives on the implications of newer practices

A number of authors discuss their interpretations on the implications of newer practices: that quality may be compromised if reviewers are paid ( Barroga, 2020); also, that innovations in peer review such as post-publication reviews and statistical reviews may reinforce a particular perspective in how scientific knowledge is perceived ( Horbach & Halffman, 2018). Rather than research outputs being seen as a snapshot of discovery, capturing one moment in time, which will be built on with new research, these innovations can lead to publications being edited with the aim of arriving at a set of inviolable facts. The authors suggest this is a new perspective, favoured by those with a realist or positivist view of knowledge perceiving the research literature as a ‘database of facts’ rather than a ‘library’ ( Horbach & Halffman, 2018, p. 13). Two reviews also cite specific barriers to change within the peer review system: that new models will never become mainstream whilst there is so much prestige to be gained from publishing in journals ( Tennant, 2018). In addition, new ways of working are often not in line with the behaviours required for job security and career progression, which results in very little motivation for researchers to change ( Tennant et al., 2017; Tennant, 2018).

Calls for action

Numerous calls for action are found in the conclusions drawn by the authors of the review articles. Three reviews ( Bruce et al., 2016; Horbach & Halffman, 2018; Tennant et al., 2017) conclude that there is a lack of empirical evidence to assess the effectiveness of innovations in peer review. On a more nuanced but similar point, Barroga (2020) and Horbach & Halffman (2018) point to the role of peer review to identify malpractice or errors in research, with Barroga calling for research to measure how far new innovations can deliver this, and Horbach and Halffman highlighting the tension between the practice of peer review and its ability to fulfil this role. Another conclusion that is shared is the need for a wider reflection on the peer review process as a research community, with both Barroga (2020) and Tennant (2018) underscoring the need to consider what different stakeholders bring to the peer review process and which role they inhabit. Tennant (2018) goes on to suggest that alongside appropriate use of technology and research incentives this reflection is necessary for the success of a new system of peer review. Finally, Tennant et al. (2017) call for the decoupling of peer review from commercial interests in order to return to a community-led process. Table 5 provides a summary of innovations described in this review.

Table 5. Summary of innovations.

OA=open access.

Review article
Innovations Barroga (2020) Bruce et al. (2016) Burley (2017) Horbach & Halffman (2018) Tennant (2018) Tennant et al. (2017)
Open peer review
Transparent peer review  
Blinded / masked peer review  
Pre-peer review commenting
Pre-publication peer review
Post-publication peer review
Post-publication commenting
Post publication peer review  
Post publication peer review  
Registered reports  
Collaborative review
Interactive peer review
Discussion during review  
Cascading peer review
Peer review as a separate
service
 
Recommendation services
Portable review
Independent peer review
Decoupled post-publication
review
Review by third parties  
Specialisation  
Results free  
Soundness only  
Non-financial  
Crediting  
Financial  
Reviewer credit  
Rewarding peer review  
Guidelines and training  
Core competencies  
Training and mentoring  
Checklists  
Platforms/servers/OA journals  
PR services  
Applications and tools  
AI-assisted peer review  
Reddit model
Stack exchange model
Amazon model
GitHub model
Hypothesis model
Wikipedia model
Blockchain model
Hybrid peer review platform

Conclusion

This review of innovations in peer review is based on papers identified in Web of Science and Scopus, limited from 2010 to 2021. A total of 291 papers were screened, with six recent review articles being included. These review articles comprise a mixture of narrative reviews, meta-analysis, state of the art and summary articles. They describe numerous approaches to peer review. In our overview we collated these descriptions of peer review into four subcategories: open/masked, pre/post publication, collaboration and decoupling, focussed and specialised. We also collated mentions of reviewer focussed initiatives and presented these in the subcategories of reviewer support and reviewer incentives. We recorded and extracted references to the use of technology to aid peer review and summarised these practices noting current uses and potential models as reported in our included papers.

The fact that there are enough review articles to warrant a review of reviews, indicates the growing maturity of the field of peer review research. One review focussed on efficacy of peer review methods in a particular field ( Bruce et al., 2016), and effectiveness evidence, testing and measuring how well particular innovations meet their objective continues to be a growing form of research in the field. However, given the size of the field and the inherent complexity of analysing the peer review system, which spans numerous disciplines and includes varied professions in its conduct, descriptive research in any form will always be essential to record the development in innovations. This literature review is a contribution in this vein. We hope that our overview of peer review innovations will support future work in this area.

Acknowledgements

An earlier version of this article can be found on SocArXiv (doi: https://doi.org/10.31235/osf.io/qaksd).

Funding Statement

This work was supported by Wellcome [221297] as part of its core funding of the Research on Research Institute (RoRI).

The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

[version 2; peer review: 2 approved, 1 not approved]

Data availability

No data are associated with this article.

References

  1. ASAPbio: Reimagine Review.2021. Reference Source [Google Scholar]
  2. Barnett R: University knowledge in an age of supercomplexity. High Educ. 2000;40(4):409–422. 10.1023/A:1004159513741 [DOI] [Google Scholar]
  3. Barroga E: Innovative strategies for peer review. J Korean Med Sci. 2020;35(20):e138. 10.3346/jkms.2020.35.e138 [DOI] [PMC free article] [PubMed] [Google Scholar]
  4. Bendiscioli S, Garfinkel M: Dealing with the limits of peer review with innovative approaches to allocating research funding. EMBO Science Policy Programme,2021. Reference Source [Google Scholar]
  5. Black N, van Rooyen S, Godlee F, et al. : What makes a good reviewer and a good review for a general medical journal? JAMA. 1998;280(3):231–233. 10.1001/jama.280.3.231 [DOI] [PubMed] [Google Scholar]
  6. Booth A, Sutton A, Clowes M, et al. : Systematic approaches to a successful literature review (Third edition.). SAGE.2021. Reference Source [Google Scholar]
  7. Bruce R, Chauvin A, Trinquart L, et al. : Impact of interventions to improve the quality of peer review of biomedical journals: a systematic review and meta-analysis. BMC Med. 2016;14(1):85. 10.1186/s12916-016-0631-5 [DOI] [PMC free article] [PubMed] [Google Scholar]
  8. Burley R: Peer review in the 21st century. Inf Serv Use. 2017;37(3):259–261. 10.3233/ISU-170850 [DOI] [Google Scholar]
  9. CASP.2018. Reference Source [Google Scholar]
  10. Choi JP: Herd behavior, the "penguin effect," and the suppression of informational diffusion: an analysis of informational externalities and payoff interdependency. Rand J Econ. 1997;28(3):407–425. Reference Source [Google Scholar]
  11. Crewe R: Assuring the quality of scholarly South African journals: An experiment in journal peer review. S Afr J Sci. 2020;116(9–10):8795. 10.17159/sajs.2020/8795 [DOI] [Google Scholar]
  12. Feinstein J: Plagiarism and CrossCheck: a cross-publisher initiative. Editors' Bulletin. 2008;4(3):124–127. 10.1080/17521740802651260 [DOI] [Google Scholar]
  13. Goodman SN, Berlin J, Fletcher SW, et al. : Manuscript quality before and after peer review and editing at Annals of Internal Medicine. Ann Intern Med. 1994;121(1):11–21. 10.7326/0003-4819-121-1-199407010-00003 [DOI] [PubMed] [Google Scholar]
  14. Grant MJ, Booth A: A typology of reviews: an analysis of 14 review types and associated methodologies. Health Info Libr J. 2009;26(2):91–108. 10.1111/j.1471-1842.2009.00848.x [DOI] [PubMed] [Google Scholar]
  15. Higgins JPT, Altman DG, Gøtzsche PC, et al. : The Cochrane Collaboration’s tool for assessing risk of bias in randomised trials. BMJ. 2011;343:d5928. 10.1136/bmj.d5928 [DOI] [PMC free article] [PubMed] [Google Scholar]
  16. Hoepner J: Silencing behaviours in contested research and their implications for academic freedom. Australian Universities' Review,2019;61(1):31–41. Reference Source [Google Scholar]
  17. Horbach SPJMS, Halffman WW: The changing forms and expectations of peer review. Res Integr Peer Rev. 2018;3:8. 10.1186/s41073-018-0051-5 [DOI] [PMC free article] [PubMed] [Google Scholar]
  18. Kaltenbrunner W, Pinfield S, Waltman L, et al. : Innovating peer review, reconfiguring scholarly communication: An analytical overview of ongoing peer review innovation activities. SocArXiv. 2022. 10.31235/osf.io/8hdxu [DOI] [Google Scholar]
  19. Moher D, Liberati A, Tetzlaff J, et al. : Preferred Reporting Items for Systematic Reviews and meta-analyses: The PRISMA Statement. PLoS Med. 2009;6(7):e1000097. 10.1371/journal.pmed.1000097 [DOI] [PMC free article] [PubMed] [Google Scholar]
  20. Ortega JL: Are peer-review activities related to reviewer bibliometric performance? A scientometric analysis of Publons. Scientometrics. 2017;112(2):947–962. 10.1007/s11192-017-2399-6 [DOI] [Google Scholar]
  21. Ouzzani M, Hammady H, Fedorowicz Z, et al. : Rayyan-a web and mobile app for systematic reviews. Syst Rev. 2016;5(1):210. 10.1186/s13643-016-0384-4 [DOI] [PMC free article] [PubMed] [Google Scholar]
  22. Priem J, Hemminger BM: Decoupling the scholarly journal. Front Comput Neurosci. 2012;6:19. 10.3389/fncom.2012.00019 [DOI] [PMC free article] [PubMed] [Google Scholar]
  23. Ragone A, Mirylenka K, Casati F, et al. : On peer review in computer science: Analysis of its effectiveness and suggestions for improvement. Scientometrics. 2013;97(2):317–356. 10.1007/s11192-013-1002-z [DOI] [Google Scholar]
  24. Rogers EM: Diffusion of Innovations.5th Edition: Free Press,2003. Reference Source [Google Scholar]
  25. Ross-Hellauer T: What is open peer review? A systematic review [version 2; peer review: 4 approved]. F1000Res. 2017;6:588. 10.12688/f1000research.11369.2 [DOI] [PMC free article] [PubMed] [Google Scholar]
  26. Smith R: Peer review: a flawed process at the heart of science and journals. J R Soc Med. 2006;99(4):178–82. 10.1258/jrsm.99.4.178 [DOI] [PMC free article] [PubMed] [Google Scholar]
  27. Sutton A, Clowes M, Preston L, et al. : Meeting the review family: exploring review types and associated information retrieval requirements. Health Info Libr J. 2019;36(3):202–222. 10.1111/hir.12276 [DOI] [PubMed] [Google Scholar]
  28. Tennant JP: The state of the art in peer review. FEMS Microbiol Lett. 2018;365(19):fny204. 10.1093/femsle/fny204 [DOI] [PMC free article] [PubMed] [Google Scholar]
  29. Tennant JP, Dugan JM, Graziotin D, et al. : A multi-disciplinary perspective on emergent and future innovations in peer review [version 3; peer review: 2 approved]. F1000Res. 2017;6:1151. 10.12688/f1000research.12037.3 [DOI] [PMC free article] [PubMed] [Google Scholar]
  30. Thorburn A: F1000 Prime is now Faculty Opinions.2020. Reference Source [Google Scholar]
  31. Walker R, Rocha da Silva P: Emerging trends in peer review-a survey. Front Neurosci. 2015;9:169. 10.3389/fnins.2015.00169 [DOI] [PMC free article] [PubMed] [Google Scholar]
  32. Waters AM, LeBeau RT, Young KS, et al. : Towards the enhancement of quality publication practices in clinical psychological science. Behav Res Ther. 2020;124:103499. 10.1016/j.brat.2019.103499 [DOI] [PubMed] [Google Scholar]
  33. Wilsdon J: The metric tide: Independent review of the role of metrics in research assessment and management.2015. 10.4135/9781473978782 [DOI] [Google Scholar]
  34. Woods HB: Knowledge production and disciplinary practices in a British University: A qualitative cross-disciplinary case study.EdD thesis, University of Sheffield,2018. Reference Source [Google Scholar]
Wellcome Open Res. 2023 Mar 3. doi: 10.21956/wellcomeopenres.20892.r54540

Reviewer response for version 2

Bahar Mehmani 1

I thank the authors for adjusting their manuscript to address my previous concerns and comments. I do not have any further comments.

Is the review written in accessible language?

Yes

Are all factual statements correct and adequately supported by citations?

Yes

Are the conclusions drawn appropriate in the context of the current research literature?

Partly

Is the topic of the review discussed comprehensively in the context of the current literature?

Partly

Reviewer Expertise:

Peer review studies, research in peer review, peer review innovation

I confirm that I have read this submission and believe that I have an appropriate level of expertise to confirm that it is of an acceptable scientific standard.

Wellcome Open Res. 2023 Jan 30. doi: 10.21956/wellcomeopenres.20892.r54541

Reviewer response for version 2

Mario Malički 1

Thank you for the updated version, but for the effort I invested in the initial review, I expected a point by point response to the raised comments. The naming/classification I see has remained as before without effort to align this with STM or other initiatives, the inclusion of a narrative summary, a paper without references remains questionable as any other research paper or opinion paper might have mentioned more "innovations" than that one, and future potential models remain included with too weak a rationale behind it as proposals listed in many papers on peer review could have fallen under the definition of potential future models.

Is the review written in accessible language?

Yes

Are all factual statements correct and adequately supported by citations?

No

Are the conclusions drawn appropriate in the context of the current research literature?

No

Is the topic of the review discussed comprehensively in the context of the current literature?

No

Reviewer Expertise:

Peer review

I confirm that I have read this submission and believe that I have an appropriate level of expertise to state that I do not consider it to be of an acceptable scientific standard, for reasons outlined above.

Wellcome Open Res. 2023 Jan 30. doi: 10.21956/wellcomeopenres.20892.r54539

Reviewer response for version 2

Erik Cobo 1

The authors have responded to my comments and those of the other reviewers. The article has improved and, in my opinion, should be indexed.

Is the review written in accessible language?

Yes

Are all factual statements correct and adequately supported by citations?

Yes

Are the conclusions drawn appropriate in the context of the current research literature?

Yes

Is the topic of the review discussed comprehensively in the context of the current literature?

Yes

Reviewer Expertise:

Statistician-methodologist expert in pivotal clinical trials; and with experience in editorial boards

I confirm that I have read this submission and believe that I have an appropriate level of expertise to confirm that it is of an acceptable scientific standard.

Wellcome Open Res. 2022 Jun 8. doi: 10.21956/wellcomeopenres.19603.r49098

Reviewer response for version 1

Bahar Mehmani 1

Thank you for the opportunity of reviewing this insightful review.

Woods et al. have analyzed the body of published 'peer-reviewed' papers on innovation in peer review in scholarly publishing and report their findings in this paper. 

It is important to note that by default innovation entails failure. No results and negative results of innovative ideas organized by journal publishers and societies as well as funders might have a hard time ending in the body of published literature if ever tried for and as such, it is even more relevant to consider non-peer-reviewed articles for such studies.

Nevertheless, while the title and abstract mention innovations in scholarly publishing authors only discuss innovation in the process of reviewing journal manuscripts, and even there, only the external peer-review process through the lens of 4 previous reviews whose authors had the same focus. In this way, authors miss the opportunity to list innovations within the wider editorial process during the peer review, innovative ideas in books and monographs peer review, grant proposal peer review, and research elements such as data, code, protocol, software, and hardware peer review. 

Also by limiting the search strategy to peer-reviewed and published in indexed journals, authors miss a few innovative ideas published by journals and publishers that do not have an incentive to publish a peer-reviewed paper about their initiative. These stakeholders usually announce their new initiatives on their web pages, or as blog posts. Below you will find a few examples.

Examples of some other innovations in peer review that have not been listed in this paper are: 

to name a few of the reviewer-focused innovations

There is also a missed opportunity of narrowing the focus only to innovations in the external peer-review process. There have been several innovative streams to improve the editorial side of the peer review process that help speed up the process and improve the quality of feedback. A good example is using AI for matching manuscripts with reviewer areas of expertise resulting in several reviewer recommender tools such as:

To summarize: 

I understand addressing the above critics means changing the methods and scope of this study which is not preferable. I suggest authors change their title and abstract to reflect the above-mentioned items and add a limitation section to their paper listing these limitations.

Is the review written in accessible language?

Yes

Are all factual statements correct and adequately supported by citations?

Yes

Are the conclusions drawn appropriate in the context of the current research literature?

Partly

Is the topic of the review discussed comprehensively in the context of the current literature?

Partly

Reviewer Expertise:

peer review studies, research in peer review, peer review innovation

I confirm that I have read this submission and believe that I have an appropriate level of expertise to confirm that it is of an acceptable scientific standard, however I have significant reservations, as outlined above.

Wellcome Open Res. 2023 Jan 6.
Helen B Woods 1

Thank you for taking time to review this article and to present additional experiments with peer review in scholarly publishing not included in this review. Whilst these specific experiments were not included in our review, some of the approaches are included in a generic way. For example, the idea of co-reviewing or mentoring a new reviewer is a training intervention and we include some examples of this kind of innovation in the ‘Reviewer support’ section. In a similar way ‘Select Crowd Review’ Is an example of collaborative review. Some of the other examples represent approaches that are not described in our review, and this has been stated in the limitations section. In addition, innovations in other areas of the publishing process, such as speeding the efficiency of matching reviewers with manuscripts have not been included as you point out. We have made a statement to this effect in the limitations section to make this clearer to the reader and make a more transparent link between our data sources and the types of interventions found and included in our article.

Wellcome Open Res. 2022 May 26. doi: 10.21956/wellcomeopenres.19603.r50488

Reviewer response for version 1

Mario Malički 1

Thank you for the opportunity to review this manuscript. I struggled with the paper a lot, and you will see below why. To me, this text read like an internal report, and not as a scholarly manuscript with elements and structure found in typical systematic, umbrella, or narrative reviews. It was also not clear what is your main contribution – a summary of terms you created and defined or a summary of terms others used. I expected the former.

  •  Title – the meta-summary term should be avoided. If this is an umbrella review, use that, or specify why this is not one.

Abstract

  • quality assurance – I recommend deleting this term or specifying in the methods/introduction what falls under this term and how it differs or overlaps with peer review. If it is important it should be included in the title – e.g. Overview (or umbrella review) of innovations in peer review and quality assurance in sc. publishing. If as you say later it's captured by your definition, avoid it here – as it implies it is not captured.  

  • This literature review – why then use the term meta-summary in the title

  • The aim of this literature review – please rephrase, and say what is the aim of this paper. Its relation and origin, and later use for the project can be described in the introduction/methods.

  • Please rewrite the methods, and include information expected for lit/sys. review, who screened, how, in which software, language limits, and so on, as well as how, was summary conducted, was there a meta-analysis, and so on.

  • Say our initial strategy found that all individual interventions we identified were captured in 6 reviews (and even some we did not), and so we adapted this project to be an umbrella. We grouped the intervention into….

  • From a simple synthesis of the review authors’ conclusions – please mention first in the methods how you synthesized results and defined high level categories. I also hoped you would tell us which interventions were most commonly attempted.

Introduction

  • I recommend a complete rewrite and expansion of the introduction, see comments on the abstract. What is the aim of this paper? The aims of the overall project, and how it came to be are secondary.  

Definitions and framing.

  • I have difficulties with your definition of peer and would require it to be changed. A peer (reviewer) is an individual, not a software, or an AI. I understand that for the purpose of the article, and that in other published articles, the peer review term is used to describe all checks/suggestions (including those done by software) but that does not mean that the word peer needs to be redefined. Peer review is a compound noun, and its meaning can incorporate what you need, without redefining the word peer. Furthermore, later in the article, you use the term reviewer focused initiatives – which implies reviewer, is not used interchangeably with peer reviewer, nor does it include your peer definition.

  • Similarly, peer review also includes informal responses, questions, and comments posted on social media – this goes against the new taxonomy of peer review – which specifically differentiates comments from peer review. So please state, in our study peer review also incorporates …

  • These types of informal responses were found in a study examining disciplinary knowledge production ( Woods, 2018), where examples of researchers’ peer review practice were identified. – Please rephrase this I did not understand the meaning of this sentence. Did you want to say- those informal, or social media comments are similar to comments seen in peer review? See our study on preprint comments 1

  • Particularly in applied fields, this is coming to fruition, - please provide a reference

  • This literature review describes innovations – would recommend deleting this sentence, it is redundant

  • The next section will describe the methodology used in this literature review and how we will present the results. This is followed by the results themselves in which we classify different types of innovations. We conclude with some discursive reflections on the current situation. – Would recommend deleting this, the subtitles below are sufficient

Methods

  • This review was undertaken - please use the same term everywhere, and move this sentence to the background.

  •  I find that the whole first section - Overall approach should be deleted, and search of literature stated first. Statements like - , it was not necessary to identify – if the paragraph is not deleted I recommend it is changed to:  We felt it was not – this is an issue or a limitation that a reviewer can oppose, and I would. – We did not limit the search – please first define, what you did search for and when, and provide a full search strategy. And then explain inclusion, and exclusion criteria, as well as bias assessment, or check to see how those papers matched your definition of peer review and innovations. If none were excluded, then state here that you included all papers that described any type of intervention/innovation.

  •  Details of Reimagine Review are not necessary and should be deleted. Just cite it and say how many of their intervention made it to your final inventory.

  •  Data collection – I would recommend using PRISMA 2020 subtitles, i.e. information sources, and that you use all applicable subtitles despite this not being sys. review You did use the Flow chart.

  • If as you say in the end review articles were only included, then perhaps umbrella review is the best term

  •  Presentation of results – would recommend deleting this paragraph

  •  I find the description of 6 reviews unnecessary and would move them to the appendix. It is also not clear from the methods, what is the difference between a summary article, state-of-the art review, and narrative review – are these terms the authors used or you, how would you classify them, and why?

  • Open/masked peer review – You could consider using the STM taxonomy rather than open/masked - https://osf.io/68rnz/

  • In pre and post publication commenting, this amounts to comments or feedback by any interested party, - this is not true, there are specialized preprint review platforms actively inviting specific individuals. What would you call the platform where your article was posted and is peer-reviewed by me – who was invited to review?

  • Registered reports or study protocols in my view should not be a sub-category under pre-post pub commenting. They are calls for review of study designs/ideas, rather than completed studies. And they can happen either as a pre-or post, as open or masked, as collaborative, or as decoupling. As can grant peer review. This makes me also question the term approach to peer review – what is approached here? In Burley's paper, she calls these models of peer review. How do other reviews call these innovations? Types, modes, approaches, descriptors? These approaches are descriptions of elements of peer review, which can one day be meta-data for peer review. A registered report is a type of article, not peer review.

  • Looking at the titles of these categories, I fear a lot has been missed by not using the STM approach for subtitles, i.e. I would rename your “approaches” to Review focused Innovations or Review process interventions, and would call the sub-categories: Reviewer identity (innovations), Review timing (innovations), Reviewer interaction (innovations), Content focus (innovations),

  • Technology support – rather than Technology to support peer review. – but see my comment 15

  • As table 1 includes 576 included studies, I also hoped to see a table with these 576 studies classified by which innovations they cover. Was information based only on summaries in the 6 studies, without checking if something was missed by the original authors?

  • Reviewer focused initiatives +- should rather be Reviewer focused innovations. But I believe you need to mention that AI checks and software cheeks are also Editor/publisher focused and can also be given to authors to self-evaluate their manuscripts. Like language software, self-evaluation checklists, statcheck, and so on.  Incentives can also be used to make editors do a better job. You also stated: that Barroga (2020) 2 cites informal training that researchers do for themselves – this is then an intervention on authors. So perhaps this category should be revised, and if the first high level is focused on the review process, then these here are perhaps focusing on review skill/expertise.  

  • I find the weakest part of the summary to be the technical support, and the names assigned to them. GitHub and many other models are already captured in the “approaches section” they are rather platforms where a manuscript or peer review reports could be stored. And so there is nothing in the GitHub model that was not covered before. Amazon model refers to the rating of reviews – and this belongs to reviewer-focused initiatives not here. It should have been made clear when technology is used to enable other approaches, versus when it provides a new type of innovation. I find it also very strange that AI approach was listed primarily under focused review. Why does it matter who performs it? AI review can be pre-or post, it can be open or masked (in theory, if the type of AI used is not disclosed), it can be focused or for all. Why isn’t AI also mentioned under reviewer support, they can save time for reviewers to check many things. Finally, why use Baraga’s models and summarize them here. Can they be called an innovation, if they were never attempted in peer review– e.g. GitHub model, and amazon model. They are theories/models. You stated “what makes an innovation is if the practice is new to those who are implementing it. – if it was not implemented at least once, it should not have been included. It is a theory/model. What is it in this model that is not an innovation captured elsewhere? Just the use of a different term?

  • My last comment is not directly about this paper, but more on the 4 schools described in How to improve scientific peer review: Four schools of thought - https://osf.io/preprints/socarxiv/v8ghj/. The categorization of high-level innovations here, and the schools showcase quite a different approach to classifications, and maybe both need to be revised in light of one another and my comments here.   

I will refrain from commenting on the discussion as I believe my above comments are already extensive and ask for a lot of changes, that should then also be reflected by rewriting the discussion.  

In hopes that my comments may help you improve your manuscript,

Mario Malički

Is the review written in accessible language?

Yes

Are all factual statements correct and adequately supported by citations?

No

Are the conclusions drawn appropriate in the context of the current research literature?

No

Is the topic of the review discussed comprehensively in the context of the current literature?

No

Reviewer Expertise:

Peer review

I confirm that I have read this submission and believe that I have an appropriate level of expertise to state that I do not consider it to be of an acceptable scientific standard, for reasons outlined above.

References

  • 1. : Analysis of single comments left for bioRxiv preprints till September 2019. Biochem Med (Zagreb) .2021;31(2) : 10.11613/BM.2021.020201 020201 10.11613/BM.2021.020201 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2. : Innovative Strategies for Peer Review. J Korean Med Sci .2020;35(20) : 10.3346/jkms.2020.35.e138 e138 10.3346/jkms.2020.35.e138 [DOI] [PMC free article] [PubMed] [Google Scholar]
Wellcome Open Res. 2023 Jan 6.
Helen B Woods 1

Thank you for reviewing our article and for alerting us to areas for improvement. What was particularly useful was to highlight instances where our use of language was less precise than it could have been, for example in the use of ‘peer review’, ‘peer’ and ‘innovation’. We have addressed these issues in the text. In response to other reviewers' comments we have added a limitations section and updated the title and abstract to more accurately reflect the scope of the article, which we think will also address some of your concerns regarding a more accurate description of the scope of the article. We have also provided additional citations in the text which have strengthened the article, for example to define the different review types discussed. With reference to the ‘Four Schools’ article, this has now been updated and is published here:  https://osf.io/preprints/socarxiv/v8ghj/

Wellcome Open Res. 2022 Apr 22. doi: 10.21956/wellcomeopenres.19603.r49097

Reviewer response for version 1

Erik Cobo 1

I’m thankful to be able to review the Wellcome OR manuscript titled "Innovations in peer review in scholarly publishing: a meta-summary”.

I just have two major suggestions for authors. And one minor comment -that the authors might consider or not.

Major suggestions

  • Please, on Methodology, provide operational definitions for the terms “meta-summary”, “narrative review”, “summary article” and “state-of-the-art review”.

  • Please, revise the numbers on the flowchart. Note your actual numbers did not add (223 excluded +291 screened should equal 247 after duplicates are removed?).

  • Please, consider further explanations on the figure foot or the text.

Minor comment

  • Your text and your conclusion seem to avoid statements about the methods to provide evidence of the peer review processes. In particular, I wonder if there is any causal link between those peer review methods and the ultimate goal of Science, reproducibility.

Is the review written in accessible language?

Yes

Are all factual statements correct and adequately supported by citations?

Yes

Are the conclusions drawn appropriate in the context of the current research literature?

Yes

Is the topic of the review discussed comprehensively in the context of the current literature?

Yes

Reviewer Expertise:

Methodologist expert in pivotal clinical trials

I confirm that I have read this submission and believe that I have an appropriate level of expertise to confirm that it is of an acceptable scientific standard, however I have significant reservations, as outlined above.

Wellcome Open Res. 2023 Jan 6.
Helen B Woods 1

Thank you for taking the time to review our article, and for providing us with feedback to improve our work. This is much appreciated. We have addressed the major suggestions including a new simplified and accurate flow chart (hence no additional notes in a footer have been added to Figure 1). With reference to your comments about research methods and causal links, we have described the review papers that the article is based on, this includes the characteristics of each article, including research methods where these were employed. In our article, we didn’t set out to establish any links between reproducibility and particular types of peer review or particular methods to research peer review processes. However, the desire to improve the quality of research, including more detailed and accurate reporting, is clearly a motivating factor for many of the initiatives we describe.

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Data Availability Statement

    No data are associated with this article.


    Articles from Wellcome Open Research are provided here courtesy of The Wellcome Trust

    RESOURCES