Skip to main content
F1000Research logoLink to F1000Research
. 2016 Apr 14;5:683. [Version 1] doi: 10.12688/f1000research.8452.1

The effects of an editor serving as one of the reviewers during the peer-review process

Marco Giordan 1, Attila Csikasz-Nagy 2,4, Andrew M Collings 3, Federico Vaggi 1,a
PMCID: PMC4962294  PMID: 27508056

Abstract

Background

Publishing in scientific journals is one of the most important ways in which scientists disseminate research to their peers and to the wider public. Pre-publication peer review underpins this process, but peer review is subject to various criticisms and is under pressure from growth in the number of scientific publications.

Methods

Here we examine an element of the editorial process at eLife, in which the Reviewing Editor usually serves as one of the referees, to see what effect this has on decision times, decision type, and the number of citations. We analysed a dataset of 8,905 research submissions to eLife since June 2012, of which 2,750 were sent for peer review, using R and Python to perform the statistical analysis.

Results

The Reviewing Editor serving as one of the peer reviewers results in faster decision times on average, with the time to final decision ten days faster for accepted submissions (n=1,405) and 5 days faster for papers that were rejected after peer review (n=1,099). There was no effect on whether submissions were accepted or rejected, and a very small (but significant) effect on citation rates for published articles where the Reviewing Editor served as one of the peer reviewers.

Conclusions

An important aspect of eLife’s peer-review process is shown to be effective, given that decision times are faster when the Reviewing Editor serves as a reviewer. Other journals hoping to improve decision times could consider adopting a similar approach.

Keywords: peer review, decision times, eLife

Background

Although pre-publication peer review has been strongly criticised – for its inefficiencies, lack of speed, and potential for bias (for example, see 1 and 2) – it remains the gold standard for the assessment and publication of research 3. eLife was launched to “improve [...] the peer-review process” 4 in the life and biomedical sciences, and one of the journal’s founding principles is that “decisions about the fate of submitted papers should be fair, constructive, and provided in a timely manner” 5. However, peer review is under pressure from the growth in the number of scientific publications, which increased by 8–9% annually from the 1940s to 2012 6, and growth in submissions to eLife would inevitably challenge the capacity of their editors and procedures.

eLife’s editorial process has been described before 7, 8. In brief, each new submission is assessed by a Senior Editor, usually in consultation with one or more members of the Board of Reviewing Editors, to identify whether it is appropriate for in-depth peer review. Traditionally, editors recruit peer reviewers and, based on their input, make a decision about the fate of a paper. Once a submission is sent for in-depth peer review, however, the Reviewing Editor at eLife has extra responsibility. First, the Reviewing Editor is expected to serve as one of the peer reviewers. Second, once the reviews have been submitted independently, the Reviewing Editor should engage in discussions with the other reviewers to reach a decision they can all agree with. Third, when asking for revisions, the Reviewing Editor should synthesise the separate reviews into a single set of revision requirements. Fourth, wherever possible, the Reviewing Editor is expected to make a decision on the revised submission without re-review. At other journals, the Reviewing Editor may instead be known as an Academic Editor or Associate Editor.

Since editors have extra responsibility in eLife’s peer-review process, here we focus our analysis on the effect of the Reviewing Editor serving as one of the peer reviewers, and we examine three outcomes: 1) the effect on decision times; 2) the effect on the decision type (accept, reject or revise); and 3) the citation rate of published papers. The results of the analysis are broken down by the round of revision and the overall fate of the submission. We do not consider the effect of the discussion between the reviewers or the effect of whether the Reviewing Editor synthesizes the reviews or not.

Methods

We analyzed a dataset containing information about 9,589 papers submitted to eLife since June 2012 in an anonymised format. The dataset contained the date each paper was first submitted, and, if it was sent for peer review, the dates and decisions taken at each step in the peer-review process. Information about authors had been removed, and the identity of reviewers and editors was obfuscated to preserve confidentiality.

As a pre-processing step, we removed papers that had been voluntarily withdrawn, or where the authors appealed a decision, as well as papers where the records were corrupted or otherwise unavailable. After clean up, our dataset consisted of a total of 8,905 submissions, of which 2,750 were sent for peer review. For the rest of the paper, we focus our analysis on this subset of 2,750 papers, of which 1,405 had been accepted, 1,099 had been rejected, and the rest were still under consideration. The article types included are Research Articles (MS type 1), Short Reports (MS type 14), Tools and Resources (MS type 19), and Research Advances (MS type 15). Registered Reports are subject to a slightly different review process and have not been included.

Before discussing the results, we introduce a few definitions: the “ eLife Decision Time” is the amount of time taken by eLife from that version of the submission being received until a decision has been reached for a particular round of review. The“Author Time” is the amount of time taken by the authors to revise their article for that round of revision. The “Total Time” is the time from first submission to acceptance, or amount of time taken for eLife to publish a paper from the moment it was first received for consideration. By definition, the “Total Time” is equal to the sum of the “ eLife Decision Time” and the “Author Time” across all rounds, including the initial submission step. “Revision Number” indicates the round of revision. We distinguish between Reviewing Editors who served as one of the reviewers during the first round of review and Reviewing Editors who did not serve as one of the reviewers (i.e., those who undertook more of a supervisory role during the review process) with the “Editor_As_Reviewer” variable (True or False).

We illustrate the variables with a real example taken from the dataset ( Table 1).

Table 1. An example from the dataset.

MS
Type
Revision
Number
Received
Date
Decision
Date
eLife
Decision
Time
Total
Time
Author
Time
Editor_As_
Reviewer
5 1 2012-06-20 2012-06-21 1 N/A N/A
1 1 2012-06-27 2012-07-25 28 N/A 6 True
2 2012-09-05 2012-09-05 0 77 42 True

The example submission from Table 1 was received as an “initial submission” (MS TYPE 5) on 20 June 2012. One day later, the authors were encouraged to submit a “full submission” (MS TYPE 1) that would be sent for in-depth peer review. The full submission was received on 27 June 2012, when the Reviewing Editor was assigned and reviewers were contacted. In this example, the Reviewing Editor also served as one of the reviewers (indicated by the “Editor_As_Reviewer” variable).

On 25 July (28 days later), the Reviewing Editor sent out a decision asking for revisions to the authors, who submitted their revised manuscript on 5 September. The paper was accepted on the same day that it was resubmitted. In this case, the total eLife Decision Time was 29 days (including the pre-review stage), the Author Time was 48 days, and the Total Time ( eLife Decision Time plus Author Time) was 77 days. Total Time refers only to the total time across all rounds and revisions for each paper - and does not vary across rounds. Since we are focusing on the role of the editors in the peer review process, in the rest of the paper we will ignore the time spent in the pre-review stage.

All of the statistical analyses were performed using R and Python. On the Python side, we used statsmodels, scipy, numpy, and pandas for the data manipulation and analysis. To plot the results we used bokeh, matplotlib, and seaborn. Details of all the analysis, together with code to reproduce all image and tables in the paper are available on the companion repository of this paper here: https://github.com/FedericoV/eLife_Editorial_Process.

To obtain the citation numbers, we used a BeautifulSoup to scrape the eLife website, which provides detailed information about citations for each published paper.

Results and discussion

First, we examined the effect of the Reviewing Editors serving as one of the reviewers on the time from submission to acceptance or from submission to rejection after peer review (Total Time). When the Reviewing Editor served as a reviewer (Editor_As_Reviewer = True), the total processing time was 10 days faster in the case of accepted papers and more than 5 days faster in the case of papers rejected after peer review ( Figure 1). Both differences are statistically significant (see Table 2 for details). Intuitively, regardless of the role of the Reviewing Editor, rejection decisions are typically much faster than acceptance decisions, as they go through fewer rounds of revision, and are not usually subject to revisions from the authors.

Figure 1. Decision times are faster when the Reviewing Editor serves as one of the reviewers.

Figure 1.

We compare the total time from submission to acceptance and submission to rejection after peer review. Orange indicates papers submissions where the Reviewing Editor served as one of the peer reviewers, while light blue indicates submissions where the Reviewing Editor did not serve as one of the reviewers (i.e., the editors had more of a supervisory role).

Table 2. Effect of a Reviewing Editor serving as a reviewer (Editor_As_Reviewer) on eLife Decision Time and Author Time.

Counts eLife Decision Time Author Time
Editor_As_Reviewer False True False True M-W False True M-W
Decision_Type Revision Number
Accept Full
Submission
0 5.000 12.000 21.600 23.333 1.000 4.000 4.833 0.915
1 440.000 650.000 8.802 6.949 0.006 52.209 51.168 0.402
2 115.000 164.000 4.339 3.238 0.175 16.487 14.652 0.402
3 6.000 11.000 3.667 2.455 0.747 6.833 9.909 0.811
Reject Full
Submission
0 461.000 616.000 36.104 30.981 0.000 6.182 6.430 0.267
1 10.000 10.000 22.200 31.600 0.148 64.900 101.800 0.402
2 1.000 N/A 17.000 N/A 0.000 60.000 N/A 0.000
3 N/A N/A N/A N/A N/A N/A N/A N/A
Revise Full
Submission
0 705.000 946.000 36.018 31.053 0.000 5.786 5.744 0.402
1 129.000 182.000 19.651 15.747 0.006 66.930 64.110 0.915
2 6.000 12.000 7.833 7.167 0.747 21.333 35.250 0.730
3 N/A N/A N/A N/A N/A N/A N/A N/A

One possible reason why submissions reviewed by the Reviewing Editor have a faster turnaround is because fewer people are involved (e.g., the Reviewing Editor in addition to two external reviewers, rather than the Reviewing Editor recruiting three external reviewers), and review times are limited by the slowest person. To test this, we built a linear model to predict the total review time as a function of editor type (whether the Reviewing Editor served as a reviewer or not), decision (accept or reject), and the number of unique reviewers across all rounds (see Table S1). Indeed, the total review time did increase with each reviewer (7.4 extra days per reviewer, p < 0.001) and the effect of a Reviewing Editor serving as one of the reviewers remained significant (–9.3 days when a Reviewing Editor served as one of the reviewers, p < 0.0001).

Next, we examined this effect across all rounds of review (rounds 0, 1, 2) and decision types (accept, reject and revise). The results are shown in Figure 2 and summarized in Table 2. Again, we see that processing times are consistently faster across almost every round, when the editors serves as one of the peer reviewers, except in the cases where the sample size was very small.

Figure 2. Decision times are faster when the Reviewing Editor serves as one of the reviewers across different rounds of review.

Figure 2.

Boxplots showing decision times for different rounds of review, depending on decision type and whether the Reviewing Editor served as one of the reviewers (orange) or not (light blue).

Interestingly, when the Reviewing Editor serves as one of the peer reviewers, the eLife Decision Time is reduced, but the time spent on revisions (Author Time) does not change. This suggests that the actual review process is more efficient when the Reviewing Editor serves as a reviewer, but the extent of revisions being requested from the authors remains constant.

We next examined the chances of a paper being accepted, rejected or revised when a Reviewing Editor served as one of the reviewers. We found no significant difference when examining the decision type on a round-by-round basis ( Table 3) (chi-squared test, p = 0.33).

Table 3. Effect of a Reviewing Editor serving as one of the reviewers on editorial decisions.

Editor_As_Reviewer False True Totals
Accept Full Submission 566 837 1403
Reject Full Submission 472 626 1098
Revise Full Submission 840 1140 1980
Totals 1878 2603 4481

To test whether eLife’s acceptance criteria changed over time, we built a logit model including as a predictive variable the number of days since eLife began accepting papers and whether the Reviewing Editor served as one of the reviewers. The number of days since publication had a very small (–0.003) but significant effect (p < 0.02) while the effect of the Reviewing Editor serving as a reviewer was not significant (see Table S2). We also tested whether a Reviewing Editor serving as a reviewer had an effect on the number of rounds of revision before the final decision and found no significant effect (see Table S3).

The final outcome we examined was the number of citations (as tracked by Scopus) received by papers published by eLife. Papers accumulate citations over time, and, as such, papers published earlier tend to have more citations ( Figure 3).

Figure 3. Papers accumulate citations over time.

Figure 3.

log1p of the number of citations (indexed by Scopus) received by each paper accepted by eLife versus the number of days between when eLife first started accepting papers. Green dots represent papers where the Reviewing Editor served as one of the reviewers, while blue dots represent papers where the Reviewing Editor did not serve as one of the reviewers.

We examined this effect using a generalized linear model. As variables, we considered whether the Reviewing Editor served as a reviewer (Editor_As_Reviewer, true or false), as well as the number of days between eLife publishing its first manuscript and the day the Scopus database was queried. The presence of a Reviewing Editor serving as a reviewer had no significant effect on the number of citations (see Table S4). Papers with longer total review times tended to be cited less (this effect is small but significant).

One of the most noticeable effects of a Reviewing Editor serving as one of the peer reviewers at eLife is the faster decision times. However, serving as a Reviewing Editor and one of the reviewers for the same submission is a significant amount of work. As the volume of papers received by eLife has increased, the fraction of editors willing to serve as a reviewer has decreased. While in 2012 almost all editors also served as reviewers, that percentage decreased in 2013 and 2014. There are signs of a mild increase in the percentage of editors willing to serve as reviewers in 2015 ( Figure 4).

Figure 4. A decreasing proportion of Reviewing Editors served as one of the reviewers as submission volumes increased.

Figure 4.

Time series plots for the number of active editors who served as one of the reviewers (Unique RE, blue) or not (Unique SE, green) in a given month. The total number of papers sent for peer review is shown in red. The % of papers is in black.

Conclusions

Due to an increasingly competitive funding environment, scientists are under immense pressure to publish in scientific journals, yet the peer-review process remains relatively opaque at many journals. In a systematic review from 2012, the authors conclude that “Editorial peer review, although widely used, is largely untested and its effects are uncertain” 9. Recently, journals and conferences (e.g., 10) have launched initiatives to improve the fairness and transparency of the review process. eLife is one such example. Meanwhile, scientists are frustrated by the time it takes to publish their work 11.

We report the analysis of a dataset consisting of articles received by eLife since launch and examine factors that affect the duration of the peer-review process, the chances of a paper being accepted, and the number of citations that a paper receives. In our analysis, when an editor serves as one of the reviewers, the time taken during peer review is significantly decreased. Although there is additional work and responsibility for the editor, this could serve as a model for other journals that want to improve the speed of the review process.

Journals and editors should also think carefully about the optimum number of peer reviewers per paper. With each extra reviewer, we found that an extra 7.4 days are added to the review process. Editors should of course consider subject coverage and ensure that reviewers with different expertise can collectively comment on all parts of a paper, but where possible there may be advantages, certainly in terms of speed and easing the pressure on the broader reviewer pool, of using fewer reviewers per paper overall.

Insofar as the editor serving as a reviewer is concerned, we did not observe any difference in the chances of a paper being accepted or rejected, but we did notice a modest increase in the overall number of citations that a paper receives when an editor serves as one of the reviewers, although this effect is very small. An interesting result from our analysis is that a longer peer-review process or more referees does not lead to an increase in citations, so this is another reason for journals and editors to carefully consider the impact of the number of reviewers involved, and to strive to communicate the results presented in a timely manner for others to build upon. As eLife is a relatively young journal, we can verify if the citations trend we observe will hold over longer periods as different papers accumulate citations.

Data and software availability

All code for the analysis as well as the datasets: https://github.com/FedericoV/eLife_Editorial_Process

Archived version as at the time of publication: http://dx.doi.org/10.5281/zenodo.48544 15

To reproduce Figure 4, we pre-processed the raw dataset that contained the identity of the editors to avoid disclosing any information about the identity of reviewers.

Acknowledgements

We gratefully acknowledge discussions and input from Mark Patterson (eLife’s Executive Director) and Peter Rodgers (eLife’s Features Editor). We thank James Gilbert (Senior Production Assistant at eLife) for extracting data from the submission system for analysis.

Funding Statement

Andy Collings is employed by eLife Sciences Publications Ltd. eLife is supported by the Howard Hughes Medical Institute, the Max Planck Society and the Wellcome Trust.

The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

[version 1; referees: 1 approved

Supplementary material

Table S1. Linear model for Total Time.

We built a linear model of the Total Time as a function of whether the Reviewing Editor served as a reviewer (Editor_As_Reviewer) (categorical variable, two levels), the final decision made on a paper (Decision_No), and the number of unique reviewers. The revision time increased with the number of reviewers, but it decreased when a Reviewing Editor served as one of the reviewers.

coef std err z P>|z| [95.0% Conf. Int.]
Intercept 85.7868 4.462 19.225 0.000 77.041 94.533
C(Editor_As_
Reviewer)[T.True]
-9.2265 1.701 -5.425 0.000 -12.560 -5.893
C(Decision_Type)
[T.Reject Full
Submission]
-65.3500 1.622 -40.285 0.000 -68.529 -62.171
Unique_Reviewers 7.1892 1.620 4.438 0.000 4.014 10.364

Table S2. Linear model for the chances of a paper being accepted.

We used logit regression to estimate the chances of a paper being accepted as a function of whether the Reviewing Editor served as one of the reviewers (Editor_As_Reviewer), the number of unique reviewers, and the number of days between when a paper was published and the first published paper by eLife. The only significant variable is the days since eLife started accepting papers for publication (although the effect on the chances of a paper being accepted is very small).

coef std err z P>|z| [95.0% Conf. Int.]
Intercept 0.6080 0.257 2.368 0.018 0.105 1.111
C(Editor_As_Reviewer)[T.True] 0.0822 0.087 0.946 0.344 -0.088 0.252
Publication_Since_Start -0.0003 0.000 -2.393 0.017 -0.001 -6.08e-05
Unique_Reviewers -0.0380 0.081 -0.468 0.640 -0.197 0.121

Table S3. Effect of a Reviewing Editor serving as a reviewer on the number of rounds of revision.

We used a GLM with a log link function to model the number of revisions that a paper undergoes prior to a final decision as a function of whether a Reviewing Editor served as one of the reviewers (Editor_As_Reviewer), the number of unique reviewers, the decision type, and the number of days since eLife started accepting papers. The only variable that had a significant effect was the decision type, as papers that are rejected tend to be overwhelmingly rejected early on and thus undergo fewer rounds of revision.

coef std err z P>|z| [95.0% Conf. Int.]
Intercept 0.6856 0.099 6.951 0.000 0.492 0.879
Editor_As_Reviewer[T.True] -0.0134 0.033 -0.403 0.687 -0.078 0.052
C(Decision_Type)[T.Reject
Full Submission]
-0.7762 0.035 -22.216 0.000 -0.845 -0.708
Publication_Since_Start 1.058e-05 5.33e-05 0.198 0.843 -9.4e-05 0.000
Unique_Reviewers 0.0382 0.031 1.230 0.219 -0.023 0.099

Table S4. Citation rates.

Linear model for the citations (as indexed by Scopus) after log1p transform. Publications where the Reviewing Editor served as one of the reviewers have a slightly higher number of citations, although this effect is very moderate. Papers also accumulate citations over time, thus papers that have been published longer ago (Publication_Since_Start) tend to accumulate more citations. The number of unique reviewers, as well as the total decision time before acceptance, had no significant effect.

coef std err z P>|z| [95.0% Conf. Int.]
Intercept 2.7603 0.127 21.713 0.000 2.511 3.010
Editor_As_Reviewer[T.True] 0.0960 0.043 2.240 0.025 0.012 0.180
Decision_Time -0.0005 0.000 -1.126 0.260 -0.001 0.000
Publication_Since_Start -0.0025 6.96e-05 -35.521 0.000 -0.003 -0.002
Unique_Reviewers 0.0599 0.040 1.503 0.133 -0.018 0.138

References

  • 1. Smith R: Peer review: a flawed process at the heart of science and journals. J R Soc Med. 2006;99(4):178–182. 10.1258/jrsm.99.4.178 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2. Smith R: Classical peer review: an empty gun. Breast Cancer Res. 2010;12(Suppl 4):S13. 10.1186/bcr2742 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3. Mayden KD: Peer Review: Publication’s Gold Standard. J Adv Pract Oncol. 2012;3(2):117–122. 10.6004/jadpro.2012.3.2.8 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4. Schekman R, Patterson M, Watt F, et al. : Scientific publishing: Launching eLife, Part 1. eLife. 2012;1:e00270. 10.7554/eLife.00270 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5. Schekman R, Watt F, Weigel D: Scientific publishing: Launching eLife, Part 2. eLife. 2012;1:e00365. 10.7554/eLife.00365 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6. Bornmann L, Mutz R: Growth rates of modern science: A bibliometric analysis based on the number of publications and cited references.2014; arXiv:1402.4578. Reference Source [Google Scholar]
  • 7. Schekman R, Watt F, Weigel D: Scientific publishing: The eLife approach to peer review. eLife. 2013;2:e00799. 10.7554/eLife.00799 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8. Schekman R, Watt FM, Weigel D: Scientific publishing: A year in the life of eLife. eLife. 2013;2:e01516. 10.7554/eLife.01516 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9. Jefferson T, Alderson P, Wager E, et al. : Effects of editorial peer review: a systematic review. JAMA. 2002;287(21):2784–2786. 10.1001/jama.287.21.2784 [DOI] [PubMed] [Google Scholar]
  • 10. Francois O: Arbitrariness of peer review: A Bayesian analysis of the NIPS experiment.arXiv:1507.06411.2015. Reference Source [Google Scholar]
  • 11. Powell K: Does it take too long to publish research? Nature. 2016;530(7589):148-51. 10.1038/530148a [DOI] [PubMed] [Google Scholar]
  • 12. Jones E, Oliphant E, Peterson P, et al. : SciPy: Open Source Scientific Tools for Python.2001. Reference Source [Google Scholar]
  • 13. Waskom M, Botvinnik O, Okane D, et al. : Seaborn Plotting Library.2016. 10.5281/zenodo.45133 [DOI] [Google Scholar]
  • 14. Seabold S, Perktold J: “Statsmodels: Econometric and statistical modeling with python.” Proceedings of the 9th Python in Science Conference.2010. Reference Source [Google Scholar]
  • 15. Vaggi F: eLife_Editorial_Process: Review_Version. Zenodo. 2016. Data Source
F1000Res. 2016 Jul 26. doi: 10.5256/f1000research.9100.r13397

Referee response for version 1

Ivan Oransky 1, Alison Abritis 1

Thank you for the opportunity to review this manuscript. This is a well-done study, and the conclusions follow from the results. We would recommend accepting the article once all clarifications and revisions have been made, or the lack of doing so adequately justified.

A. While brevity is generally to be admired, we would recommend a bit more detail about the statistical analyses. These are critical, but are reduced to 3 sentences and a referral to the programming language through an external link. We would suggest that the main text include the (brief) discussion of the analyses done, and rationale for them, rather than have those relegated to the external link.

B. The interpretation of the findings seems to be attributing causal factors - an A leads to B consideration - for which the control of variables is too limited. We believe that interpreting these as associations would be more consistent with the findings.

Consider the statement: "Journals and editors should also think carefully about the optimum number of peer reviewers per paper. With each extra reviewer, we found that an extra 7.4 days are added to the review process." Given that there appeared to be no inclusion of either article quality or complexity in the evaluation, is it not possible that issues within the article itself required the use of additional reviewers (i.e. a B leads to A perspective)? Perhaps extra reviewers with specific expertise was required, or concerns with potential problems in the manuscript led to consultations with other reviewers. It does not seem safe to assume that it was the addition of the reviewer that added extra days.

Similarly, the study centers around the role of the editor in the reviewing process, and the discussion suggested that the involvement of the reviewing editor as a peer reviewer expedited the process. There was little discussion of other factors that could have accounted for the statistical results. For example, perhaps the reviewing editor selected articles that piqued his or her interest, or were more clearly presented. Perhaps the reviewing editor selected to review at times more convenient to his or her workload, while other reviewers did not have such an option. The reviewing editor might select to review articles perceived to be of greater or timelier value to the journal itself, which may increase the speed of the review. 

Specific questions:

A. According to their method section, the authors state that they began with an initial N=9,589. After purging other articles they had an N=8,905. They then isolated a total of 2,750 articles subjected to the peer review process for the study: "For the rest of the paper, we focus our analysis on this subset of 2,750 papers, of which 1,405 had been accepted, 1,099 had been rejected, and the rest [which would equal 246] were still under consideration."

Looking at the Excel spreadsheet for citation counts, there are 1407 lines with entry numbers. For peer-reviewed papers, excel spreadsheet has 2747 (after removing duplicate entries based on the MS NO column) entries for manuscripts numbered up to 12621. The excel spreadsheet for unique reviewers has 2747 entries, with a final MS NO of 12621.

The numbers do not appear to match, and there is no explanation for that in methods. Exactly how many manuscripts were reviewed, how many rejected and why, and how many were tracked?

B. In the Excel spreadsheet for citations, the second column was titled "Citations," but these figures do not appear to have any relation to the Scopus citation numbers. What numbers were used for the actual citation counts?

We also note that we find the suggestions by other reviewers compelling, and would be happy to review a revision of this manuscript should that be considered useful.

We have read this submission. We believe that we have an appropriate level of expertise to confirm that it is of an acceptable scientific standard, however we have significant reservations, as outlined above.

F1000Res. 2016 Jul 5. doi: 10.5256/f1000research.9100.r14223

Referee response for version 1

Bernd Pulverer 1

Giordan et al analyzed 2,750 manuscripts sent out at the journal eLife for peer review (of which 1,405 ended up published). The authors compare papers in which the editors functions as ‘reviewing editor’ that is as one of three referees. Globally, and at almost every decision stage the process is accelerate significantly if the reviewing editor functions as one of the referees, with no or very small impact in author revision time and citation rates, respectively.

The authors calculate that every additional external referee adds 7.4 days to the process and suggest that journals strive to balance the need for covering all the required expertises carefully with the negative effect on the speed of evaluation.

The quality and speed of the peer review process are topics of active debate. Despite widespread criticism, the publication in certain peer reviewed journals continues to directly impact research assessment by both funders and institutions. The quality and fairness of the process is therefore paramount not only to assure the reliability of the literature, but also to inform research assessment in a balanced manner. Notwithstanding the slow delivery of this particular referee report, speed matters in particular in fast moving and highly competitive research areas like the biosciences.

Quantitative evidence that well defined aspects of an editorial process has positive effect on quality and or speed is therefore of significant importance.

The authors have carefully analyzed a decent sized dataset and report a statistically significant effect of a well defined change in the editorial process, while also showing evidence that this change has no detrimental effect on the quality of the editorial assessment, at least as far as the outcome is analyzed (here, in terms of two parameters: revision time and citation rate).

While this manuscript makes a significant contribution, I have a number of suggestions I would invite the authors to consider in revision:

Textural:

  1. Abstract/main text; Background: It is not merely the growth of the number publications that puts the system under pressure (after all, in principle the editorial/peer review process may well be able to scale with increased research output), but rather the increased pressure publish in a small number of high Impact Factor journals in an effort to optimize chances of a positive impact on research assessment.

  2. Please introduce the journal eLife, including the scientific scope, as different communities have widely different peer review and citation cultures and this will likely affect the findings reported here.

  3. Abstract: Results. As presented, I found it confusing that the first sentence describes an apparently sizeable difference between accepted (10 days faster) and rejected (5 days faster), while the next sentence states ‘there was no effect on whether submissions were accepted or rejected)’.

  4. Abstract: the dataset is described as consisting of an analysis of 8,905 submissions, when in reality the 2,750 papers sent for review were analyzed. This could be formulated more clearly.

  5. I would suggest for clarity to remove the ‘False’ and ‘True’ nomenclature and change to ‘reviewing editor’ and ‘editor’ assessment or similar.

 

Analysis:

  1. It is unclear to me if the authors can exclude any biases in terms of which manuscripts were selected for formal review by an editor vs. outside only refereeing. Have the authors attempted to assess possible bias? For example, maybe the reviewing editors tend to review the manuscripts themselves that strike them as the more interesting ones, or maybe certain subject areas are preferentially subject to one approach.

  2. I am missing a clear definition of what editorial peer review is and what is not. It is clear that this is likely not a completely binary situation and the authors do not describe how the decisions were parsed so clearly into the two groups (on p3. ‘more of a supervisory role’ vs. less of such a role sounds quite vague).

  3. Why were appealed manuscripts removed form the analysis? This may introduce a bias as possibly erroneous decisions are excluded. How many appeals were excluded?

  4. It is unclear to me if all paper invariably had three referees (i.e. 3 outside or 2 outside+reviewing ed.). I assume some papers only had two referees. Were they excluded? If not, how did these score for speed, revision time  and citation?

  5. The definition of ‘Total Time’ on p3. Is unclear: it is stated to be both ‘first submission-acceptance’ and ‘first submission to publication’: which one is it?

  6. Please state that Scopus citations were assessed when fist introducing the topic. I would recommend to use the same time window for all papers (e.g. 12 months after publication) as this renders citation rates more comparable. Why was the eLife website scraped for citation rates, and not the primary Scopus database – that data may in principle be more reliable.

  7. The 10 days (accepted) vs. 5 days (rejected) faster: is this simply the additive effect of 2 rounds vs. 1 round of review?

  8. Please include basic stats information in the figure legends – in particular fig 2, where the numbers will decrease dramatically for ‘revision 1’ and ‘revision 2’.

  9. It is unclear to me if reviewing editors were invariably faster the outside referees. It would be useful to quantify this and assuming there is a striking difference to speculate why – it is the individuals selected by e-Life or due to policing or incentives provided by the journal? After all, similar strategies could be applied to outside referees. On a related point, it would be useful to quantify if the reports by the reviewing editors were qualitatively different (e.g. length). One assumes the ultimate decision on the manuscript was as also much better correlated with the reports by reviewing editors than those of the outside referees.

  10. I am confused: in fig 2 ‘Reject full submission’ in revision 0 and revision 1 is slower than ‘accept’. This seems to be the opposite to fig 1 and in fact less intuitive than the results in fig 1.  Since manuscripts are rarely re-reviewed (see p1), are all the datapoints displayed in ‘revision 1’ and ‘revision 2’ for re-review processes?

  11. For fig 4, I would suggest to plot the manuscript load/editor.

 

Non-essential further reaching analysis (suggestions):

  1. it would have been useful to measure and present the acceptance/rejection rates of manuscripts assess by three outside referees compared with two referees + reviewing editor.

  2. it would have been useful to quantify the % of agreement between the reviewing editor and the outside referees, compared with agreement between the outside referees.

 

I have read this submission. I believe that I have an appropriate level of expertise to confirm that it is of an acceptable scientific standard, however I have significant reservations, as outlined above.

F1000Res. 2016 May 12. doi: 10.5256/f1000research.9100.r13794

Referee response for version 1

Alesia Zuccala 1

I enjoyed reading this article, and I liked the fact that it was concisely written. I would only like to share a few comments.

The authors of this paper have chosen to exam the peer review process for  eLife, which is an electronic journal for the life and biomedical sciences.  They note that two papers have previously described the editorial process of  eLife; however for their own study, I think it would have been useful to include a link to the  eLife website ( https://elifesciences.org/).   I was curious about whether or not it was an open access journal, and I looked for the website to obtain this information.   More importantly, I also looked to the website, because I wanted to know how many persons serve as Senior Editors or sit on the editorial board of  eLife.  What is interesting here is that the term “editor” for this journal is stretched to include one Editor-in-Chief, three Deputy-Editors, thirty-two Senior Editors and a 282-member Board of Reviewing Editors.   The peer review system for this journal is quite different from the ‘traditional’ journal, but to be more precise, it differs because “Reviewing Editors” are specialists who have agreed to review for the journal on a regular basis, and may in some cases call upon additional ‘outside’ reviewers. 

What we do not know from this paper, is whether or not two or more of the 282 Reviewing Editors sometimes choose to review the same paper.  At the  eLife website, the following is noted:  “The Reviewing editor usually reviews the article him or herself, calling on one or two additional reviewers as needed”.  Are the additional reviewers  always from the outside?  If not, how would this change the authors’ hypothesis related to the ‘effects of an editor serving as one of the reviewers’?

The methods used for the data analysis are explained very well, with the exception of one detail:  How did the authors acquire the initial dataset of 9,589 papers?   This information is presented in the ‘Acknowledgements’ section, but could have also been added to the Methods section, for more clarity.

The graphs related to the authors’ findings are clear and present interesting information, but I am not sure how the citation data were collected from Scopus for the peer-reviewed papers in  eLife and whether or not ‘citation windows’ were used for the papers depending on the year in which they were published.  Essentially the authors are correct in saying that “ papers accumulate citations over time, and, as such, papers published earlier tend to have more citations”, hence citation windows are used to correct for this.  The highest rates of citation (especially in the life sciences and biomedicine) will appear within three-to-five years following an article’s date of publication.  For this reason, bibliometricians usually count citations within this three-to-five year time-frame to determine an article’s initial impact.   Since the articles used in this study had been “ submitted to eLife since June 2012” the authors should have focus on three things:  1) the involvement of a Reviewing Editor as a peer reviewer or not, 2) the number of days between start of the submitted paper’s acceptance and publication, and 3) the papers’ citation rate following 3-5 years after final publication.

I have read this submission. I believe that I have an appropriate level of expertise to confirm that it is of an acceptable scientific standard.

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Data Citations

    1. Vaggi F: eLife_Editorial_Process: Review_Version. Zenodo. 2016. Data Source

    Articles from F1000Research are provided here courtesy of F1000 Research Ltd

    RESOURCES