Abstract
The COVID-19 pandemic elicited a substantial hike in journal submissions and a global push to get medical evidence quickly through the review process. Editorial decisions and peer-assessments were made under intensified time constraints, which may have amplified social disparities in the outcomes of peer-reviewing, especially for COVID-19 related research. This study quantifies the differential impact of the pandemic on the duration of the peer-review process for women and men and for scientists at different strata of the institutional-prestige hierarchy. Using mixed-effects regression models with observations clustered at the journal level, we analysed newly available data on the submission and acceptance dates of 78,085 medical research articles published in 2019 and 2020. We found that institution-related disparities in the average time from manuscript submission to acceptance increased marginally in 2020, although half of the observed change was driven by speedy reviews of COVID-19 research. For COVID-19 papers, we found more substantial institution-related disparities in review times in favour of authors from highly-ranked institutions. Descriptive survival plots also indicated that scientists with prestigious affiliations benefitted more from fast-track peer reviewing than did colleagues from less reputed institutions. This difference was more pronounced for journals with a single-blind review procedure compared to journals with a double-blind review procedure. Gender-related changes in the duration of the peer-review process were small and inconsistent, although we observed a minor difference in the average review time of COVID-19 papers first authored by women and men.
1. Introduction
The COVID-19 pandemic has exacerbated inequalities in science. Gender gaps in authorships, publication outputs, submission rates and invited journal contributions have widened [1–11], and young scientists, especially those with children, have seen substantial decreases in their weekly time for research [12, 13]. Studies also indicate intensified global disparities in publication outputs, with a decline in articles from developing countries [14, 15] and with few developing countries directly involved in COVID-19 related research [16].
The scientific community reacted quickly to the pandemic. Journal submissions surged [17–19], and editors and reviewers were fast to respond to the need for evidence on the prevention, diagnosis and treatment of the virus [14, 20]. However, diligent peer reviewing is time-demanding, and in the wake of the pandemic, the average review time–from submission to publication–decreased by 49% according to an analysis of articles published in 14 medical journals, prior to and during COVID-19 [21]. This decrease was primarily driven by speedy reviews of COVID-19 related submissions, while no acceleration was observed for manuscripts on other topics [22, 23].
The disruption of the publication system imposed by COVID-19 may have led to gender and institution-related disparities in peer reviewing for two reasons. First, editorial decisions during the pandemic were made under intensified time constraints with a substantial hike in journal-submissions and a global push to get COVID-19 research quickly through the review process. Research suggests that such time constraints can exacerbate status biases in complex decision-making processes [24–28], and both gender and institutional affiliation have been highlighted as “status signals” that may implicitly influence evaluative judgments in favour of male researchers and scientists from high status institutions [29–35]. Second, evidence suggests that women’s working conditions have been disproportionately affected by the pandemic due to disparities in teaching, service and caregiving loads [13, 36–40]. The increase in time constraints is likely reflected in the publication speed of women scientists, who, on average, may have been slower to respond to resubmission requests, hence delaying the duration of the peer-review process. Similarly, also institution-related disparities may have been exacerbated by the pandemic. Scientists at less affluent research institutions may have lacked the necessary resources, infrastructure and support to adapt quickly to the changing working conditions imposed by COVID-19, which may also have prolonged the review process.
In this study, we adopt a new perspective on the widening disparities during COVID-19 by using recently available data on the peer-review duration of medical papers. We aim to examine whether the pandemic has widened gender and institution-related differences in the average time from manuscript submission to acceptance. Additionally, given that some peer review processes differ in terms of whether institutional status queues and/or gender are visible to reviewers (due to having either single or double blinded review), we also descriptively investigate whether double blind peer review reduces institution/gender biases in publication speed during the pandemic.
2. Materials and methods
On January 11, 2021, we searched PubMed Medline for papers including COVID-19 related keywords (‘COVID-19’, ‘COVID’, ‘SARS-CoV-2’, ‘severe acute respiratory syndrome coronavirus 2’, ‘2019-nCoV’, ‘2019 novel coronavirus’, ‘Wuhan coronavirus’) [41] in their titles or abstracts. These articles needed to be published between March 11, 2020 (the date when the World Health Organization declared COVID-19 a global pandemic) and December 31, 2020. This query returned 106,116 articles (S1 Fig in S1 Appendix). After removing the articles with missing information about submission-acceptance dates (N = 64,530), we restricted our focus to outlets that had published > 50 articles with COVID-19 related keywords resulting in a core sample of 80 medical and health-science journals and a final sample of 8,828 COVID-19 related research articles. Next, we ran queries to establish two control samples: one consisting of articles without COVID-19 related keywords published in the same 80 journals during the pandemic (March 11, 2020, to December 31, 2020, N = 82,226) and another consisting of articles without COVID-19 related keywords published in the same journals during 2019 (March 11, 2019, to December 31, 2019, N = 64,685). In these datasets, we also removed entries with missing information. In total, our final sample consists of 78,085 papers distributed across the following three subsamples: 2020 COVID-19 related papers (N = 8,828), 2020 non-COVID-19 related papers (N = 40,125) and 2019 control papers (N = 29,132).
Data on the full names and affiliations of first- and last authors were obtained by matching the 78,085 PubMed records to article meta-data from a structured version of Web of Science hosted at the Centre for Science and Technology Studies (CWTS), Leiden University. The CWTS version of the WoS database also provides an estimated gender for individual researchers based on three name-to-gender assignment algorithms (Gender API, Gender Guesser and Genderize.io). This approach estimates a given author’s likelihood of being a man or a woman based on full names and country affiliations, and has been shown to be ~96% accurate [42]. We were able to use this database to assign an estimated gender of 90% of the single authors, 71% of the first authors and 82% of the last authors.
We used information from the Leiden Ranking (2019) and the QS World University Ranking (2019) to infer the reputation of each first and last author’s institutional affiliation. From the former, we relied on the PP top-10% indicator, which ranks universities based on their proportion of publications belonging to the top 10% most cited globally. PP top-10% is a yearly measure calculated based on field-normalized citations [43]. From the latter, we adopted the QS comprehensive ranking (averaged across the last three years), which provides a proxy of a university’s worldwide reputation, with 40% of the underlying ranking data derived from a global reputation survey. Of the 73,902 articles with eligible metadata in WoS, 48,270 had either first or last authors with institutional affiliations covered by the Leiden Ranking, while 30,900 had either first or last authors with affiliations covered by the QS-ranking.
2.1 Outcome measure
Our outcome variable is a count measure of the number of days from the initial submission of a manuscript to its acceptance for publication (i.e., peer-review duration). Information on the submission and acceptance dates of each article was retrieved from PubMed Medline.
2.2 Main predictors
Gender was computed as a three-factor variable (man, woman, unknown gender). This third gender category covers all the authors for which the name-to-gender assignment algorithms did not provide reliable results. Institution status (Leiden) and Institution status (QS) are continuous variables measuring the status of the first and last authors’ institutional affiliations according to the PP-top-10% and QS-rankings. These variables have been re-scaled to obtain standardized measures centred on their overall mean. Following Gelman [44], we divided each variable by two standard deviations to allow the numeric inputs to be interpreted on the same scale as the binary case variables (i.e. year and COVID-19). A one unit change on the rescaled measure of institution status should be interpreted as a two-standard deviation change on a scale from a lower ranking to a higher ranking institution.
2.3 Covariates
Year is a binary variable that measures whether a research article was published in 2019 or 2020 (2019 = 0, 2020 = 1). We interacted Year with the Gender and Institution Status in Models 1 to 4, since we were interested in changes occurring in the wake of the pandemic. COVID-19 was computed as a binary variable that measures whether a research article includes COVID-19 related keywords in its title/abstract or not (No COVID-19 keywords = 0, COVID-19 keywords = 1). COVID-19 was used as a covariate in Models 2 and 4, and interacted with Gender and Institution Status in Models 5 and 6. Journal Impact Factor is a continuous variable that measures the average number of citations of articles published in each medical journal within the preceding two years. This variable was used as a covariate in all models to adjust for variations in the prestige of the various outlets, and has been rescaled by dividing by two standard deviations [44]. Peer-review Procedure is a dichotomous measure that specifies if a journal operates with a single-blind or a double-blind review (double-blind = 1, single-blind = 0) procedure. In the double-blind procedure, the identity of the author is kept hidden from the reviewers, while, in the single-blind condition, both editors and reviewers know the identity of the author. Information on peer-review procedure was obtained by screening journal websites and by contacting journal editors by email. Descriptive statistics for all variables can be found in S1 Table in S1 Appendix. To measure the relationship between our main predictors (Institution Status and Gender) and the number of days from manuscript submission to acceptance (peer-review duration) we used linear-mixed effects models [45], with papers nested in journals focusing on first author (Model 1; 2; 5) and last author (Model 3; 4; 6) information. This approach allowed us to account for journal-level variation in average review times [46]. To examine the sensitivity of the regression results to alternative model specifications, we carried out robustness checks using Poisson generalized mixed models–appropriate for right-skewed count outcomes like ours [47]–and linear regression models with cluster-robust standard errors at the journal level.
We used descriptive survival plots to examine the within-group variability in the publication speed of articles published by authors at highly ranked and lower-ranked institutions. Specifically, we used the Kaplan-Meier non-parametric estimator [48]. This approach visualizes the expected duration of the peer-review process (in days and across groups) up to the point where a paper is accepted for publication. In the Kaplan-Meier survival analysis, we created dummies for authors affiliated with top-ranked and lower-ranked institutions. Authors affiliated with top-ranked institutions are those whose affiliation fall within the 10th decile of the Leiden and QS ranking indicators. Lower-ranked universities refer to institutions that fall within the 5th decile or lower on the same measures. We also included peer-review procedure as a factor in the survival plots.
The statistical analyses were conducted in R version 4.0.2. For the linear mixed effects models and the Poisson generalized mixed model, we used the ‘lme4’ 1.1–27.1 package [49], and for the linear regression models with cluster-robust standard errors, we used the ‘estimatr’ v. 0.30.2 package [50]. We used ‘Sjplot’ v. 2.8.9 package [51] to produce Fig 1, and the ‘survminer’ v. 0.4.9 package [52] to produce Figs 2 and 3.
3. Results
3.1. Institution status
Our analysis suggests that institution-related disparities in the average time from manuscript submission to acceptance have slightly increased during the pandemic. As indicated by the interaction between Institution Status and Year in Table 1 (Models 1 and 3), the marginal difference between the publication speed of papers from higher ranked and lower ranked universities (measured by two standard deviations difference on the Leiden Ranking) increased by 1.98 days (99% CI: -3.48 to -0.48) for first authors and by 2.15 days (99% CI: -3.71 to—0.6) for last authors in 2020 compared to 2019. Comparing review times for top-and bottom-ranked institutions that are two standard deviations below and above the mean of the ranking distribution, this corresponds to a marginal difference of ∼ 3.96 days for first authors and ∼ 4.3 days for last authors (S2 Fig in S1 Appendix). Given that the average duration of the peer-review process was ∼70 days for first authors and last authors in 2020, these marginal differences seem relatively small. The observed change appears to be partially explained by speedy reviews of COVID-19 research. Indeed, when COVID-19 is factored into the analysis (Table 1, Models 2 and 4), the coefficients for the interaction decrease to -0.83 days (99% CI: -2.30 to 0.65) for first authors and to -1.10 days (99% CI: -2.64 to 0.43) for last authors.
Table 1. Mixed linear regression models predicting the change in peer-review duration from 2019 to 2020.
Linear mixed effect models | ||||||||
---|---|---|---|---|---|---|---|---|
M1 peer-review time (First authors) | M2 peer-review time (First authors) | M3 peer-review time (Last authors) | M4 peer-review time (Last authors) | |||||
Predictors | Estimates | 99% CI | Estimates | 99%CI | Estimates | 99%CI | Estimates | 99%CI |
Intercept | 76.34 | 68.90 – 83.78 | 79.77 | 72.41 – 87.12 | 76.45 | 68.94 – 83.96 | 79.94 | 72.54 – 87.33 |
Institution Status (LR) | 0.61 | -0.58 – 1.81 | 0.12 | -1.06 – 1.29 | 0.82 | -0.41 – 2.05 | 0.37 | -0.84 – 1.58 |
Year | -6.78 | -7.98 – -5.59 | -4.09 | -5.28 – -2.91 | -6.33 | -7.78 – -4.88 | -3.74 | -5.18 – -2.31 |
Journal Impact Factor | -0.66 | -6.81 – 5.49 | -0.53 | -6.61 – 5.55 | -1.52 | -7.74 – 4.70 | -1.24 | -7.36 – 4.88 |
Gender (male) | -0.87 | -2.20 – 0.45 | -1.05 | -2.35 – 0.25 | -0.54 | -1.94 – 0.86 | -0.60 | -1.98 – 0.78 |
Gender (unknown) | 0.71 | -0.82 – 2.25 | 0.96 | -0.55 – 2.47 | 1.47 | -0.30 – 3.24 | 1.77 | 0.03 – 3.51 |
Institution Status (LR)*Year | -1.98 | -3.48 – -0.48 | -0.83 | -2.30 – 0.65 | -2.15 | -3.71 – -0.60 | -1.10 | -2.64 – 0.43 |
Year*Gender (male) | -1.02 | -2.71 – 0.67 | 0.23 | -1.43 – 1.88 | -1.07 | -2.86 – 0.73 | -0.37 | -2.14 – 1.39 |
Year*Gender (unknown) | 0.45 | -1.55 – 2.45 | 0.46 | -1.51 – 2.42 | 0.12 | -2.18 – 2.42 | 0.22 | -2.04 – 2.48 |
COVID-19 | -22.55 | -23.96 – -21.14 | -22.51 | -24.01 – -21.02 | ||||
Random Effects | ||||||||
σ2 | 966.70 | 934.02 | 954.03 | 922.91 | ||||
τ00 | 623journal | 605.53 journal | 626.27 journal | 606.54 journal | ||||
ICC | 0.39 | 0.39 | 0.40 | 0.40 | ||||
N | 80 journals | 80 journals | 80 journals | 80 journals | ||||
Observations | 48269 | 48269 | 44571 | 44571 | ||||
Marginal R2 / Conditional R2 | 0.008 / 0.397 | 0.034 / 0.415 | 0.008 / 0.403 | 0.033 / 0.418 | ||||
AIC | 469188.389 | 467530.623 | 432676.812 | 431199.268 |
We obtain comparable results when Institution Status is calculated based on the QS ranking as opposed to the Leiden Ranking (S2 Table in S1 Appendix). Robustness checks based on Poisson generalized mixed models and linear regression models with cluster-robust standard errors at the journal level yield qualitatively similar results (S4 and S5 Tables in S1 Appendix).
Zeroing in on papers published in 2020 (Table 2), we observe an 18.01 days (99% CI: -20.25 to– 15.78), and 18.6 days (99% CI: -21.34 to -15.88) shorter average review time for first and last authored papers from higher ranked universities that focus on COVID-19 compared to papers on other topics (Models 5 and 6). As indicated by the interaction between Institution Status (measured by the Leiden ranking) and COVID-19, two standard deviations move-up in the institutional prestige hierarchy reduces this average review time with an additional 5.49 days (99% CI: -7.94 to -3.05) for first author and 5.43 days (99% CI: -8.03 to -2.83) for last authors.
Table 2. Mixed linear regression models predicting the moderating effect of the COVID-19 topic on peer-review duration.
Linear mixed effect models | ||||
---|---|---|---|---|
M5 peer-review time (First authors) | M6 peer-review time (Last authors) | |||
Predictors | Estimates | 99%CI | Estimates | 99%CI |
Intercept | 71.62 | 64.09–79.14 | 72.06 | 64.50 – 79.61 |
Institution Status (LR) | 0.60 | -0.40 – 1.61 | 0.55 | -0.49 – 1.59 |
COVID-19 | -18.01 | -20.25 – -15.78 | -18.60 | -21.32 – -15.88 |
Journal Impact Factor | -0.79 | -7.05 – 5.46 | -1.55 | -7.87 – 4.77 |
Gender (male) | -0.27 | -1.34 – 0.81 | -0.47 | -1.61 – 0.66 |
Gender (unknown) | 1.83 | 0.50 – 3.15 | 2.00 | 0.46 – 3.53 |
Institution Status* COVID-19 | -5.49 | -7.94 – -3.05 | -5.43 | -8.03 – -2.83 |
COVID-19 *Gender (male) | -3.20 | -5.97 – -0.43 | -2.22 | -5.35 – 0.90 |
COVID-19 *Gender (unknown) | -1.97 | -5.53 – 1.59 | 1.02 | -3.08 – 5.11 |
Random Effects | ||||
σ2 | 840.97 | 833.47 | ||
τ00 | 636.20 journal | 636.78 journal | ||
ICC | 0.43 | 0.43 | ||
N | 80 journals | 80 journals | ||
Observations | 28970 | 26686 | ||
Marginal R2 / Conditional R2 | 0.037 / 0.452 | 0.035 / 0.453 | ||
AIC | 277689.924 | 255576.412 |
Fig 1 plots this interaction and indicates that the review time of papers on other topics than COVID-19 (red line) does not vary substantively by institution status. However, in the case of COVID-19 papers (blue line), the status variation is obvious. Comparing universities that are two standard deviations below and above the mean of our measure of institution status, we find that COVID-19 papers with first authors and last authors at highly-ranked universities see an 18–20% shorter average review time than COVID-19 papers with first and last authors at low-ranked universities.
These marginal differences are slightly more pronounced when institution status is calculated based on the QS ranking, with an estimated gain of 5.8 days (99% CI: -8.65- -2.96) for first authors and -6.58 days (99% CI: -9.75- -3.41) for last authors for every two standard deviations increase in institution status (S3 Table in S1 Appendix). Robustness checks based on alternative model specifications yield comparable results (S4 and S5 Tables in S1 Appendix).
3.2. Gender
Our analysis of gender-related changes in the duration of the peer-review process indicates small and inconsistent effects. As demonstrated by the interaction between Gender and Year in Table 1 (Models 1 and 3), the change in the average gender difference in review time from 2019 and 2020 is negligible (first authors, marginal gender difference = -1.02 days, 95% CI: -2.71 to 0.67; last authors, marginal gender difference = 1.07 days, 95% CI: -2.86 to 0.73). However, according to the interaction between Gender and COVID-19 in the subsample of papers published in 2020 (Table 2, Model 5 and 6), the average review time of COVID-19 papers by men first authors is 3.20 days (99% CI: -5.97 to -0.43) shorter than the average review time of COVID-19 papers by women first authors. This marginal gender difference is less pronounced in the subsample of last authors, with an estimated reduction of 2.22 days for men (99% CI: -5.35 to 0.90) (Table 2, Model 6). Robustness checks indicate comparable results across alternative model specifications (S4 and S5 Tables in S1 Appendix).
3.3. Survival plots
Further analyses indicate that scientists with prestigious affiliations have benefitted slightly more from fast-track peer reviewing during COVID-19 than colleagues from less reputed institutions. In Fig 2, we used survival plots to visualize within-group variability in the publication speed of COVID-19 articles by authors at highly ranked and lower-ranked institutions in 2020. The X-axes in the plots denote the length of the peer-review process in number of days, while the Y-axes show the proportion of papers still under review over the total number of submitted papers (e.g., 0.75 means that 75% of papers are still under review while 25% have been published).
As shown in the figure, we observe a slightly faster decline in the proportion of papers from highly-ranked institutions (teal line) on the left side of the X-axis, compared to papers from lower-ranked institutions (red line). Indeed, while ∼25% of papers from highly ranked institutions were published within 15 days from submission, this was only the case for ∼16–17% of the papers from lower ranked institutions. Trends are similar when the distinction between highly-ranked and lower-ranked institutions is based on the QS-ranking as opposed to the Leiden ranking (see S5 Fig in S1 Appendix).
In Fig 3, we explore how the publication speed of COVID-19 papers varies by peer-review procedure. Specifically, we compare the trajectories of papers from highly ranked and lower ranked institutions that have been submitted to journals with single- or double-blind peer-review procedures. As shown in the upper left and right panels, the trajectories of first-authored papers from highly ranked and lower ranked institutions (measured by the Leiden ranking) differ slightly across the single-blind and double-blind review procedures. The median difference in peer-review between higher and lower ranked institutions is ∼7 days for papers in single-blind journals (left panel) and ∼1 day for papers in double-blind journals (right panel). As shown in the left panel, the trend line for papers from highly ranked institutions shows a steep decline from day 0 to day 10 and is below the trend line for papers from lower-ranked institutions until day 110. In comparison, the trend lines for papers from highly ranked and lower ranked institutions undergoing double-blind review intersect at day 32–33. Results for the sample of last authors (lower panels) are comparable. In the complementary analysis, where top-ranked and lower-ranked universities are identified based on the QS-ranking, the advantage for top-ranked universities under the single-blind procedure compared to the double-blind procedure is slightly more pronounced (S6 Fig in S1 Appendix).
As a counterfactual analysis, we ran the same survival analysis as presented in Fig 3 but based on the 2019 and 2020 samples of papers on other topics than COVID-19. In these comparisons status-related differences in the publication speed across double blind and single-blind journals were miniscule (S8-S11 Figs in S1 Appendix).
4. Discussion
The current study aimed to assess whether the pandemic had widened existing institution and gender-related differences in publishing speed, by using bibliometric metadata on the duration of peer-review. While prior studies have used such data to document increases in the average publication speed during COVID-19 [17–19, 53], none have examined how changes in journal turn-around times varied by gender and institution status.
Using paper-level data from 80 medical journals, we show that researchers at highly ranked institutions have increased their publication speed slightly more in 2020 than have scientists at less reputed institutions. However, the average difference is small, and should be interpreted with caution (more on this below). We observe a more substantial status-related difference for COVID-19 research (compared to research on other topics), with papers from top-ranked institutions seeing faster review times than papers from lower-ranked institutions. Moreover, survival plots indicate that scientists with prestigious affiliations have benefitted the most from fast-track peer reviewing and especially so in journals with single-blind review procedures. Finally, our analysis of gender-related changes in publishing speed indicates small and inconsistent effects although we observe a slight difference in the average review time of COVID-19 papers first authored by women and men.
Consistent with previous evidence on COVID-19 [3, 14–20] and earlier health emergencies [54, 55], our analysis shows that scientists (and editors) have responded quickly to the pandemic by speeding up the publication process. On average, we observe a 23 day shorter review time for COVID-19 research compared to papers on other topics (Table 1). The need for quick information to support policy makers and practitioners in making evidence-based choices [18], and the resulting increase in publication speed may have accentuated status biases in peer-reviewing by advantaging researchers at prestigious locations.
At the same time, the observed institution-related disparities may be driven by (average) differences in the quality and immediate relevance of COVID-19 papers coming out of more and less affluent institutions. As mentioned earlier, scientists at less affluent research institutions may have lacked the necessary resources, infrastructure and support to adapt quickly to the changing working conditions imposed by COVID-19. This may have lowered the average quality of their submissions, while also making them slower to respond to resubmission requests, hence delaying the duration of the peer-review process.
While the present study does not allow us to disentangle the influence of these proposed mechanisms, it highlights a previously overlooked disparity in peer-reviewing during COVID-19 that has enabled research from high-status institutions to enter more quickly into the public domain.
As for gender, our finding that especially women first-authors (who on average tend to be earlier in their career than women last authors) of COVID-19 papers had a slightly slower publishing speed, align with previous research on gender disparities during the pandemic. According to this literature, lockdowns of schools, universities and daycares took the greatest toll on early career women scientists by considerably reducing their weekly time for research [32–37]. Further, evidence suggests that initial funding for COVID-19 related research was biased toward male applicants [36] and that women were less likely than men to first-author solicited COVID-19 papers [10]. Hence, lower access to resources and fewer (informal) connections to journal editors may also have slowed down early-career women’s publishing speed compared to men’s.
Our study has some limitations. First, we lack information on the duration of the peer-review process for rejected papers, which introduces a possible selection bias. If institution and gender-related disparities in journal rejection-rates have widened during COVID-19, this may bias our estimations of disparities in review times downwards.
Second, given the sensitivity of our analysis to uncertainties in the models and data, the small increase in institution- and gender related disparities for papers published in 2020 compared to 2019 should be interpreted with caution. Such uncertainties include database-related errors in the categorization and registration of medical research articles (e.g., spelling mistakes in author names and errors in institutional affiliations in PubMed and Web of Science) and inherent problems associated with non-random sampling procedures [56].
Third, by restricting our analysis to papers from institutions in the Leiden and QS university rankings, we reduced our initial sample by 34% and 58%, respectively. While many of the papers not covered in our analysis likely come out of private organizations, hospitals and health centers, some will be from less affluent research institutions of relevance to our study.
More generally, there are some areas that research may benefit from expanding upon. For example, while the current study accounted for variations at the journal level through a nested design, we did not directly examine how journal characteristics such as open-access status or print vs. online publication models may have influenced the institutional and gender-related differences observed in our study.
Similarly, lack of individual-level researcher data constrained what was possible in the current paper, but future research may benefit from running matched studies (e.g., using propensity-score matching) or in the optimal case within-subjects studies. This would help to reduce implicit differences between groups of interest.
Finally, COVID-19 related changes in editorial workloads, practices and decision criteria that we have not been able to measure may have impacted our findings [18]. Similarly, differences in scientists’ willingness and time to take on extra reviewer responsibilities in 2020 may have skewed the reviewer composition towards male evaluators and evaluators from more affluent countries and research institutions [6]. In the future, studies could consider how such changes may have affected journal turn-around times for different author groups.
In summary, our study indicates notable institution-related disparities in the peer-review duration of COVID-19 research compared to medical research on other topics. Future studies should examine how this disparity has affected the available evidence pool on variations in the transmission, prevention and treatment of the Corona virus across geographical regions.
Supporting information
Data Availability
Data are available on the Open Science Framework (https://osf.io/zebka/).
Funding Statement
This study was funded by Carlsbergfondet (the Carlsberg foundation) – Award # CF19-0566. P.I. M.W.N] The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
References
- 1.Kibbe MR. Consequences of the COVID-19 Pandemic on Manuscript Submissions by Women. JAMA Surg [Internet]. 2020. Sep 1;155(9):803–4. Available from: doi: 10.1001/jamasurg.2020.3917 [DOI] [PubMed] [Google Scholar]
- 2.Lerchenmüller C, Schmallenbach L, Jena A, Lerchenmueller M. Longitudinal analyses of gender differences in first authorship publications related to COVID-19. BMJ Open [Internet]. 2021. Apr 5 [cited 2021 Nov 3];11(4). Available from: https://pubmed.ncbi.nlm.nih.gov/33820790/ doi: 10.1136/bmjopen-2020-045176 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Madsen EB, Nielsen MW, Bjørnholm J, Jagsi R, Andersen JP. Author-level data confirm the widening gender gap in publishing rates during COVID-19. Elife. 2022. Mar 1;11. doi: 10.7554/eLife.76559 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.King MM, Frederickson ME. The Pandemic Penalty: The Gendered Effects of COVID-19 on Scientific Productivity. Socius [Internet]. 2021. Jan 1;7:23780231211006976. Available from: 10.1177/23780231211006977 [DOI] [Google Scholar]
- 5.Pinho-Gomes A-C, Peters S, Thompson K, Hockham C, Ripullone K, Woodward M, et al. Where are the women? Gender inequalities in COVID-19 research authorship. BMJ Glob Heal [Internet]. 2020. Jul 1;5(7):e002922. Available from: http://gh.bmj.com/content/5/7/e002922.abstract doi: 10.1136/bmjgh-2020-002922 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.Squazzoni F, Bravo G, Grimaldo F, García-Costa D, Farjam M, Mehmani B. Gender gap in journal submissions and peer review during the first wave of the COVID-19 pandemic. A study on 2329 Elsevier journals. PLoS One [Internet]. 2021. Oct 20 [cited 2021 Nov 3];16(10):e0257919. Available from: https://journals.plos.org/plosone/article?id=10.1371/journal.pone.0257919 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Andersen JP, Nielsen MW, Simone NL, Lewiss RE, Jagsi R. COVID-19 medical papers have fewer women first authors than expected. Elife [Internet]. 2020. Jun 1 [cited 2021 Apr 26];9:1–7. Available from: /pmc/articles/PMC7304994/ doi: 10.7554/eLife.58807 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Cui R, Ding H, Zhu F. Gender Inequality in Research Productivity During the COVID-19 Pandemic. Manuf Serv Oper Manag [Internet]. 2021. Jun 16; Available from: 10.1287/msom.2021.0991 [DOI] [Google Scholar]
- 9.Vincent-Lamarre P, Sugimoto CR, Larivière V. Monitoring Women’s Scholarly Production during the COVID‐19 Pandemic. Nature Index. (blogpost). 2020. [Google Scholar]
- 10.Brown C, Novick TK, Jacobs EA. Gender Disparities in Authorship of Invited Manuscripts During the COVID-19 Pandemic. Women’s Heal Reports [Internet]. 2021. Jun 1 [cited 2022 Jun 28];2(1):149–53. Available from: https://www.liebertpub.com/doi/10.1089/whr.2021.0023 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Anabaraonye N, Tsai CJ, Saeed H, Chino F, Ekpo E, Ahuja S, et al. Impact of the Early COVID-19 Pandemic on Gender Participation in Academic Publishing in Radiation Oncology. Adv Radiat Oncol. 2022. Mar 1;7(2):100845. doi: 10.1016/j.adro.2021.100845 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Deryugina T, Shurchkov O, Stearns JE. COVID-19 Disruptions Disproportionately Affect Female Academics. SSRN [Internet]. 2021. Jan 25 [cited 2021 Nov 3]; Available from: https://www.nber.org/papers/w28360 [Google Scholar]
- 13.Myers KR, Tham WY, Yin Y, Cohodes N, Thursby JG, Thursby MC, et al. Unequal effects of the COVID-19 pandemic on scientists. Nat Hum Behav [Internet]. 2020;4(9):880–3. Available from: doi: 10.1038/s41562-020-0921-y [DOI] [PubMed] [Google Scholar]
- 14.Cai X, Fry C V., Wagner CS. International collaboration during the COVID-19 crisis: autumn 2020 developments [Internet]. Vol. 126, Scientometrics. Springer Science and Business Media B.V.; 2021. [cited 2021 Apr 29]. p. 3683–92. Available from: 10.1007/s11192-021-03873-7 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Malekpour M-R, Abbasi-Kangevari M, Azadnajafabad S, Ghamari S-H, Rezaei N, Rezazadeh-Khadem S, et al. How the scientific community responded to the COVID-19 pandemic: A subject-level time-trend bibliometric analysis. Radfar A, editor. PLoS One [Internet]. 2021. Sep 30 [cited 2022 Jun 28];16(9):e0258064. Available from: doi: 10.1371/journal.pone.0258064 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Fry C V., Cai X, Zhang Y, Wagner CS. Consolidation in a crisis: Patterns of international collaboration in early COVID-19 research. Bornmann L, editor. PLoS One [Internet]. 2020. Jul 21 [cited 2021 Apr 19];15(7):e0236307. Available from: https://dx.plos.org/10.1371/journal.pone.0236307 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Else H. How a torrent of COVID science changed research publishing—in seven charts. Vol. 588, Nature. NLM (Medline); 2020. p. 553. [DOI] [PubMed] [Google Scholar]
- 18.Kun Á. Time to Acceptance of 3 Days for Papers About COVID-19. Publications [Internet]. 2020. Jun 3 [cited 2021 Apr 19];8(2):30. Available from: https://www.mdpi.com/2304-6775/8/2/30 [Google Scholar]
- 19.Zhang L, Zhao W, Sun B, Huang Y, Glänzel W. How scientific research reacts to international public health emergencies: a global analysis of response patterns. Scientometrics [Internet]. 2020. Jul 1 [cited 2021 Apr 19];124(1):747–73. Available from: /pmc/articles/PMC7282204/ doi: 10.1007/s11192-020-03531-4 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Aggarwal A, Agosti E, Singh PM, Varshini A, Garg K, Chaurasia B, et al. Scientometric analysis of medical publications during COVID-19 pandemic: The twenty-twenty research boom. Minerva Med. 2021. Oct 1;112(5):631–40. doi: 10.23736/S0026-4806.21.07489-9 [DOI] [PubMed] [Google Scholar]
- 21.Horbach SPJM. Pandemic publishing: Medical journals strongly speed up their publication process for COVID-19. Quant Sci Stud [Internet]. 2020. Aug 1 [cited 2021 Apr 29];1(3):1056–67. Available from: 10.1162/qss_a_00076 [DOI] [Google Scholar]
- 22.Horbach SPJM. No time for that now! Qualitative changes in manuscript peer review during the Covid-19 pandemic. Res Eval [Internet]. 2021. Jan 5 [cited 2021 May 3]; Available from: https://academic.oup.com/rev/advance-article/doi/10.1093/reseval/rvaa037/6064166 [Google Scholar]
- 23.Homolak J, Kodvanj I, Virag D. Preliminary analysis of COVID-19 academic information patterns: a call for open science in the times of closed borders. Scientometrics [Internet]. 2020. Sep 1 [cited 2021 Apr 29];124(3):2687–701. Available from: doi: 10.1007/s11192-020-03587-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24.Bodenhausen G, Lichtenstein M. Social stereotypes and information-processing strategies: the impact of task complexity. J Pers Soc Psychol [Internet]. 1987. [cited 2021 Nov 3];52(5):871–80. Available from: https://pubmed.ncbi.nlm.nih.gov/3585699/ doi: 10.1037//0022-3514.52.5.871 [DOI] [PubMed] [Google Scholar]
- 25.Maule AJ, Hockey GRJ, Bdzola L. Effects of time-pressure on decision-making under uncertainty: Changes in affective state and information processing strategy. Acta Psychol (Amst) [Internet]. 2000. [cited 2021 Nov 3];104(3):283–301. Available from: /record/2000-08548-001 doi: 10.1016/s0001-6918(00)00033-0 [DOI] [PubMed] [Google Scholar]
- 26.Stepanikova I. Racial-ethnic biases, time pressure, and medical decisions. J Health Soc Behav [Internet]. 2012. Sep [cited 2021 Nov 3];53(3):329–43. Available from: https://pubmed.ncbi.nlm.nih.gov/22811465/ [DOI] [PubMed] [Google Scholar]
- 27.Yu R. Stress potentiates decision biases: A stress induced deliberation-to-intuition (SIDI) model. Neurobiol Stress. 2016. Jun 1;3:83–95. doi: 10.1016/j.ynstr.2015.12.006 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28.Van Knippenberg A, Dijksterhuis A, Vermeulen D. Judgement and memory of a criminal act: the effects of stereotypes and cognitive load. Eur J Soc Psychol [Internet]. 1999. Mar 1;29(2–3):191–201. Available from: [DOI] [Google Scholar]
- 29.Budden AE, Tregenza T, Aarssen LW, Koricheva J, Leimu R, Lortie CJ. Double-blind review favours increased representation of female authors. Trends Ecol Evol. 2008. Jan 1;23(1):4–6. doi: 10.1016/j.tree.2007.07.008 [DOI] [PubMed] [Google Scholar]
- 30.Crane D. The Gatekeepers of Science: Some Factors Affecting the Selection of Articles for Scientific Journals. Am Sociol [Internet]. 1967. Nov 3;2(4):195–201. Available from: http://www.jstor.org/stable/27701277 [Google Scholar]
- 31.Tomkins A, Zhang M, Heavlin WD. Reviewer bias in single- versus double-blind peer review. Proc Natl Acad Sci U S A [Internet]. 2017. Nov 28 [cited 2021 Jan 11];114(48):12708–13. Available from: www.pnas.org/lookup/suppl/doi:10.1073/pnas.1707323114/-/DCSupplemental.www.pnas.org/cgi/doi/10.1073/pnas.1707323114 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 32.Ross JS, Gross CP, Desai MM, Hong Y, Grant AO, Daniels SR, et al. Effect of Blinded Peer Review on Abstract Acceptance. JAMA [Internet]. 2006. Apr 12;295(14):1675–80. Available from: doi: 10.1001/jama.295.14.1675 [DOI] [PubMed] [Google Scholar]
- 33.Knobloch-Westerwick S, Glynn CJ, Huge M. The Matilda Effect in Science Communication: An Experiment on Gender Bias in Publication Quality Perceptions and Collaboration Interest. Sci Commun [Internet]. 2013. Feb 6;35(5):603–25. Available from: 10.1177/1075547012472684 [DOI] [Google Scholar]
- 34.Moss-Racusin CA, Dovidio JF, Brescoll VL, Graham MJ, Handelsman J. Science faculty’s subtle gender biases favor male students. Proc Natl Acad Sci U S A [Internet]. 2012. Oct 9 [cited 2021 Nov 3];109(41):16474–9. Available from: http://www.ncbi.nlm.nih.gov/pubmed/22988126 doi: 10.1073/pnas.1211286109 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35.Hengel E. Publishing while Female. Are women held to higher standards? Evidence from peer review. Cambridge working papers in economics. 2017. [Google Scholar]
- 36.Breuning M, Fattore C, Ramos J, Scalera J. The Great Equalizer? Gender, Parenting, and Scholarly Productivity During the Global Pandemic. PS Polit Sci Polit [Internet]. 2021. Jul 22 [cited 2021 Nov 3];54(3):427–31. Available from: https://www.cambridge.org/core/product/identifier/S1049096520002036/type/journal_article [Google Scholar]
- 37.Krukowski RA, Jagsi R, Cardel MI. Academic Productivity Differences by Gender and Child Age in Science, Technology, Engineering, Mathematics, and Medicine Faculty During the COVID-19 Pandemic. J Women’s Heal [Internet]. 2020. Nov 18;30(3):341–7. Available from: 10.1089/jwh.2020.8710 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 38.Staniscuaski F, Kmetzsch L, Soletti RC, Reichert F, Zandonà E, Ludwig ZMC, et al. Gender, Race and Parenthood Impact Academic Productivity During the COVID-19 Pandemic: From Survey to Action [Internet]. Vol. 12, Frontiers in Psychology. 2021. p. 1640. Available from: https://www.frontiersin.org/article/10.3389/fpsyg.2021.663252 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39.Witteman HO, Haverfield J, Tannenbaum C. COVID-19 gender policy changes support female scientists and improve research quality. Proc Natl Acad Sci U S A [Internet]. 2021. Feb 9 [cited 2021 Nov 3];118(6). Available from: http://www.ncbi.nlm.nih.gov/pubmed/33531366 doi: 10.1073/pnas.2023476118 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40.Yildirim TM, Eslen-Ziya H. The differential impact of COVID-19 on the work conditions of women and men academics during the lockdown. Gender, Work Organ [Internet]. 2021. Jan 1;28(S1):243–9. Available from: doi: 10.1111/gwao.12529 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 41.CADTH. Canadian Agency for Drugs and Technologies in Health [Internet]. 2020 [cited 2021 Sep 29]. Available from: https://covid.cadth.ca/literature-searching-tools/cadth-covid-19-search-strings/#covid-19-medline.
- 42.Boekhout H, van der Weijden I, Waltman L. Gender differences in scientific careers: A large-scale bibliometric analysis. 18th Int Conf Sci Inf ISSI 2021. 2021;145–56. [Google Scholar]
- 43.CWTS. CWTS Leiden Ranking—Information—Indicators [Internet]. 2020 [cited 2021 Sep 29]. Available from: https://www.leidenranking.com/information/indicators
- 44.Gelman A. Scaling regression inputs by dividing by two standard deviations. Stat Med. 2008. Jul;27(15):2865–73. doi: 10.1002/sim.3107 [DOI] [PubMed] [Google Scholar]
- 45.West BT, Welch KB, Galecki AT. Linear Mixed Models [Internet]. Linear Mixed Models. Chapman and Hall/CRC; 2014. [cited 2021 May 17]. Available from: https://www-taylorfrancis-com.ep.fjernadgang.kb.dk/https://www-taylorfrancis-com.ep.fjernadgang.kb.dk/books/mono/10.1201/b17198/linear-mixed-models-brady-west-kathleen-welch-andrzej-galecki [Google Scholar]
- 46.Gelman A, Hill J. Data Analysis Using Regression and Multilevel/Hierarchical Models [Internet]. Cambridge: Cambridge University Press; 2006. [cited 2021 Sep 1]. Available from: http://ebooks.cambridge.org/ref/id/CBO9780511790942 [Google Scholar]
- 47.Allison PD, Long JS. Departmental Effects on Scientific Productivity. Am Sociol Rev [Internet]. 1990. Nov 15;55(4):469–78. Available from: http://www.jstor.org/stable/2095801 [Google Scholar]
- 48.Liu Xian. Introduction. In: Survival Analysis [Internet]. 2012. p. 1–19. (Wiley Online Books; ). Available from: 10.1002/9781118307656.ch1 [DOI] [Google Scholar]
- 49.Bates D, Mächler M, Bolker B, Walker S. Fitting Linear Mixed-Effects Models Using lme4. J Stat Softw [Internet]. 2015. Oct 7;67(1 SE-Articles):1–48. Available from: https://www.jstatsoft.org/index.php/jss/article/view/v067i01 [Google Scholar]
- 50.Blair G, Cooper J, Coppock A, Humphreys M. Declaring and Diagnosing Research Designs. Am Polit Sci Rev [Internet]. 2019. Aug 1 [cited 2021 Nov 2];113(3):838–59. Available from: https://www.cambridge.org/core/journals/american-political-science-review/article/declaring-and-diagnosing-research-designs/3CB0C0BB0810AEF8FF65446B3E2E4926 doi: 10.1017/s0003055419000194 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 51.Lüdecke D. sjPlot: Data Visualization for Statistics in Social Science. 2021. [Google Scholar]
- 52.Kassambara A, Kosinski M, Biecek P, Fabian S. survminer: Drawing Survival Curves using “ggplot2.” 2021. [Google Scholar]
- 53.Helliwell JA, Bolton WS, Burke JR, Tiernan JP, Jayne DG, Chapman SJ. Global academic response to COVID-19: Cross-sectional study. Learn Publ [Internet]. 2020. Oct 1;33(4):385–93. Available from: doi: 10.1002/leap.1317 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 54.Palayew A, Norgaard O, Safreed-Harmon K, Andersen TH, Rasmussen LN, Lazarus J V. Pandemic publishing poses a new COVID-19 challenge [Internet]. Vol. 4, Nature Human Behaviour. Nature Research; 2020. [cited 2021 Apr 29]. p. 666–9. Available from: www.nature.com/nathumbehav [DOI] [PubMed] [Google Scholar]
- 55.Khanali J, Malekpour MR, Kolahi AA. Improved dynamics of sharing research findings in the COVID-19 epidemic compared with the SARS and Ebola epidemics. BMC Public Health [Internet]. 2021. Dec 1 [cited 2022 Jun 28];21(1):105. Available from: https://bmcpublichealth.biomedcentral.com/articles/10.1186/s12889-020-10116-6 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 56.Berk RA, Freedman DA. Statistical assumptions as empirical commitments. Law, punishment, Soc Control Essays Honor Sheldon Messin. 2003;2:235–54. [Google Scholar]