Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2021 Jun 4.
Published in final edited form as: Nat Med. 2021 Jan;27(1):6–7. doi: 10.1038/s41591-020-01203-7

Automated screening of COVID-19 preprints: can we help authors to improve transparency and reproducibility?

Tracey Weissgerber 1,2,, Nico Riedel 1, Halil Kilicoglu 3, Cyril Labbé 4, Peter Eckmann 5,6, Gerben ter Riet  7,8, Jennifer Byrne 9,10, Guillaume Cabanac 11, Amanda Capes-Davis 12, Bertrand Favier 13, Shyam Saladi 14, Peter Grabitz 1,2, Alexandra Bannach-Brown 1, Robert Schulz 1,2, Sarah McCann 1,2, Rene Bernard 15, Anita Bandrowski 5,6
PMCID: PMC8177099  NIHMSID: NIHMS1704244  PMID: 33432174

To the Editor—The COVID-19 pandemic has thrust preprints into the spotlight, attracting attention from the media and the public, as well as from scientists. Preprints are articles not yet published in a peer-reviewed journal, and as such they offer a unique opportunity to improve reporting. The Automated Screening Working Group (https://scicrunch.org/ASWG/about/COVIDPreprint) aims to provide rapid feedback that may help authors of COVID-19 preprints to improve their transparency and reproducibility.

One quarter of COVID-19 papers published have been preprints. Most of these appear on medRxiv; others appear on bioRxiv or other servers1. Although publishing results in preprints allows them to be posted rapidly, the absence of traditional peer review has raised concerns about preprint quality. Unfortunately, it has been impossible for scientists to keep pace with the thousands of COVID-19 preprints published since February. Preprints are vetted before posting to confirm that they describe scientific studies and to prevent posting on topics that could damage public health; however, routine assessment of manuscript quality or flagging of common reporting problems is not feasible at this scale.

Although automated screening is not a replacement for peer review, automated tools can identify common problems. Examples include failure to state whether experiments were blinded or randomized2, failure to report the sex of participants2 and misuse of bar graphs to display continuous data3. We have been using six tools48 to screen all new medRxiv and bioRxiv COVID-19 preprints (Table 1). New preprints are screened daily9. By this means, reports on more than 8,000 COVID preprints have been shared using the web annotation tool hypothes.is (RRID:SCR_000430) and have been tweeted out via @SciScoreReports (https://hypothes.is/users/sciscore). Readers can access these reports in two ways. The first option is to find the link to the report in the @SciScoreReports tweet in the preprint’s Twitter feed, located in the metrics tab. The second option is to download the hypothes. is bookmarklet. In addition, readers and authors can reply to the reports, which also contain information on solutions.

Table 1 |.

Tools used to screen COVID-19 preprints

Tool Screens for Link and RRID
SciScore4 Blinding, randomization, sample-size calculations, sex/gender, ethics and consent statements, resources, RRIDs http://sciscore.com; RRID:SCR_016251
ODDPub5 Open data, open code https://github.com/quest-bih/oddpub; RRID:SCR_018385
Limitation- Recognizer6 Author-acknowledged limitations https://github.com/kilicogluh/limitation-recognizer; RRID:SCR_018748
Barzooka Bar graphs of continuous data https://quest-barzooka.bihealth.org; RRID:SCR_018508
JetFighter7 Rainbow color maps https://jetfighter.ecrlife.org; RRID:SCR_018498
Seek and Blastn8 Correct identification of nucleotide sequencesa http://scigendetection.imag.fr/TPD52/; RRID:SCR_016625
a

Semiautomated tool—requires human confirmation of results. rrID, research resource identifier.

Screening of 6,570 medRxiv and bioRxiv COVID-19 preprints posted before 19 July revealed several interesting results. 13.6% of preprints shared open data and 14.3% shared open code, making it easier for others to reuse data or reproduce results. Approximately one third (34.4%) of COVID-19 preprints acknowledged at least one study limitation. 7.3% of preprints included bar graphs of continuous data. This is problematic because many different datasets can lead to the same bar graph, and the actual data may suggest different conclusions from those implied by the summary statistics alone3. Therefore, authors should use dot plots, box plots or violin plots instead3. Among papers with color maps, 7.6% used rainbow colormaps, which are not colorblind safe and also create visual artifacts for viewers with normal vision7. Rainbow color maps should be replaced with more-informative color maps that are perceptually uniform and colorblind accessible, such as viridis7. 1,775 preprints (27%) contained an ethics approval statement for human or animal research. This suggests that nearly three quarters of COVID-19 preprints are secondary or tertiary analyses, modeling studies or cell line studies that do not require approval. Although there are known sex differences in COVID-1910, only 20% of all COVID-19 preprints, and 38% of preprints with an ethics approval statement, address sex as a biological variable. Statements regarding sample size calculations (1.4%), blinding (2.7%) and randomization (11.4%) were uncommon, even among studies that contained a human ethics statement (present in 2.4%, 5.4% and 12.6%, respectively). Many COVID-19 preprints are modeling studies, however, and hence these criteria are not always relevant. 6.1% of preprints used nonhuman organisms, mainly mice. Among the 552 preprints that included cell lines, 7% described how the cell lines were authenticated (e.g., short tandem repeat profiling) or were kept free of contamination (e.g., mycoplasma detection tests).

Our work shows that it is feasible to conduct large-scale automated screening of preprints and provide rapid feedback to authors and readers. Automated tools are not perfect—they make mistakes, and they cannot always determine whether a problem is relevant to a given paper. Moreover, some problems are too complex for automated tools to detect. Despite these limitations, automated tools can quickly flag potential problems and may complement peer reviews. We hope that these reports will raise awareness about factors that affect transparency and reproducibility, while helping authors to improve their manuscripts. Further research is needed to determine whether automated tools improve reporting.

Acknowledgements

SciScore was funded by the US National Institutes of Health (OD024432, DA039832, DK097771, MH119094, and HHSN276201700124P). The development and application of Seek & Blastn is supported by grants from the US Office of Research Integrity, grant ID ORIIR180038-01-00 (J.A.B., C.L.), and from the National Health and Medical Research Council of Australia, Ideas grant ID APP1184263 (J.A.B., C.L., A.C.D.). Development of Limitation-Recognizer was partially supported by the intramural research program of the NIH and National Library of Medicine.

Footnotes

Competing interests

A.B. is a cofounder of SciCrunch Inc.

References

RESOURCES