Skip to main content
PLOS ONE logoLink to PLOS ONE
. 2021 Aug 17;16(8):e0256352. doi: 10.1371/journal.pone.0256352

Clinical validation of an open-access SARS-COV-2 antigen detection lateral flow assay, compared to commercially available assays

Christine M Bachman 1,*,#, Benjamin D Grant 1,#, Caitlin E Anderson 1, Luis F Alonzo 1, Spencer Garing 1, Sam A Byrnes 1,¤, Rafael Rivera 1, Stephen Burkot 1, Alexey Ball 1, James W Stafford 1, Wenbo Wang 1, Dipayan Banik 1, Matthew D Keller 1, David M Cate 1,¤, Kevin P Nichols 1,¤, Bernhard H Weigl 1, Puneet Dewan 1
Editor: Etsuro Ito2
PMCID: PMC8370603  PMID: 34403456

Abstract

Rapid tests for SARS-COV-2 infection are important tools for pandemic control, but current rapid tests are based on proprietary designs and reagents. We report clinical validation results of an open-access lateral flow assay (OA-LFA) design using commercially available materials and reagents, along with RT-qPCR and commercially available comparators (BinaxNOW® and Sofia®). Adult patients with suspected COVID-19 based on clinical signs and symptoms, and with symptoms ≤7 days duration, underwent anterior nares (AN) sampling for the OA-LFA, Sofia®, BinaxNOW ™, and RT-qPCR, along with nasopharyngeal (NP) RT-qPCR. Results indicate a positive predictive agreement with NP sampling as 69% (60% -78%) OA-LFA, 74% (64% - 82%) Sofia®, and 82% (73% - 88%) BinaxNOW™. The implication for these results is that we provide an open-access LFA design that meets the minimum WHO target product profile for a rapid test, that virtually any diagnostic manufacturer could produce.

Introduction

With more than 3,250,000 deaths worldwide [1] due to the Covid-19 pandemic, early detection of infectious patients is required to aid in efforts to interrupt transmission. One of the most effective tools to detect potentially infectious cases of COVID-19 are rapid diagnostic tests for the virus that causes the disease, SARS-COV-2. Transmission models have predicted early detection in the community coupled with isolation may accelerate reductions in transmission [2, 3].

At present, most health facilities rely on polymerase chain reaction (PCR) -based tests, which are based on amplification of genetic material for detection of the virus. While enormously sensitive, PCR poses several challenges. For example, time to results may be prolonged, cold chain may be required to transfer samples to lab facilities, expensive reagents are utilized, and trained lab technicians are needed to perform the testing [4]. Simpler and less-expensive antigen-based rapid tests have been developed to address these issues. While they may not be as sensitive as PCR, antigen-based rapid tests are generally inexpensive, amenable to mass-production, can be conducted by minimally trained individuals, and offer nearly immediate results for public health actions [5]. The Food and Drug Administration (FDA) has provided Emergency Use Authorizations (EUA) to (as of 5/17/2021) 24 rapid antigen diagnostic tests, most of which are based on detection of SARS-COV2 nucleocapsid protein [6] All of these assays are proprietary and are produced by commercial entities.

We previously published the development of a rapid test in the lateral flow assay (LFA) format to detect the nucleocapsid protein of the SARS-CoV-2 virus [7]. This assay is distinct from available commercial assays in that the assay design and architecture were released as open access, and that all reagents are commercially available. In-vitro performance of the assay met the World Health Organization (WHO) target product profile (TPP) for a rapid diagnostic test [8]. Specifically, the analytical sensitivity achieved the target WHO target of 106 genomic copies/mL. Here we report the subsequent clinical validation procedures and results from this prototype lateral flow assay, in comparison to RT-qPCR results from swabs taken from nasopharyngeal and anterior nares sampling. Two additional experimental approaches were evaluated in this study–a mobile phone application to image and interpret a completed LFA, and a system to evaluate a patient’s sense of smell. We include on-site rapid test results from two commercially available rapid tests as comparators, the Quidel® Sofia®SARS Antigen FIA and the Abbot® BinaxNOW™ COVID-19 Ag card to compare performance.

Methods

Study design

An open-label prospective study design of adults aged 18 and above presenting with symptoms of Covid-19 in 2 private clinics in Los Angeles County, California. These sites offer drive-through testing services as a part of standard clinical care. Ethics approval was obtained by IntegReview IRB (GHLPOC-01), USA. The IRB committee approved of consent obtained electronically.

Participants

Eligibility criteria include symptomatic adult patients (age ≥18 years) with an onset of symptoms within 7 days. Symptoms included at least one of the following: fever, cough, shortness of breath, fatigue, muscle or body aches, loss of taste or smell, headache, congestion of nose, sore throat, nausea, vomiting, or diarrhea. Participants were consecutively offered voluntary enrollment in the study.

Point-of-care test methods

The index test being evaluated is the OA-LFA. Briefly, the OA-LFA is a rapid diagnostic test with time to results in 30 minutes and uses anterior nares (AN) samples. Comparator point-of-care (POC) tests were also used to evaluate against the OA-LFA performance. These tests included the Sofia® SARS Antigen Fluorescent Immunoassay (FIA) (Quidel Corporation, San Diego, California, US) and the BinaxNOW™ Rapid Antigen Test for SARS-CoV2 (Abbott, Abbott Park, Illinois, US). One nasopharyngeal swab (NP) and four AN swabs were taken per patient. The NP swab was collected for reference RT-qPCR testing, and one AN swab was collected for each POC test. An additional AN swab was collected for RT-qPCR and reference antigen-testing. The rapid test results were blinded between research staff.

The OA-LFA utilized a PurFlock Ultra Elongated swab (Puritan Medical Products, Guilford, Maine, US, catalog no. 25-3806-U BT) for AN sampling. The BinaxNOW™ and Sofia® tests utilized swabs provided with each respective kit. For each test, both nostrils were swabbed by a healthcare professional. The swab was inserted into the anterior nares portion of the nostril until resistance was met and rotated five times in one direction before removing and repeating the process on the other nostril. First, the Sofia® and OA-LFA tests were each swabbed in opposite nostrils to allow each test a sample from a previously unsampled nostril. Specifically, the OA-LFA swab was used in the right nostril and the Sofia® swab was used in the left nostril prior to sampling the other nostril. After samples were taken for both Sofia® and OA-LFA tests, the BinaxNOW™ swab and reference AN swab were performed in a similar manner. The three LFAs were tested immediately after collection. The reference AN swab, also a Puritan PurFlock swab, was placed back in its dry transport container and placed on ice. The NP swab (Puritan Medical Products, Guilford, Maine, US, catalog no. 25-3306-H) was placed into a 15 mL screw top tube. At the end of each day, the reference AN and NP swabs were transported to a -20°C unit where they remained for up to two weeks before transport and storage at -80°C.

Immediately after sampling, the OA-LFA swab was placed into a sample tube containing 500 μL of running buffer (2% IGEPAL CA-630 [Sigma, St Louis, MO, I8896] in 1X PBS [Thermo Fisher, Waltham, MA 10010023]). The swab was rotated in the tube three times and then allowed to sit for 1 minute, after which the swab was discarded. An exact volume transfer pipette (Electron Microscopy Science, 70969–29) was used to remove 150 μL from the sample tube and gently pipetted onto the sample port of the LFA. After 30 minutes, the OA-LFA was read visually followed by quantitation of the test line intensity via commercial LFA reader (AX-2X-S, Axxin, Australia). Line intensity was also quantified via a GH Labs developed mobile LFA reader application (described below). The Sofia® and BinaxNOW™ were run per kit insert without modifications.

Reference method testing

The AN swab and NP swab were both used for reference RT-qPCR testing at GH Labs after being stored and shipped at –80°C. Additionally, antigen levels were measured utilizing an electrochemiluminescence immunoassay on the Meso Scale Discovery (MSD) MESO QuickPlex SQ 120 (MSD, Rockville, Maryland).

Nasopharyngeal swab rehydration

Swabs were removed from the –80°C freezer and set to warm to room temperature for 10 minutes prior to rehydration. Then, 500 μL of Tris-EDTA buffer solution, pH 7.4 was added to NP swab for 10 minutes and incubated at room temperature. Swab was vortexed and buffer solution was removed from tube for RNA purification.

Anterior nares swab rehydration

The AN swabs were placed in a tube containing 500 μL of Viral Transport Medium (VTM). The VTM was made using the Center for Disease Control Recipe sterile Hanks Balanced Salt Solution (1X HBSS with calcium and magnesium ions, no phenol red) with 2% FBS (sterile heat inactivated fetal bovine serum), 100μg/mL Gentamicin and 0.5μg/mL Amphotericin B final concentration). The swab was rotated in the VTM three times and then allowed to sit for 10 minutes, after which the swab was discarded. 81 μL of the sample was set aside for the MSD immunoassay and the remainder was handed off for RT-qPCR.

PCR testing

The RT-qPCR uses the CDC 2019-nCoV Real-Time RT-PCR Diagnostic Panel (SOP: CDC-006-00006). RNA is extracted from nasal swab samples using the QIAamp Viral Mini Kit as recommended by the CDC. The extracted RNA is amplified using a multiplexed version of the CDC 2019-nCoV Real-Time RT-qPCR targets (three targets: N1, N2, and RP) with the QuantaBio qScript XLT 1-step mix on a BioRad CFX96 Real-Time PCR Detection System [9]. The N1 gene is used to quantify SARS-CoV-2 viral load and the RP gene is used to quantify human RNA load. The RT-qPCR standards are control plasmids for both SARS-CoV-2 and Human RP gene which are procured from ITD. Both control plasmids are quantified in-house using the BioRad Digital PCR system. Samples were considered positive with a CT value of <40 [10] A standard curve for the N1 gene is run on each plate and utilized to calculate the viral-load from the CT value [9]. The N1 gene is used to quantify SARS-CoV-2 viral load and the RP gene is used to quantify human RNA load. The RT-qPCR standards are control plasmids for both SARS-CoV-2 and Human RP gene which are procured from ITD. Both control plasmids are quantified in-house using the BioRad Digital PCR system. Samples were considered positive with a CT value of <40 [10] A standard curve for the N1 gene is run on each plate and utilized to calculate the viral load from the CT value. Scientists performing RT-qPCR were blinded to antigen test results obtained at the clinical site.

Anterior nares antigen level testing

Nucleocapsid antigen levels were determined using a GH Labs developed Meso Scale Discovery (MSD) immunoassay. The sandwich assay utilizes two monoclonal antibodies, 40143-MM08 and 40143-MM05 (Sino Biologicals), that exhibit high specificity to SARS and SARS-CoV-2 nucleocapsid proteins [11]. Sample lysis is achieved utilizing Igepal CA-630 (Sigma I8896). In addition to Igepal, the lysis buffer contains 1.5 mg/mL HBR-1 (Heterophilic Blocking Reagent 1, Part 3KC533, Scantibodies Laboratory Inc., USA) to reduce background signal. Quantification was performed using a standard curve prepared with recombinantly produced SARS-CoV-2 nucleocapsid (Acro Biosystems, Nun-C5227). A cut-off point of 10 pg/mL was chosen as the threshold for MSD positivity. This value was chosen to provide a minimum of 95% negative predicative agreement against both NP and AN RT-qPCR.

Mobile reader application

An Android application (“reader app”) was developed to capture images of the completed LFA tests and to determine whether control and/or test lines could be detected in the images. The reader app was based on the “rdt-scan” framework originally developed by Park et al. [12] and licensed under an open-source BSD-3 clause. Several features were added to or modified from the original framework to customize the reader app for use in this study. As detailed further in (see S1 File), this included placing a sticker with two ArUco codes on each cassette (see S1 Fig) to provide visual features (“keypoints”) for the app to recognize on the otherwise-blank cassettes, as well as modifying how the app determines which keypoints are the most reliable. Further updates included configuring the app to search for control and test lines in the proper locations, optimizing the preprocessing steps prior to peak detection, and determining the proper threshold to distinguish the test line from background noise after converting the cropped test strip region into 1-dimensional intensity profiles along the direction of flow, averaged across the width of the strip.

After the app was initiated, the user had a live video view from the camera, and the app ran several real-time analyses on each video frame that enable prompts to the user to adjust various factors related to the image acquisition (see S2 Fig). Once the app recognized that the live image contained the designated LFA (by identifying the ArUco codes) and passed all quality checks, including a new check for flatness of the phone, it saved the current video frame for processing. The authors also added a trigger at this point to immediately acquire a new, high-resolution still image, which was used for results presented in this manuscript.

In this study, the reader app was used across four Moto G7 Power mobile phones to acquire images from 155 of the 170 GH Labs tests. Based on findings from pre-trial testing with spiked serial dilutions, the metric reported to the user for control and test lines was the prominence, or peak height above baseline, calculated from the 1D intensity profile of the red color channel extracted from the RGB mobile phone images (see S1 File for further details). If no such peak greater than a set threshold could be found in the pre-defined control or test line regions of the image, the app reported “no line found.” For computing receiver operating characteristic (ROC) curves and concordance values against visual read and PCR results, the peak height above baseline threshold was adjusted to find the optimal cutoff point.

Clinical symptoms and standardized anosmia assessment

Patients self-identified for signs and symptoms of COVID-19, including fever or chills, cough, shortness of breath or difficulty breathing, fatigue, muscle or body aches, headache loss of taste or smell, sore throat, congestion or runny nose, nausea, or vomiting, and/or diarrhea. Standardized assessment for anosmia was conducted in parallel. USMELLIT is an integrated card and mobile application to systematically assess patients for anosmia loss of smell [13]. Cards present a random sequence of scent challenges in a ‘scratch and sniff’ format. The health care provider administered a card to the patient and input the patient ID into the app after scanning the QR code on the card. Patients were requested to scratch 5 sections of the card and associate an object with every section they smelled. Examples of smells associated with the card include orange, banana, roses, mint, and no-scent. Patients were scored on the number of correct answers they received out of 5. This number was captured in the case report form later for data analysis.

Analysis

Sensitivity and specificity were calculated for each POC test compared to AN and NP RT-qPCR. Exact binomial confidence intervals were calculated in R using epiR [14]. Sensitivity was also calculated for tests stratified by AN and NP RT-qPCR viral load. Differences between the sensitivity and specificity of each LFA relative to AN and NP RT-qPCR were assessed in R using McNemar’s mid-P test [15].

Results

In the month of December 2020, a total of 170 subjects were enrolled at two drive-in sites in Riverside and Orange County. As shown in Fig 1, 110 of the 170 individuals (64%) were positive by NP PCR and 92 (54%) were positive by AN PCR. The logarithmic value of NP and AN viral load showed strong correlation (Pearson’s Correlation Coefficient of 0.897), as shown in Fig 2. All subjects positive by AN PCR were also positive by NP PCR. Expectedly, discordant results (negative AN PCR and positive NP PCR) occurred when NP PCR viral loads were low. Overall, 84% of NP PCR positive cases were also detected by AN PCR.

Fig 1. Flowchart of samples positive and negative for study tests.

Fig 1

Fig 2. Nasopharyngeal vs anterior nares PCR viral load results.

Fig 2

Cases where both swabs were positive are blue, both negative are red, and nasopharyngeal are in green. There is a clear positive correlation; however, the anterior nares swabs from the same individual have a lower viral load.

The sensitivity and specificity for the OA-LFA, Sofia®, BinaxNOW™ tests and the MSD antigen reference test relative to both NP and AN PCR are shown in Table 1. Overall and in this patient population, amongst the rapid tests, BinaxNOW™ showed the highest sensitivity, followed by Sofia® and OA-LFA. All three rapid tests met the WHO TPP acceptable sensitivity target of 80% relative to the AN PCR. Relative to the NP PCR, only the BinaxNOW™ met this clinical sensitivity threshold, with a sensitivity of 82%. All of the rapid tests exceeded the WHO TPP acceptable specificity target of 97%, relative to the NP PCR.

Table 1. Positive percent agreement and negative percent agreement values, including 95% confidence intervals, for all assays with either nasopharyngeal (NP) and anterior nares (AN) PCR as reference.

Assay NP Swab PCR AN Swab PCR
Sensitivity (95% CI) Specificity (95% CI) Sensitivity (95% CI) Specificity (95% CI)
OA-LFA 69% (60%-78%), 75/108 97% (88%-100%), 57/59 83% (74%-90%) 75/90 97% (91%-100%), 75/77
Sofia® 74% (64%-82%), 81/110 98% (91%-100%), 59/60 86% (77%-92%), 79/92 96% (89%-99%), 75/78
BinaxNOW™ 82% (73%-88%), 89/109 100% (94%-100%), 60/60 91% (84%-96%), 84/92 94% (85%-98%), 72/77
MSD Ag ELISA 79% (70%-86%), 87/110 98% (91%-100%), 59/60 91% (84%-96%), 84/92 95% (89%-99%), 74/78
AN Swab PCR 84% (75%-90%), 92/110 100% (94%-100%), 60/60 NA NA

We stratified the data based on a cutoff of 1000 copies/μL to understand the performance of each test at higher and lower viral loads. This cut-off corresponds to the acceptable analytical sensitivity outlined in the WHO TPP [8]. The results are shown in Table 2. In the lower viral load cases detected by NP PCR, all rapid tests demonstrated low sensitivity. Amongst the rapid tests, sensitivity among lower viral load cases was greatest for BinaxNOW™, but was still only 53%. In higher viral load cases, all rapid tests performed well, with sensitivity exceeding the 90% relative to both AN and NP reference PCR. In Fig 3, test results for PCR positive cases are shown in relation to AN and NP PCR viral loads. The WHO TPP acceptable analytical sensitivity target of 10^3 copies/μL is shown to highlight test performance above and below this threshold.

Table 2. Positive percent agreement, including 95% confidence intervals, for all assays with reference to both nasopharyngeal and anterior nares PCR.

Assay NP Swab PCR AN Swab PCR
Sensitivity < = 1000 copies/μL (95% CI) Sensitivity >1000 copies/μL (95% CI) Sensitivity < = 1000 copies/μL (95% CI) Sensitivity >1000 copies/μL (95% CI)
OA-LFA 25% (12%-42%), 9/36 92% (83%-97%), 66/72 55% (36%-73%), 17/31 98% (91%-100%), 58/59
Sofia® 38% (22%-55%), 14/37 92% (83%-97%), 67/73 64% (45%-80%), 21/33 98% (91%-100%), 58/59
BinaxNOW™ 53% (35%-70%), 19/36 96% (88%-99%), 70/73 79% (61%-91%), 26/33 98% (91%-100%), 58/59
MSD Ag ELISA 46% (29%-63%), 17/37 96% (88%-99%), 70/73 76% (58%-89%), 25/33 100% (94%-100%), 59/59
AN Swab PCR 57% (39%-73%), 21/37 97% (90%-100%), 71/73 NA NA

Results are binned into two groups based on if the reference method measured less than or equal or above 1000 copies per microliter.

Fig 3. Test results as a function of NP (top) and AN (bottom) PCR positive tests.

Fig 3

All tests perform well above the WHO TPP analytical sensitivity target of 103 copies/μL.

Using the McNemar mid-P value test for paired proportions, we compared the sensitivity and specificity between the three tests. Due to the small number of discordant results between the three tests it was determined that the mid-P variant of McNemar’s test was the most appropriate [16] and results are shown in S1 Table. Pairwise comparison demonstrated that BinaxNOW™ had a significantly higher sensitivity (p < .05) relative to the OA-LFA or the Sofia® assays, using either AN or NP PCR reference. Amongst patients with AN PCR viral loads greater than 1000 copies/μL, there were no significant difference in sensitivity between any LFA. Although the study was not powered to detect statistical significance between specificities, the specificities of all three tests were similar and high.

As seen in Fig 4A, the reader app performance could largely mirror visual interpretation of the GH Labs test, with a single discordant case that appeared to stem from an improper visual interpretation. When compared against PCR reference values (Fig 4B and 4C), the reader app ROC curves produced AUC values of 0.95 and 0.88 versus AN swab and NP swab results, respectively. Optimal sensitivity and specificity from these ROC curves were 86% and 99% for AN swabs and 72% and 97% for NP swabs. Compared with visual performance (Table 1), the reader app produced equivalent to slightly higher (up to 3%) concordance values with PCR.

Fig 4.

Fig 4

ROC curves for reader app performance versus (A) visual interpretation of GH labs assay, (B) AN swab PCR, and (C) NP swab PCR results. All represent use of the test line’s peak height above background metric extracted from the red channel of the high-resolution images acquired by the reader app. AUC values were 0.99, 0.95, and 0.88 for panels A, B, and C, respectively.

ROC curves were generated for the standardized anosmia assessment results versus PCR result (S3 Fig). The resulting area under the curves (AUC) are 0.69 relative to NP PCR and 0.63 relative to AN swab PCR. Classifying individuals as positive based on either a zero USMELLIT score or a positive LFA result increases the sensitivity with NP PCR, without affecting the specificity. Specifically, the sensitivity for the OA-LFA, Sofia® and BinaxNOW™ increases from 69% to 75%, 74% to 80% and 82% to 84%, respectively. Similar improvements were not seen relative to AN PCR. Sensitivity and specificity for each test in combination with USMELLIT is provided in S2 Table.

Discussion

COVID-19 has created an unprecedented demand for rapid diagnostic testing to reduce transmission. Existing commercial tests have addressed some of the demand, yet wider use of more frequent testing will require innovative approaches to expand supply and access to effective diagnostics. Here we report the clinical validation results of an open-access LFA design that meets the minimum WHO analytic sensitivity requirements and operational characteristics for a point of care test for suspected COVID-19 cases and their close contacts, that virtually any diagnostic manufacturer could produce.

Clinical validation in this population of symptomatic adults, with duration of symptoms <7 days, shows that the OA-LFA met the WHO TPP acceptable specificity relative to both AN and NP PCR, and the sensitivity target relative to AN PCR. However, the OA-LFA, along with the two commercially available tests, failed to detect the subset of patients with low viral loads, which represented less than one-fifth of patients with positive results in the population sampled. The reader app was able to reproduce visually read results by trained clinical research staff with near-perfect efficiency, and perhaps better in a subset of cases, which could be an important benefit for settings where users have less training.

This study elicits several important insights. One is the importance of the choice of reference standard for evaluating antigen rapid tests for SARS-CoV-2. In this study, NP PCR detected additional cases as compared to AN PCR a majority of those cases had very low viral loads. As described, the extraction methods differed slightly between NP and AN swabs. However, the relative performance is consistent with that reported elsewhere [17, 18]. The authors believe that AN PCR is a sufficient reference assay in the evaluation of rapid tests, where the intended use is to detect the majority of patients with early, acute SARS-CoV-2 infection, allowing for immediate implementation of isolation and other efforts to arrest transmission of the virus. A second item is that the sensitivity will be dependent on the population tested and the distribution of the viral burden. In a different use case, such as screening of asymptomatic patients where viral load may be shifted lower, then we would expect all the rapid tests, including the OA-LFA, to perform worse than reported in this study. Performance at higher viral loads is perhaps most relevant for the use case of detecting infectious individuals who are more likely to transmit the virus [19, 20].

In this study OA-LFA comparison to commercial assays showed similar performance, although pairwise comparison demonstrated the superiority of the BinaxNOW™ format. We speculate this may be related to the innovative form factor of the BinaxNOW™, with minimal sample dilution and high transfer efficiency. Lastly, the reference assay on MSD did not perform any better than rapid antigen tests. We speculate that this may be due to specimen degradation and loss or alteration of target antigen due to freeze-thaw. These hypotheses require further exploration.

A limitation to this study is that participants self-identified with having symptoms within 7 days of self-reported onset. There remains a strong need for serial testing cohort studies among exposed individuals to characterize how rapid test performs over the time patients are infectious. This study also has limited generalizability, because it was conducted in a single population, by trained health care personnel, at a peak epidemic time. A final limitation of this study is the lack of a viral culture reference; this reference would allow us to the opportunity to more directly assess each RDT’s ability to detect individuals that pose a significant transmission risk.

The OA-LFA information assay architecture is publicly disseminated as open access. Researchers can replicate it without proprietary research and development on their own. If selected antibodies become unavailable by supply or cost, we have reported alternatives, and perhaps better performers, which can be substituted in the development process [21]. The design can also be submitted to WHO’s Covid Technology Access Pool (C-TAP) and shared with manufacturers in Low- and Middle- Income Countries (LMICs) for their consideration. Additional collaboration and technical assistant may be warranted to assist global LFA manufacturing capacity to meet the demand of the pandemic.

Supporting information

S1 Fig

(A) Reference image of cassette used in this study, including the added sticker of ArUco codes. (B) Blue circles indicate the keypoints used in the SIFT algorithm for image recognition and transformation. Keypoints were filtered to ensure that pixels in the read window or sample pad were excluded from the matching process.

(TIF)

S2 Fig. Selected screenshots from the various stages of the user interacting with the reader app.

(A) Attempting to capture an image of the cassette—note the instructions to the user on top to move the phone and then keep it steady, as well as some but not all quality checks being passed. (B) Representative view of the cassette passing all quality checks, but just out of flatness range to facilitate screen grab. (C) Result window showing the full captured image, the extracted read window, and locations and intensities (peak height of red channel) of control and test lines, if found.

(TIF)

S3 Fig

ROC curves for USmellIt relative to NP (A) and AN (B) PCR. A cut-off value of a USMELLIT score of 0 retains 100% specificity while providing 22% sensitivity for NP PCR while only 92% specificity is seen relative to AN PCR. The combination of an antigen-based test with a score of 0 offers potential improvement over an antigen test alone, as discussed in the main text.

(TIF)

S1 Table. Results from McNemar’s mid-P value test comparing the sensitivity and specificity for all three tests against AN and NP swab PCR.

Additionally, results from the McNemar mid-P test are shown for the sensitivity for cases with viral loads greater than and less than 1000 copies/μL.

(DOCX)

S2 Table. Sensitivity and specificity of each test relative to AN and NP PCR with or without the inclusion of the zero USMELLIT criterion.

(DOCX)

S1 Data

(CSV)

S1 File

(DOCX)

Acknowledgments

The authors would like to acknowledge CureBase and the CovidClinic for their support in setting up and implementing this study. A special thanks to the creators of USMELLIT for their involvement in the study.

Abbreviations

AN

Anterior Nares

AUC

Area under the Curve

C-PAP

Covid Technology Access Pool

FDA

Food and Drug Administration

LMIC

Low- and Middle-Income Countries

LFA

Lateral Flow Assay

MSD

Meso Scale Discovery

NP

Nasopharyngeal

OA-LFA

Open Access Lateral Flow Assay

PCR

Polymerase Chain Reaction

TPP

Target Product Profile

VTM

Viral Transport Medium

WHO

World Health Organization

Data Availability

All relevant data are within the paper and its Supporting Information files.

Funding Statement

All Authors of the manuscript work for the funding organization - Global Health Labs. https://www.ghlabs.org/.

References

  • 1.Tracking—Johns Hopkins Coronavirus Resource Center. [cited 6 May 2021]. Available: https://coronavirus.jhu.edu/data
  • 2.Ricks S, Kendall EA, Dowdy DW, Sacks JA, Schumacher SG, Arinaminpathy N. Quantifying the potential value of antigen-detection rapid diagnostic tests for COVID-19: a modelling analysis. BMC Medicine. 2021;19: 75. doi: 10.1186/s12916-021-01948-z [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Du Z, Pandey A, Bai Y, Fitzpatrick MC, Chinazzi M, Pastore y Piontti A, et al. Comparative cost-effectiveness of SARS-CoV-2 testing strategies in the USA: a modelling study. The Lancet Public Health. 2021;6: e184–e191. doi: 10.1016/S2468-2667(21)00002-5 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Afzal A. Molecular diagnostic technologies for COVID-19: Limitations and challenges. Journal of Advanced Research. 2020;26: 149–159. doi: 10.1016/j.jare.2020.08.002 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Larremore DB, Wilder B, Lester E, Shehata S, Burke JM, Hay JA, et al. Test sensitivity is secondary to frequency and turnaround time for COVID-19 screening. Science Advances. 2021;7. doi: 10.1126/sciadv.abd5393 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Coronavirus Disease 2019 (COVID-19) Emergency Use Authorizations for Medical Devices | FDA. [cited 5 Apr 2021]. Available: https://www.fda.gov/medical-devices/emergency-use-authorizations-medical-devices/coronavirus-disease-2019-covid-19-emergency-use-authorizations-medical-devices
  • 7.Grant BD, Anderson CE, Garing SH, Alonzo LF, Williford JR, Baughman TA, et al. A SARS-CoV-2 Coronavirus Nucleocapsid Protein Antigen-Detecting Lateral Flow Assay Licence: CC BY 4.0 Citation information: Grant. [cited 20 May 2021]. doi: 10.26434/chemrxiv.12794825.v1 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.COVID-19 Target product profiles for priority diagnostics to support response to the COVID-19 pandemic v.1.0. [cited 29 Apr 2021]. Available: https://www.who.int/publications/m/item/covid-19-target-product-profiles-for-priority-diagnostics-to-support-response-to-the-covid-19-pandemic-v.0.1
  • 9.Byrnes SA, Gallagher R, Steadman A, Bennett C, Rivera R, Ortega C, et al. Multiplexed and Extraction-Free Amplification for Simplified SARS-CoV-2 RT-PCR Tests. Cite This: Anal Chem. 2021;93: 4160–4165. doi: 10.1021/acs.analchem.0c03918 [DOI] [PubMed] [Google Scholar]
  • 10.Bustin SA, Benes V, Garson JA, Hellemans J, Huggett J, Kubista M, et al. The MIQE guidelines: Minimum information for publication of quantitative real-time PCR experiments. Clinical Chemistry. 2009;55: 611–622. doi: 10.1373/clinchem.2008.112797 [DOI] [PubMed] [Google Scholar]
  • 11.Anti-Coronavirus Nucleocapsid Antibody, 40143-MM05 | Sino Biological. [cited 6 Apr 2021]. Available: https://www.sinobiological.com/antibodies/cov-nucleocapsid-40143-mm05
  • 12.Park C, Mariakakis A, Yang J, Lassala D, Djiguiba Y, Keita Y, et al. Supporting smartphone-based image capture of rapid diagnostic tests in low-resource settings. ACM International Conference Proceeding Series. New York, NY, USA: Association for Computing Machinery; 2020. pp. 1–11. doi: 10.1145/3392561.3394630 [DOI]
  • 13.u-Smell-it. [cited 3 May 2021]. Available: https://www.usmellit.com/
  • 14.CRAN—Package epiR. [cited 3 May 2021]. Available: https://cran.r-project.org/web/packages/epiR/index.html
  • 15.Fagerland MW, Lydersen S, Laake P. The McNemar test for binary matched-pairs data: Mid-p and asymptotic are better than exact conditional. BMC Medical Research Methodology. 2013;13: 91. doi: 10.1186/1471-2288-13-91 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Pembury Smith MQR, Ruxton GD. Effective use of the McNemar test. Behavioral Ecology and Sociobiology. 2020;74. doi: 10.1007/s00265-020-02916-y [DOI] [Google Scholar]
  • 17.Marx GE, Smith-Jeffcoat SE, Biggerstaff BJ, Koh M, Nawrocki CC, Travanty EA, et al. SARS-CoV-2 detection by rRT-PCR on self-collected anterior nares swabs or saliva compared with clinician-collected nasopharyngeal swabs—Denver and Atlanta, August–November, 2020. medRxiv. 2021; 2021.02.16.21251521. doi: 10.1101/2021.02.16.21251521 [DOI] [Google Scholar]
  • 18.Berenger BM, Fonseca K, Schneider AR, Hu J, Zelyas N. Sensitivity of nasopharyngeal, nasal and throat swab for the detection of SARS-CoV-2. medRxiv. medRxiv; 2020. p. 2020.05.05.20084889..doi: 10.1101/2020.05.05.20084889 [DOI] [Google Scholar]
  • 19.Marks M, Millat-Martinez P, Ouchi D, Roberts C h., Alemany A, Corbacho-Monné M, et al. Transmission of COVID-19 in 282 clusters in Catalonia, Spain: a cohort study. The Lancet Infectious Diseases. 2021;0. doi: 10.1016/S1473-3099(20)30985-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Kawasuji H, Takegoshi Y, Kaneda M, Ueno A, Miyajima Y, Kawago K, et al. Transmissibility of COVID-19 depends on the viral load around onset in adult and symptomatic patients. Pan X, editor. PLOS ONE. 2020;15: e0243597. doi: 10.1371/journal.pone.0243597 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Cate D, Hsieh H, Glukhova V, Bishop JD, Hermansky HG, Barrios-Lopez B, et al. Antibody Screening Results for Anti-Nucleocapsid Antibodies Towards the Development of a SARS-CoV-2 Nucleocapsid Protein Antigen Detecting Lateral Flow Assay. 2021. [cited 2 Aug 2021]. doi: 10.26434/CHEMRXIV.12709538.V2 [DOI] [PMC free article] [PubMed] [Google Scholar]

Decision Letter 0

Etsuro Ito

21 Jul 2021

PONE-D-21-19695

Clinical validation of an open-access SARS-COV-2 antigen detection lateral flow assay, compared to commercially available assays.

PLOS ONE

Dear Dr. Bachman,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please pay special attention to the methods that the reviewers suggested.

Please submit your revised manuscript by Sep 04 2021 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Etsuro Ito

Academic Editor

PLOS ONE

Journal requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at 

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and 

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. Please review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice.

3. You indicated that you had ethical approval for your study. In your Methods section, please ensure you have also stated whether the IRB committee approved the format of written consent via a phone app.

4. Your ethics statement should only appear in the Methods section of your manuscript. If your ethics statement is written in any section besides the Methods, please move it to the Methods section and delete it from any other section. Please ensure that your ethics statement is included in your manuscript, as the ethics statement entered into the online submission form will not be published alongside your manuscript. 

5. Please upload a copy of Supporting Information Figure 3 which you refer to in your text on page 13.

6. Please include captions for your Supporting Information files at the end of your manuscript, and update any in-text citations to match accordingly. Please see our Supporting Information guidelines for more information: http://journals.plos.org/plosone/s/supporting-information.

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: I Don't Know

Reviewer #3: Yes

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: The study was designed scientifically and the manuscript was prepared following a standard format. The results and conclusions are valuable for the researchers in the field of lateral flow assay and medical technologists as potential end users. I do feel that the data capturing and analyses through ArUco codes are quite new to the researchers in the field and would like to expect the authors to introduce more details separately.

Reviewer #2: This paper reports a prospective diagnostic accuracy study comparing an open access lateral flow antigen detection assay to two different commercially available assays. An evaluation of a mobile phone reader and the addition of a system to evaluate sense of smell. The study is generally well conducted and reported. The sample size is small, particular for precise estimation of specificity, however the results for the open access assay are promising, particularly for manufacturers in low and middle income countries and when the assay is used in combination with the ‘USMELLIT’ device. Much of the supplementary material seems to be missing, e.g. Figure S2 and all of the Supplementary tables, which is unfortunate. I have a few comments that the authors may wish to consider.

Pg 4. Methods, Participants – please provide a definition of ‘symptomatic’, i.e. was one of the listed symptoms sufficient or was more than one required

Pg 4. Methods, Point-of-care test methods – Please state whether the Sofia and BinaxNOW assays are validated for use with AN swabs. Please also describe who collected the swab and carried out the test (self-collected, health care professional etc), whether the NP swab was obtained first or after the 4 AN swabs, and importantly whether the antigen tests were interpreted blinded to the result of the others.

Pg 7. Methods, Reference Method testing – Were those carrying out the PCR testing blinded to the results of the antigen detection tests?

Pg 8, Methods, Mobile reader application – Line 193-196 are difficult to follow for the general reader, especially as the Figure is not accessible. Please consider explaining in less technical language. Also, although the study does not develop the detection algorithm used in the app, modifications to it are made for the study purposes. ‘Pre-trial testing’ is mentioned – this appears to have been used to set the threshold at which the app considers a test line to be present. The supplementary materials should provide further details about this as one would usually expect to see an algorithm (or modification of an existing algorithm) to be trained on one set of images and tested on another independent set. Perhaps the authors could comment on whether this was done and whether one can expect the result observed here to be repeated using different sets of images.

Pg 9, Methods, Analysis – Please use the terms sensitivity and specificity in preference to PPA and NPA as the latter can easily be confused with positive and negative predictive values.

Pg 10 Results, line 241-242 – The threshold chosen for comparing results in higher/lower viral load is quite low – most studies use a threshold of 10^5 or 10^6 copies/ml. The WHO acceptable sensitivity of 80% refers to results in all samples and not restricted to higher viral load

Pg 12 line 262 – It is currently recommended not to rely on the ‘statistical significance’ of comparisons between tests but to point to differences that might be within or beyond that expected by chance. I suspect the study was not powered to detect differences in specificity therefore I would remove the statement about significance (line 264-265)

Pg 13 line 287-288. The empirical data underlying the sensitivity estimates for the antigen tests combined with the USMELLIT should be reported.

Discussion - The authors do not clearly make the case for conducting AN PCR as well as NP PCR. To my knowledge the WHO do not specify a sample type for the reference method against which antigen tests should be compared for establishing performance characteristics however there is no evidence that the cases detected by AN PCR are the more clinically relevant (infectious cases), therefore care should be taken when interpreting these results otherwise it does look a bit like cherry picking the results to make the tests look better.

Reviewer #3: Bacham and colleagues have submitted an article presenting a clinical validation of an open-access lateral flow assay (OA-LFA) design using commercially available materials and reagents, along with RT-qPCR, BinaxNOW® and Sofia® rapid diagnostic tests. They found that the open-access LFA meets the minimum WHO target product profile for a rapid test with a positive predictive agreement with NP sampling of 69% (60% -78%) for OA-LFA as compared to 74% (64% - 82%) for Sofia®, and 82% (73% - 88%) BinaxNOW™ in adults patients with COVID-19 symptoms less than 7 days. Bacham et al include the evaluation of a reader app that could largely mirror visual interpretation of Labs

Test.

There are however few some points to address before being considered for publication in PlosOne

1. The authors have rightly chosen to calculate the positive and negative predictive agreement to assess the performance of the OA-LFA. However, most of the evaluation of diagnostic tests use Sensitivity, specificity, PPV and NPV. We recommend the authors to include those parameters using the RT-qPCR as the reference as done in most rapid diagnostic tests evaluation. This will make easier the comparison with other evaluations. Moreover in line 233-235 the authors refer to the WHO TPP requirement of 80% sensitivity and 97 specificity.

2. In table 1 the authors provide important data to appreciate the performance of each test. It would be better to also provide the 2x2 table for each vs RT-qPCR results. If this is not possible in the main text they could be added in the supplementary

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: Yes: Jianfu Jeffrey Wang

Reviewer #2: Yes: Jacqueline Dinnes

Reviewer #3: Yes: Yap Boum

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

Attachment

Submitted filename: Reviewer PlosOne PONE-D-21-19695 yb.docx

PLoS One. 2021 Aug 17;16(8):e0256352. doi: 10.1371/journal.pone.0256352.r002

Author response to Decision Letter 0


3 Aug 2021

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

COMPLETE

2. Please review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice. COMPLETE

3. You indicated that you had ethical approval for your study. In your Methods section, please ensure you have also stated whether the IRB committee approved the format of written consent via a phone app. COMPLETE

4. Your ethics statement should only appear in the Methods section of your manuscript. If your ethics statement is written in any section besides the Methods, please move it to the Methods section and delete it from any other section. Please ensure that your ethics statement is included in your manuscript, as the ethics statement entered into the online submission form will not be published alongside your manuscript. COMPLETE

5. Please upload a copy of Supporting Information Figure 3 which you refer to in your text on page 13. COMPLETE

6. Please include captions for your Supporting Information files at the end of your manuscript, and update any in-text citations to match accordingly. Please see our Supporting Information guidelines for more I think we are good on this – will review at the end. information: http://journals.plos.org/plosone/s/supporting-information.

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

________________________________________

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: I Don't Know

Reviewer #3: Yes

________________________________________

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

________________________________________

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

________________________________________

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: The study was designed scientifically and the manuscript was prepared following a standard format. The results and conclusions are valuable for the researchers in the field of lateral flow assay and medical technologists as potential end users. I do feel that the data capturing and analyses through ArUco codes are quite new to the researchers in the field and would like to expect the authors to introduce more details separately. The details are included in Supplemental Methods and is now more clearly introduced on Page 7.

Reviewer #2: This paper reports a prospective diagnostic accuracy study comparing an open access lateral flow antigen detection assay to two different commercially available assays. An evaluation of a mobile phone reader and the addition of a system to evaluate sense of smell. The study is generally well conducted and reported. The sample size is small, particular for precise estimation of specificity, however the results for the open access assay are promising, particularly for manufacturers in low and middle income countries and when the assay is used in combination with the ‘USMELLIT’ device. Much of the supplementary material seems to be missing, e.g. Figure S2 and all of the Supplementary tables, which is unfortunate. I have a few comments that the authors may wish to consider.

Pg 4. Methods, Participants – please provide a definition of ‘symptomatic’, i.e. was one of the listed symptoms sufficient or was more than one required COMPLETE

Pg 4. Methods, Point-of-care test methods – Please state whether the Sofia and BinaxNOW assays are validated for use with AN swabs. INDICATED on PAGE 5, per following kit instruction. Please also describe who collected the swab and carried out the test (self-collected, health care professional etc), COMPLETE whether the NP swab was obtained first or after the 4 AN swabs, and importantly whether the antigen tests were interpreted blinded to the result of the others. INCLUDED

Pg 7. Methods, Reference Method testing – Were those carrying out the PCR testing blinded to the results of the antigen detection tests? YES, ADDED

Pg 8, Methods, Mobile reader application – Line 193-196 are difficult to follow for the general reader, especially as the Figure is not accessible. Please consider explaining in less technical language. Also, although the study does not develop the detection algorithm used in the app, modifications to it are made for the study purposes. ‘Pre-trial testing’ is mentioned– this appears to have been used to set the threshold at which the app considers a test line to be present. The supplementary materials should provide further details about this as one would usually expect to see an algorithm (or modification of an existing algorithm) to be trained on one set of images and tested on another independent set. Perhaps the authors could comment on whether this was done and whether one can expect the result observed here to be repeated using different sets of images. Further details on the detection of the test line are provided in Supplemental Methods section 1.4, although this portion may have accidentally been excluded by the submission process initially. The portion of the manuscript called out by the reviewer now explicitly states the presence of these additional details and was further modified slightly for clarity. In general, the peak detection algorithm is deterministic so does not need to be “trained” like a machine learning algorithm – rather, it just needs an appropriate peak height cutoff value.

Pg 9, Methods, Analysis – Please use the terms sensitivity and specificity in preference to PPA and NPA as the latter can easily be confused with positive and negative predictive values.

We have switched to sensitivity and specificity as recommended.

Pg 10 Results, line 241-242 – The threshold chosen for comparing results in higher/lower viral load is quite low – most studies use a threshold of 10^5 or 10^6 copies/ml. The WHO acceptable sensitivity of 80% refers to results in all samples and not restricted to higher viral load

We have adjusted our cut-off for higher/lower viral load to be 10^6 copies/mL in order to align with the WHO TPP acceptable analytical sensitivity cutoff. Additionally, we agree the acceptable clinical sensitivity refers to all samples. We have edited the manuscript to report on the clinical sensitivity relative to the TPP based on both AN and NP reference PCR swabs, but not based on viral load stratification.

Pg 12 line 262 – It is currently recommended not to rely on the ‘statistical significance’ of comparisons between tests but to point to differences that might be within or beyond that expected by chance. I suspect the study was not powered to detect differences in specificity therefore I would remove the statement about significance (line 264-265) New language added to this section.

Pg 13 line 287-288. The empirical data underlying the sensitivity estimates for the antigen tests combined with the USMELLIT should be reported.

The data has been added to supplemental table 2.

Discussion - The authors do not clearly make the case for conducting AN PCR as well as NP PCR. To my knowledge the WHO do not specify a sample type for the reference method against which antigen tests should be compared for establishing performance characteristics however there is no evidence that the cases detected by AN PCR are the more clinically relevant (infectious cases), therefore care should be taken when interpreting these results otherwise it does look a bit like cherry picking the results to make the tests look better.

The decision to include analysis both against AN and NP PCR is to reflect the lack of consensus in the appropriate reference standard. As correctly pointed out by the reviewer, the WHO does not specify a reference test. For their FDA EUA submission, Abbott BinaxNOW utilized an AN swab for reference PCR testing. Quidel Sofia used AN swabs for some subjects and NP swabs for others. We reported performance against both reference swabs in this manuscript to highlight the impact of the reference method choice. Ideally, viral culture would be used as a metric to evaluate sensitivity in patients most likely to transmit the disease.

Reviewer #3: Bacham and colleagues have submitted an article presenting a clinical validation of an open-access lateral flow assay (OA-LFA) design using commercially available materials and reagents, along with RT-qPCR, BinaxNOW® and Sofia® rapid diagnostic tests. They found that the open-access LFA meets the minimum WHO target product profile for a rapid test with a positive predictive agreement with NP sampling of 69% (60% -78%) for OA-LFA as compared to 74% (64% - 82%) for Sofia®, and 82% (73% - 88%) BinaxNOW™ in adults patients with COVID-19 symptoms less than 7 days. Bacham et al include the evaluation of a reader app that could largely mirror visual interpretation of Labs

Test.

There are however few some points to address before being considered for publication in PlosOne

1. The authors have rightly chosen to calculate the positive and negative predictive agreement to assess the performance of the OA-LFA. However, most of the evaluation of diagnostic tests use Sensitivity, specificity, PPV and NPV. We recommend the authors to include those parameters using the RT-qPCR as the reference as done in most rapid diagnostic tests evaluation. This will make easier the comparison with other evaluations. Moreover in line 233-235 the authors refer to the WHO TPP requirement of 80% sensitivity and 97 specificity.

We have replaced PPA and NPA with sensitivity and specificity, based on the suggestion of reviewer 2 and 3, and to be consistent with the WHO TPP.

2. In table 1 the authors provide important data to appreciate the performance of each test. It would be better to also provide the 2x2 table for each vs RT-qPCR results. If this is not possible in the main text they could be added in the supplementary Thoughts, Because we have 3 tables on the results, we found this sufficient. We can supply a 2X2 table upon request.

________________________________________

Attachment

Submitted filename: reviewer comments and responses Aug 3 2021.docx

Decision Letter 1

Etsuro Ito

5 Aug 2021

Clinical validation of an open-access SARS-COV-2 antigen detection lateral flow assay, compared to commercially available assays.

PONE-D-21-19695R1

Dear Dr. Bachman,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Etsuro Ito

Academic Editor

PLOS ONE

Additional Editor Comments:

Thank you for your appropriate revision.

Acceptance letter

Etsuro Ito

9 Aug 2021

PONE-D-21-19695R1

Clinical validation of an open-access SARS-COV-2 antigen detection lateral flow assay, compared to commercially available assays.

Dear Dr. Bachman:

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department.

If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org.

If we can help with anything else, please email us at plosone@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Prof. Etsuro Ito

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 Fig

    (A) Reference image of cassette used in this study, including the added sticker of ArUco codes. (B) Blue circles indicate the keypoints used in the SIFT algorithm for image recognition and transformation. Keypoints were filtered to ensure that pixels in the read window or sample pad were excluded from the matching process.

    (TIF)

    S2 Fig. Selected screenshots from the various stages of the user interacting with the reader app.

    (A) Attempting to capture an image of the cassette—note the instructions to the user on top to move the phone and then keep it steady, as well as some but not all quality checks being passed. (B) Representative view of the cassette passing all quality checks, but just out of flatness range to facilitate screen grab. (C) Result window showing the full captured image, the extracted read window, and locations and intensities (peak height of red channel) of control and test lines, if found.

    (TIF)

    S3 Fig

    ROC curves for USmellIt relative to NP (A) and AN (B) PCR. A cut-off value of a USMELLIT score of 0 retains 100% specificity while providing 22% sensitivity for NP PCR while only 92% specificity is seen relative to AN PCR. The combination of an antigen-based test with a score of 0 offers potential improvement over an antigen test alone, as discussed in the main text.

    (TIF)

    S1 Table. Results from McNemar’s mid-P value test comparing the sensitivity and specificity for all three tests against AN and NP swab PCR.

    Additionally, results from the McNemar mid-P test are shown for the sensitivity for cases with viral loads greater than and less than 1000 copies/μL.

    (DOCX)

    S2 Table. Sensitivity and specificity of each test relative to AN and NP PCR with or without the inclusion of the zero USMELLIT criterion.

    (DOCX)

    S1 Data

    (CSV)

    S1 File

    (DOCX)

    Attachment

    Submitted filename: Reviewer PlosOne PONE-D-21-19695 yb.docx

    Attachment

    Submitted filename: reviewer comments and responses Aug 3 2021.docx

    Data Availability Statement

    All relevant data are within the paper and its Supporting Information files.


    Articles from PLoS ONE are provided here courtesy of PLOS

    RESOURCES