Abstract
Introduction
Scale‐up of HIV self‐testing (HIVST) will play a key role in meeting the United Nation's 90‐90‐90 targets. Delayed re‐reading of used HIVST devices has been used by early implementation studies to validate the performance of self‐test kits and to estimate HIV positivity among self‐testers. We investigated the stability of results on used devices under controlled conditions to assess its potential as a quality assurance approach for HIVST scale‐up.
Methods
444 OraQuick® HIV‐1/2 rapid antibody tests were conducted using commercial plasma from two HIV‐positive donors and HIV‐negative plasma (high‐reactive n = 148, weak‐reactive n = 148 and non‐reactive n = 148) and incubated them for six months under four conditions (combinations of high and low temperatures and humidity). Devices were re‐read daily for one week, weekly for one subsequent month and then once a month by independent readers unaware of the previous results. We used multistage transition models to investigate rates of change in device results, and between storage conditions.
Results and discussion
There was a high incidence of device instability. Forty‐three (29%) of 148 initially non‐reactive results became false weak‐reactive results. These changes were observed across all incubation conditions, the earliest on Day 4 (n = 9 kits). No initially HIV‐reactive results changed to a non‐reactive result. There were no significant associations between storage conditions and hazard of results transition. We observed substantial statistical agreement between independent re‐readers over time (agreement range: 0.74 to 0.96).
Conclusions
Delayed re‐reading of used OraQuick® HIV‐1/2 rapid antibody tests is not currently a valid methodological approach to quality assurance and monitoring as we observed a high incidence (29%) of true non‐reactive tests changing to false weak‐reactive and therefore its use may overestimate true HIV positivity.
Keywords: HIV self‐testing, Quality assurance, Delayed re‐reading, Visual stability, False reactive, Misdiagnosis, HIV testing
1. Introduction
HIV self‐testing (HIVST) is being scaled‐up using a variety of distribution models throughout Africa, the Americas, Asia and Europe 1, 2, 3, 4. No clear monitoring and evaluation or external quality assurance (EQA) systems exist for HIVST devices and this raises concern for national reference laboratories, regulators and policymakers 5, 6, 7.
While previous studies report acceptable sensitivity and specificity when HIVST is conducted by intended users 8, 9, it is unclear whether this will be maintained once HIVST programmes are implemented at scale. Observation and in‐depth interviews reveal that without a demonstration, operator errors are common in both conducting and interpreting self‐tests 10, 11. Scale‐up will have to be accompanied by a robust quality assurance system.
A reactive HIVST indicates that HIV antibodies are present in the oral or fingerstick/blood sample of the user. Further testing to confirm a positive diagnosis following linkage to care acts as an active system for detecting false‐reactive results and ensures individuals are not incorrectly started on antiretroviral therapy (ART). In most contexts, however, the prevalence of false‐reactive results prior to ART clinic enrolment (whether or not individuals came from HIVST) is not tracked and rates of linkage remain highly variable and can be very low without active support 12, 13, 14. Self‐testers with non‐reactive results, unless linking to voluntary male medical circumcision or pre‐exposure prophylaxis services, would not typically seek or receive further testing and confirmation, meaning a false non‐reactive result would not be detected.
Methods that detect incorrect results and misinterpretation are required for individual care as well as for quality assurance. One approach, which has been utilized in early HIVST implementation studies, is for self‐testers to return used devices for delayed re‐reading by trained staff in parallel with self‐reported interpretation of results 15, 16. However, delays between device use and re‐reading, and environmental storage conditions during this period could impair the validity of this method. We therefore set out to investigate the stability of OraQuick® HIV‐1/2 rapid antibody test (OraQuick HIV) results with delayed re‐reading stored under controlled incubation conditions for prolonged periods. We selected the OraQuick® HIV‐1/2 rapid antibody test kit, which is the same product (in different packaging) as the OraQuick® HIV Self‐Test which is prequalified by the World Health Organization (WHO) 17.
2. Methods
2.1. Materials and equipment
Two different batches (HIVCO‐4308 and HIVCO‐4309) of OraQuick® HIV‐1/2 rapid antibody test kits (assembled in Thailand for OraSure Technologies, Inc. Bethlehem, PA, USA) were obtained from the manufacturer. Human HIV seroconversion panel plasma samples from two donors (Donor No. 73695 panel number 12007‐08 and 09, 18 and 75018 panel number 9077‐24 and 25 19) were purchased from ZeptoMetrix Corporation (Buffalo, NY, USA). Human plasma negative for HIV, hepatitis B, C, E and syphilis was purchased from the National Blood Service (Liverpool, UK).
2.2. Sample preparation
The OraQuick® HIV‐1/2 Rapid Antibody Test is WHO prequalified for use with oral fluid, whole blood, serum or plasma. The matrix of the sample (i.e. plasma rather than an oral crevicular fluid sample) was not a crucial factor in this investigation, as we were not investigating specificity or sensitivity. What was important was the basis of the immuno‐chromatographic stability of the test. The use of HIV antibody‐positive and ‐negative plasma allowed us to investigate this.
Four panel samples from two donors were combined to produce an HIV‐reactive “mini‐pool” of stock serum. This was checked to ensure the correct result and intensity of test line on the OraQuick HIV device. From this stock, an HIV‐reactive sample was prepared with the addition of HIV‐negative plasma (1:8 dilution factor). An HIV weak‐reactive sample was prepared with a 1:16 dilution factor.
2.3. Sample size calculation
To estimate sample size, we assumed that 0.2% of all used tests would change over six months. To estimate accuracy of rate of change within ± 1% with 95% confidence, 77 tests were required to be read for each condition, with a total of four conditions, giving a minimum sample of 308 kits. With available resources, were able to include more samples (444 total).
2.4. Conducting the tests
The study was conducted in the laboratory under controlled conditions rather than using actual patient‐used HIVST devices. This eliminated the risk that the test had not been performed correctly which could have influenced the study results.
A total of 444 OraQuick HIV tests were conducted in the laboratory following the manufacturer's instructions for use (IFU). Five microlitres of the prepared samples (HIV reactive n = 148, HIV weak‐reactive n = 148 or HIV non‐reactive n = 148) was delivered into the developer solution before mixing gently. The test device was labelled with an identification number on the back and inserted “pad end” into the developer solution. Devices were read within the 20‐ to 40‐minute reading window (measured using a digital timer) by three different laboratorians, trained in the reading of the devices and blinded to each other's interpretation.
2.5. Read definitions/interpretation
On the test device there is a window next to which there is a letter “T” for test line and a “C” for control line. As per the IFU, a non‐reactive result was recorded when only a single quality control line was visible adjacent to the letter “C” on the test device. A weak‐reactive was recorded when there were two visible lines on the test device, the first adjacent to the letter “C” (control) and the second adjacent to the letter “T” (test) but the test line was not as intense as the control line. A reactive test was recorded when both “C” and “T” lines were visible and the “T” line was at least as intense as the “C” line. An invalid result was defined as no line present adjacent to the letter “C.”
2.6. Incubation conditions
Following initial reads, devices were placed in one of four laboratory benchtop incubators (Benchmark Scientific), each set to a different incubation condition: (a) control temperature (30°C) with high humidity (70%); (b) control temperature (30°C) with low humidity (20%); (c) high temperature (40°C) with high humidity (70%); (d) or high temperature (40°C) with low humidity (20%). Each condition had 37 HIV non‐reactive, 37 HIV weak‐reactive and 37 HIV reactive devices allocated to it (Figure 1).
2.7. Re‐reading intervals
Devices were re‐read by either two or three blinded and independent readers daily for one week, weekly for one subsequent month and then once a month for the following five months, giving a total of 13 reads over the 6‐month study period (November 2016 through to April 2017). Each of the laboratorians interpreted the test face up, recorded the test result (check box non‐reactive, weak‐reactive or reactive) on the data log sheet and then turned the test over to record the test identification number along with any additional comments. Data were input onto a blinded (of previous re‐read result) electronic log. Data were unblinded and analysed after 6 months.
2.8. Data analysis
Two laboratorians had to be in agreement for a “final” test result interpretation. We compared agreement between readers at each time point using the kappa statistic with bootstrapped 95% confidence intervals for three readers and Scott's pi for two readers. To estimate the hazard of transition between device states (non‐reactive, weak‐reactive, and reactive) over time, and the effects of incubation storage conditions, we fitted a multistage transition model using a hidden Markov process. Model fit was evaluated by visually comparing the fitted hazard function within each condition over time with observed transition events. In the final model, terms for piecewise intensities were fitted at Day 1 to 2, Day 2 to 3; Day 3 to 4; Day 8 to 15; and Day 15 to 181 to account for the high intensity of transition. Analysis was done using R version 3.3.2 (R Foundation for Statistical Computing, Vienna).
3. Results and discussion
Devices were first read following the manufacturer IFU, after 20 minutes and within 40 minutes of conducting the test (Day 0) for control purposes. On Day 0, all reactive devices gave the expected dilution results (reactive or weak‐reactive) and a following masked re‐read showed all three independent readers in agreement (100%). Statistical agreement between independent readers over the six‐month period ranged from 0.70 (95% confidence interval (CI): 0.66 to 0.74) to 0.96 (95% CI: 0.94 to 0.98) 20.
There was a high incidence of OraQuick HIV result transition between states over time (Figure 2). A total of 43 of the 148 true non‐reactive devices (29%) changed to a false weak‐reactive result with the earliest change observed on Day 4 (n = 9 kits) incubated at high temperature and low humidity (Figure 1). Transition between states over time was also observed, with tests changing from true non‐reactive to false weak‐reactive and then back to true non‐reactive (77 instances out of a total of 1776) and weak‐reactive results changing to strong reactive and then back to weak‐reactive (135 instances). The majority of these true reactive transitions occurred early (from Day 1) with the greatest intensity of transition occurring up to Day 15. Transitions continued to occur throughout the six‐month follow‐up period. No devices with an initial reactive result changed to a weak‐reactive or non‐reactive result over the six‐month period. The test control line showed 100% stability throughout the study.
Changes occurred across all controlled incubation conditions with the earliest transition from a true non‐reactive to a false weak‐reactive occurring under high temperature and low humidity conditions on Day 4. However, in our final model, there was no significant association between the incubation condition under which devices were stored and the hazard of transition between stages (Table 1).
Table 1.
Incubation condition and transition stage (From → To) | Hazard ratio for transition intensity (vs. cool/dry incubation condition) | 95% confidence interval |
---|---|---|
Cool/humid | ||
HIV non‐reactive → HIV weak‐reactive | 0.88 | 0.47 to 1.64 |
HIV weak‐reactive → HIV non‐reactive | 0.95 | 0.33 to 2.71 |
HIV weak‐reactive → HIV reactive | 1.27 | 0.80 to 2.03 |
HIV reactive → HIV weak‐reactive | 1.41 | 0.87 to 2.28 |
Warm/humid | ||
HIV non‐reactive → HIV weak‐reactive | 0.81 | 0.43 to 1.56 |
HIV weak‐reactive → HIV non‐reactive | 1.27 | 0.47 to 3.42 |
HIV weak‐reactive → HIV reactive | 0.87 | 0.53 to 1.44 |
HIV reactive → HIV weak‐reactive | 0.94 | 0.56 to 1.59 |
Warm/Dry | ||
HIV non‐reactive → HIV weak‐reactive | 1.06 | 0.57 to 1.95 |
HIV weak‐reactive → HIV non‐reactive | 1.24 | 0.46 to 3.35 |
HIV weak‐reactive → HIV reactive | 0.94 | 0.58 to 1.52 |
HIV reactive → HIV weak‐reactive | 1.15 | 0.69 to 1.89 |
Estimated by fitting multistage transition model for each test read condition with hidden Markov process, and with terms for incubation condition and piecewise transition intensities between Day 1 to 2, Day 2 to 3, Day 3 to 4, Day 8 to 15 and Day 15 to 181.
Our key finding shows the OraQuick HIV device can have a result change from a true non‐reactive to a false weak‐reactive result when reading is extended beyond the manufacturer reading time window. The reasons underlying our finding are not clear and we did not find any association with different temperature and humidity conditions. Explanations for the appearance of the false weak‐reactive lines may be due to nonspecific antibody binding at the HIV antigen test site on the devices nitrocellulose test strip 21 or nonspecific binding of protein‐A gold conjugate which the test uses as the colorimetric indicator or perhaps a lateral back flow or “settling effect” over time and further investigation into these hypotheses is required.
The observed change in result raises concerns over the use of delayed re‐reading of devices for monitoring HIVST interpretation, as well as for programmatic monitoring, evaluation and EQA. Research studies utilizing delayed re‐reading of returned OraQuick® HIV Self‐Test for establishing positivity may overestimate the true HIV positivity amongst a self‐testing population.
A previous study conducted in Malawi examined the pre‐use stability of OraQuick® HIV test kits 16. 371 optimally stored and 375 pre‐incubated used devices were re‐read over a 12‐month period. A 0.2% change from an initial reactive result to a later non‐reactive was observed (one in the pre‐incubated and one in the optimally stored group). These results suggested that HIVST device results remained stable over time. However, the focus of this study was its effect on pre‐use storage conditions. Post‐use storage conditions were not rigorously monitored and so cannot be reliable compared with the results from our study.
During this controlled study, our trained laboratorians could correctly distinguish false weak‐reactive test lines from true weak‐reactive test lines as they have a greyish appearance compared with the pinker true‐reactive. Implementation of this more nuanced approach may however prove challenging in programmatic settings where previous reports show that providers struggle to identify and interpret weak reactives 22 and other factors, such as interferents, and tests used among people with HIV using ART can cause weak reactives 23.
In addition to a false weak‐reactive line causing uncertainty to an EQA model, when testing a population, it is likely that more “true negative” samples will change to “false weak reactive” and delayed re‐reading by self‐testers themselves could lead to individual misinterpretation and misunderstandings. Our study showed that the OraQuick HIV device was stable up to four days after the sample was applied, suggesting that the risk of this is low but nevertheless self‐testers need clear messages about the read window and the importance of reading the device according to manufacturer instructions.
A limitation of this study was that on some re‐read days only two individual re‐reads were conducted (23%) and therefore a third “tie breaker” re‐read was not available. The very nature of self‐testing (conducting the test privately at home) means conventional facility/laboratory‐based QA systems of test devices are eluded, and an alternative approach is required. Digital photography and immediate re‐reading are two other options that are being further explored for QA during HIVST scale‐up, but these also have their limitations. National reference laboratories should play an integral role in external quality control measures by conducting batch testing at the actual sites of distribution of HIVST to ensure that the integrity of the devices is not compromised during transport and storage.
4. Conclusions
The use of re‐reading used OraQuick HIVST devices as an approach to quality assurance and monitoring test results is not advised. The instability observed in true non‐reactive tests changing to false weak reactive test results in our study demonstrates that re‐reading is not a reliable method to assess user interpretation of the OraQuick HIVST and measurement of HIV positivity rates among self‐testers.
Competing interests
PM is funded by the Wellcome Trust (206575/Z/17/Z). ELC is funded by the Wellcome Trust (WT200901/Z/16/Z). The remaining authors have no conflicts of interest to disclose.
Authors’ contributions
VW, RD and MT formulated and designed the experiments. VW, RD, CW, TE and EA performed the experiments. VW, PM and RD analysed the data. VW, RD, CW, TE, EA, CJ, MM, EC, FC, HA, KH, PM and MT wrote the concise communication.
Acknowledgements
The authors thank Dr Elliot Cowan, Dr Namuunda Mutombo and Mr Richard Chilongosi for their insightful comments on study design, and Mr Mohammed Majam for providing data on levels of weak‐reactive HSTAR results.
Funding
The study was undertaken in collaboration with Unitaid, Population Services International, WHO, Liverpool School of Tropical Medicine, London School of Hygiene and Tropical Medicine, and the rest of the STAR Consortium. The current work is supported by Unitaid, grant number: PO#10140‐0‐600. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
Watson, V. , Dacombe, R. J. , Williams, C. , Edwards, T. , Adams, E. R. , Johnson, C. , Mutseta, M. N. , Corbett, E. L. , Cowan, F. M. , Ayles, H. , Hatzold, K. , MacPherson, P. and Taegtmeyer, M. Re‐reading of OraQuick HIV‐1/2 rapid antibody test results: quality assurance implications for HIV self‐testing programmes. . J Int AIDS Soc. 2019; 22(S1):e25234
References
- 1. Thirumurthy H, Masters SH, Napierala Mavedzenge S, Maman S, Omanga E, Agot K. Promoting male partner testing and safer sexual decision making through secondary distribution of HIV self‐tests by HIV uninfected female sex workers and women receiving antenatal and postpartum care in Kenya: a cohort study. Lancet HIV. 2016;3(6):266–74. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2. Marlin R, Young S, Bristow C, Wilson G, Rodriguez J, Ortiz J, et al. Feasibility of HIV self‐test vouchers to raise community‐level serostatus awareness, Los Angeles. BMC Public Health. 2014;14:1226. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3. Green K, Thu H. In the hands of the community: accelerating key population‐led HIV lay and self‐testing in Viet Nam. [Abstract THSA1804] 21st International AIDS Conference 18‐22 July 2016.
- 4. Brady M. Self‐testing for HIV: initial experience of the UK's first kit. [Abstract 19] 22nd Annual Conference of the British HIV Association 21 April 2016.
- 5. Makusha T, Knight L, Taegtmeyer M, Tulloch O, Davids A, Lim J, et al. HIV self‐testing could “revolutionize testing in South Africa, but it has got to be done properly”: perceptions of key stakeholders. PLoS One. 2015;10(3):e0122783. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6. Van Rooyen H, Tulloch O, Mukoma W, Makusha T, Chepuka L, Knight LC, et al. What are the constraints and opportunities for HIVST scale‐up in Africa? Evidence from Kenya, Malawi and South Africa. J. Int AIDS Soc. 2015;18:19445. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7. Dacombe RJ, Watson V, Sibanda E, Nyirenda L, Simwinga M, Cowen E, et al. A Baseline Assessment of the Policy and Regulatory Environment for HIV Self‐testing in Malawi, Zimbabwe and Zambia. [Oral poster] African Society for Laboratory Medicine 3rd International Conference 3‐8 December 2016.
- 8. Johnson C, Figueroa C, Cambiano V, Phillips A, Sands A, Meurant R, et al. A clinical utility risk‐benefit analysis for HIV self‐testing. [Abstract TUPEC0834]. 9th IAS Conference on HIV Science 23‐26 July 2017.
- 9. Figueroa C, Johnson C, Ford N, Sands A, Dalal S, Meurant R, et al. Reliability of HIV rapid diagnostic tests for self‐testing compared with testing by health‐care workers: a systematic review and meta‐analysis. Lancet HIV. 2018;5(6):277–290. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10. Gotsche CI, Simwinga M, Muzumara A, Kapaku KN, Sigande L, Neuman M, et al. HIV self‐testing in Zambia: User ability to follow the manufacturer's instructions for use. [Abstract MOPED1167]. 9th IAS Conference on HIV Science 23‐26 July 2017.
- 11. Mwau M, Achieng L, Bwana P. Performance and usability of INSTI, a blood‐based rapid HIV self test for qualitative detection of HIV antibodies in intended use populations in Kenya. [Abstract MOAX0106LB] 9th IAS Conference on HIV Science 23‐26 July 2017.
- 12. Ortblad K, Musoke DK, Ngabirano T, Nakitende A, Magoola J, Kayiira P, et al. Direct provision versus facility collection of HIV self‐tests among female sex workers in Uganda: a cluster‐randomized controlled health systems trial. PLoS One. 2017;14(11):e1002458. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13. MacPherson P, Lalloo DG, Webb EL, Maheswaran H, Choko AT, Makombe SD, et al. Effect of optional home initiation of HIV care following HIV self‐testing on antiretroviral therapy initiation among adults in Malawi. A randomized clinical trial. JAMA. 2014;312(4):372–379. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14. Choko AT, Feilding K, Stallard N, Maheswaran H, Lepine A, Desmond N, et al. Investigating interventions to increase uptake of HIV testing and linkage into care or prevention for male partners of pregnant women in antenatal clinics in Blantyre, Malawi: study protocol for a cluster randomised trial. Trials. 2017;18:349. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15. Unitaid PSI HIV STAR Consortium . Zambia clinical performance study protocol. 2016; Available from: http://hivstar.lshtm.ac.uk/files/2016/11/STAR-Protocol-for-Zambia-Clinical-Performance-Study-web.pdf
- 16. Choko AT, Taegtmeyer M, MacPherson P, Cocker D, Khundi M, Thindwa D, et al. Initial accuracy of HIV rapid test kits stored in suboptimal conditions and validity of delayed reading of oral fluid tests. PLoS One. 2016;11(6):e0158107. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17. WHO Prequalification of In Vitro Diagnostics . Public Report. Product: OraQuick HIV Self‐Test WHO reference number: PQDx 0159‐055‐01. 2017. [cited 2018 Aug 08]. Available from: http://www.who.int/diagnostics_laboratory/evaluations/pq-list/170720_final_amended_pqdx_0159_055_01_oraquick_hiv_self_test_v2.pdf
- 18. Zeptometrix . HIV Seroconversion Panel Donor No. 73695. Zeptometrix. 2019; URL:https://www.zeptometrix.com/media/documents/PIHIV12007.pdf. Accessed: 2019‐01‐04. http://www.webcitation.org/75Ajt4f0y (Archived by WebCite® at http://www.webcitation.org/75Ajt4f0y)
- 19. Zeptometrix . HIV Seroconversion Panel Donor No. 75018. Zeptometrix. 2019; URL: https://www.zeptometrix.com/media/documents/PIHIV9077.pdf. Accessed: 2019‐01‐04. http://www.webcitation.org/75AjSCRd9. (Archived by WebCite® at http://www.webcitation.org/75AjSCRd9)
- 20. Hartling L, Hamm M, Milne A, Vandermeer B, Santaguida P, Ansari M, et al. Validity and Inter‐Rater Reliability Testing of Quality Assessment Instruments. [Internet]. Rockville, MD: Agency for Healthcare Research and Quality (US); 2012[cited Accessed October 2016]. Table 2, Interpretation of Fleiss’ kappa (κ) (from Landis and Koch 1977) Available from: https://www.ncbi.nlm.nih.gov/books/NBK92295/table/methods.t2/ [PubMed] [Google Scholar]
- 21. Klarkowski D, O'Brien DP, Shanks L, Singh KP. Causes of false‐positive HIV rapid diagnostic test results. Expert Rev Anti Infect Ther. 2014;12(1):49–62. [DOI] [PubMed] [Google Scholar]
- 22. Johnson CC, Fonner V, Sands A, Ford N, Obermeyer CM, Tsui S, et al. To err is human, to correct is public health: a systematic review examining poor quality testing and misdiagnosis of HIV status. J Int AIDS Soc. 2017;20 (Suppl 6):21755. doi: 10.7448/IAS.20.7.21755. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23. Fogel JM, Piwowar‐Manning E, Debevec B, Walsky T, Schlusser K, Laeyendecker O, et al. Brief report: impact of early antiretroviral therapy on the performance of HIV rapid tests and HIV incidence assays. J Acquir Immune Defic Syndr. 2017;75(4):426–430. [DOI] [PMC free article] [PubMed] [Google Scholar]