Abstract
Background
Microscopic examination of Giemsa-stained blood films remains a major form of diagnosis in malaria case management, and is a reference standard for research. However, as with other visualization-based diagnoses, accuracy depends on individual technician performance, making standardization difficult and reliability poor. Automated image recognition based on machine-learning, utilizing convolutional neural networks, offers potential to overcome these drawbacks. A prototype digital microscope device employing an algorithm based on machine-learning, the Autoscope, was assessed for its potential in malaria microscopy. Autoscope was tested in the Iquitos region of Peru in 2016 at two peripheral health facilities, with routine microscopy and PCR as reference standards. The main outcome measures include sensitivity and specificity of diagnosis of malaria from Giemsa-stained blood films, using PCR as reference.
Methods
A cross-sectional, observational trial was conducted at two peripheral primary health facilities in Peru. 700 participants were enrolled with the criteria: (1) age between 5 and 75 years, (2) history of fever in the last 3 days or elevated temperature on admission, (3) informed consent. The main outcome measures included sensitivity and specificity of diagnosis of malaria from Giemsa-stained blood films, using PCR as reference.
Results
At the San Juan clinic, sensitivity of Autoscope for diagnosing malaria was 72% (95% CI 64–80%), and specificity was 85% (95% CI 79–90%). Microscopy performance was similar to Autoscope, with sensitivity 68% (95% CI 59–76%) and specificity 100% (95% CI 98–100%). At San Juan, 85% of prepared slides had a minimum of 600 WBCs imaged, thus meeting Autoscope’s design assumptions. At the second clinic, Santa Clara, the sensitivity of Autoscope was 52% (95% CI 44–60%) and specificity was 70% (95% CI 64–76%). Microscopy performance at Santa Clara was 42% (95% CI 34–51) and specificity was 97% (95% CI 94–99). Only 39% of slides from Santa Clara met Autoscope’s design assumptions regarding WBCs imaged.
Conclusions
Autoscope’s diagnostic performance was on par with routine microscopy when slides had adequate blood volume to meet its design assumptions, as represented by results from the San Juan clinic. Autoscope’s diagnostic performance was poorer than routine microscopy on slides from the Santa Clara clinic, which generated slides with lower blood volumes. Results of the study reflect both the potential for artificial intelligence to perform tasks currently conducted by highly-trained experts, and the challenges of replicating the adaptiveness of human thought processes.
Electronic supplementary material
The online version of this article (10.1186/s12936-018-2493-0) contains supplementary material, which is available to authorized users.
Keyword: Malaria, Convolutional neural networks, Microscopy, Digital microscopy, Artificial intelligence
Background
Timely and accurate diagnosis of malaria is critical for effective treatment and to prevent transmission in communities. Microscopic examination remains the gold standard for laboratory confirmation of malaria. Specifically, light microscopy of Romanowski-stained blood films introduced by Laveran over 130 years ago [1, 2] remains critical in malaria case management and as a reference standard for research and disease monitoring. However, light microscopy remains a source of uncertainty because it is highly dependent on individual technician performance [3]. Antigen-detecting rapid diagnostic tests (RDTs) have become the predominant diagnostic tool for case management since their introduction over 20 years ago [4, 5], but they too have limitations in specificity and sensitivity, in species differentiation, and in their inability to provide accurate quantitation [6]. Molecular methods are improving, becoming cheaper and simpler [7], but remain confined to specialized laboratories.
Visualization of parasites is not subject to issues of antigen persistence, or to genetic mutations that limit the applicability of the major histidine-rich protein 2 (HRP2)-detecting antigen-based assay for Plasmodium falciparum, for example in Peru [8]. However, microscopy has major shortfalls in identifying, quantifying, and differentiating parasites from artifacts, and differentiating species. Furthermore, microscopy requires training, patience, competence, and well-managed workflows [9]. Even with skilled technicians, there is variability, which presents challenges in making reliable comparisons over time and geography [10, 11]. Low accuracy in case management leads to direct costs in morbidity and potential death, while inaccuracy in research leads to misleading results and resultant poor decision-making in product evaluations, resource allocations, and planning. While considerable efforts have been made in recent years to address quality assurance inadequacies [10], such programmes are expensive and limited in scope; others, such as the essential activities in the World Health Organization (WHO) therapeutic efficacy survey (TES) monitoring programme [12], rely on highly accurate film interpretation, and thus are difficult to scale.
The application of digital image recognition to malaria microscopy, using artificial intelligence algorithms to replace or supplement the human factor in blood film interpretation, have been attempted, usually on thin films [13, 14]. Some have coupled this with automated non-Romanowski staining techniques aimed at reducing variability in film preparation [15–17] in efforts to address the barriers in training, facility, and quality reagent requirements of manual Giemsa staining [18].
Standardization of blood films can simplify development and operation of image-recognition algorithms. However, automated staining with Romanowski stains cannot be achieved with simple equipment. Fluorescent staining, although it enables automation, adds new challenges in species differentiation, and also adds costs in new reagents, devices and supply lines. Devices based on non-Romanowski staining will be unable to read the 200 million plus slides currently produced each year for malaria case management [19], and will, therefore, require major changes in procedures closer to patient care in order to reach scaled implementation. As an alternative to the methods described above, recent developments in machine-learning techniques, based on the use of convolutional neural networks (a form of artificial intelligence in which an algorithm automatically extracts useful visual features to analyse images), open new possibilities for automated recognition of malaria parasites in standard Giemsa-stained blood films and for coping with inherent variability in film preparation.
The core of the Autoscope algorithm is based on convolutional neural networks (CNN)—a new and rapidly evolving field. Convolutional neural networks have been applied to a number of image detection and classification tasks on which they have achieved human-level performance [20–25]. CNN algorithms ‘learn’ based on analysing large numbers of objects classified by humans into different categories, and make ‘decisions’ on which features or patterns best distinguish these object categories. A full description of the Autoscope image analysis architecture is provide in the supplemental information section. Algorithm performance depends on the accuracy of the original object classification by trained microscopists, the number of training objects, and the similarity of training images to test images.
This paper reports diagnostic accuracy of Autoscope, a prototype parasite recognition system for images of Giemsa-stained malaria blood films, based on deep neural networks and other machine learning methods [26]. The Autoscope incorporates automated scanning of Giemsa-stained blood films and parasite-detection software to identify and count malaria parasites, and uses an algorithm trained on a broad range of blood films from geographically-diverse sites. The Autoscope scans with a digital camera based on a standard 100× oil-immersion objective optical train. Details of the automated scanning microscope hardware are given in the Additional files section, and have been described previously [26]. Annotation methods are described in the Additional file 1. The algorithm was trained on two species, P. falciparum and Plasmodium vivax, confined to thick film interpretation. The algorithm is intended to form the foundation for a more broadly applicable system to consistently mimic the competency of an expert microscopist and standardize film interpretation between sites and across time.
The Autoscope algorithm had previously undergone one round of field optimization at Shoklo Malaria Research Unit (SMRU) in Thailand, which also conducted the external quality control (EQC) for this study. The device scans approximately 2.5 mm2 of Giemsa-stained thick blood film which corresponds to approximately 0.1 µL of blood on a thick blood film conforming to the WHO standard [10]. The algorithm calibration assumes ≥ 0.1 µL of blood imaged and the device issues a warning of unreliable results if it counts a low number of white blood cells (WBCs). When imaging thick blood films, Autoscope images 9 vertically-stacked slices for each of 324 Fields-of-View (FoV). A portion of a typical thick film is shown in Fig. 1a. Thumb nails of malaria parasites are shown in Fig. 1b, c. The Autoscope algorithm then processes the captured images to detect both ring and late-stage malaria parasites and returns a diagnosis and parasite quantitation. Although Autoscope can image thick or thin films, the algorithm version used in this field study only operated on thick films, distinguishing species only as P. falciparum or P. vivax.
This paper describes a study to determine the equivalency of the diagnostic performance of the prototype Autoscope compared to a microscopist in a field setting, relative to polymerase chain reaction (PCR) as reference.
Methods
This prospective study was conducted in the district of San Juan Bautista in the province of Maynas, department of Loreto in Peru. San Juan, with population 145,238 [27], is an area of high risk of malaria transmission in Loreto and consists of three zones: urban, peri-urban, and rural populations, the latter two of which are high risk areas for malaria transmission due to the presence of the primary malaria vector Anopheles darlingi.
Two health care sites with diagnostic services enrolled patients for the study: San Juan de Miraflores Health Centre (San Juan), and Santa Clara de Nanay Health Post (Santa Clara). San Juan is a larger facility serving patients from various communities, including more distant, malaria-endemic communities, whereas Santa Clara provides more basic health care to a smaller, more remote population.
A total of 700 individuals (400 from Santa Clara and 300 from San Juan) were recruited using consecutive sampling from April 2016 to July 2016 with the following inclusion criteria: (1) age between 5 and 75 years; (2) history of fever within the last 3 days, or elevated temperature (≥ 37.5 °C) on the day of admission; (3) informed consent. A finger prick blood sample was taken to create blood films for microscopy diagnosis, and additional drops of blood were spotted onto filter paper for subsequent qPCR analysis. After sampling, any patients with a malaria diagnosis using manual microscopy received treatment, administered by the Regional Malaria Control programme in accordance with Ministry of Health guidelines [28]. All samples were analysed with microscopy and Autoscope. A subset of 10% of samples were sent to SMRU for EQC (Fig. 2).
Blood film reading by microscopy and Autoscope
Blood slides, each with both thin and thick films were stained with 10% Giemsa in phosphate buffer (pH 7.2) for 10 min, and air-dried as described in the national standard protocols [28]. At each clinic, three slides were prepared from each patient. Two of these slides were set aside for the participating health centre and for internal quality control. The third slide was used in the study and examined by both the microscopists and the Autoscope device.
At the clinics, detection of the malaria parasite, quantitation, and differentiation of the species were first performed by microscopy by clinic technicians (using both thick and thin films, according to protocols), and then by an Autoscope (using the same blood film and thick film only). Clinic technicians were blinded to Autoscope results. The slides were then sent to the Universidad Peruana Cayetano Heredia (UPCH) laboratory based in Iquitos (‘Morona’) for a quality-control reading by expert microscopists and a reading by a different Autoscope. The expert microscopists at Morona were blind to initial microscopy results. Expert microscopists were certified by the National Institute of Health from Peru.
Malaria diagnosis by Autoscope involves imaging the thick blood film, then running the diagnostic algorithm on the collected images. For imaging, the Giemsa-stained slide is inserted into the device, in the same way as a slide would be inserted into a 100× oil-immersion manual microscope. Diagnosis, species identification, quantitation, and mosaics of thumbnails of suspected parasite objects are automatically generated and outputted in an html report. Diagnosis thresholds on Autoscope are pre-specified, using results on a set of validation samples, with the goal of achieving > 90% specificity.
Real-time PCR
Real-time qPCR was performed using a CFX Connect Real Time System (BioRad). Approximately 80 μL of blood was applied to a filter paper 3 M (Watman) and allowed to dry completely. Parasite DNA from a double 10 mm hole punch was extracted using the E.Z.N.A® Blood DNA Mini Kit (Omega Bio-tek) according to the manufacturer’s instructions, except that Buffer ATL (Qiagen) and Proteinase K at 20 mg/mL (Qiagen) were used in the lysis step. Amplification was carried out in a 25 µL reaction volume containing 5 µL DNA, 12.5 µL of SYBR® Green PERFECTA® Fastmix (Quanta Biosciences) and 0.3 µM of each primer. A pair of primers was used to amplify the 18S rRNA gene sequences: PL1473F18 [5′-TAA CGA ACG AGA TCT TAA-3′] and PL1679R18 [5′-GTT CCT CTA AGA AGC TTT-3′]) [29]. The conditions for the qPCR consisted of an initial denaturation at 95 °C for 2 min followed by 45 cycles of amplification at 95 °C for 20 s, 52 °C for 20 s and 68 °C for 30 s and a final extension at 68 °C for 3 min, with fluorescence acquisition at the end of each extension step. After PCR amplification, a melting curve was performed to check the specificity of the amplicon, this consisted of a cycle at 95 °C for 10 s, followed by 68 °C for 2 min, and a gradual temperature increase of 0.5 °C/s up to 90 °C, with fluorescence measurement at each temperature transition. Each run included 2 positive controls (DNA sample from P. vivax and P. falciparum), a negative control (DNA from uninfected human), and a blank (nuclease-free water).
For all samples, the baseline and threshold were assigned for each run by the software BioRad CFX Manager 3.1. Samples were considered negative if the cycle threshold (Ct) value was greater than 39 or if the melt curve did not exceed the threshold or did not align with the positive control. Samples were considered positive when amplification curves crossed the threshold before cycle 39 and had the expected melt curve profile for each species (P. falciparum 73.5 °C ± 0.5 °C; P. vivax 77.5 °C ± 0.5 °C). Technicians performing qPCR were blinded to all results obtained from clinic sites.
Quality control
SMRU received 70 paired field slides representing 10% of the Autoscope evaluation. Slides came from both San Juan and Santa Clara clinics. SMRU has been an early evaluator of the Autoscope and has expert microscopy capacity to diagnose multiple species.
Analysis
Autoscope performance measures that were assessed included sensitivity, specificity, positive predictive value (PPV), negative predictive value (NPV), quantitation accuracy, and species ID accuracy. Performance was evaluated in two ways: (1) Diagnosis was compared to expert microscopy using PCR as reference; (2) Quantitation was directly compared to expert microscopy. Autoscope performance was parsed by WBC count because the detection algorithm assumes that 0.1 µL of blood has been imaged. Indeterminate tests were handled using the Obare Method [30], which is an excel-based tool to tabulate concordance of samples, and a technical data quality coordinator guaranteed completeness of the data.
For Autoscope performance, the Autoscopes at the two clinics of origin (San Juan, Santa Clara) were used. For expert microscopy performance, the microscopy readings from Morona were used. Microscopists at the clinics of origin produced very similar results; agreement between clinic microscopy diagnosis (including species) vs. Morona microscopy diagnosis was 97% (95% CI 95–98%) for the Santa Clara clinic, with 387 of 400 sample diagnoses in agreement. Agreement between clinic microscopy diagnosis versus Morona diagnosis was 99% (95% CI 98–100%) for the San Juan clinic, with 298 of 300 sample diagnoses in agreement.
Results from the two clinics are presented separately, to demonstrate variability of diagnostic accuracy per clinic. Point estimates and exact binomial confidence intervals for sensitivity, specificity, accuracy, PPV and NPV were calculated using the EpiR Package [31] in the R statistical environment [32].
The sample size of 700 was calculated as the number of samples required to characterize diagnostic performance (sensitivity and specificity) of the Autoscope with 5% margin of error, assuming 95% sensitivity and specificity, and 25% prevalence (one-sided exact binomial test, significance level α = 0.05, power 1 − β = 0.80).
Results
Peru field data
Autoscope performance differed between the clinics of origin of the slides processed, and by volume of blood imaged. San Juan thick films generally had greater volume of blood per unit area than Santa Clara thick films. Autoscope results are given for two groups of slides: (1) in the main text, results are given for all slides, regardless of Autoscope WBC counts; (2) results are provided for slides where Autoscope imaged at least 0.1 µL of blood (i.e. WBC count ≥ 600), assuming 6000 WBCs/µL in Additional file 2. Manual microscopy results from Morona are reported for all slides, i.e. (1) above.
WBC counts by clinic
Because Autoscope images a fixed area of the thick film (324 FoVs, approximately 2.5 mm2), the film thickness (volume of blood/mm2) determines the total volume imaged. This is approximated here by the white blood cell count. San Juan slides had WBC counts consistent with Autoscope’s design assumptions (i.e. WBC count ≥ 600 in 324 FoVs), while Santa Clara slides often had WBC counts lower than this (Fig. 3). Only 39% of Santa Clara slides (156/400) contained at least 600 WBCs in the blood imaged, compared to 84% of San Juan slides (253/300). Macro images of slides from the two clinics are shown in Additional file 3.
Autoscope versus microscopy, using PCR as reference
All samples were analysed by Autoscope and microscopy (Fig. 4). Samples were batched and PCR was performed at the end of the trial. Diagnostic performance of both microscopy (Table 1) and Autoscope (Table 2) differed according to clinic of origin. Microscopy and Autoscope performed better on San Juan slides. Poorer diagnostic performance by Autoscope on Santa Clara slides, relative to San Juan slides, is also evident even when samples with < 600 WBC are excluded (Fig. 5). Limits of Detection (LoD) were similar for microscopy and Autoscope on San Juan slides, ~ 100 p/uL (this estimate is limited because there were no San Juan slides with manually-counted parasitaemia < 100 p/µL). On Santa Clara slides, Autoscope LoD was ~ 200, while microscopy LoD was ~ 50 p/µL (a few Santa Clara slides had < 100 p/µL, accounting for the difference from LoD on San Juan slides). In general, the Autoscope and microscopists missed the same low-parasitaemia slides that were flagged as positive by PCR. LoD applies to thick films for both microscopy and Autoscope, since microscopy necessarily uses the thick film is used to diagnose low-parasitaemia samples.
Table 1.
n slides (pos, neg) | Diagnostic performance % (95% CI) | |||||
---|---|---|---|---|---|---|
Sensitivity | Specificity | PPV | NPV | Accuracy | ||
San Juan | ||||||
ll species | 300 (123, 177) | 68 (59–76) | 100 (98–100) | 100 (96–100) | 82 (76–87) | 87 (83–91) |
P. vivax | 300 (102, 198) | 70 (60–78) | 100 (98–100) | 100 (95–100) | 86 (81–91) | 90 (86–93) |
P. falciparum | 300 (21, 279) | 62 (38–82) | 100 (99–100) | 100 (75–100) | 97 (95–99) | 97 (95–99) |
Santa Clara | ||||||
All speciesa | 400 (151, 249) | 42 (34–51) | 97 (94–99) | 90 (81–96) | 74 (68–78) | 77 (72–81) |
P. vivax | 400 (100, 300) | 46 (36–56) | 98 (96–99) | 88 (77–96) | 84 (80–88) | 85 (81–88) |
P. falciparum | 400 (52, 348) | 31 (19–45) | 100 (99–100) | 100 (79–100) | 91 (87–93) | 91 (88–94) |
Results for microscopy, separated by clinic. Sensitivity was stronger on San Juan slides. Specificity was strong on slides from both clinics
aOne mixed species sample was detected
Table 2.
n slides (pos, neg) | Diagnostic performance % (95% CI) | |||||
---|---|---|---|---|---|---|
Sensitivity | Specificity | PPV | NPV | Accuracy | ||
San Juan | ||||||
All species | 300 (123, 177) | 72 (64–80) | 85 (79–90) | 77 (68–84) | 82 (75–87) | 80 (75–84) |
P. vivax | 300 (102, 198) | 72 (62–80) | 83 (77–88) | 69 (59–77) | 85 (79–90) | 79 (74–83) |
P. falciparum | 300 (21, 279) | 33 (15–57) | 99 (97–100) | 78 (40–97) | 95 (92–97) | 95 (91–97) |
Santa Clara | ||||||
All speciesa | 400 (151, 249) | 52 (44–60) | 70 (64–76) | 52 (43–59) | 71 (65–76) | 64 (58–68) |
P. vivax | 400 (100, 300) | 60 (50–70) | 71 (66–76) | 41 (33–49) | 84 (79–89) | 69 (64–73) |
P. falciparum | 400 (52, 348) | 4 (0–13) | 99 (98–100) | 50 (7–93) | 87 (84–90) | 87 (83–90) |
Results for autoscope on all slides, separated by clinic
aOnly one mixed species sample was detected
Comparison between Autoscope and microscopy given adequate blood volume can be seen in the results for San Juan, Additional file 3: Figure S3a. When sensitivities are similar (~ 68%), microscopy specificity was higher (98% vs 90%). Autoscope sensitivity at higher specificity (i.e. at a different point on the sens/spec tuning curve) can be inferred from values on slides with > 1000 WBCs: At this point, Autoscope has similar specificity to microscopy (98%), but lower sensitivity (60%). So trained microscopy had a performance edge.
Autoscope quantitation concordance with expert microscopy
Similarity of parasite quantitation determined by Autoscope versus Morona expert microscopy varied according to the clinic of origin. To evaluate the degree of similarity, Pearson’s correlation coefficient was calculated on the log-transformed parasitaemia values of true positive slides, using PCR as the reference standard. Autoscope and microscopy parasite densities were more strongly correlated on San Juan slides, r = 0.86 (95% CI 0.79–0.91), compared to Santa Clara slides, r = 0.59 (95% CI 0.37–0.75) (Fig. 5). On the Santa Clara slides, Autoscope underestimated parasite density relative to microscopy on most slides (Fig. 5). If microscopy quantitation was done on thin films, this would naturally lead to relative underestimation by Autoscope, since thick films having lower parasitaemias [33].
Inter-device quantitation concordance
The concordance of the Autoscope at Morona and the Autoscopes at the two clinics depended on clinic and on the number of WBCs imaged. For San Juan slides, 48 of 300 slides (16%) were discordant. For Santa Clara slides, 168 of 400 (42%) were discordant. On concordant slides, Pearson’s correlation coefficient was computed to assess the correlation between log-transformed parasitaemia values. Restricted to concordant slides, the Autoscopes at San Juan and at Morona correlated with r =0.90 (95% CI 0.85–0.93) on all slides, and 0.95 (95% CI 0.92–0.97) on slides with > 600 WBCS. Autoscopes at Santa Clara correlated with r = 0.41 (95% CI 0.26–0.55) on all slides, and r = 0.71 (95% CI 0.37–0.88) on slides with > 600 WBCs (Fig. 6).
External quality control
EQC was performed to provide a subjective scale of readability. In December 2016, SMRU reviewed 70 paired field slides representing 10% of the overall study slides, selected randomly, following WHO guidelines for EQC.
Of the 70 slides, 59 slides (84.3%) were considered adequate for evaluation. Three slides (4.3%) were invalid because of stain deterioration over time (precipitate). Readable slides were cross-checked. Of the 59 readable slides, 3 slides (5.1%) diagnosed as ‘negative’ by microscopy were P. falciparum positive. Discarding the three invalid slides, slide quality issues identified included 16 slides (22.9%) which had thick films with flaws, mostly insufficient blood, and 6 slides (8.6%) that were poorly stained.
The proportion of slides with inadequate blood volume varied by clinic. Excluding invalid slides, 4 of 33 San Juan slides (12%) had inadequate volume. Of Santa Clara slides, 13 out of 34 (38%) had inadequate volume. These proportions correspond to the proportions of slides for which Autoscope counted less than 600 WBCs (San Juan) and less than 400 WBCs (Santa Clara), further confirming lower blood density on Santa Clara slides.
Discussion
This study examined the diagnostic performance of a prototype device that uses an image recognition algorithm to identify and classify malaria parasites on routinely-prepared Giemsa-stained blood films. Results of the study reflect both the potential for artificial intelligence to perform tasks currently conducted by highly-trained experts, and the challenges of replicating the broad depth and adaptiveness of human thought processes. The prototype device was not able to compensate as well as expert microscopists on blood films with inadequate blood volume. On slides with adequate blood volume, the prototype device demonstrated that machine learning-based algorithms have potential to achieve equivalence to expert human microscopists and thus offer a viable approach to mitigating the persistent problem of poor reliability of routine malaria microscopy.
A key finding of this study was that it evaluated diagnostic performance of a prototype machine-learning device under field conditions that represent a realistic use-case for the technology. Blood films prepared at the clinics demonstrated variability, in slide preparation quality and blood volume, likely to be routinely observed were the prototype device to be used at peripheral health clinics as intended. Slides were prepared in a routine manner intended to optimize human performance, and scanned at a magnification intended for human-based identification.
Because only two clinics were included in the study, it is impossible to generalize broadly from the differences observed between the two. Diagnostic performance of the prototype device was considerably poorer, well below that of expert microscopists, on slides from one of the two clinics. It is not possible from this study to know how prevalent such conditions are, and how representative the two clinics in the study are of clinics that would use the device overall.
In addition, the algorithm in this study was trained on high-quality blood films, and it is not possible to know from this study what the diagnostic performance would have been had the algorithm been trained on films of variable quality. The high performance at San Juan suggests that the algorithm had been sufficiently trained in terms of number of correctly annotated objects parasites to correctly classify new objects, given blood films with presentation similar to that of training films. The algorithm used in this trial was trained on approximately 150 high-quality thick films with 75,000 parasites, obtained from several sites in endemic and non-endemic countries. Poorer performance at Santa Clara is likely due in part to the device not having been trained on slides with low blood volume. Autoscope thresholds were calibrated assuming approximately 0.1 µL of blood scanned; less blood scanned will lead to a larger standard error for discovered objects per WBC parasites and distractors. Low blood volume slides violate Autoscope’s design assumptions, resulting in lower specificity. Error generated by low blood volume slides could be mitigated by altering the scanning protocol to ensure sufficient blood is imaged, for example by basing scanning volume on the WBC count as a human technician is trained to do.
The differential performance of microscopists and the AI-based algorithm between the clinics is important for understanding the potential of such algorithms in diagnostic practice. An algorithm is limited by the images on which it has been trained. A microscopist takes into account a wide set of contextual conditions. Knowing the particulars of the clinic and technicians who prepared a film, noting sometimes subtle changes in overall slide color, and the ability to select certain areas of a thick film for greater scrutiny, could be included in a software algorithm. However, in practice the complexity and the often localized nature of the knowledge required make this impractical.
The Autoscope hardware performed well. Failures on Santa Clara slides were due to the software, in particular the algorithm’s lack of adaptability to irregular blood films prepared in the Santa Clara clinic. However, the Peruvian malaria control program maintains strong quality protocols. Thus the presence of difficult (for the algorithm) blood films even within the top-notch Peruvian system indicates that the issue of slide variability will occur throughout the world.
The primary use case for the Autoscope is as a research tool. While the algorithm is designed to standardize microscopy for research, the results suggest that with further improvement it may be possible to meet sufficient accuracy for case management using these methods. Given the relatively early version of the algorithm used in this study, the results hold considerable promise for replacing or enhancing certain aspects of human-based malaria microscopy with software, without changing well-established and low-cost methods for blood film preparation.
A high-performing algorithm could serve to standardize film interpretation across geography and across time, of particular relevance to clinical studies of drugs, vaccines and other diagnostics. It could also serve as a cross-checking on microscopist performance; machines do not tire, and cross-checking of slides, while recommended in any Quality Assurance (QA) programme, is rarely performed adequately [10]. In non-endemic areas, such as diagnosis of returned travelers, the device could supplement diagnosis provided by technicians who rarely see real parasites.
If the challenges of variable slide quality can be overcome, the prototype device described herein may serve as a primary diagnostic tool in areas where microscopy is still the basis for case management. Recognition and flagging of poor quality films can partially address this problem. As with other machine-learning algorithms, exposure to a larger number of well-annotated images is a proven way to improve performance, and should be considered in future research.
Conclusion
Autoscope’s diagnostic performance was on par with routine microscopy when slides had adequate blood volume to meet its design assumptions, as represented by results from the San Juan clinic. Autoscope’s diagnostic performance was poorer than routine microscopy on slides from the Santa Clara clinic, which generated slides with lower blood volumes. Results of the study reflect both the potential for artificial intelligence to perform tasks currently conducted by highly-trained experts, and the challenges of replicating the adaptiveness of human thought processes.
Additional files
Authors’ contributions
KT supported lab aspects and wrote the manuscript, CMB and DB contributed to study design and wrote the manuscript, CBD wrote and reviewed the manuscript, performed data analysis, and created the computer vision algorithm, JA and TO lab aspects, FA coordinator, DGV study design, SP external quality control, CM wrote and reviewed the manuscript and created computer vision algorithm, SKM developed hardware, CMT and LU created the computer vision algorithm, VMH reviewed the manuscript and performed data analysis. All authors read and approved the final manuscript.
Acknowledgements
We gratefully acknowledge the institutions that provided blood films to Global Good to train the Autoscope algorithm (Amref Health Africa’s Regional Laboratory Programme, the Hospital for Tropical Diseases, Universidad Peruana Cayetano Heredia, Shoklo Malaria Research Unit, the WorldWide Antimalarial Resistance Network, Kenya Medical Research Institute, the University of Lagos, the Defense Services Medical Academy, and the Research Institute for Tropical Medicine) and the Shoklo Malaria Research Unit for performing the EQC on the slide set. We are grateful to the Directors and staff of San Juan and Santa Clara health centers and the Morona laboratory for their assistance and collaboration.
Competing interests
The authors declare that they have no competing interests.
Availability of data and materials
Data from the study and a combined dataset used for this analysis are available from the corresponding author on request.
Consent for publication
Not applicable.
Ethics approval and consent to participate
Approval for the study was granted by the Universidad Peruana Cayetano Heredia Ethics Committee. Registration Number: 066047. Evaluación en Campo del Autoscope (Nuevo Método para el Diagnóstico de Malaria). Consent was obtained from all participants enrolled in this study.
Funding
Funding was provided by Bill and Melinda Gates Foundation Trust through Intellectual Ventures Global Good Fund. The funders had no role in the study design, implementation, analysis, manuscript preparation, or decision to submit this article for publication. The findings and conclusions in this report are those of the authors and do not necessarily represent the official position of funders.
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Abbreviations
- AI
artificial intelligence
- CNN
convolutional neural networks
- EQC
external quality control
- FOV
fields of view
- NPV
negative predictive value
- PCR
polymerase chain reaction
- PPV
positive predictive value
- QA
quality assurance
- RDT
rapid diagnostic test
- SMRU
Shoklo Malaria Research Unit
- TES
WHO Therapeutic Efficacy Survey
- UPCH
Universidad Peruana Cayetano Heredia
- WBC
white blood cell
- WHO
World Health Organisation
References
- 1.Laveran A. Traité des fiévres palustres, avec la description des microbes du paludisme. Paris: Octave Doin; 1884. [Google Scholar]
- 2.Bruce-Chwatt LJ. Alphonse Laveran’s discovery 100 years ago and today’s global fight against malaria. J R Soc Med. 1981;74:531–536. doi: 10.1177/014107688107400715. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Zurovac D, Midia B, Ochola SA, English M, Snow RW. Microscopy and outpatient malaria case management among older children and adults in Kenya. Trop Med Int Health. 2006;11:432–440. doi: 10.1111/j.1365-3156.2006.01587.x. [DOI] [PubMed] [Google Scholar]
- 4.Shiff CJ, Minjas J, Premji Z. The ParaSight-F test: a simple rapid manual dipstick test to detect Plasmodium falciparum infection. Parasitol Today. 1994;10:494–495. doi: 10.1016/0169-4758(94)90164-3. [DOI] [PubMed] [Google Scholar]
- 5.Dietze R, Perkins M, Boulos M, Luz F, Reller B, Corey GR. The diagnosis of Plasmodium falciparum infection using a new antigen detection system. Am J Trop Med Hyg. 1995;52:45–49. doi: 10.4269/ajtmh.1995.52.45. [DOI] [PubMed] [Google Scholar]
- 6.Bell D, Peeling RW. Evaluation of rapid diagnostic tests: malaria. Nat Rev Microbiol. 2006;4:S34–S38. doi: 10.1038/nrmicro1524. [DOI] [PubMed] [Google Scholar]
- 7.Poon LLM. Sensitive and inexpensive molecular test for Falciparum malaria: detecting Plasmodium falciparum DNA directly from heat-treated blood by loop-mediated isothermal amplification. Clin Chem. 2005;52:303–306. doi: 10.1373/clinchem.2005.057901. [DOI] [PubMed] [Google Scholar]
- 8.Gamboa D, Ho M-F, Bendezu J, Torres K, Chiodini PL, Barnwell JW, et al. A large proportion of P. falciparum isolates in the Amazon region of Peru lack pfhrp2 and pfhrp3: implications for malaria rapid diagnostic tests. PLoS ONE. 2010;5:8091. doi: 10.1371/journal.pone.0008091. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.Wongsrichanalai C, Barcus MJ, Muth S, Sutamihardja A, Wernsdorfer WH. A review of malaria diagnostic tools: microscopy and rapid diagnostic test (RDT) Am J Trop Med Hyg. 2007;77:119–127. doi: 10.4269/ajtmh.2007.77.119. [DOI] [PubMed] [Google Scholar]
- 10.WHO. Malaria microscopy quality assurance manual. Version 2. Geneva: World Health Organization; 2016.
- 11.Maguire JD, Lederman ER, Barcus MJ, O’Meara WAP, Jordon RG, Duong S, et al. Production and validation of durable, high quality standardized malaria microscopy slides for teaching, testing and quality assurance during an era of declining diagnostic proficiency. Malar J. 2006;5:92. doi: 10.1186/1475-2875-5-92. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.WHO. Methods for surveillance of antimalarial drug effiacy. Geneva: World Health Organization; 2009.
- 13.Ross NE, Pritchard CJ, Rubin DM, Dusé AG. Automated image processing method for the diagnosis and classification of malaria on thin blood smears. Med Biol Eng Comput. 2006;44:427–436. doi: 10.1007/s11517-006-0044-2. [DOI] [PubMed] [Google Scholar]
- 14.Tek FB, Dempster AG, Kale I. Computer vision for microscopy diagnosis of malaria. Malar J. 2009;8:153. doi: 10.1186/1475-2875-8-153. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Gillet P, Bosselaers K, Cnops L, Bottieau E, Van Esbroeck M, Jacobs J. Evaluation of the SD FK70 malaria Ag Plasmodium vivax rapid diagnostic test in a non-endemic setting. Malar J. 2009;8:129. doi: 10.1186/1475-2875-8-129. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Uguen C, Rabodonirina M, De Pina JJ, Vigier JP, Martet G, Maret M, et al. ParaSight-F rapid manual diagnostic test of Plasmodium falciparum infection. Bull World Health Organ. 1995;73:643–649. [PMC free article] [PubMed] [Google Scholar]
- 17.Vink JP, Laubscher M, Vlutters R, Silamut K, Maude RJ, Hasan MU, et al. An automatic vision-based malaria diagnosis system. J Microsc. 2013;250:166–178. doi: 10.1111/jmi.12032. [DOI] [PubMed] [Google Scholar]
- 18.WHO. Basic malaria microscopy. 2nd ed. Geneva: World Health Organization; 2010.
- 19.WHO. World malaria report 2016. Geneva: World Health Organization; 2017.
- 20.Poostchi M, Silamut K, Maude R, Jaeger S, Thoma G. Image analysis and machine learning for detecting malaria. Transl Res. 2018;194:36–55. doi: 10.1016/j.trsl.2017.12.004. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Loddo A, Di Ruberto C, Kocher M. Recent advances of malaria parasites detection systems based on mathematical morphology. Sensors. 2018;18:513. doi: 10.3390/s18020513. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22.Gopakumar GP, Swetha M, Sai Siva G, Sai Subrahmanyam GRK. Convolutional neural network-based malaria diagnosis from focus stack of blood smear images acquired using custom-built slide scanner. J Biophotonics. 2018;11:3. doi: 10.1002/jbio.201700003. [DOI] [PubMed] [Google Scholar]
- 23.Abbas N, Saba T, Mohamad D, Rehman A, Almazyad AS, Al-Ghamdi JS. Machine aided malaria parasitemia detection in Giemsa-stained thin blood smears. Neural Comput Appl. 2018;29:803–818. doi: 10.1007/s00521-016-2474-6. [DOI] [Google Scholar]
- 24.Le M-T, Bretschneider TR, Kuss C, Preiser PR. A novel semi-automatic image processing approach to determine Plasmodium falciparum parasitemia in Giemsa-stained thin blood smears. BMC Cell Biol. 2008;9:15. doi: 10.1186/1471-2121-9-15. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.Linder N, Turkki R, Walliander M, Mårtensson A, Diwan V, Rahtu E, et al. A malaria diagnostic tool based on computer vision screening and visualization of Plasmodium falciparum candidate areas in digitized blood smears. PLoS ONE. 2014;9:e104855. doi: 10.1371/journal.pone.0104855. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Mehanian C, Jaiswal M, Thompson C, Horning M, Ostbye T, McGuire S, et al. Computer-automated malaria diagnosis and quantitation using convolutional neural networks. In: IEEE Int Conf Comput Vis ICCV. 2017;116–25.
- 27.Instituto Nacional de Estadistica e Informatica (INEI). Poblacion 2000–2015. https://proyectos.inei.gob.pe/web/poblacion/. Accessed 5 Sep 2018.
- 28.Ministerio de Salud del Perú . Manual de Procedimientos de Laboratorio Para El Diagnóstico de Malaria (MINSA) Lima: Ministerio de Salud del Perú; 2003. [Google Scholar]
- 29.Mangold KA, Manson RU, Koay ESC, Stephens L, Regner M, Thomson RB, et al. Real-time PCR for detection and identification of Plasmodium spp. J Clin Microbiol. 2005;43:2435–2440. doi: 10.1128/JCM.43.5.2435-2440.2005. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30.Rosen, M, Dhorda, M, Bell, D, Gatton, M. Obare method calculator. www.wwarn.org. Accessed 5 Sep 2018.
- 31.Stevenson M, Nunes T, Heuer C, Marshall J, Sanchez J, Thornton R, et al. epiR: tools for the analysis of epidemiological data. https://CRAN.Rproject.org/package=epiR. Accessed 5 Sep 2018.
- 32.R Core Team. R: A language and environment for statistical computing. https://www.R-project.org. Accessed 5 Sep 2018.
- 33.Alexander N, Schellenberg D, Ngasala B, Petzold M, Drakeley C, Sutherland C. Assessing agreement between malaria slide density readings. Malar J. 2010;9:4. doi: 10.1186/1475-2875-9-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
Data from the study and a combined dataset used for this analysis are available from the corresponding author on request.