Abstract
Background
Machine learning has been an emerging tool for various aspects of infectious diseases including tuberculosis surveillance and detection. However, the World Health Organization (WHO) provided no recommendations on using computer-aided tuberculosis detection software because of a small number of studies, methodological limitations, and limited generalizability of the findings.
Methods
To quantify the generalizability of the machine-learning model, we developed a Deep Convolutional Neural Network (DCNN) model using a Tuberculosis (TB)-specific chest x-ray (CXR) dataset of one population (National Library of Medicine Shenzhen No.3 Hospital) and tested it with non-TB-specific CXR dataset of another population (National Institute of Health Clinical Centers).
Results
In the training and intramural test sets using the Shenzhen hospital database, the DCCN model exhibited an AUC of 0.9845 and 0.8502 for detecting TB, respectively. However, the AUC of the supervised DCNN model in the ChestX-ray8 dataset was dramatically dropped to 0.7054. Using the cut points at 0.90, which suggested 72% sensitivity and 82% specificity in the Shenzhen dataset, the final DCNN model estimated that 36.51% of abnormal radiographs in the ChestX-ray8 dataset were related to TB.
Conclusion
A supervised deep learning model developed by using the training dataset from one population may not have the same diagnostic performance in another population. Conclusion: Technical specification of CXR images, disease severity distribution, dataset distribution shift, and overdiagnosis should be examined before implementation in other settings.
Keywords: Computer science, Applied computing in medical science, Tuberculosis, Deep learning, Chest X-Ray
Computer science; Applied computing in medical science; Tuberculosis; Deep learning; Chest X-Ray.
1. Introduction
Tuberculosis (TB) is a major health problem in many regions of the world, especially in developing countries. While a majority of patients have their lungs infected with TB, some individuals might have TB infection in other body organs. Hence, unlike other infectious diseases, the diagnosis of TB is relatively more difficult and several tests are usually needed. Although Chest X-ray (CXR) is one of the primary tools for TB screening, a suspected individual requires clinical, biological, and genetic investigations before the actual diagnosis can be made and the medications are readily prescribed. As part of the World Health Organization (WHO) systematic screening strategy to ensure early and correct diagnosis for all people with TB, CXR is one of the primary tools for triaging and screening for TB because of its relatively high sensitivity, depending on how the CXR is interpreted [1]. However, significant intra- and inter-observer variations in the reading of CXR can lead to overdiagnosis or underdiagnosis of tuberculosis.
Deep convolutional neural network (DCNN) has emerged as an attractive technique for TB surveillance and detection. Besides, one study indicated that this approach can accurately detect TB cases in less than 3 min at minimal expense (https://www.digitalcreed.in/ai-for-tb/). This ‘supervised’ machine learning algorithm learns a mapping from a set of covariates to the outcome of interest by using the training data then applies this mapping to the new test data for identification or prediction tasks [2]. In a common deep learning model, the covariates are the color pixel values of the CXR images whereas the outcome is the radiologist's interpretation and impression of the corresponding CXR.
Recently, the National Institute of Health (NIH) released the ChestX-ray8 dataset with more than 100,000 anonymized CXR images and their associated data which compiled from more than 32,000 patients [3]. These data allow researchers to further develop an algorithm for classifying lung abnormalities labelled from the radiological reports using the National Language Processing technique [4, 5]. Nonetheless, according to the ChestX-ray8 criteria, a CXR image with minimal lung lesions could be incorrectly labelled as normal whereas these radiologic labels are not specific to TB [3].
The Computer-Aided Detection for Tuberculosis (CAD4TB), SemanticMD, and Qure.ai are selected examples of currently available computer-aided detection (CAD) software. CAD4TB is TB specific and had demonstrated a good diagnostic performance [6, 7] but still inferior to that of expert readers [8]. As of 2016, WHO provided no recommendations on using CAD for TB because of the small number of studies, methodological limitations, and limited generalizability of the findings [1]. The other two software cover a broader range of clinical conditions but available only for investigational use in the U.S.
Unlike other diagnostic tests, the technical specification of CXR images and disease severity distribution could affect the diagnostic performance of a supervised machine learning model. That is, the picture archive and communication system (PACS) data has more information but requires more time, storage, and processing power than low bit-depth formats (.jpg) whereas the supervised deep learning model developed based on CXR images of hospitalized patients would be smarter at detecting severe case than another community-based model. To quantify the extent of dataset specificity that limits the generalizability of CAD for TB, we developed a DCNN model using a TB-specific CXR dataset of one population and tested it with non-TB-specific CXR dataset of another population.
2. Methods
Two de-identified Health Insurance Portability and Accountability Act (HIPAA)-compliant datasets, the National Library of Medicine (NLM) Shenzhen No.3 Hospital X-ray set, and the NIH ChestX-ray8 database were included in this study. The Shenzhen dataset collected 336 normal and 326 abnormal CXR of bacteriologically and/or clinically confirmed TB with various manifestations as part of the standard TB care at Shenzhen No.3 Hospital in Shenzhen, Guangdong providence, China. The Shenzhen dataset was provided in PNG format with a resolution of 3000 × 3000 pixels. The ChestX-ray8 comprises of CXR images acquired as a part of routine care at NIH Clinical Center, Bethesda, Maryland, USA. The images were directly extracted from the DICOM file and resized as 1024 × 1024 bitmap images without significantly losing the detailed contents. It comprised of 112,120 frontal-view CXR images (60,360 normal and 51,760 abnormal images) of 30,805 unique de-identified patients with the text mined eight disease image labels (including normal, atelectasis, cardiomegaly, effusion, infiltration, mass, nodule, pneumonia, and pneumothorax), from their corresponding radiological reports using natural language processing.
Firstly, the Shenzhen dataset was split into training (75%), validation (15%), and intramural test (10%) sets. Based on the TensorFlow framework, Inception V3, a novel pre-trained DCNN, was augmented with color-space, crop-flip, rotational methods to classify each image as having TB characteristics or as healthy. The learning rate was 0.01, training and validating batch sizes were both 100 with five EPOCHs (https://github.com/panasun/dac4tb). The desktop computer was equipped with Intel i7-7700 CPU @ 3.60GHz, 32G RAM, and NVIDIA GeForce GTX 1080 @ 8GB GPU.
Given no classification of TB within the Chest X-ray8 dataset, we adopted the WHO guideline on Chest Radiography in Tuberculosis Detection which described the standard characteristics of CXR associated with active TB disease by using only 5,000 images of normal CXR and 5,000 images of CXR with active TB associated features according to the WHO guideline (infiltration, pneumonia, atelectasis, and effusion) (1) from Chest X-ray8 dataset to create a valid extramural test set to examine the generalizability of the DCNN model to classify normal and other CXR in addition to test set from intramural (Shenzhen) dataset.
Lastly, the prevalence of TB-associated CXR in the ChestX-ray8 dataset was estimated by using the final DCNN model. Receiver operating characteristic (ROC) curves and areas under the curve (AUC) were used to assess model performance and to define the optimal cut point for TB detection.
3. Results
In the training and intramural test sets using the Shenzhen hospital database, the DCNN model exhibited an AUC of 0.9845 and 0.8502 for detecting TB, respectively (Figures 1 and 2). However, the AUC of the supervised DCNN model in the ChestX-ray8 dataset was dramatically dropped to 0.7054 (Figure 3). Using the cut points at 0.90 which suggested 72% sensitivity and 82% specificity in the Shenzhen dataset, the final DCNN model estimated that 36.51% of abnormal radiographs in the ChestX-ray8 dataset were related to TB.
4. Discussion
Our findings suggested that the diagnostic performance of a supervised machine learning model is dataset specific, because of the varying technical specification of CXR images and disease severity distribution in different populations. In this study, the Inception V3 TensorFlow model maintained the diagnostic accuracy for the training, validation, and intramural test images in the Shenzhen dataset. This ability to correctly classified unseen test images within the same dataset is known as generalization. However, the significant drop in the diagnostic performance in the extramural test images reflects poor generalizability.
The falling of performance when the distribution of the data changed is known as ‘distribution shift’, and still an active area of research in the machine learning community. One might argue that the distribution shift becomes more apparent in our study because of insufficient model preparation and optimization. However, we had augmented our model with several techniques that have been known to improve generalization of deep learning in the other area of research including color-space, crop-flip, and rotational methods. Although, there are other techniques [9,10] that may help improve our model performance. By far the best, the discrepancy in CXR readings pattern between the two datasets according to differences in prevalence, manifestations of common lung abnormalities, as well as health care system in China and the U.S. are unavoidably having an important role in performance shift in this study. Thus, it is more likely our findings indicated that dataset distribution shifts in deep learning in some specific research area is inevitable. Careful implementation of AI and further researches to examine the problem will improve the development and adoption of AI in medicine and public health.
The 36.51% of abnormal CXR in the NIH ChestX-ray8 dataset was associated with TB might demonstrate the ‘overdiagnosis’ of deep learning since many CXR abnormalities that are compatible with pulmonary TB are seen also in several lung pathologies and, therefore, are indicative not only of TB but also of other pathologies.
Some limitations of this study should be noted. Like other deep learning studies, the images were resized to a manageable dimension before being fed into the model as a larger file will increase the training time and will require more robust central and graphical processing power. Accuracy may be improved by using higher resolution images, particularly for subtle findings, and more research is needed in this regard. Secondly, we did not comparatively explore the effect of using other pre-trained models such as the VGGNet or ResNet as the former is a very slow pre-trained sequential model and the latter is a relatively less efficient model than the Inception V3. Lastly, this retrospective study was conducted using the datasets that were available at the time. Further investigation on the use of deep learning in a real-world large-scale screening program for pulmonary TB in prevalent regions is essential.
5. Conclusions
A supervised deep learning model developed by using the training dataset from one population may not always have the same diagnostic performance in another population. Technical specification of CXR images, disease severity distribution, dataset distribution shift, and overdiagnosis should be examined before implementation in other settings.
Declarations
Author contribution statement
K. Pongpirul, S. Sathitratanacheewin: Conceived and designed the experiments; Performed the experiments; Analyzed and interpreted the data; Wrote the paper.
P. Sunanta: Conceived and designed the experiments; Performed the experiments; Analyzed and interpreted the data; Contributed reagents, materials, analysis tools or data.
Funding statement
This study was supported by the Health Systems Research Institute (HSRI 62-103) and Ratchadapiseksompotch Matching Fund, Faculty of Medicine, Chulalongkorn University (RA-MF-12/62).
Competing interest statement
The authors declare no conflict of interest.
Additional information
No additional information is available for this paper.
Acknowledgements
We also thank Dr. Ekapol Chuangsuwanich, Faculty of Engineering, Chulalongkorn University and Dr. Phalin Kamolwat, Bureau of Tuberculosis, Department of Disease Control, Ministry of Public Health for kind advice.
References
- 1.World Health Organization . World Health Organization; Geneva: 2016. Chest Radiography in Tuberculosis Detection: Summary of Current WHO Recommendations and Guidance on Programmatic Approaches. [Google Scholar]
- 2.Wiens J., Shenoy E.S. Machine learning for healthcare: on the verge of a major shift in healthcare epidemiology. Clin. Infect. Dis. 2018;66(1):149–153. doi: 10.1093/cid/cix731. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Wang X. 2017. ChestX-ray8: Hospital-Scale Chest X-ray Database and Benchmarks on Weakly-Supervised Classification and Localization of Common Thorax Diseases. arXiv:1705.02315. [Google Scholar]
- 4.Rajpurkar P. 2017. CheXNet: Radiologist-Level Pneumonia Detection on Chest X-Rays with Deep Learning. arXiv:1711.05225. [Google Scholar]
- 5.Yao L. 2017. Learning to Diagnose from Scratch by Exploiting Dependencies Among Labels. arXiv:1710.10501. [Google Scholar]
- 6.Maduskar P. Performance evaluation of automatic chest radiograph reading for detection of tuberculosis (TB): a comparative study with a clinical officers and certified readers on TB suspects in sub-Saharan Africa. Insight Imag. 2013;4(Suppl 1) [Google Scholar]
- 7.Muyoyeta M. The sensitivity and specificity of using a computer aided diagnosis program for automatically scoring chest X-rays of presumptive TB patients compared with Xpert MTB/RIF in Lusaka Zambia. PloS One. 2014;9(4) doi: 10.1371/journal.pone.0093757. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Steiner A. Screening for pulmonary tuberculosis in a Tanzanian prison and computer-aided interpretation of chest X-rays. Public Health Action. 2015;5(4):249–254. doi: 10.5588/pha.15.0037. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.Zhou P., Han J., Cheng G., Zhang B. Learning compact and discriminative stacked autoencoder for hyperspectral image classification. IEEE Trans. Geosci. Rem. Sens. July 2019;57(7):4823–4833. [Google Scholar]
- 10.Cheng G., Yang C., Yao X., Guo L., Han J. When deep learning meets metric learning: remote sensing image scene classification via learning discriminative CNNs. IEEE Trans. Geosci. Rem. Sens. May 2018;56(5):2811–2821. [Google Scholar]