Skip to main content
Journal of Clinical Medicine logoLink to Journal of Clinical Medicine
. 2022 Dec 30;12(1):303. doi: 10.3390/jcm12010303

Diagnostic Accuracy of the Artificial Intelligence Methods in Medical Imaging for Pulmonary Tuberculosis: A Systematic Review and Meta-Analysis

Yuejuan Zhan 1, Yuqi Wang 1, Wendi Zhang 1, Binwu Ying 2,*, Chengdi Wang 1,*
Editor: Sukhwinder Singh Sohal
PMCID: PMC9820940  PMID: 36615102

Abstract

Tuberculosis (TB) remains one of the leading causes of death among infectious diseases worldwide. Early screening and diagnosis of pulmonary tuberculosis (PTB) is crucial in TB control, and tend to benefit from artificial intelligence. Here, we aimed to evaluate the diagnostic efficacy of a variety of artificial intelligence methods in medical imaging for PTB. We searched MEDLINE and Embase with the OVID platform to identify trials published update to November 2022 that evaluated the effectiveness of artificial-intelligence-based software in medical imaging of patients with PTB. After data extraction, the quality of studies was assessed using quality assessment of diagnostic accuracy studies 2 (QUADAS-2). Pooled sensitivity and specificity were estimated using a bivariate random-effects model. In total, 3987 references were initially identified and 61 studies were finally included, covering a wide range of 124,959 individuals. The pooled sensitivity and the specificity were 91% (95% confidence interval (CI), 89–93%) and 65% (54–75%), respectively, in clinical trials, and 94% (89–96%) and 95% (91–97%), respectively, in model-development studies. These findings have demonstrated that artificial-intelligence-based software could serve as an accurate tool to diagnose PTB in medical imaging. However, standardized reporting guidance regarding AI-specific trials and multicenter clinical trials is urgently needed to truly transform this cutting-edge technology into clinical practice.

Keywords: pulmonary tuberculosis, artificial intelligence, medical imaging, diagnostic accuracy, sensitivity, specificity

1. Introduction

Tuberculosis (TB) is one of the major communicable diseases that seriously endanger human health primarily in developing countries [1], and at least 5.8 million people were estimated to have contracted tuberculosis in 2020. However, around one-sixth of people with active tuberculosis are left undetected or not officially reported each year, which may delay the progress of elimination of this disease before 2035 [2]. Timely diagnosis and treatment could benefit a wide range of tuberculosis patients and minimize the transmission of pathogen in the whole population.

Mycobacterium tuberculosis culture on solid and/or liquid media is still the golden standard for diagnosis. However, the efficiency of culture-based diagnosis in clinical practice is diminished due to long turnaround times and lack of laboratory infrastructure, especially in resource-limited countries. To solve this, the Xpert MTB/RIF assay has emerged as a maturely implemented tool in many countries haunted greatly by TB disease, which is a semiautomated rapid molecular method allowing for rapid diagnosis based upon detection of Mycobacterium tuberculosis DNA and resistance to rifampicin [3], but the application of such rapid tests remains far too limited, with only 1.9 million (33%) people having taken it as an initial diagnostic test in 2022. Simultaneously, the World Health Organization (WHO) has recommended using chest X-ray (CXR) images as a screening technique to better target individuals needing a further microbiological test, which has been proved to be relatively easy to operate, low-cost and highly sensitive [4]. However, an accurate diagnosis with CXRs extremely depends on the clinical experience of radiologists, which poses a huge challenge in the aforementioned countries. As such, there has been increasing interest in using artificial-intelligence-based (AI-based) software in medical imaging for pulmonary tuberculosis (PTB) detection, achieving diagnostic accuracy improvement and cost reduction at the same time. Currently, more than 40 AI-based software programs certified for CXR or computed tomography (CT) examination are available, among which only five are certified for CXR tuberculosis detection [5]. In 2021, Creswell and colleagues conducted a study that tested the five certified software programs (CAD4TB (v6), InferRead®DR (v2), Lunit INSIGHT CXR (v4.9.0), JF CXR-1 (v2), and qXR (v3)) with cohorts in Bangladesh and found that AI-based software significantly outperformed radiologists in TB detection [6]. However, poor reporting and wide variations in design and methodology limit the reliable interpretation of reported diagnostic accuracy [7]. Furthermore, systematic reviews [8,9] of the diagnostic accuracy of this software also identified several limitations in the available evidence, and uncertainty remains regarding its performance in PTB diagnosis.

Hence, we conducted a systematic review and meta-analysis to synthesize evidence of the accuracy of AI-based software in medical imaging for PTB and to provide new insights for future research.

2. Materials and Methods

2.1. Data Source and Search Strategy

A MEDLINE and Embase search through the OVID platform was performed on update to November 2022 without any restriction of country. The search terms were built as follows: ‘artificial intelligence’ (deep learning, machine learning, computer assisted, or cnn), ‘imaging’ (radiography, computed tomography, CT, photograph, or X-ray), ‘diagnostic accuracy metrics’ (sensitivity or specificity), and ‘pulmonary tuberculosis’ (Tuberculosis or tb). The full search strategy is laid out in Supplementary Materials File S2. This systematic review was registered in PROSPERO with the number CRD42022379114 and followed the preferred reporting items for systematic reviews and meta-analyses (PRISMA) guidelines (Supplementary Materials File S3).

2.2. Study Selection

Two researchers independently assessed the candidate studies for inclusion via screening of titles and abstracts, followed by the full text. Any discrepancy between the two researchers was resolved by a third researcher to achieve a consensus. We included all published studies that used AI-based software to analyze medical imaging in PTB diagnosis. Studies that met the following criteria were included in the final group: (1) Any study that analyzed medical imaging for PTB diagnosis with AI-based software; (2) Studies that provided raw diagnostic accuracy data, sensitivity, specificity, area under curve (AUC), accuracy, negative predictive values (NPVs), or positive predictive values (PPVs). Studies were excluded when they met the following criteria: (1) Case reports, conference reports, reviews, meta-analyses, abstracts without full articles, commentaries/editorials, mathematical modeling studies, and economic analyses; (2) Studies that investigated the accuracy of image segmentation or disease prediction; (3) Triage studies; (4) Studies without outcomes or separate data; (5) Studies that failed to report the source of the included population.

2.3. Data Extraction

Two researchers independently extracted demographic and diagnostic-accuracy data using a standardized extraction form from the included studies. When disagreements could not be resolved, we consulted with a third researcher. We extracted data that included study characteristics (first author name, country, year, study design, patient selection methods), demographic information (gender, age, human immunodeficiency virus (HIV) status, drug resistance, history of TB, treatment, imaging modality), AI-based software descriptions (type of artificial intelligence, model, data set, validation methods, threshold score), reference standards, and diagnostic accuracy measures (true and false positives and negatives (TP, FP, FN, TN), AUC, accuracy, sensitivity, specificity, PPV, NPV, and other reported metrics). If there were more than one reported accuracy data set for the same software, with other conditions consistent except for the threshold, the data set with the highest summed sensitivity and specificity would be extracted.

2.4. Quality Assessment

The risk of bias and applicability concerns of the included studies were assessed by two researchers separately, with a revised tool developed for diagnostic studies: QUADAS-2. Any disagreement between the two researchers was resolved through discussion with a third researcher.

2.5. Data Synthesis and Analysis

Data from development studies and clinical studies were analyzed separately. We first obtained the accuracy data that corresponded to TP, FP, FN, and TN in each included study and calculated the estimated pooled sensitivity, specificity, and AUC associated with the 95% CI, using bivariate random-effects models. Additionally, forest plots of sensitivity and specificity were generated for each study. We also used the model to create a summary receiver operating characteristic (SROC) curve. The I2 index was used to assess the heterogeneity between the studies. Values greater than 50% were indicative of substantial heterogeneity [10]. We subsequently chose different study designs, software, reference standards, and AI types as potential sources of heterogeneity, using subgroup analyses to explore the results. A sensitivity analysis was also performed to assess the robustness of the results and identify possible sources of heterogeneity. According to the PRISMA-DTA statement, neither a systematic review nor a meta-analysis of diagnostic accuracy studies is required to assess publication bias [11]. Analyses were conducted in Review Manager version 5.7 and Stata version 17.0 (Stata Corp., College Station, TX, USA), with the midas and metaninf command packages.

3. Results

3.1. Identification of Studies and Study Characteristics

A total of 3987 articles were identified, of which 404 duplicates and 2628 articles were excluded based on screening of titles and abstracts initially. We then excluded 894 studies upon viewing the full-text articles. Finally, 61 studies (23 clinical and 38 development studies) were included in our descriptive analysis (Table 1) [6,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33,34,35,36,37,38,39,40,41,42,43,44,45,46,47,48,49,50,51,52,53,54,55,56,57,58,59,60,61,62,63,64,65,66,67,68,69,70,71]. Due to missing information about the raw diagnostic data from the development studies, we only included 13 development studies, with 18 test evaluation results, in the quantitative analysis (Figure 1) [38,42,43,47,51,54,63,64,66,67,68,69,70].

Table 1.

Methods of studies included in descriptive analysis.

First Author, Year Imaging Modality Computer Software/Model Reference Standard Accuracy Measures
Maduskar et al., 2013 [12] CXR CAD4TB (v 1.08) AFB smear, MTB culture TP, FP, TN, FN, AUC, ACC, Sn, Sp, PPV, NPV
Muyoyeta et al., 2014 [13] CXR CAD4TB (v 1.08) Xpert MTB/RIF, human reader TP, FP, TN, FN, AUC, ACC, Sn, Sp, PPV, NPV
Steiner et al., 2015 [14] CXR CAD4TB (v 3.07) Human reader AUC
Melendez et al., 2018 [15] CXR CAD4TB (v 5) Human reader TP, FP, TN, FN, AUC, ACC, Sn, Sp, PPV, NPV
Zaidi et al., 2018 [16] CXR CAD4TB (v 3.07) Xpert MTB/RIF TP, FP, TN, FN, AUC, ACC, Sn, Sp, PPV, NPV
Qin et al., 2019 [17] CXR CAD4TB (v 6),
qXR (v 2),
Lunit INSIGHT CXR (v 4.7.2)
Xpert MTB/RIF TP, FP, TN, FN, AUC, ACC, Sn, Sp
Philipsen et al., 2019 [18] CXR CAD4TB (v 5) Xpert MTB/RIF, human reader TP, FP, TN, FN, AUC, ACC, Sn, Sp, PPV, NPV
Murphy et al., 2020 [19] CXR CAD4TB (v 6) Xpert MTB/RIF TP, FP, TN, FN, AUC, Sn, Sp
Nash et al., 2020 [20] CXR qXR (v 2) AFB smear, Xpert MTB/RIF or MTB culture AUC, Sn, Sp
Soares et al., 2023 [21] CXR CAD4TB (v 6),
Lunit INSIGHT CXR (v 3.1.0.0),
qXR (v 3)
Xpert MTB/RIF, MTB culture AUC, Sn, Sp, PPV, NPV
Qin et al., 2021 [6] CXR CAD4TB (v 7),
InferRead DR (v 2),
Lunit INSIGHT CXR (v 4.9.0),
JF CXR-1 (v 2),
qXR, (v 3)
Xpert MTB/RIF AUC, Sn, Sp
Breuninger et al., 2014 [22] CXR CAD4TB (v 3.07) AFB smear, MTB culture Sn, Sp, PPV, NPV
Khan et al., 2020 [23] CXR qXR (v 2),
CAD4TB (v 6)
MTB culture ACC, Sn, Sp, PPV, NPV
Young et al., 2020 [24] CXR Not named Human reader AUC, Sn, Sp
Liao et al., 2022 [25] CXR JF CXR-1 (v 2) Human reader TP, FP, TN, FN, AUC, ACC, Sn, Sp, PPV, NPV
Codlin et al., 2021 [26] CXR qXR (v 3),
CAD4TB (v 7),
Genki (v 2),
Lunit INSIGHT CXR (v 3.1.0.0),
JF CXR-1 (v 3.0),
InferRead DR Chest (v 1.0.0.0),
ChestEye (v 1),
T-Xnet (v 1),
XrayAME (v 1),
COTO (v 1),
SemanticMD (v 1),
Dr CADx (v 0.1)
Xpert MTB/RIF TP, FP, TN, FN, AUC, ACC, Sn, Sp, PPV, NPV
Habib et al., 2020 [27] CXR CAD4TB (v 3.07) Xpert MTB/RIF AUC, Sn, Sp, PPV, NPV
Koesoemadinata et al., 2018 [28] CXR CAD4TB (v 5) Composite reference standard(s) AUC, Sn, Sp
Lee et al., 2020 [29] CXR Lunit INSIGHT CXR (v 4.7.2) MTB culture, AFB smear, TB polymerase chain reaction, human reader TP, FP, TN, FN, AUC, ACC, Sn, Sp, PPV, NPV
Gelaw et al., 2022 [30] CXR CAD4TB (v 6),
Lunit INSIGHT CXR (v 4.9.0),
qXR (v 2)
Xpert MTB/RIF, Mycobacterium tuberculosis (MTB) culture TP, FP, TN, FN, Sn, Sp
Ehrlich et al., 2022 [31] CXR CAD4TB (v 7) Human reader TP, FP, TN, FN, AUC, Sn, Sp
Kagujje et al., 2022 [32] CXR CAD4TB (v 7),
qXR (v 3)
Xpert MTB/RIF TP, FP, TN, FN, AUC, Sn, Sp
Tavaziva et al., 2022 [33] CXR Lunit INSIGHT CXR (v 4.9.0) Xpert MTB/RIF, Mycobacterium tuberculosis (MTB) culture TP, FP, TN, FN, AUC, ACC, Sn, Sp
Shen et al., 2010 [34] CXR Bayesian classifier Human reader ACC
Melendez et al., 2015 [35] CXR si-miSVM+PEDD Human reader AUC
Pasa et al., 2019 [36] CXR CNN Human reader AUC, ACC
Xie et al., 2020 [37] CXR RCNN Human reader AUC, ACC, Sn, Sp
Ma et al., 2020 [38] CT U-Net Sputum smear AUC, ACC, Sn, Sp, PPV, NPV
Rajpurkar et al., 2020 [39] CXR DenseNet Xpert MTB/RIF, MTB culture ACC, Sn, Sp
Oloko-Oba et al., 2021 [40] CXR EfficientNets Human reader AUC, ACC, Sn, Sp
Mamalakis et al., 2021 [41] CXR DenseNet-121,
ResNet-50
Human reader AUC, F1, precision, recall
Rajakumar et al., 2021 [42] CXR VGG16,
VGG19,
KNN
Human reader ACC, Sn, Sp, NPV
Sharma et al., 2021 [43] CXR Tree,
SVM,
Naïve Bayes
Composite reference standard(s) AUC, F1, CA, precision, recall
Wang et al., 2021 [44] CT 3D-ResNet AFB smear, MTB culture AUC, Sn, Sp, ACC, F1
Showkatian et al., 2022 [45] CXR ConvNet Human reader AUC, ACC, F1, precision, recall
Zhou et al., 2022 [46] CXR ResNet Human reader AUC, ACC, Sn, Sp, PPV, NPV
Rajaraman et al., 2021 [47] CXR ImageNet,
VGG-16
Human reader AUC, ACC, Sn, Sp, F1, precision
Yan et al., 2021 [48] CT SeNet-ResNet-18 Human reader ACC, precision, recall
Zhang et al., 2021 [49] CT CBIR-CSNN Composite reference standard(s) AUC, ACC
Arzhaeva et al., 2009 [50] CXR MVDB Human reader AUC
Jaeger et al., 2014 [51] CXR SVM Human reader AUC, ACC
Chauhan et al., 2014 [52] CXR SVM Human reader AUC, ACC, Sn, Sp, F1, precision
Hogeweg et al., 2015 [53] CXR RF50,
GB50,
LDA,
KNN13
MTB culture, human reader AUC
Lakhani et al., 2017 [54] CXR AlexNet,
GoogLeNet
Human reader AUC, ACC, Sn, Sp
Han et al., 2021 [55] CXR VGG16 Human reader AUC, Sn, Sp
An et al., 2022 [56] CXR E-TBNet (ResNet) Human reader ACC, Sn, Sp, NPV, ppv, F1
Lee et al., 2021 [57] CXR EfficientNet Xpert MTB/RIF, MTB culture, human reader AUC
Khatibi et al., 2021 [58] CXR CNN,
CCNSE
Human reader AUC, ACC
Kim et al., 2020 [59] CXR DCNN Human reader AUC, Sn, Sp, NPV, PPV, F1
Feng et al., 2020 [60] CT CNN Composite reference standard(s) AUC, ACC, Sn, Sp
Hwang et al., 2019 [61] CXR CNN Human reader AUC, Sn, Sp
Heo et al., 2019 [62] CXR I-CNN(VGG19),
D-CNN(VGG19)
Human reader AUC
Aguiar et al., 2016 [63] CXR MLP Human reader AUC, Sn, Sp, PPV, NPV
Faruk et al., 2021 [64] CT Xception,
InceptionV3,
InceptionResNetV2,
MobileNetV2
Human reader Sn, precision, recall, F1
Karki et al., 2021 [65] CXR InceptionV3,
Xception
Human reader AUC
Dasanayaka et al., 2021 [66] CXR VGG16,
InceptionV3,
Ensemble
Human reader ACC, Sn, Sp
Govindarajan et al., 2021 [67] CXR ELM,
OSELM
Human reader Sn, Sp, precision, F1
Acharya et al., 2022 [68] CXR ImageNet fine-tuned normalization-free networks Human reader Sn, Sp, AUC, ACC, precision, recall
Kadry et al., 2022 [69] CXR VGG16, Fine Tree Xpert MTB/RIF, Mycobacterium tuberculosis (MTB) culture, human reader Sn, Sp, ACC, NPV
Kazemzadeh et al., 2023 [70] CXR NR Human reader Sn, Sp, AUC
Margarat et al., 2022 [71] CXR DBN-AMBO Human reader Sp, ACC, precision, recall, NPV

Abbreviations: CXR, chest X-ray; CT, computed tomography; CAD, computer-aided detection; CNN, convolutional neural networks; RCNN, regions with CNN features; KNN, K-nearest neighbor; VGG, visual geometry group; SVM, support vector machine; HIV, human immunodeficiency virus; DLAD, deep-learning-based automatic detection; AFB, acid-fast bacilli; MTB, Mycobacterium tuberculosis; TP, true positive; FP, false positive; TN, true negative; FN, false negative; AUC, area under the receiver operating curve; ACC, accuracy; Sn, sensitivity; Sp, specificity; CA, cluster accuracy; DBN-AMBO, deep belief network with adaptive monarch butterfly optimization.

Figure 1.

Figure 1

Study flow diagram. Computer-aided detection (CAD).

A total of 50 trials, described in 23 clinical studies, were included in the review, with 124,959 people reporting the diagnostic accuracy of the software used for CXR. No study provided prespecified sample-size calculations. In total, twelve studies [6,13,14,19,21,23,25,26,27,28,32,33] used prospectively collected data, and nine studies [6,17,19,20,21,23,25,26,29] used deep-learning-based versions. Additionally, twelve studies [6,12,14,17,18,19,20,25,26,28,29,30] compared software performance with human readers. Reference standards varied greatly; six studies [14,15,24,25,29,31] compared diagnostic performance with human readers, and fourteen studies [6,12,16,17,19,20,21,22,23,26,27,30,32,33] used microbiological references, while three studies [13,18,28] used both. Notably, some studies evaluated the diagnostic accuracy of AI-based software for special populations. Two studies were conducted on diabetic populations [27,28] and one study included only people from prison [14]. Only fourteen studies [6,13,16,17,20,21,22,23,26,27,28,31,32,33] included their own study populations, and the rest were collected from other studies.

Within the model-development studies, thirty reported diagnostic accuracy for PTB identification with deep-learning-based algorithms, compared with eight studies [34,35,43,50,51,52,53,67] that used machine-learning models. Altogether, twenty-seven out of thirty-eight of the available studies were based on public data sets. Several data sets (Montgomery (NIH), Shenzhen (NIH), and Belarus) were analyzed in most studies, but dataset demographic details were not described in most of the studies. Only one article explicitly described the use of semiautomatic lesion delineation for training data. To validate model performance, nine studies [44,46,48,49,59,60,61,68,70] validated algorithms for external data, while the remaining only implemented internal validation. Considering the economics of practical use, thirty-two out of the thirty-eight studies used CXRs as a diagnostic tool, with CT remaining to be further developed. In addition, eleven studies [36,39,40,41,42,48,52,53,57,60,70] made all of the code used in their implementation freely available to the public. As an important step in the radiomic pipeline, feature extraction played a decisive role in the whole process. Hogeweg, L. et al. [53] combined the results of shape analysis, texture analysis, and focal lesion detection into one combined TB score.

3.2. Quality Assessment of Studies

The overall results of the methodological-quality assessment of the included clinical and development studies are summarized respectively in Figure 2 and Figure 3. For clinical studies, the main sources of bias included index tests, flow, and timing. Most development studies were classified as high-risk, particularly with deficiencies in their methods of patient selection, the reference standards used, and their index tests.

Figure 2.

Figure 2

Quality assessment (QUADAS) graph of clinical studies.

Figure 3.

Figure 3

Quality assessment (QUADAS) graph of development studies.

For the patient-selection domain, a high or unclear risk of bias was observed in 84% (thirty-two out of thirty-eight) of the development studies, which was mainly related to missing information in the CXR/CT databases. For the index test, a prespecified threshold was reported only in 30% (seven out of twenty-three) of the clinical studies, and 18% (seven out of thirty-eight) of the development studies had a prespecified threshold, while the other studies had a high risk of bias, since the threshold was determined after the analysis in each. For the reference standard domain, a high or unclear risk of bias was seen in 76% (twenty-nine out of thirty-eight) of the development studies, with regards to assessment by radiologists as the reference standard. For flow and timing, there was a high or unclear risk of bias in 39% (nine out of twenty-three) of the clinical studies and 50% (nineteen out of thirty-eight) of the development studies due to the inconsistency of the reference standards and a lack of inclusion of all patients.

3.3. Diagnostic Accuracy Reported in AI-Based Software Assay for PTB

We found that only 13 development studies reported TP, FP, FN, and TN for index tests. Of all the 38 articles that included accuracy assessments, the sensitivity ranged from 0.580 to 0.993 and the specificity from 0.570 to 0.996. It is worth noting that CT showed a higher sensitivity in diagnosis with AI (0.750–0.993 of CT vs. 0.580–0.993 of CXR). The reported performance is summarized in Figure 4. The pooled sensitivity of all included studies was 94% (95% CI 89–96%), with I2 = 93.22 (95% CI 91.07–95.37), and the pooled specificity was 95% (95% CI 91–97%), with I2 = 97.52 (95% CI 96.94–98.09). After excluding the CT-based study, we obtained pooled sensitivity and specificity values of 93% (95% CI 87–96%) and 94% (95% CI 90–97%), respectively.

Figure 4.

Figure 4

Forest plot of development-study sensitivity and specificity for PTB [38,42,43,47,51,54,63,64,66,67,68,69,70].

In total, 23 clinical studies, including 124,959 patients, evaluated the diagnostic efficacy of AI programs for PTB. The sensitivity ranged from 0.487 to 1.00, and the pooled sensitivity was 91% (95% CI 89–93%), with I2 = 93.05 (95% CI 91.74–94.36). The specificity ranged from 0.063 to 0.997, and the pooled specificity was 65% (95% CI 54–75%), with I2 = 99.87 (95% CI 99.86–99.88) (Figure 5).

Figure 5.

Figure 5

Forest plot of clinical-study sensitivity and specificity for PTB [6,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33].

There was significant heterogeneity in both sensitivity and specificity. We also constructed SROC curves and calculated the AUC for the included studies. The overall diagnostic performance of the clinical studies and the development studies was comparable [AUC 0.91 (95% CI 0.89–0.94) and 0.98 (95% CI 0.97–0.99), respectively] (Supplementary Materials File S1).

3.4. Subgroup and Sensitivity Analyses

Considering the variability of the methods and models tested in the development studies, we only performed a subgroup analysis in the clinical studies, based on predefined parameters, including study design, software, reference standard, and AI type. Some studies were excluded from the relevant subgroup analyses due to missing information or not being categorized into specific groups.

Compared to different study designs, the pooled specificity was 48% (95% CI 34–62%, I2 = 99.87; 99.86–99.88) in the prospective assay versus 75% (95% CI 53–89%, I2 = 99.94; 99.93–99.94) in the nonprospective assay. When Xpert MTB/RIF was used as the reference standard, the pooled specificity of the Xpert MTB/RIF assay [36% (95% CI 24–50%, I2 = 99.93; 99.93–99.94)] was much lower than that of the studies that used human readers [90% (95% CI 80–95%, I2 = 98.70; 98.32–99.08)]. Furthermore, the sensitivity and the specificity of various AI-based software (CAD4TB, qXR, Lunit INSIGHT CXR) evidently differed. The results of the subgroup analyses are summarized in detail in Table 2. There was still a substantial level of heterogeneity among each subgroup analysis.

Table 2.

Subgroup analysis based on different standards.

Studies Sensitivity (95%CI) Specificity (95%CI) DOR (95%CI) AUC (95%CI)
All (23) 0.91(0.89–0.93) 0.65(0.55–0.75) 20(13–29) 0.91(0.89–0.94)
Study Design
Prospective (12) 0.91(0.87–0.94) 0.48(0.34–0.62) 9(4–20) 0.85(0.82–0.88)
Nonprospective (11) 0.87(0.78–0.93) 0.75(0.53–0.89) 20(5–84) 0.90(0.87–0.92)
Software
CAD4TB (18) 0.89(0.82–0.94) 0.57(0.42–0.70) 11(4–30) 0.83(0.80–0.86)
qXR (8) 0.79(0.61–0.90) 0.55(0.24–0.83) 5(1–38) 0.77(0.73–0.80)
Lunit INSIGHT CXR (8) 0.88(0.75–0.94) 0.78(0.40–0.95) 25(3–211) 0.91(0.88–0.93)
Reference standard
Human reader (5) 0.90(0.84–0.94) 0.90(0.80–0.95) 77(22–269) 0.95(0.93–0.97)
Xpert MTB/RIF (9) 0.90(0.85–0.93) 0.36(0.24–0.50) 5(2–12) 0.79(0.75–0.82)
AI type
Deep learning (13) 0.91(0.89–0.92) 0.62(0.48–0.74) 16(10–23) 0.91(0.88–0.93)
Machine learning (9) 0.93(0.85–0.97) 0.61(0.46–0.75) 21(11–42) 0.87(0.83–0.89)

Abbreviation: DOR, diagnostic odds ratio; AUC, area under curve.

We subsequently performed sensitivity analyses on the clinical and development studies, respectively. Results of our sensitivity analyses are provided in Supplementary Materials File S1. In the clinical studies, we found three articles that had great effects on the overall results. After removal of the corresponding articles, we obtained a still-high heterogeneity (I2 = 92.97, 91.55–94.39 for sensitivity, I2 = 99.83, 99.82–99.84 for specificity).

4. Discussion

This study sought to (1) evaluate the diagnostic efficacy of AI-based software for PTB and (2) describe the study characteristics, and evaluate the study methodology and the quality of reporting of AI-based software for PTB diagnosis, as well as providing some advice for future software development and clinical applications. Meta-analysis demonstrated that AI-based software has high accuracy in both clinical applications and development studies, indicating that it can assist the physicians in improving the accuracy of PTB diagnosis. However, due to the high heterogeneity and variability between studies, relevant results must be treated with caution when the result of AI-based software is used as a reference standard.

In this systematic review and meta-analysis, we included 23 clinical studies and 38 development studies of PTB diagnosis. Since some missing data were reported, the final count was 13 development studies and 23 clinical studies eligible for quantitative synthesis. Our results show that AI-based software has an excellent ability to diagnose PTB in medical imaging, with pooled sensitivities greater than 0.9 [clinical studies: 91% (95% CI 89%–93%); development studies: 94% (95% CI 89%–96%)]. Additionally, the pooled specificity of the software in the clinical studies was only modest [65% (95% CI 54%–75%)], while that in the development studies was relatively high [95% (95% CI 91%–97%)], which may have been caused by the application of the same test-data set for diagnostic performance assessment. However, a high level of heterogeneity was observed in all the results. Subgroup analysis revealed that nonprospective studies had significantly higher specificity and lower sensitivity than prospective studies had, which might have been due to the inclusion of identified PTB patients in the nonprospective studies. Additionally, studies that used Xpert MTB/RIF as a reference standard had much lower specificity compared to studies that used human readers, possibly because human readers were weaker than Xpert MTB/RIF in correctly identifying negative patients. Furthermore, all commercially available software (CAD4TB, Lunit INSIGHT CXR, and qXR) showed its advantages in improvement of diagnostic accuracy, but we found evident differences in sensitivity and specificity among various AI-based software. The level of heterogeneity between the subgroups remained high, suggesting that study design, software type, AI type, and different reference standards might not be source of heterogeneity. Our follow-up sensitivity analysis indicated that different types of medical imaging might be the sources of heterogeneity, as CT could offer enhanced sensitivity [72].

A number of methodological limitations in the existing evidence were identified, as were study-level factors associated with the reported accuracy, which should all be taken into consideration.

In development studies, most of the current AI-based software was developed for CXR, and only six studies were applied to CT. Because of the deficiency of accuracy data, we performed no subgroup analysis for CT versus CXR. In addition, specific accuracy results, threshold establishment, and inclusion criteria may not have been described well enough to allow emulation for further comparison and may cause greater clinical and methodological heterogeneity. A large proportion of the articles used human readers as the reference standard, meaning systematic overestimation of the diagnostic accuracy of the software. Furthermore, the lack of external validation made it very difficult to formally evaluate algorithm performance. Although most of the experiments used publicly available data sets for model training, few experiments fully disclosed their model details and codes. In addition, almost all of the development articles used manual-lesion-depiction data sets. Semiautomated approaches are known to have greater advantages in lesion delineation, as has been demonstrated with other lung diseases [73], so we encourage more studies in the future to adopt this approach. Several aspects mentioned above lead to the inability to guarantee reproducibility of these experiments. Much of the existing work focuses on multiparametric classification models, ignoring the influence of individual features. Accumulating evidence has confirmed the important role of individual features in discrimination of benign and malignant lung lesions [74,75]; this has great potential for improvement of accuracy and disease identification, and is also informative for research of automated classification models for PTB.

All of the clinical studies evaluated commercially available software developed for CXR. A total of 11 software types were tested, but the version and threshold reported varied among studies. There were varying methodologies of threshold determination and population inclusion, potentially resulting in a high level of heterogeneity. It is worth noting that 13 articles also compared the diagnostic accuracy of AI-based software with human clinicians, which would provide a more objective criterion allowing for a better comparison of models between studies.

Our study had several limitations. Although we searched the relevant literature as comprehensively as possible, some of the literature might have been missed. In addition, some studies failed to report demographic information in detail, and the corresponding subgroup analysis could not be performed. Furthermore, the limited number of studies included for different versions of the software allowed for no further analysis. When AI-based software was used to diagnose PTB, there was significant heterogeneity among studies, so it is difficult to determine whether the software is clinically applicable. Lastly, because current clinical software requires the inclusion of patients over 15 years of age, the diagnostic efficiency for children needs to be further determined.

To improve the future clinical applicability of AI-based software, we recommend that studies include detailed reporting of demographic information, and hope that existing reporting guidelines for diagnostic accuracy studies (STARD) [76] and prediction models (TRIPOD) [77] can be improved as soon as possible to conduct AI-specific amendments. In addition, some model training and validations were performed on CXRs from data sets or sites, potentially resulting in an overestimation of diagnosis power. As such, we suggest that different data sets should be used for model training and testing. Moreover, research teams can collaborate with multiple clinical centers for clinical trials and external validation to make results superior and investigate the stability and heterogeneity of their performance in clinical scenarios. What is more, we appealed to a large number of open, multi-source, and anonymous databases, along with detailed reporting of all of the information needed, such as reference standard, age, HIV status, etc., to fulfill the need for an adequate amount of data with high quality. At the same time, we recommend that development studies make their model details and all of the code used for their experiments freely available to the public to make it possible to reproduce these studies. It is also noteworthy that the diagnostic accuracy of AI-based software should be evaluated against a microbiological reference standard. Lastly, we found a lack of use of AI-based software in CT, and more studies may be needed to explore its superiority in early diagnosis of PTB. In addition, the influence of parameters such as intensity quantization, on imaging and final diagnosis in particular, could be considered.

5. Conclusions

In summary, there were relatively high pooled sensitivity and specificity values of AI-based software, which indicates that AI-based software has potential to facilitate diagnosis of PTB in medical imaging, especially in large-scale screening. Heterogeneity was significantly high and extensive variation in reporting, design, and methodology was observed. Thus, standardized reporting guidance around AI-specific trials and multicenter clinical trials is urgently needed to further confirm their stability and heterogeneity in various populations and settings. In the future, we expect more AI-based software with high accuracy to be comprehensively applied for early clinical detection of PTB.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/jcm12010303/s1, Supplementary Materials File S1: Figure S1 Quality assessment (QUADAS 2) summary of clinical studies: risk of bias & applicability concerns; Figure S2 Quality assessment (QUADAS 2) summary of development studies: Risk of bias and applicability concerns; Figure S3 Sensitivity analysis of clinical studies; Figure S4 Sensitivity analysis of development studies; Figure S5 Summary receiver operating characteristic (SOC) curve of clinical studies; Figure S6 Summary receiver operating characteristic (SROC) curve of development studies; Table S1 Demographics of clinical studies; Table S2 Accuracy information of clinical studies included; Table S3 Accuracy measures reported by development studies; Table S4 Accuracy measures reported by development studies; Supplementary Materials File S2: Search strategy; Supplementary Materials File S3: PRISMA checklist [78].

Author Contributions

Conceptualization and design, C.W. and B.Y.; data curation and data analysis, Y.Z., Y.W. and W.Z.; manuscript editing and manuscript review, C.W., B.Y., Y.Z., Y.W. and W.Z. All authors have read and agreed to the published version of the manuscript.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Funding Statement

This research was supported by the National Natural Science Foundation of China (82100119), Chinese Postdoctoral Science Foundation (2022T150451, 2021M692309), the Science and Technology Project of Sichuan (2020YFG0473, 2022ZDZX0018), National College Students’ innovation and entrepreneurship training program of Sichuan University (20231261L).

Footnotes

Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

References

  • 1.Pai M., Behr M.A., Dowdy D., Dheda K., Divangahi M., Boehme C.C., Ginsberg A., Swaminathan S., Spigelman M., Getahun H., et al. Tuberculosis. Nat. Rev. Dis. Prim. 2016;2:16076. doi: 10.1038/nrdp.2016.76. [DOI] [PubMed] [Google Scholar]
  • 2.World Health Organization Global Tuberculosis Report 2021. [(accessed on 10 November 2022)]. Available online: https://www.who.int/teams/global-tuberculosis-programme/tb-reports.
  • 3.Chen X., Hu T.Y. Strategies for advanced personalized tuberculosis diagnosis: Current technologies and clinical approaches. Precis. Clin.Med. 2021;2:35–44. doi: 10.1093/pcmedi/pbaa041. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Hoog A.H.V., Meme H.K., Laserson K.F., Agaya J.A., Muchiri B.G., Githui W.A., Odeny L.O., Marston B.J., Borgdorff M.W. Screening strategies for tuberculosis prevalence surveys: The value of chest radiography and symptoms. PLoS ONE. 2012;7:e38691. doi: 10.1371/journal.pone.0038691. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Diagnostic Image Analysis Group AI for radiology: An implementation guide 2020. [(accessed on 10 November 2022)]. Available online: https://grand-challenge.org/aiforradiology/
  • 6.Qin Z.Z., Ahmed S., Sarker M.S., Paul K., Adel A.S.S., Naheyan T., Barrett R., Banu S., Creswell J. Tuberculosis detection from chest X-rays for triaging in a high tuberculosis-burden setting: An evaluation of five artificial intelligence algorithms. Lancet Digit. Health. 2021;3:e543–e554. doi: 10.1016/S2589-7500(21)00116-3. [DOI] [PubMed] [Google Scholar]
  • 7.Liu X., Faes L., Kale A.U., Wagner S.K., Fu D.J., Bruynseels A., Mahendiran T., Moraes G., Shamdas M., Kern C., et al. A comparison of deep learning performance against health-care professionals in detecting diseases from medical imaging: A systematic review and meta-analysis. Lancet Digit. Health. 2019;1:e271–e297. doi: 10.1016/S2589-7500(19)30123-2. [DOI] [PubMed] [Google Scholar]
  • 8.Harris M., Qi A., JeaGal L., Torabi N., Menzies D., Korobitsyn A., Pai M., Nathavitharana R.R., Khan F.A. A systematic review of the diagnostic accuracy of artificial intelligence-based computer programs to analyze chest X-rays for pulmonary tuberculosis. PLoS ONE. 2019;14:e0221339. doi: 10.1371/journal.pone.0221339. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Pande T., Cohen C., Pai M., Khan F.A. Computer-aided detection of pulmonary tuberculosis on digital chest radiographs: A systematic review. Int. J. Tuberc. Lung Dis. 2016;20:1226–1230. doi: 10.5588/ijtld.15.0926. [DOI] [PubMed] [Google Scholar]
  • 10.Puhan M.A., Gimeno-Santos E., Cates C.J., Troosters T. Pulmonary rehabilitation following exacerbations of chronic obstructive pulmonary disease. Cochrane Database Syst. Rev. 2016;2019:CD005305. doi: 10.1002/14651858.CD005305.pub4. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.McInnes M.D.F., Moher D., Thombs B.D., McGrath T.A., Bossuyt P.M., Clifford T., Cohen J.F., Deeks J.J., Gatsonis C., Hooft L., et al. Preferred reporting items for a systematic review and meta-analysis of diagnostic test accuracy studies. The PRISMA-DTA Statement. JAMA. 2018;319:388–396. doi: 10.1001/jama.2017.19163. [DOI] [PubMed] [Google Scholar]
  • 12.Maduskar P., Muyoyeta M., Ayles H., Hogeweg L., Peters-Bax L., Van Ginneken B. Detection of tuberculosis using digital chest radiography: Automated reading vs. interpretation by clinical officers. Int. J. Tuberc. Lung Dis. 2013;17:1613–1620. doi: 10.5588/ijtld.13.0325. [DOI] [PubMed] [Google Scholar]
  • 13.Muyoyeta M., Maduskar P., Moyo M., Kasese N., Milimo D., Spooner R., Kapata N., Hogeweg L., Van Ginneken B., Ayles H. The sensitivity and specificity of using a computer aided diagnosis program for automatically scoring chest X-rays of presumptive TB patients compared with Xpert MTB/RIF in Lusaka Zambia. PLoS ONE. 2014;9:e93757. doi: 10.1371/journal.pone.0093757. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Steiner A., Mangu C., Hombergh J.V.D., van Deutekom H., van Ginneken B., Clowes P., Mhimbira F., Mfinanga S., Rachow A., Reither K., et al. Screening for pulmonary tuberculosis in a Tanzanian prison and computer-aided interpretation of chest X-rays. Public Health Action. 2015;5:249–254. doi: 10.5588/pha.15.0037. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Melendez J., Hogeweg L., Sánchez C.I., Philipsen R.H.H.M., Aldridge R., Hayward A.C., Abubakar I., van Ginneken B., Story A. Accuracy of an automated system for tuberculosis detection on chest radiographs in high-risk screening. Int. J. Tuberc. Lung Dis. 2018;22:567–571. doi: 10.5588/ijtld.17.0492. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Zaidi S.M.A., Habib S.S., Van Ginneken B., Ferrand R.A., Creswell J., Khowaja S., Khan A. Evaluation of the diagnostic accuracy of computer-aided detection of tuberculosis on chest radiography among private sector patients in Pakistan. Sci. Rep. 2018;8:12339. doi: 10.1038/s41598-018-30810-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Qin Z.Z., Sander M.S., Rai B., Titahong C.N., Sudrungrot S., Laah S.N., Adhikari L.M., Carter E.J., Puri L., Codlin A.J., et al. Using artificial intelligence to read chest radiographs for tuberculosis detection: A multi-site evaluation of the diagnostic accuracy of three deep learning systems. Sci. Rep. 2019;9:15000. doi: 10.1038/s41598-019-51503-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Philipsen R.H.H.M., Sánchez C.I., Melendez J., Lew W.J., Van Ginneken B. Automated chest X-ray reading for tuberculosis in the Philippines to improve case detection: A cohort study. Int. J. Tuberc. Lung Dis. 2019;23:805–810. doi: 10.5588/ijtld.18.0004. [DOI] [PubMed] [Google Scholar]
  • 19.Murphy K., Habib S.S., Zaidi S.M.A., Khowaja S., Khan A., Melendez J., Scholten E.T., Amad F., Schalekamp S., Verhagen M., et al. Computer aided detection of tuberculosis on chest radiographs: An evaluation of the CAD4TB v6 system. Sci. Rep. 2020;10:5492. doi: 10.1038/s41598-020-62148-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Nash M., Kadavigere R., Andrade J., Sukumar C.A., Chawla K., Shenoy V.P., Pande T., Huddart S., Pai M., Saravu K. Deep learning, computer-aided radiography reading for tuberculosis: A diagnostic accuracy study from a tertiary hospital in India. Sci. Rep. 2020;10:210. doi: 10.1038/s41598-019-56589-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Soares T.R., de Oliveira R.D., Liu Y.E., Santos A.D.S., dos Santos P.C.P., Monte L.R.S., de Oliveira L.M., Park C.M., Hwang E.J., Andrews J.R., et al. Evaluation of chest X-ray with automated interpretation algorithms for mass tuberculosis screening in prisons: A cross-sectional study. Lancet Reg. Health-Am. 2023;17:100388. doi: 10.1016/j.lana.2022.100388. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Breuninger M., Van Ginneken B., Philipsen R.H.H.M., Mhimbira F., Hella J.J., Lwilla F., Hombergh J.V.D., Ross A., Jugheli L., Wagner D., et al. Diagnostic accuracy of computer-aided detection of pulmonary tuberculosis in chest radiographs: A validation study from Sub-Saharan Africa. PLoS ONE. 2014;9:e106381. doi: 10.1371/journal.pone.0106381. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Khan F.A., Majidulla A., Tavaziva G., Nazish A., Abidi S.K., Benedetti A., Menzies D., Johnston J.C., Khan A.J., Saeed S. Chest X-ray analysis with deep learning-based software as a triage test for pulmonary tuberculosis: A prospective study of diagnostic accuracy for culture-confirmed disease. Lancet Digit. Health. 2020;2:e573–e581. doi: 10.1016/S2589-7500(20)30221-1. [DOI] [PubMed] [Google Scholar]
  • 24.Young C., Barker S., Ehrlich R., Kistnasamy B., Yassi A. Computer-aided detection for tuberculosis and silicosis in chest radiographs of gold miners of South Africa. Int. J. Tuberc. Lung Dis. 2020;24:444–451. doi: 10.5588/ijtld.19.0624. [DOI] [PubMed] [Google Scholar]
  • 25.Liao Q., Feng H., Li Y., Lai X., Pan J., Zhou F., Zhou L., Chen L. Evaluation of an artificial intelligence (AI) system to detect tuberculosis on chest X-ray at a pilot active screening project in Guangdong, China in 2019. J. X-ray Sci. Technol. 2022;30:221–230. doi: 10.3233/XST-211019. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Codlin A.J., Dao T.P., Vo L.N.Q., Forse R.J., Van Truong V., Dang H.M., Nguyen L.H., Nguyen H.B., Nguyen N.V., Sidney-Annerstedt K., et al. Independent evaluation of 12 artificial intelligence solutions for the detection of tuberculosis. Sci. Rep. 2021;11:23895. doi: 10.1038/s41598-021-03265-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Habib S.S., Rafiq S., Zaidi S.M.A., Ferrand R.A., Creswell J., Van Ginneken B., Jamal W.Z., Azeemi K.S., Khowaja S., Khan A. Evaluation of computer aided detection of tuberculosis on chest radiography among people with diabetes in Karachi Pakistan. Sci. Rep. 2020;10:6276. doi: 10.1038/s41598-020-63084-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Koesoemadinata R.C., Kranzer K., Livia R., Susilawati N., Annisa J., Soetedjo N.N.M., Ruslami R., Philipsen R., van Ginneken B., Soetikno R.D., et al. Computer-assisted chest radiography reading for tuberculosis screening in people living with diabetes mellitus. Int. J. Tuberc. Lung Dis. 2018;22:1088–1094. doi: 10.5588/ijtld.17.0827. [DOI] [PubMed] [Google Scholar]
  • 29.Lee J.H., Park S., Hwang E.J., Goo J.M., Lee W.Y., Lee S., Kim H., Andrews J.R., Park C.M. Deep learning–based automated detection algorithm for active pulmonary tuberculosis on chest radiographs: Diagnostic performance in systematic screening of asymptomatic individuals. Eur. Radiol. 2020;31:1069–1080. doi: 10.1007/s00330-020-07219-4. [DOI] [PubMed] [Google Scholar]
  • 30.Gelaw S.M., Kik S.V., Ruhwald M., Ongarello S., Egzertegegne T.S., Gorbacheva O., Gilpin C., Marano N., Lee S., Phares C.R. Diagnostic accuracy of three computer-aided detection systems for detecting pulmonary tuberculosis on chest radiography when used for screening: Analysis of an international, multicenter migrants screening study. medRxiv. 2022 doi: 10.1101/2022.03.30.22273191. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31.Ehrlich R., Barker S., Naude J.T.W., Rees D., Kistnasamy B., Naidoo J., Yassi A. Accuracy of computer-aided detection of occupational lung disease: Silicosis and pulmonary tuberculosis in Ex-Miners from the South African gold mines. Int. J. Environ. Res. Public Health. 2022;19:12402. doi: 10.3390/ijerph191912402. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Kagujje M., Kerkhoff A.D., Nteeni M., Dunn I., Mateyo K., Muyoyeta M. The performance of computer-aided detection digital chest X-ray reading technologies for triage of active tuberculosis among persons with a history of previous tuberculosis. Clin. Infect. Dis. 2022:ciac679. doi: 10.1093/cid/ciac679. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Tavaziva G., Majidulla A., Nazish A., Saeed S., Benedetti A., Khan A.J., Khan F.A. Diagnostic accuracy of a commercially available, deep learning-based chest X-ray interpretation software for detecting culture-confirmed pulmonary tuberculosis. Int. J. Infect. Dis. 2022;122:15–20. doi: 10.1016/j.ijid.2022.05.037. [DOI] [PubMed] [Google Scholar]
  • 34.Shen R., Cheng I., Basu A. A hybrid knowledge-guided detection technique for screening of infectious pulmonary tuberculosis from chest radiographs. IEEE Trans. Biomed. Eng. 2010;57:2646–2656. doi: 10.1109/TBME.2010.2057509. [DOI] [PubMed] [Google Scholar]
  • 35.Melendez J., van Ginneken B., Maduskar P., Philipsen R.H.H.M., Reither K., Breuninger M., Adetifa I.M.O., Maane R., Ayles H., Sanchez C.I. A novel multiple-instance learning-based approach to computer-aided detection of tuberculosis on chest X-rays. IEEE Trans. Med. Imaging. 2015;34:179–192. doi: 10.1109/TMI.2014.2350539. [DOI] [PubMed] [Google Scholar]
  • 36.Pasa F., Golkov V., Pfeiffer F., Cremers D. Efficient deep network architectures for fast chest X-ray tuberculosis screening and visualization. Sci. Rep. 2019;9:6268. doi: 10.1038/s41598-019-42557-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37.Xie Y., Wu Z., Han X., Wang H., Wu Y., Cui L., Feng J., Zhu Z., Chen Z. Computer-aided system for the detection of multicategory pulmonary tuberculosis in radiographs. J. Healthc. Eng. 2020;2020:9205082. doi: 10.1155/2020/9205082. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 38.Ma L., Wang Y., Guo L., Zhang Y., Wang P., Pei X., Qian L., Jaeger S., Ke X., Yin X., et al. Developing and verifying automatic detection of active pulmonary tuberculosis from multi-slice spiral CT images based on deep learning. J. X-ray Sci. Technol. 2020;28:939–951. doi: 10.3233/XST-200662. [DOI] [PubMed] [Google Scholar]
  • 39.Rajpurkar P., O’Connell C., Schechter A., Asnani N., Li J., Kiani A., Ball R.L., Mendelson M., Maartens G., van Hoving D.J., et al. CheXaid: Deep learning assistance for physician diagnosis of tuberculosis using chest X-rays in patients with HIV. npj Digit. Med. 2020;3:115. doi: 10.1038/s41746-020-00322-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40.Oloko-Oba M., Viriri S. Ensemble of EfficientNets for the diagnosis of tuberculosis. Comput. Intell. Neurosci. 2021;2021:9790894. doi: 10.1155/2021/9790894. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41.Mamalakis M., Swift A.J., Vorselaars B., Ray S., Weeks S., Ding W., Clayton R.H., Mackenzie L.S., Banerjee A. DenResCov-19: A deep transfer learning network for robust automatic classification of COVID-19, pneumonia, and tuberculosis from X-rays. Comput. Med. Imaging Graph. 2021;94:102008. doi: 10.1016/j.compmedimag.2021.102008. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42.Rajakumar M., Sonia R., Maheswari B.U., Karuppiah S. Tuberculosis detection in chest X-ray using Mayfly-algorithm optimized dual-deep-learning features. J. X-ray Sci. Technol. 2021;29:961–974. doi: 10.3233/XST-210976. [DOI] [PubMed] [Google Scholar]
  • 43.Sharma A., Sharma A., Malhotra R., Singh P., Chakrabortty R.K., Mahajan S., Pandit A.K. An accurate artificial intelligence system for the detection of pulmonary and extra pulmonary tuberculosis. Tuberculosis. 2021;131:102143. doi: 10.1016/j.tube.2021.102143. [DOI] [PubMed] [Google Scholar]
  • 44.Wang L., Ding W., Mo Y., Shi D., Zhang S., Zhong L., Wang K., Wang J., Huang C., Ye Z., et al. Distinguishing nontuberculous mycobacteria from Mycobacterium tuberculosis lung disease from CT images using a deep learning framework. Eur. J. Nucl. Med. 2021;48:4293–4306. doi: 10.1007/s00259-021-05432-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 45.Showkatian E., Salehi M., Ghaffari H., Reiazi R., Sadighi N. Deep learning-based automatic detection of tuberculosis disease in chest X-ray images. Pol. J. Radiol. 2022;87:118–124. doi: 10.5114/pjr.2022.113435. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 46.Zhou W., Cheng G., Zhang Z., Zhu L., Jaeger S., Lure F.Y.M., Guo L. Deep learning-based pulmonary tuberculosis automated detection on chest radiography: Large-scale independent testing. Quant. Imaging Med. Surg. 2022;12:2344. doi: 10.21037/qims-21-676. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 47.Rajaraman S., Zamzmi G., Folio L., Alderson P., Antani S. Chest X-ray bone suppression for improving classification of tuberculosis-consistent findings. Diagnostics. 2021;11:840. doi: 10.3390/diagnostics11050840. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 48.Yan C., Wang L., Lin J., Xu J., Zhang T., Qi J., Li X., Ni W., Wu G., Huang J., et al. A fully automatic artificial intelligence–based CT image analysis system for accurate detection, diagnosis, and quantitative severity evaluation of pulmonary tuberculosis. Eur. Radiol. 2021;32:2188–2199. doi: 10.1007/s00330-021-08365-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 49.Zhang K., Qi S., Cai J., Zhao D., Yu T., Yue Y., Yao Y., Qian W. Content-based image retrieval with a Convolutional Siamese Neural Network: Distinguishing lung cancer and tuberculosis in CT images. Comput. Biol. Med. 2021;140:105096. doi: 10.1016/j.compbiomed.2021.105096. [DOI] [PubMed] [Google Scholar]
  • 50.Arzhaeva Y., Hogeweg L., De Jong P.A., Viergever M.A., Van Ginneken B. Global and local multi-valued dissimilarity-based classification: Application to computer-aided detection of tuberculosis. Pt 2Med. Image Comput. Comput. Assist. Interv. 2009;12:724–731. doi: 10.1007/978-3-642-04271-3_88. [DOI] [PubMed] [Google Scholar]
  • 51.Jaeger S., Karargyris A., Candemir S., Folio L., Siegelman J., Callaghan F., Xue Z., Palaniappan K., Singh R.K., Antani S., et al. Automatic tuberculosis screening using chest radiographs. IEEE Trans. Med. Imaging. 2014;33:233–245. doi: 10.1109/TMI.2013.2284099. [DOI] [PubMed] [Google Scholar]
  • 52.Chauhan A., Chauhan D., Rout C. Role of Gist and PHOG features in computer-aided diagnosis of tuberculosis without segmentation. PLoS ONE. 2014;9:e112980. doi: 10.1371/journal.pone.0112980. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 53.Hogeweg L., Sanchez C.I., Maduskar P., Philipsen R., Story A., Dawson R., Theron G., Dheda K., Peters-Bax L., van Ginneken B. Automatic detection of tuberculosis in chest radiographs using a combination of textural, focal, and shape abnormality analysis. IEEE Trans. Med. Imaging. 2015;34:2429–2442. doi: 10.1109/TMI.2015.2405761. [DOI] [PubMed] [Google Scholar]
  • 54.Lakhani P., Sundaram B. Deep learning at chest radiography: Automated classification of pulmonary tuberculosis by using Convolutional Neural Networks. Radiology. 2017;284:574–582. doi: 10.1148/radiol.2017162326. [DOI] [PubMed] [Google Scholar]
  • 55.Han D., He T., Yu Y., Guo Y., Chen Y., Duan H., Yu N. Diagnosis of active pulmonary tuberculosis and community acquired pneumonia using Convolution Neural Network based on transfer learning. Acad. Radiol. 2022;29:1486–1492. doi: 10.1016/j.acra.2021.12.025. [DOI] [PubMed] [Google Scholar]
  • 56.An L., Peng K., Yang X., Huang P., Luo Y., Feng P., Wei B. E-TBNet: Light Deep Neural Network for automatic detection of tuberculosis with X-ray DR Imaging. Sensors. 2022;22:821. doi: 10.3390/s22030821. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 57.Lee S., Yim J.-J., Kwak N., Lee Y.J., Lee J.-K., Lee J.Y., Kim J.S., Kang Y.A., Jeon D., Jang M.-J., et al. Deep learning to determine the activity of pulmonary tuberculosis on chest radiographs. Radiology. 2021;301:435–442. doi: 10.1148/radiol.2021210063. [DOI] [PubMed] [Google Scholar]
  • 58.Khatibi T., Shahsavari A., Farahani A. Proposing a novel multi-instance learning model for tuberculosis recognition from chest X-ray images based on CNNs, complex networks and stacked ensemble. Phys. Eng. Sci. Med. 2021;44:291–311. doi: 10.1007/s13246-021-00980-w. [DOI] [PubMed] [Google Scholar]
  • 59.Kim T.K., Yi P.H., Hager G.D., Lin C.T. Refining dataset curation methods for deep learning-based automated tuberculosis screening. J. Thorac. Dis. 2020;12:5078–5085. doi: 10.21037/jtd.2019.08.34. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 60.Feng B., Chen X., Chen Y., Lu S., Liu K., Li K., Liu Z., Hao Y., Li Z., Zhu Z., et al. Solitary solid pulmonary nodules: A CT-based deep learning nomogram helps differentiate tuberculosis granulomas from lung adenocarcinomas. Eur. Radiol. 2020;30:6497–6507. doi: 10.1007/s00330-020-07024-z. [DOI] [PubMed] [Google Scholar]
  • 61.Hwang E.J., Park S., Jin K.-N., Kim J.I., Choi S.Y., Lee J.H., Goo J.M., Aum J., Yim J.-J., Park C.M., et al. Development and validation of a deep learning–based automatic detection algorithm for active pulmonary tuberculosis on chest radiographs. Clin. Infect. Dis. 2019;69:739–747. doi: 10.1093/cid/ciy967. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 62.Heo S.-J., Kim Y., Yun S., Lim S.-S., Kim J., Nam C.-M., Park E.-C., Jung I., Yoon J.-H. Deep learning algorithms with demographic information help to detect tuberculosis in chest radiographs in annual workers’ health examination data. Int. J. Environ. Res. Public Health. 2019;16:250. doi: 10.3390/ijerph16020250. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 63.Aguiar F.S., Torres R.C., Pinto J.V.F., Kritski A.L., Seixas J.M., Mello F.C.Q. Development of two artificial neural network models to support the diagnosis of pulmonary tuberculosis in hospitalized patients in Rio de Janeiro, Brazil. Med. Biol. Eng. Comput. 2016;54:1751–1759. doi: 10.1007/s11517-016-1465-1. [DOI] [PubMed] [Google Scholar]
  • 64.Faruk O., Ahmed E., Ahmed S., Tabassum A., Tazin T., Bourouis S., Khan M.M. A novel and robust approach to detect tuberculosis using transfer learning. J. Healthc. Eng. 2021;2021:1002799. doi: 10.1155/2021/1002799. [DOI] [PMC free article] [PubMed] [Google Scholar] [Retracted]
  • 65.Karki M., Kantipudi K., Yu H., Yang F., Kassim Y.M., Yaniv Z., Jaeger S. Identifying drug-resistant tuberculosis in chest radiographs: Evaluation of CNN architectures and training strategies; Proceedings of the 2021 43rd Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC); Mexico City, Mexico. 1–5 November 2021; pp. 2964–2967. [DOI] [PubMed] [Google Scholar]
  • 66.Dasanayaka C., Dissanayake M.B. Deep learning methods for screening pulmonary tuberculosis using chest X-rays. Comput. Methods Biomech. Biomed. Eng. Imaging Vis. 2021;9:39–49. doi: 10.1080/21681163.2020.1808532. [DOI] [Google Scholar]
  • 67.Govindarajan S., Swaminathan R. Extreme learning machine based differentiation of pulmonary tuberculosis in chest radiographs using integrated local feature descriptors. Comput. Methods Programs Biomed. 2021;204:106058. doi: 10.1016/j.cmpb.2021.106058. [DOI] [PubMed] [Google Scholar]
  • 68.Acharya V., Dhiman G., Prakasha K., Bahadur P., Choraria A., Prabhu S., Chadaga K., Viriyasitavat W., Kautish S. AI-assisted tuberculosis detection and classification from chest X-rays using a deep learning normalization-free network model. Comput. Intell. Neurosci. 2022;2022:2399428. doi: 10.1155/2022/2399428. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 69.Kadry S., Srivastava G., Rajinikanth V., Rho S., Kim Y. Tuberculosis detection in chest radiographs using spotted hyena algorithm optimized deep and handcrafted features. Comput. Intell. Neurosci. 2022;2022:9263379. doi: 10.1155/2022/9263379. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 70.Kazemzadeh S., Yu J., Jamshy S., Pilgrim R., Nabulsi Z., Chen C., Beladia N., Lau C., McKinney S.M., Hughes T., et al. Deep learning detection of active pulmonary tuberculosis at chest radiography matched the clinical performance of radiologists. Radiology. 2023;306:124–137. doi: 10.1148/radiol.212213. [DOI] [PubMed] [Google Scholar]
  • 71.Margarat G.S., Hemalatha G., Mishra A., Shaheen H., Maheswari K., Tamijeselvan S., Kumar U.P., Banupriya V., Ferede A.W. Early diagnosis of tuberculosis using deep learning approach for iot based healthcare applications. Comput. Intell. Neurosci. 2022;2022:3357508. doi: 10.1155/2022/3357508. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 72.Skoura E., Zumla A., Bomanji J. Imaging in tuberculosis. Int. J. Infect. Dis. 2015;32:87–93. doi: 10.1016/j.ijid.2014.12.007. [DOI] [PubMed] [Google Scholar]
  • 73.Owens C.A., Peterson C., Tang C., Koay E.J., Yu W., Mackin D.S., Li J., Salehpour M.R., Fuentes D.T., Court L., et al. Lung tumor segmentation methods: Impact on the uncertainty of radiomics features for non-small cell lung cancer. PLoS ONE. 2018;13:e0205003. doi: 10.1371/journal.pone.0205003. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 74.Bianconi F., Palumbo I., Fravolini M.L., Rondini M., Minestrini M., Pascoletti G., Nuvoli S., Spanu A., Scialpi M., Aristei C., et al. Form factors as potential imaging biomarkers to differentiate benign vs. malignant lung lesions on CT scans. Sensors. 2022;22:5044. doi: 10.3390/s22135044. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 75.Peikert T., Duan F., Rajagopalan S., Karwoski R.A., Clay R., Robb R.A., Qin Z., Sicks J., Bartholmai B.J., Maldonado F. Novel high-resolution computed tomography-based radiomic classifier for screen-identified pulmonary nodules in the national lung screening trial. PLoS ONE. 2018;13:e0196910. doi: 10.1371/journal.pone.0196910. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 76.Bossuyt P.M., Reitsma J.B., Bruns D.E., Gatsonis C.A., Glasziou P.P., Irwig L., Lijmer J.G., Moher D., Rennie D., de Vet H.C.W., et al. STARD 2015: An updated list of essential items for reporting diagnostic accuracy studies. Clin. Chem. 2015;61:1446–1452. doi: 10.1373/clinchem.2015.246280. [DOI] [PubMed] [Google Scholar]
  • 77.Collins G.S., Reitsma J.B., Altman D.G., Moons K.G.M. Transparent reporting of a multivariable prediction model for individual prognosis or diagnosis (TRIPOD): The TRIPOD statement. BMJ. 2014;350:g7594. doi: 10.1136/bmj.g7594. [DOI] [PubMed] [Google Scholar]
  • 78.Page M.J., McKenzie J.E., Bossuyt P.M., Boutron I., Hoffmann T.C., Mulrow C.D., Shamseer L., Tetzlaff J.M., Akl E.A., Brennan S.E., et al. The PRISMA 2020 statement: An updated guideline for reporting systematic reviews. BMJ. 2021;372:n71. doi: 10.1136/bmj.n71. [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Data Availability Statement

Not applicable.


Articles from Journal of Clinical Medicine are provided here courtesy of Multidisciplinary Digital Publishing Institute (MDPI)

RESOURCES