Abstract
Deep learning (DL) has the potential to transform medical diagnostics. However, the diagnostic accuracy of DL is uncertain. Our aim was to evaluate the diagnostic accuracy of DL algorithms to identify pathology in medical imaging. Searches were conducted in Medline and EMBASE up to January 2020. We identified 11,921 studies, of which 503 were included in the systematic review. Eighty-two studies in ophthalmology, 82 in breast disease and 115 in respiratory disease were included for meta-analysis. Two hundred twenty-four studies in other specialities were included for qualitative review. Peer-reviewed studies that reported on the diagnostic accuracy of DL algorithms to identify pathology using medical imaging were included. Primary outcomes were measures of diagnostic accuracy, study design and reporting standards in the literature. Estimates were pooled using random-effects meta-analysis. In ophthalmology, AUC’s ranged between 0.933 and 1 for diagnosing diabetic retinopathy, age-related macular degeneration and glaucoma on retinal fundus photographs and optical coherence tomography. In respiratory imaging, AUC’s ranged between 0.864 and 0.937 for diagnosing lung nodules or lung cancer on chest X-ray or CT scan. For breast imaging, AUC’s ranged between 0.868 and 0.909 for diagnosing breast cancer on mammogram, ultrasound, MRI and digital breast tomosynthesis. Heterogeneity was high between studies and extensive variation in methodology, terminology and outcome measures was noted. This can lead to an overestimation of the diagnostic accuracy of DL algorithms on medical imaging. There is an immediate need for the development of artificial intelligence-specific EQUATOR guidelines, particularly STARD, in order to provide guidance around key issues in this field.
Subject terms: Whole body imaging, Translational research
Introduction
Artificial Intelligence (AI), and its subfield of deep learning (DL)1, offers the prospect of descriptive, predictive and prescriptive analysis, in order to attain insight that would otherwise be untenable through manual analyses2. DL-based algorithms, using architectures such as convolutional neural networks (CNNs), are distinct from traditional machine learning approaches. They are distinguished by their ability to learn complex representations in order to improve pattern recognition from raw data, rather than requiring human engineering and domain expertise to structure data and design feature extractors3.
Of all avenues through which DL may be applied to healthcare; medical imaging, part of the wider remit of diagnostics, is seen as the largest and most promising field4,5. Currently, radiological investigations, regardless of modality, require interpretation by a human radiologist in order to attain a diagnosis in a timely fashion. With increasing demands upon existing radiologists (especially in low-to-middle-income countries)6–8, there is a growing need for diagnosis automation. This is an issue that DL is able to address9.
Successful integration of DL technology into routine clinical practice relies upon achieving diagnostic accuracy that is non-inferior to healthcare professionals. In addition, it must provide other benefits, such as speed, efficiency, cost, bolstering accessibility and the maintenance of ethical conduct.
Although regulatory approval has already been granted by the Food and Drug Administration for select DL-powered diagnostic software to be used in clinical practice10,11, many note that the critical appraisal and independent evaluation of these technologies are still in their infancy12. Even within seminal studies in the field, there remains wide variation in design, methodology and reporting that limits the generalisability and applicability of their findings13. Moreover, it is noted that there has been no overarching medical specialty-specific meta-analysis assessing diagnostic accuracy of DL performance, particularly in ophthalmology, respiratory medicine and breast surgery, which have the most diagnostic studies to date13.
Therefore, the aim of this review is to (1) quantify the diagnostic accuracy of DL in speciality-specific radiological imaging modalities to identify or classify disease, and (2) to appraise the variation in methodology and reporting of DL-based radiological diagnosis, in order to highlight the most common flaws that are pervasive across the field.
Results
Search and study selection
Our search identified 11,921 abstracts, of which 9484 were screened after duplicates were removed. Of these, 8721 did not fulfil inclusion criteria based on title and abstract. Seven hundred sixty-three full manuscripts were individually assessed and 260 were excluded at this step. Five hundred three papers fulfilled inclusion criteria for the systematic review and contained data required for sensitivity, specificity or AUC. Two hundred seventy-three studies were included for meta-analysis, 82 in ophthalmology, 115 in respiratory medicine and 82 in breast cancer (see Fig. 1). These three fields were chosen to meta-analyse as they had the largest numbers of studies with available data. Two hundred twenty-four other studies were included for qualitative synthesis in other medical specialities. Summary estimates of imaging and speciality-specific diagnostic accuracy metrics are described in Table 1. Units of analysis for each speciality and modality are indicated in Tables 2–4.
Fig. 1. PRISMA flow diagram of included studies.

PRISMA (preferred reporting items for systematic reviews and meta-analyses) flow diagram of included studies.
Table 1.
Summary estimates of pooled speciality and imaging modality specific diagnostic accuracy metrics.
| Imaging modality | Diagnosis | AUC | 95% CI | I2 | Sensitivity | 95% CI | I2 | Specificity | 95% CI | I2 | PPV | 95% CI | I2 | NPV | 95% CI | I2 | Accuracy | 95% CI | I2 | F1 score | 95% CI | I2 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| Ophthalmology imaging | ||||||||||||||||||||||
| RFP | DR | 0.939 | 0.920–0.958 | 99.9 | 0.976 | 0.975–0.977 | 99.9 | 0.902 | 0.889–0.916 | 99.7 | 0.389 | 0.166–0.612 | 99.7 | 1 | 1 | 90.6 | 0.927 | 0.899–0.955 | 96.3 | |||
| RFP | AMD | 0.963 | 0.948–0.979 | 99.3 | 0.973 | 0.971–0.974 | 99.9 | 0.924 | 0.896–0.952 | 99.6 | 0.797 | 0.719–0.875 | 99.9 | |||||||||
| RFP | Glaucoma | 0.933 | 0.924–0.942 | 99.6 | 0.883 | 0.862–0.904 | 99.9 | 0.918 | 0.898–0.938 | 99.7 | 0.881 | 0.847–0.915 | 97.7 | |||||||||
| RFP | ROP | 0.96 | 0.913–1.008 | 99.5 | 0.907 | 0.749–1.066 | 99.8 | |||||||||||||||
| OCT | DR | 1 | 0.999–1.0 | 98.1 | 0.954 | 0.937–0.972 | 98.9 | 0.993 | 0.991–0.994 | 98.2 | 0.97 | 0.959–0.981 | 97.5 | |||||||||
| OCT | AMD | 0.969 | 0.955–0.983 | 99.4 | 0.997 | 0.996–0.997 | 99.7 | 0.932 | 0.914–0.950 | 98.9 | 0.936 | 0.906–0.965 | 99.6 | |||||||||
| OCT | Glaucoma | 0.964 | 0.941–0.986 | 77.7 | ||||||||||||||||||
| Respiratory imaging | ||||||||||||||||||||||
| CT | Lung nodules | 0.937 | 0.924–0.949 | 97 | 0.86 | 0.831–0.890 | 99.7 | 0.896 | 0.871–0.921 | 99.2 | 0.785 | 0.711–0.858 | 99.2 | 0.889 | 0.870–0.908 | 98.4 | 0.79 | 0.747–0.834 | 97.9 | |||
| CT | Lung cancer | 0.887 | 0.847–0.928 | 95.9 | 0.837 | 0.780–0.894 | 94.6 | 0.826 | 0.735–0.918 | 98.1 | 0.827 | 0.784–0.870 | 81.7 | |||||||||
| X-ray | Nodules | 0.884 | 0.842–0.925 | 99.6 | 0.75 | 0.634–0.866 | 99 | 0.944 | 0.912–0.976 | 98.4 | 0.86 | 0.736–0.984 | 99.8 | 0.894 | 0.842–0.945 | 81.4 | ||||||
| X-ray | Mass | 0.864 | 0.827–0.901 | 99.7 | 0.801 | 0.683–0.919 | 99.7 | |||||||||||||||
| X-ray | Abnormal | 0.917 | 0.869–0.966 | 99.9 | 0.873 | 0.762–0.985 | 99.9 | 0.894 | 0.860–0.929 | 98.7 | 0.85 | 0.567–1.133 | 100 | 0.859 | 0.736–0.983 | 99 | 0.76 | 0.558–0.962 | 99.7 | |||
| X-ray | Atelectasis | 0.824 | 0.783–0.866 | 99.7 | ||||||||||||||||||
| X-ray | Cardiomegaly | 0.905 | 0.871–0.938 | 99.7 | ||||||||||||||||||
| X-ray | Consolidation | 0.875 | 0.800–0.949 | 99.9 | 0.914 | 0.816–1.013 | 99.5 | 0.751 | 0.637–0.866 | 98.6 | 0.897 | 0.828–0.966 | 96.4 | |||||||||
| X-ray | Pulmonary oedema | 0.893 | 0.843–0.944 | 99.9 | ||||||||||||||||||
| X-ray | Effusion | 0.906 | 0.862–0.950 | 99.8 | ||||||||||||||||||
| X-ray | Emphysema | 0.885 | 0.855–0.916 | 99.7 | ||||||||||||||||||
| X-ray | Fibrosis | 0.834 | 0.796–0.872 | 99.7 | ||||||||||||||||||
| X-ray | Hiatus hernia | 0.894 | 0.858–0.930 | 99.8 | ||||||||||||||||||
| X-ray | Infiltration | 0.724 | 0.682–0.767 | 99.6 | ||||||||||||||||||
| X-ray | Pleural thickening | 0.816 | 0.762–0.870 | 99.8 | ||||||||||||||||||
| X-ray | Pneumonia | 0.845 | 0.782–0.907 | 99.9 | 0.951 | 0.936–0.965 | 96.3 | 0.716 | 0.480–0.953 | 100 | 0.681 | 0.367–0.995 | 100 | 0.763 | 0.559–0.968 | 100 | 0.889 | 0.838–0.941 | 97.6 | |||
| X-ray | Pneumothorax | 0.91 | 0.863–0.957 | 99.9 | 0.718 | 0.433–1.004 | 100 | 0.918 | 0.870–0.965 | 99.9 | 0.496 | 0.369–0.623 | 100 | |||||||||
| X-ray | Tuberculosis | 0.979 | 0.978–0.981 | 99.6 | 0.998 | 0.997–0.999 | 99.6 | 1 | 0.999–1.000 | 95.3 | 0.94 | 0.921–0.959 | 84.6 | |||||||||
| Breast imaging | ||||||||||||||||||||||
| MMG | Breast cancer | 0.873 | 0.853–0.894 | 98.8 | 0.851 | 0.779–0.923 | 99.9 | 0.882 | 0.859–0.905 | 97.2 | 0.905 | 0.880–0.930 | 97.9 | |||||||||
| Ultrasound | Breast cancer | 0.909 | 0.881–0.936 | 91.7 | 0.853 | 0.815–0.891 | 93.9 | 0.901 | 0.870–0.931 | 96.6 | 0.804 | 0.727–0.880 | 93.7 | 0.922 | 0.851–0.992 | 97.2 | 0.873 | 0.841–0.906 | 87.5 | 0.855 | 0.803–0.906 | 87.9 |
| MRI | Breast cancer | 0.868 | 0.850–0.886 | 27.8 | 0.786 | 0.710–0.861 | 80.5 | 0.788 | 0.697–0.880 | 86.2 | ||||||||||||
| DBT | Breast cancer | 0.908 | 0.880–0.937 | 63.2 | 0.831 | 0.675–0.988 | 97.6 | 0.918 | 0.905–0.930 | 0 | ||||||||||||
Table 2.
Characteristics of ophthalmic imaging studies.
| Study | Model | Prospective? | Test set | Population | Test datasets | Type of internal validation | External validation | Reference standard | AI vs clinician? | Imaging modality | Pathology |
|---|---|---|---|---|---|---|---|---|---|---|---|
| Abramoff et al. 2016 | AlexNet/VGG | No | 1748 | Photographs | Messidor-2 | NR | No | Expert consensus | No | Retinal fundus photography | Referable DR |
| Abramoff et al. 201814 | AlexNet/VGG | Yes | 819 | Patients | Prospective cohort from 10 primary care practice sites in USA | NR | Yes | Expert consensus | No | Retinal fundus photography | More than mild DR |
| Ahn et al. 2018 | (a) Inception-v3; (b) customised CNN | No | (a) 464; (b) 464 | Images | Kim’s Eye Hospital, Korea | Random split | No | Expert consensus | No | Retinal fundus photography | Early and advanced glaucoma |
| Ahn et al. 2019 | ResNet50 | No | 219 | Photographs | Kim’s Eye Hospital, Korea | Random split | No | Expert consensus | No | Retinal fundus photographs | Pseudopapilloedema |
| Al-Aswad et al. 201946 | Pegasus (ResNet50) | No | 110 | Photographs | Singapore Malay Eye Study | Random split | No | Existing diagnosis from source data | Yes | Retinal fundus photographs | Glaucoma |
| Alqudah et al. 201922 | AOCT-NET | No | 1250 | Scans | Farsiu Ophthalmology 2013 | Hold-out method | Yes | NR | No | OCT | (a) AMD; (b) DME |
| Arcadu et al. 2019 | Inception-v3 | No | (a) 1237; (b) 1798 | Images | RISE/RIDE trials | Random split | No | Expert consensus | No | Retinal fundus photography | (a) DME—central subfield thickness >400 µm; (b) DME—central fovea thickness >400 µm |
| Asaoka et al. 2016 | Deep feed-forward neural network with stacked denoising autoencoder | No | 279 | Eyes | University of Tokyo Hospital, Tokyo | Random split | No | Other imaging technique | No | Visual Fields | Preperimetric open-angle glaucoma |
| Asaoka et al. 2019 | Customised CNN | No | 196 | Images | University of Tokyo Hospital, Tokyo | Random split | No | Expert consensus | No | OCT | Early open-angle glaucoma |
| Asaoka et al. 201923 | ResNet50 | No | (a) 205; (b) 171 | Scans | (a) Iinan Hospital; (b) Hiroshiuma University | NR | Yes | Expert consensus | No | OCT | Glaucoma |
| Bellemo et al. 201915 | VGG/ResNet | Yes | 3093 | Eyes | Kitwe Central Hospital Eye Unit, Zambia | NA | Yes | Expert consensus | No | Retinal fundus photography | (a) Referable DR; (b) vision-threatening DR; (c) DME |
| Bhatia et al. 201924 | VGG-16 | No | (a) 4686; (b) 384; (c) 148; (d) 100; (e) 135; (f) 135; (g) 148; (h) 100 | Scans | (a) Shiley Eye Institute of the UCDS; (b) Devers Eye Institute; (c) Noor Eye Hospital; (d) Ophthalmica Ophthalmology Greece; (e) Cardiff University; (f) Cardiff University; (g) Noor Eye Hospital; (h) Ophthalmica Ophthalmology Greece | NA | Yes | (a) Expert consensus; (b) NR; (c) NR; (d) NR; (e) Expert consensus + further imaging; (f) expert consensus + further imaging; (g) NR; (h) NR | No | OCT | (a) Abnormal scan; (b–f) AMD; (g–h) DME |
| Brown et al. 201847 | Inception-v1 and U-Net | No | 100 | Photographs | i-ROP | Hold-out method | No | Expert consensus | Yes | Retinal fundus photography | Plus disease in ROP |
| Burlina et al. 201749 | DCNN | No | 5664 | Images | AREDS 4 dataset | NR | No | Expert consensus | Yes | Retinal fundus photography | AMD-AREDS 4 step |
| Burlina et al. 201848 | ResNet50 | No | 5000 | Images | AREDS | Random split | No | Reading centre grader | No | Retinal fundus photographs | Referable AMD |
| Burlina et al. 201850 | AlexNet | No | 13,480 | Photographs | NIH AREDS | NR | No | Reading centre grader | Yes | Retinal fundus photography | Referable AMD |
| Burlina et al. 2018 | ResNet50 | No | (a) 6654; (b) 58,978 | Images | (a) AREDS 9 dataset; (b) AREDS 4 dataset | NR | No | Reading centre grader | Yes | Retinal fundus photography | (a) AMD-AREDS 4 step; (b) AMD-AREDS 9 step |
| Chan et al. 201825 | AlexNet, VGGNet, GoogleNet | No | 4096 | Images | SERI | NR | Yes | Reading centre grader | No | OCT | DME |
| Choi et al. 2017 | VGG-19 | No | (a) 3000; (b) 3000 | Photographs | STARE database | Random split | No | Expert consensus | No | Retinal fundus photographs | (a) DR; (b) AMD |
| Christopher et al. 201816 | (a) VGG-16; (b) Inception-v3; (c) ResNet50 | Yes | 1482 | Images | ADAGES and DIGS | Random split | No | Expert consensus | No | Retinal fundus photography | Glaucomatous optic neuropathy |
| Das et al. 2019 | VGG-16 | No | 1000 | Images | UCSD | Hold-out method | No | Expert consensus | No | OCT | DME |
| De Fauw et al. 201851 | (a) U-Net (b) customised CNN | No | (a) 997; (b) 116 | (a) Scans (Topcon device); (b) scans (Spectralis device) | Moorfields, London | Random split | No | Follow up | Yes | OCT | Urgent referral eye disease |
| ElTanboly et al. 2016 | Deep fusion classification network (DFCN) | No | 12 | OCT scans | Hold-out method | No | NR | No | OCT | Early DR | |
| Gargeya et al. 201726 | CNN | No | (a) 15,000 (b) 1748; (c) 463 | Photographs | (a) EyePACS-1; (b) Messidor-2; (c) E-Opthma | Random split | Yes | Expert consensus | No | Retinal fundus photography | DR |
| Gomez-Valverde et al. 201952 | VGG-19 | No | 494 | Photographs | ESPERANZA | Random split | No | Expert consensus | Yes | Retinal fundus photographs | Glaucoma suspect or glaucoma |
| Grassman et al. 201827 | Ensemble:random forest | No | (a) 12,019; (b) 5555 | Images | (a) AREDS dataset; (b) KORA dataset | Random split | Yes | Reading centre grader | No | Retinal fundus photography | AMD-AREDS 9 step |
| Gulshan et al. 201917 | Inception-v3 | Yes | 3049 | Photographs | Prospective | NA | Yes | Expert consensus | Yes | Retinal fundus photographs | Referable DR |
| Gulshan et al. 201628 | Inception-v3 | No | (a) 8788; (b) 1745 | Photographs | (a) EyePACS-1; (b) Messidor-2 | Random split | Yes | Reading centre grader | Yes | Retinal fundus photography | Referable DR |
| Hwang et al. 201929 | (a) ResNet50; (b) VGG-16; (c) Inception-v3; (d) ResNet50; (e) VGG-16; (f) Inception-v3 | No | (a–c) 3872; (d–f) 750 | Images | (a–c) Department of Ophthalmology of Taipei Veterans General Hospital; (d–f) External validation | Random split | Yes | Expert consensus | Yes | OCT | AMD-AREDS 4 step |
| Jammal et al. 201953 | ResNet34 | No | 490 | Images | Randomly drawn from test sample | No | Reading centre grader | Yes | Retinal fundus photographs | Glaucomatous optic neuropathy | |
| Kanagasingham et al. 201821 | DCNN | Yes | 398 | Patients | Primary Care Practice, Midland, Western Australia | NA | Yes | Reading centre grader | No | Retinal fundus photography | Referable DR |
| Karri et al. 2017 | GoogLeNet | No | 21 | Scans | Duke University | Random split | No | NR | No | OCT | (a) DME; (b) dry AMD |
| Keel et al. 201818 | Inception-v3 | Yes | 93 | Images | St Vincent’s Hospital Melbourne and University Hospital Geelong, Barwon Health | NA | Yes | Reading centre grader | No | Retinal fundus photography | Referable DR |
| Keel et al. 201930 | CNN | No | 86,202 | Photographs | Melbourne Collaborative Cohort Study | Hold-out method | Yes | Expert consensus | No | Retinal fundus photographs | Neovascular AMD |
| Kermany et al. 201854 | Inception-v3 | No | (a) 1000; (b–d) 500 | Scans | Shiley Eye Institute of the University of California San Diego, the California Retinal Research Foundation, Medical Centre Ophthalmology Associates, the Shanghai First People’s Hospital, and Beijing Tongren Eye Centre | Random split | No | Consensus involving experts and non-experts | Yes | OCT | (a) Choroidal neovascularisation vs DME vs drusen vs normal; (b) choroidal neovascularisation; (c) DME; (d) AMD |
| Krause et al. 201831 | CNN | No | 1958 | Images | EyePACS-2 | Hold-out method | Yes | Expert consensus | No | Retinal fundus photographs | Referable DR |
| Lee et al. 2017 | VGG-16 | No | 2151 | Scans | Random split | No | Routine clinical notes | No | OCT | AMD | |
| Lee et al. 2019 | CNN | No | 200 | Photographs | Seoul National University Hospital | Hold-out method | No | Other imaging technique | No | Retinal fundus photographs | Glaucoma |
| Li et al. 2018108 | Inception-v3 | No | 8000 | Scans | Guangdong (China) | Random split | No | Expert graders | No | Retinal fundus photography | Glaucomatous optic neuropathy |
| Li et al. 201955 | VGG-16 | No | 1000 | Images | Shiley Eye Institute of the University of California San Diego, the California Retinal Research Foundation, Medical Centre Ophthalmology Associates, the Shanghai First People’s Hospital, and Beijing Tongren Eye Centre | Random split | No | Expert consensus | No | OCT | Choroidal neovascularisation vs DME vs drusen Vs normal |
| Li et al. 2019 | OCT-NET | No | 859 | Scans | Wenzhou Medical University | Random split | No | Expert graders | No | OCT | Early DR |
| Li et al. 201933 | Inception-v3 | No | 800 | Images | Messidor-2 | Random split | Yes | Reading centre grader | No | Retinal fundus photographs | Referable DR |
| Li et al. 2019 | ResNet50 | No | 1635 | Images | Shanghai Zhongshan Hospital and the Shanghai First People’s Hospital | Random split | No | Reading centre grader | Yes | OCT | DME |
| Lin et al. 2019109 | CC-Cruiser | Yes—multicentre RCT | 350 | Images | Multicentre RCT | NA | NA | Expert consensus | Yes | Slit-lamp photography | Childhood cataracts |
| Li F et al. 2018 | VGG-15 | No | 300 | Images | NR | Random split | No | NR | No | Visual Fields | Glaucoma |
| Li Z et al. 201833 | CNN | No | 35,201 | Photographs | NIEHS, SiMES, AusDiab | Random split | Yes | Reading centre grader | No | Retinal fundus photographs | Referable DR |
| Liu et al. 201835 | ResNet50 | No | (a) 754; (b) 30 | Photographs | (a) NR; (b) HRF | Random split | Yes | Reading centre grader | Yes | Retinal fundus photographs | Glaucomatous optic discs |
| Liu et al. 201934 | CNN | No | (a) 28,569; (b) 20,466; (c) 12,718; (d) 9305; (e) 29,676; (f) 7877 | Photographs | (a) Local Validation (Chinese Glaucoma Study Alliance); (b) Beijing Tongren Hospital; (c) Peking University Third Hospital; (d) Harbin Medical University First Hospital; (e) Handan Eye Study; (f) Hamilton Glaucoma Centre | Random split | Yes | Consensus involving experts and non-experts | No | Retinal fundus photographs | Glaucomatous optic neuropathy |
| Long et al. 201756 | DCNN | No | 57 | Images | Multihospital clinical trial | Hold-out method | No | Expert consensus | Yes | Ocular images | Congenital Cataracts |
| MacCormick et al. 201936 | DenseNet | No | (a) 130; (b) 159 | Images | (a) ORIGA; (b) RIM-ONE | Random split | Yes | (a) NR; (b) expert consensus | No | Retinal fundus photography | Glaucomatous optic discs |
| Maetshke et al. 2019 | 3D CNN | No | 110 | OCT scans | Fivefold cross validation | Random split | No | Follow up | No | OCT | Glaucomatous optic neuropathy |
| Matsuba et al. 201857 | DCNN | No | 111 | Images | Tsukazaki Hospital | NR | No | Expert consensus + further imaging | Yes | Retinal fundus photography (optos) | Exudative AMD |
| Medeiros et al. 2019 | ResNet34 | No | 6292 | Images | Duke University | Random split | No | Follow up | No | Retinal fundus photography | Glaucomatous optic neuropathy |
| Motozawa et al. 2019 | CNN | No | 382 | Images | Kobe City Medical Centre | Random split | No | Routine clinical notes | No | OCT | AMD |
| Muhammad et al. 2017 | AlexNet | No | 102 | Images | NR | NR | No | Expert consensus | No | OCT | Glaucoma suspect or glaucoma |
| Nagasato et al. 2019 | VGG-16 | No | 466 | Images | NR | K-fold cross validation | No | NR | No | Retinal fundus photography (optos) | Retinal vein occlusion |
| Nagasato et al. 201958 | DNN | No | 322 | Scans | Tsukazaki Hospital and Tokushima University Hospital | K-fold cross validation | No | Expert graders | Yes | OCT | Retinal vein occlusion |
| Nagasawa et al. 2019 | VGG-16 | No | 378 | Images | Tsukazaki Hospital and Tokushima University Hospital | K-fold cross validation | No | Expert graders | No | Retinal fundus photography (optos) | Proliferative diabetic retinopathy |
| Ohsugi et al. 2017 | DCNN | No | 166 | Images | Tsukazaki Hospital | Random split | No | Expert consensus | No | Retinal fundus photography (optos) | Rhegmatogenous retinal detachment |
| Peng et al. 201959 | Inception-v3 | No | 900 | Images | AREDS | Random split | No | Reading centre grader | Yes | Retinal fundus photography | Age-related macular degeneration-AREDS 4 step |
| Perdomo et al. 2019 | OCT-NET | No | 2816 | Images | SERI-CUHK data set | Random split | No | Expert graders | No | OCT | DME |
| Phan et al. 2019 | DenseNet201 | No | 828 | Images | Yamanashi Koseiren Hospital | No | Expert consensus + further imaging | No | Retinal fundus photography | Glaucoma | |
| Phene et al. 201937 | Inception-v3 | No | (a) 1205; (b) 9642; (c) 346 | Images | (a) EyePACS, Inoveon, the United Kingdom Biobank, the Age-Related Eye Disease Study, and Sankara Nethralaya; (b) Atlanta Veterans Affairs (VA) Eye Clinic; (c) Dr. Shroff’s Charity Eye Hospital, New Delhi, India | Random split | Yes | Reading centre grader | Yes | Retinal fundus photographs | Glaucomatous optic neuropathy |
| Prahs et al. 2017 | GoogLeNet | No | 5358 | Images | Heidelberg Eye Explorer, Heidelberg Engineering | Random split | No | Expert graders | No | OCT | Injection vs No injection for AMD |
| Raju et al. 2017 | CNN | No | 53,126 | Images | EyePACS-1 | Random split | No | NR | No | Retinal fundus photography | Referable DR |
| Ramachandran et al. 201838 | Visiona intelligent diabetic retinopathy screening platform | No | (a) 485; (b) 1200 | Photographs | (a) ODEMS; (b) Messidor | NA | Yes | Expert graders | No | Retinal fundus photographs | Referable DR |
| Raumviboonsuk et al. 201939 | Inception-v4 | No | (a–c) 25,348; (d) 24,332 | Images | National screening program for DR in Thailand | NA | Yes | Expert consensus | Yes | Retinal fundus photography | (a) Moderate non-proliferative DR or worse; (b) severe non-proliferative DR or worse; (c) proliferative DR; (d) referable DME |
| Redd et al. 2018 | Inception-v1 and U-Net | No | 4861 | Images | Multicentre i-ROP study | NR | No | Expert graders + further imaging | No | Retinal fundus photography | Plus disease in ROP |
| Rogers et al. 201945 | Pegasus (ResNet50) | No | 94 | Photographs | EODAT | NA | Yes | Reading centre grader | Yes | Retinal fundus photographs | Glaucomatous optic neuropathy |
| Sandhu et al. 201819 | Deep fusion SNCAE | Yes | 160 | Scans | University of Waikato | NA | No | Clinical diagnosis | No | Retinal fundus photographs | Non-proliferative DR |
| Sayres et al. 201940 | Inception-v4 | No | 2000 | Images | EyePACS-2 | NA | Yes | Expert consensus | Yes | Retinal fundus photographs | Referable DR |
| Shibata et al. 201860 | (a) ResNet; (b) VGG-16 | No | 110 | Images | Matsue Red Cross Hospital | Random split | No | Expert consensus | Yes | Retinal fundus photography | Glaucoma |
| Stevenson et al. 2019 | Inception-v3 | No | (a) 2333; (b) 2283; (c) 2105 | Photographs | Publicly available databases | Random split | No | Existing diagnosis from source data | No | Retinal fundus photographs | (a) Glaucoma; (b) DR; (c) AMD |
| Ting et al. 201741 | VGGNet | No | (a) 71,896; (b) 15,798; (c) 3052; (d) 4512; (e) 1936; (f) 1052; (g) 1968; (h) 2302; (i) 1172; (j) 1254; (k) 7706; (l) 35,948; (m) 35,948 | Images | (a) Singapore National Diabetic Retinopathy Screening Program 2014–2015; (b) Guangdong (China); (c) Singapore Malay Eye Study; (d) Singapore Indian Eye Study; (e) Singapore Chinese Eye Study; (f) Beijing Eye Study; (g) African American Eye Disease Study; (h) Royal Victoria Eye and Ear Hospital; (i) Mexican; (j) Chinese University of Hong Kong, (k, l) Singapore National Diabetic Retinopathy Screening Program 2014–2015 | Random split | Yes | Expert consensus | No | Retinal fundus photography | Referable DR |
| Ting et al. 201942 | VGGNet | No | 85,902 | Images | Combined eight datasets | NA | Yes | Consensus involving experts and non-experts | No | Retinal fundus photography | (a) Any DR; (b) referable DR; (c) vision-threatening DR |
| Treder et al. 2017 | Inception-v3 | No | 100 | Scans | NR | Hold-out method | No | NR | No | OCT | Exudative AMD |
| van Grinsven et al. 201644 | (a) Ses CNN 60; (b) NSesCNN170 | No | 1200 | Images | Messidor | Random split | Yes | Existing diagnosis from source data | Yes | Retinal fundus photographs | Retinal haemorrhage |
| Verbraak et al. 201943 | AlexNet/VGG | No | 1293 | Images | Netherlands Star-SHL | NA | Yes | Expert consensus | No | Retinal fundus photography | (a) DR-vision-threatening; (b) DR- more than mild |
| Xu et al. 2017 | CNN | No | 200 | Photographs | Kaggle | Random split | No | Existing diagnosis from source data | No | Retinal fundus photographs | DR |
| Yang et al. 2019 | VGGNet | No | 500 | Photographs | Intelligent Ophthalmology Database of Zhejiang Society for Mathematical Medicine in China | Hold-out method | No | Expert consensus | No | Retinal fundus photographs | Referable DR |
| Yoo et al. 2019 | VGG-19 | No | 900 | Scans | Project Macula | Random split | No | NR | No | (a) OCT; (b) retinal fundus photographs | AMD |
| Zhang et al. 201961 | VGG-16 | No | 1742 | Images | Telemed-R screening | Random split | No | Expert consensus | Yes | Retinal fundus photographs | ROP |
| Zheng et al. 201920 | Inception-v3 | Yes | 102 | Scans | Joint Shantou International Eye Centre of Shantou University and the Chinese University of Hong Kong (JSIEC) | Hold-out method | No | NR | No | OCT | Glaucomatous optic neuropathy |
Table 4.
Characteristics of breast imaging studies.
| Study | Model | Prospective? | Test Set | Population | Test datasets | Type of internal validation | External validation | Reference standard | AI vs clinician? | Imaging modality | Body system/disease |
|---|---|---|---|---|---|---|---|---|---|---|---|
| Abdelsamea et al. 2019 | CNN | No | 118 | Images | NR | Tenfold cross validation | No | NR | No | Mammogram | Breast cancer |
| Agnes et al. 2020 | Multiscale all CNN | No | 322 | Images | mini-MIAS | Random split | No | Existing labels from dataset | No | Mammogram | Breast cancer |
| Akselrod-Ballin et al. 2017 | Faster R-CNN | No | 170 | Images | Multicentre hospital data set | Random split | No | Expert reader | No | Mammogram | Breast cancer |
| Al-Antari et al. 2018 | YOLO | No | 410 | Images | INbreast | Random split | No | Expert reader, histology, | No | Mammogram | Breast cancer |
| Al-Antari et al. 2018 | DBN | No | 150 | Images | DDSM | Random split | No | Follow up, histology, expert reader | No | Mammogram | Breast cancer |
| Al-Masni et al. 2018 | YOLO | No | 120 | Images | DDSM | Random split | No | Follow up, histology, expert reader | No | Mammogram | Breast cancer |
| Antropova et al. 2017 | VGG-19 | No | (a) 690; (b) 245; (c) 1125 | (a) Lesions; (b) images; (c) lesions | Private | Random split | No | Histology | No | (a) MRI; (b) mammogram; (c) ultrasound | Breast cancer |
| Antropova et al. 2018 | VGGNet | No | 138 | Lesions | University of Chicago | Random split | No | Histology | No | MRI | Breast cancer |
| Antropova et al. 2018 | VGGNet | No | 141 | Lesions | University of Chicago | Random split | No | Histology | No | MRI | Breast cancer |
| Arevalo et al. 2016 | CNN3 | No | 736 | Images | Breast Cancer Digital Repository (BCDR), Portugal | Stratified Sampling | No | Histology | No | Mammogram | Breast cancer |
| Bandeira Diniz et al. 2018 | CNN | No | (a) 200; (b) 288 | Images | (a) DDSM Dense Breast; (b) DDSM Non Dense Breast | Random split | No | Follow up, histology, expert reader | No | Mammogram | Breast cancer |
| Becker et al. 201791 | dANN | No | 70 | Images | Breast Cancer Digital Repository (BCDR) | Random split | Yes | Expert reader | Yes | Mammogram | Breast cancer |
| Becker et al. 201862 | DNN | No | 192 | Lesions | Private | Random split | No | Follow up, histology | Yes | Ultrasound | Breast cancer |
| Bevilacqua et al. 2019 | VGG-S | No | 39 | Images | NR | NR | No | NR | No | Digital breast tomosynthesis | Breast cancer |
| Byra et al. 201999 | VGG-19 | No | (a) 150; (b) 163; (c) 100 | Images | (a) Moores Cancer Center, University of California; (b) UDIAT (c) OASBUD | Random split | No | (a) Follow up, histology; (b) expert reader; (c) expert reader, histology, follow up | Yes | Ultrasound | Breast cancer |
| Cai et al. 2019 | CNN | No | 99 | Images | SYSUCC and Foshan, China | Random split | No | Histology | No | Mammogram | Breast cancer |
| Cao et al. 2019 | SSD300 + ZFNet | No | 183 | Lesions | Sichuan Provincial People’s Hospital | Random split | No | Expert consensus | No | Ultrasound | Breast cancer |
| Cao et al. 2019 | NF-Net | No | 272 | Lesions | Sichuan Provincial People’s Hospital | Random split | No | Histology | No | Ultrasound | Breast cancer |
| Cheng et al. 2016 | Stacked denoising autoencoder | No | 520 | Lesions | Taipei Veterans General Hospital | NR | No | Histology | No | Ultrasound | Breast Nodules |
| Chiao et al. 2019 | Mask R-CNN | No | 61 | Images | China Medical University Hospital | Random split | No | Histology, routine clinical report | No | Ultrasound | Breast cancer |
| Choi et al. 2019100 | CNN | No | 253 | Lesions | Samsung Medical Centre, Seoul | NR | No | Follow up, histology | Yes | Ultrasound | Breast cancer |
| Chougrad et al. 2018 | Inception-v3 | No | (a) 5316; (b) 600; (c) 200 | Images | (a) DDSM; (b) Inbreast; (c) BCDR | Random split | No | (a) Follow up, histology, expert reader; (b) expert reader, histology; (c) clinical reports | No | Mammogram | Breast cancer |
| Ciritsis et al. 201992 | dCNN | No | (a) 101; (b) 43 | Images | (a) Internal validation; (b) external validation | Random split | Yes | Follow up, histology | Yes | Ultrasound | Breast cancer |
| Cogan et al. 201993 | ResNet-101 Faster R-CNN | No | 124 | Images | INbreast | NA | Yes | Expert reader, histology, | No | Mammogram | Breast cancer |
| Dalmis et al. 2018 | U-Net | No | 66 | Images | NR | Random split | No | Follow up, histology | No | MRI | Breast cancer |
| Dalmis et al. 2019101 | DenseNet | No | 576 | Lesions | Raboud University Medical Center | NR | No | Follow up, histology | Yes | MRI | Breast cancer |
| Dhungel et al. 2017 | CNN | No | 82 | Images | INbreast | Random split | No | Expert reader, histology, | No | Mammogram | Breast cancer |
| Duggento et al. 2019 | CNN | No | 378 | Images | Curated Breast Imaging SubSet of DDSM (CBIS-DDSM) | Random split | No | Expert reader | No | Mammogram | Breast cancer |
| Fan et al. 2019 | Faster R-CNN | No | 182 | Images | Fudan University Affiliated Cancer Centre | Random split | No | Histology | No | Digital breast tomosynthesis | Breast cancer |
| Fujioka et al. 2019102 | GoogleNet | No | 120 | Lesions | Private | Random split | No | Follow up, histology | Yes | Ultrasound | Breast cancer |
| Gao et al. 2018 | SD-CNN | No | (a) 49; (b) 89 | (a) Lesions; (b) images | (a) Mayo Clinic Arizona; (b) Inbreast | NR | No | (a) Histology; (b) expert reader, histology | No | (a) Contrast enhanced digital mammogram; (b) mammogram | Breast cancer |
| Ha et al. 2019 | CNN | No | 60 | Images | Columbia University Medical Center | Random split | No | Follow up, histology | No | Mammogram | DCIS |
| Han et al. 2017 | GoogleNet | No | 829 | Lesions | Samsung Medical Centre, Seoul | Random split | No | Histology | No | Ultrasound | Breast cancer |
| Herent et al. 2019 | ResNet50 | No | 168 | Lesions | Journees Francophones de Radiologie 2018 | Random split | No | NR | No | MRI | Breast cancer |
| Hizukuri et al. 2018 | CNN | No | 194 | Images | Mie University Hospital | Random split | No | Follow up, histology | No | Ultrasound | Breast cancer |
| Huyng et al. 2016 | AlexNet | No | 607 | Images | University of Chicago | NR | No | Histology | No | Mammogram | Breast cancer |
| Jadoon et al. 2016 | CNN-DW | No | 2976 | Images | IRMA | NR | No | Histology | No | Mammogram | Breast cancer |
| Jiao et al. 2016 | CNN | No | 300 | Images | DDSM | Random split | No | Follow up, histology, expert reader | No | Mammogram | Breast cancer |
| Jiao et al. 2018 | (a) AlexNet; (b) parasitic metric learning layers | No | (a) 150; (b) 150 | Images | DDSM | Random split | No | Follow up, histology, expert reader | No | Mammogram | Breast cancer |
| Jung et al. 2018 | RetinaNet | No | (a) 410; (b) 222 | Images | (a) Inbreast; (b) GURO | Random split | No | (a) Expert reader; (b) histology | No | Mammogram | Breast cancer |
| Kim et al. 2012103 | ANN | No | 70 | Lesions | Kangwon National University College of Medicine | Random split | No | Expert consensus | Yes | Ultrasound | Breast cancer |
| Kim et al. 2018 | ResNet | No | 1238 | Images | Yonsei University Health System | Random split | No | Follow up, histology | No | Mammogram | Breast cancer |
| Kim et al. 2018 | VGG-16 | No | 340 | Images | DDSM | Hold-out method | No | Follow up, histology, expert reader | No | Mammogram | Breast cancer |
| Kooi et al. 2017 | CNN | No | 18,182 | Images | Netherlands screening database | Random split | No | Expert reader, histology, | No | Mammogram | Breast cancer |
| Kooi et al. 2017 | CNN | No | 1523 | Images | Netherlands screening database | Random split | No | Expert reader, histology, | No | Mammogram | Breast cancer |
| Kooi T et al. 2017 | CNN | No | 1804 | Images | Netherlands screening database | Hold-out method | No | Expert reader, histology, | No | Mammogram | Breast Cancer |
| Li et al. 2019 | DenseNet-II | No | 2042 | Images | First Hospital of Shanxi Medical University | Tenfold cross validation | No | Expert reader | No | Mammogram | Breast cancer |
| Li et al. 2019 | VGG-16 | No | (a) 1854; (b) 1854 | Images | Nanfang Hospital | Fivefold cross validation | No | Follow up, histology | No | (a) Digital breast tomosynthesis; (b) mammogram | Breast cancer |
| Lin et al. 2014 | FCMNN | No | 65 | Images | Far Eastern Memorial Hospital, Taiwan | Tenfold cross validation | No | Histology | No | Ultrasound | Breast cancer |
| McKinney et al. 202094 | MobileNetV2 - ResNet-v2-50, ResNet-v1-50 | No | (a) 25,856; (b) 3097 | Images | (a) UK; (b) USA | Random split | Yes | Follow up, histology | Yes | Mammogram | Breast cancer |
| Mendel et al. 2018 | VGG-19 | No | (a) 78; (b) 78 | Images | University of Chicago | Leave-one-out method | No | Follow up, histology | No | (a) Mammogram; (b) digital breast tomosynthesis | Breast cancer |
| Peng et al. 201695 | ANN | No | (a) 100; (b) 100 | Images | (a) MIAS; (b) BancoWeb | Hold-out method | Yes | Expert reader | No | Mammogram | Breast cancer |
| Qi et al. 2019 | Inception-Resnet-v2 | No | 1359 | Images | West China Hospital, Sichuan University | Random split | No | Expert consensus | No | Ultrasound | Breast cancer |
| Qiu et al. 2017 | CNN | No | 140 | Images | Private | Random split | No | Histology | No | Mammogram | Breast cancer |
| Ragab et al. 2019 | AlexNet | No | (a) 676; (b) 1581 | Images | (a) Digital database for screening mammography (DDSM); (b) Curated Breast Imaging SubSet of DDSM (CBIS-DDSM) | Random split | No | Follow up, histology, expert reader | No | Mammogram | Breast cancer |
| Ribli et al. 201896 | VGG-16 | No | 115 | Images | INbreast | NA | Yes | Expert reader, histology | No | Mammogram | Breast cancer |
| Rodriguez-Ruiz et al. 201897 | CNN | No | 240 | Images | Two datasets combined | NA | Yes | Expert reader, histology, follow up | Yes | Mammogram | Breast cancer |
| Rodriguez-Ruiz et al. 201998 | CNN | No | 2642 | Images | Combined nine datasets | NA | Yes | Follow up, histology | Yes | Mammogram | Breast cancer |
| Samala et al. 2016 | DCNN | No | 94 | Images | University of Michigan | Random split | No | Expert reader | No | Digital breast tomosynthesis | Breast cancer |
| Samala et al. 2017 | DCNN | No | 907 | Images | DDSM + private | Random split | No | Expert reader | No | Mammogram | Breast cancer |
| Samala et al. 2018 | DCNN | No | 94 | Images | University of Michigan | Random split | No | Expert reader | No | Digital breast tomosynthesis | Breast cancer |
| Samala et al. 2019 | AlexNet | No | 94 | Images | University of Michigan | Random split | No | Expert reader | No | Digital breast tomosynthesis | Breast cancer |
| Shen et al. 2019 | (a) VGG-16; (b) ResNet; (c) ResNet-VGG | No | (a) 376; (b) 376; (c) 107 | Images | (a) Curated Breast Imaging SubSet of DDSM (CBIS-DDSM); (b) Curated Breast Imaging SubSet of DDSM (CBIS-DDSM); (c) Inbreast | Random split | No | (a) Histology; (b) histology; (c) expert reader | No | Mammogram | Breast cancer |
| Shin et al. 2019 | VGG-16 | No | (a) 600; (b) 40 | Images | (a) Seoul National University Bundang Hospital; (b) UDIAT Diagnostic Centre of the Parc Taulí Corporation | Random split | No | (a) NR; (b) expert reader | No | Ultrasound | Breast cancer |
| Stoffel et al. 2018 | CNN | No | 33 | Images | Private | Random split | No | Surgical confirmation | Yes | Ultrasound | Phyllodes tumour |
| Sun et al. 2017 | CNN | No | 758 | Images | University of Texas at El Paso | Random split | No | Expert reader | No | Mammogram | Breast cancer |
| Tanaka et al. 2019 | VGG-19, Resnet152 | No | 154 | Lesions | Japan Association of Breast and Thyroid Sonology | Random split | No | Histology | No | Ultrasound | Breast cancer |
| Tao et al. 2019 | RefineNet + DenseNet121 | No | 253 | Lesions | Huaxi Hospital and China-Japan Friendship Hospital | Random split | No | Expert reader | No | Ultrasound | Breast cancer |
| Teare et al. 2017 | Inception-v3 | No | 352 | Images | DDSM + Zebra Mammography Dataset | Random split | No | Follow up, histology | No | Mammogram | Breast cancer |
| Truhn et al. 2018104 | CNN | No | 129 | Lesions | RWTH Aachen University, | Random split | No | Follow up, histology | Yes | MRI | Breast cancer |
| Wang et al. 2016 | Inception-v3 | No | 74 | Images | Breast Cancer Digital Repository (BCDR) | Random split | No | Expert reader, histology | No | Mammogram | Breast cancer |
| Wang et al. 2016 | Stacked autoencoder | No | 204 | Images | Sun Yat-sen University Cancer Center (Guangzhou, China) and Nanhai Affiliated Hospital of Southern Medical University (Foshan, China) | Hold-out method | No | Histology | No | Mammogram | Breast cancer |
| Wang et al. 2017 | CNN | No | 292 | Images | University of Chicago | Random split | No | Histology | No | Mammogram | Breast cancer |
| Wang et al. 2018 | DNN | No | 292 | Images | University of Chicago | Random split | No | Histology | No | Mammogram | Breast cancer |
| Wu et al. 2019105 | ResNet-22 | No | (a) 401; (b) 1440 | Images | NYU | Hold-out method | No | Histology | Yes | Mammogram | Breast cancer |
| Xiao et al. 2019 | Inception-v3, ResNet50, Xception | No | 206 | Images | Third Affiliated Hospital of Sun Yat-sen University | Random split | No | Surgical confirmation, histology | No | Ultrasound | Breast cancer |
| Yala et al. 2019106 | ResNet18 | No | 26,540 | Images | Massachusetts General Hospital, Harvard Medical School, | Random split | No | Clinical reports, follow up, histology | Yes | Mammogram | Breast cancer |
| Yala et al. 2019111 | ResNet18 | No | 8751 | Images | Massachusetts General Hospital, Harvard Medical School, | Random split | No | Clinical reports, follow up, histology | No | Mammogram | Breast cancer |
| Yap et al. 2018 | FCN-AlexNet | No | (a) 306; (b) 163 | Lesions | (a) Private; (b) UDIAT | NR | No | Expert reader | No | Ultrasound | Breast cancer |
| Yap et al. 2019 | FCN-8s | No | 94 | Lesions | Two datasets combined | NR | No | Expert reader | No | Ultrasound | Breast cancer |
| Yousefi et al. 2018 | DCNN | No | 28 | Images | MGH | Random split | No | Expert consensus | No | Digital breast tomosynthesis | Breast cancer |
| Zhou et al. 2019107 | 3D DenseNet | No | 307 | Lesions | Private | Random split | No | Follow up, histology | Yes | MRI | Breast cancer |
Ophthalmology imaging
Eighty-two studies with 143 separate patient cohorts reported diagnostic accuracy data for DL in ophthalmology (see Table 2 and Supplementary References 1). Optical coherence tomography (OCT) and retinal fundus photographs (RFP) were the two imaging modalities performed in this speciality with four main pathologies being diagnosed—diabetic retinopathy (DR), age-related macular degeneration (AMD), glaucoma and retinopathy of prematurity (ROP).
Only eight studies14–21 used prospectively collected data and 29 (refs. 14,15,17,18,21–45) studies validated algorithms on external datasets. No studies provided a prespecified sample size calculation. Twenty-five studies17,28,29,35,37,39,40,44–61 compared algorithm performance against healthcare professionals. Reference standards, definitions of disease and threshold for diagnosis varied greatly as did the method of internal validation used. There was high heterogeneity across all studies (see Table 2).
Diabetic retinopathy: Twenty-five studies with 48 different patient cohorts reported diagnostic accuracy data for all, referable or vision-threatening DR on RFP. Twelve studies and 16 cohorts reported on diabetic macular oedema (DME) or early DR on OCT scans. AUC was 0.939 (95% CI 0.920–0.958) for RFP versus 1.00 (95% CI 0.999–1.000) for OCT.
Age-related macular degeneration: Twelve studies reported diagnostic accuracy data for features of varying severity of AMD on RFP (14 cohorts) and 11 studies in OCT (21 cohorts). AUC was 0.963 (95% CI 0.948–0.979) for RFP versus 0.969 (95% CI 0.955–0.983) for OCT.
Glaucoma: Seventeen studies with 30 patient cohorts reported diagnostic accuracy for features of glaucomatous optic neuropathy, optic discs or suspect glaucoma on RFP and five studies with 6 cohorts on OCT. AUC was 0.933 (95% CI 0.924–0.942) for RFP and 0.964 (95% CI 0.941–0.986) for OCT. One study34 with six cohorts on RFP provided contingency tables. When averaging across the cohorts, the pooled sensitivity was 0.94 (95% CI 0.92–0.96) and pooled specificity was 0.95 (95% CI 0.91–0.97). The AUC of the summary receiver-operating characteristic (SROC) curve was 0.98 (95% CI 0.96–0.99)—see Supplementary Fig. 1.
Retinopathy of prematurity: Three studies reported diagnostic accuracy for identifying plus diseases in ROP from RFP. Sensitivity was 0.960 (95% CI 0.913—1.008) and specificity was 0.907 (95% CI 0.907–1.066). AUC was only reported in two studies so was not pooled.
Others: Eight other studies reported on diagnostic accuracy in ophthalmology either using different imaging modalities (ocular images and visual fields) or for identifying other diagnoses (pseudopapilloedema, retinal vein occlusion and retinal detachment). These studies were not included in the meta-analysis.
Respiratory imaging
One hundred and fifteen studies with 244 separate patient cohorts report on diagnostic accuracy of DL on respiratory disease (see Table 3 and Supplementary References 2). Lung nodules were largely identified on CT scans, whereas chest X-rays (CXR) were used to diagnose a wide spectrum of conditions from simply being ‘abnormal’ to more specific diagnoses, such as pneumothorax, pneumonia and tuberculosis.
Table 3.
Characteristics of respiratory imaging studies.
| Study | Model | Prospective? | Test set | Population | Test datasets | Type of internal validation | External validation | Reference standard | AI vs clinician | Imaging modality | Body system/disease |
|---|---|---|---|---|---|---|---|---|---|---|---|
| Abiyev et al. 2018 | CNN | No | 380 | Images | Chest X-ray14 | Random split | No | Routine clinical reports | No | X-ray | Abnormal X-ray |
| Al-Shabi et al. 2019 | Local-Global | No | 848 | Nodules | LIDC-IDRI | NR | No | Expert readers | No | CT | Nodules |
| Alakwaa et al. 2017 | U-Net | No | 419 | Scans | Kaggle Data Science Bowl | Random split | No | Expert reader, existing labels in dataset | No | CT | Lung cancer |
| Ali et al. 2018 | 3D CNN | No | 668 | Nodules | LIDC-IDRI | Random split | No | Expert readers | No | CT | Nodules |
| Annarumma et al. 2019110 | CNN | No | 15,887 | Images | Kings College London | Hold-out method | No | Routine clinical reports | No | X-ray | (a) Critical radiographs; (b) normal radiographs |
| Ardila et al. 201964 | Inception-v1 | No | (a) 6716; (b) 1139 | Scans | (a) National Lung Cancer Screening Trial; (b) Northwestern Medicine | Random split | Yes | Histopathology, follow up | Yes | CT | Lung cancer |
| Baltruschat et al. 2019 | ResNet50 | No | 22,424 | X-rays | Chest X-ray14 | Random split | No | Routine clinical reports | No | X-ray | (a) Abnormal chest X-ray; (b) normal chest X-ray; (c) atelectasis; (d) cardiomegaly; (e) effusion; (f) infiltration; (g) mass; (h) nodule (i) pneumonia; (j) pneumothorax; (k) consolidation; (l) oedema; (m) emphysema; (n) fibrosis; (o) pleural thickening; (p) hernia |
| Bar et al. 2018 | CNN | No | 194 | Images | Diagnostic Imaging Department of Sheba Medical Centre, Tel Hashomer, Israel | Random split | No | Expert readers | No | X-ray | (a) Abnormal X-ray; (b) cardiomegaly |
| Becker et al. 201862 | CNN | Yes | 21 | X-rays | Infectious Diseases Institute in Kampala, Uganda | Random split | No | Expert consensus | No | X-ray | Tuberculosis |
| Behzadi-Khormouji et al. 2020 | (a) ChestNet; (b) VGG-16; (c) DenseNet121 | No | 582 | X-rays | Guangzhou Women and Children’s Medical Centre | NR | No | Expert readers | No | X-ray | Consolidation |
| Beig et al. 2019 | CNN | No | 145 | Scans | Erlangen Germany, Waukesha Wis, Cleveland Ohio, Tochigi-ken Japan | Random split | No | Histopathology | No | CT | Lung cancer |
| Causey et al. 2018 | CNN | No | (a) 424; (b) 213 | Nodules | LIDC-IDRI | Random split | No | Expert readers | No | CT | Nodules |
| Cha et al. 201976 | ResNet50 | No | (a) 1483; (b) 500 | X-rays | Samsung Medical Centre, Seoul | Random split | No | Other imaging, expert readers | Yes | X-ray | (a) Lung cancer; (b) T1 lung cancer |
| Chae et al. 201977 | Ct-LUNGNET | No | 60 | Nodules | Chonbuk National University Hospital | Random split | No | Expert readers, histopathology, follow up | Yes | CT | Nodules |
| Chakravarthy et al. 2019 | Probabilistic neural network | No | 119 | Scans | LIDC/IDRI | NR | No | NR | No | CT | Lung cancer |
| Chen et al. 2019 | 3D CNN | No | 3674 | Nodules | LIDC-IDRI | NR | No | Expert readers | No | CT | Nodules |
| Cheng et al. 2016 | Stacked denoising autoencoder | No | 1400 | Nodules | LIDC-IDRI | Random split | No | Expert readers | No | CT | Nodules |
| Cicero et al. 2017 | GoogLeNet | No | 2443 | Images | Department of Medical Imaging, St Michael’s Hospital, Toronto | Random split | No | Expert readers, routine clinical reports | No | X-ray | (a) Effusion; (b) oedema; (c) consolidation; (d) cardiomegaly; (e) pneumothorax |
| Ciompi et al. 201778 | ConvNet | No | 639 | Nodules | Danish Lung Cancer Screening Trial (DLCST) | Random split | No | Non-expert readers | Yes | CT | (a) Nodules—solid; (b) nodules—calcified; (c) nodules—part-solid; (d) nodules—non-solid; (e) nodules—perifissural; (f) nodules—spiculated |
| Correa et al. 2018 | CNN | No | 60 | Images | Lima, Peru | NR | No | Expert readers | No | Ultrasound | Paediatric pneumonia |
| da Silva et al. 2017 | Evolutionary CNN | No | 200 | Nodules | LIDC-IDRI | Hold-out method | No | Expert readers | No | CT | Nodules |
| da Silva et al. 2018 | Particle swarm optimisation algorithm within CNN | No | 2000 | Nodules | LIDC-IDRI | Random split | No | Expert readers | No | CT | Nodules |
| Dai et al. 2018 | 3D DenseNet-40 | No | 211 | Nodules | LIDC-IDRI | Random split | No | Expert readers | No | CT | Nodules |
| Dou et al. 2017 | 3D CNN | No | 1186 | Nodules | LUNA16 | NR | No | Expert readers | No | CT | Nodules |
| Dunnmon et al. 201979 | ResNet18 | No | 533 | Images | Stanford University | Hold-out method | No | Expert consensus | Yes | X-ray | Abnormal X-ray |
| Gao et al. 2018 | CNN | No | 20 | Scans | University Hospitals of Geneva | Random split | No | NR | No | CT | Interstitial lung disease |
| Gong et al. 2019 | 3D SE-ResNet | No | 1186 | Nodules | LUNA16 | NR | No | Expert readers | No | CT | Nodules |
| Gonzalez et al. 2018 | CNN | No | 1000 | Scans | ECLIPSE study | Random split | No | NR | No | CT | COPD |
| Gruetzemacher et al. 2018 | DNN | No | 1186 | Nodules | LUNA16 | Ninefold cross validation | No | NR | No | CT | Nodules |
| Gu et al. 2018 | 3D CNN | No | 1186 | Nodules | LUNA16 | Tenfold cross validation | No | Expert readers | No | CT | Nodules |
| Hamidian et al. 2017 | 3D CNN | No | 104 | Nodules | LIDC-IDRI | Random split | No | Expert readers | No | CT | Nodules |
| Han et al. 2018 | Multi-CNNs | No | 812 | Regions of interest | LIDC-IDRI | Random split | No | NR | No | CT | Ground glass opacity |
| Heo et al. 2019 | VGG-19 | No | 37,677 | X-rays | Yonsei University Hospital, South Korea | Hold-out method | No | Expert readers | No | X-ray | Tuberculosis |
| Hua et al. 2015 | (a) CNN; (b) deep belief network | No | 2545 | Nodules | LIDC-IDRI | NR | No | Expert readers | No | CT | Nodules |
| Huang et al. 2019 | R-CNN | No | 176 | Scans | LIDC-IDRI | Random split | No | Expert readers | No | CT | Nodules |
| Huang et al. 2019 | Amalgamated-CNN | No | 1795 | Nodules | LIDC/IDRI and Ali Tianchi medical | Random split | No | Expert readers | No | CT | Nodules |
| Hussein et al. 2019 | VGG | No | 1144 | Nodules | LIDC/IDRI | Random split | No | Expert readers | No | CT | Lung cancer |
| Hwang et al. 201867 | DCNN | No | (a) 450; (b) 183; (c) 140; (d) 173; (e) 170; (f) 132; (g) 646 | X-rays | (a) Internal validation; (b) Seoul National University Hospital; (c) Boromae Hospital; (d) Kyunghee University Hospital; (e) Daejeon Eulji Medical Centre; (f) Montgomery; (g) | Random split | Yes | Expert readers | Yes | X-ray | Tuberculosis |
| Hwang et al. 201965 | Lunit INSIGHT | No | 1135 | X-rays | Seoul National University Hospital | NA | Yes | Expert consensus, other imaging | Yes | X-ray | Abnormal chest X-ray |
| Hwang et al. 201966 | DCNN | No | (a) 1089; (b) 1015 | X-rays | (a) Internal validation; (b) external validation | Random split | Yes | Expert reader, other imaging, histopathology | Yes | X-ray | Neoplasm/TB/pneumonia/pneumothorax |
| Jiang et al. 2018 | CNN | No | 25,723 | Nodules | LIDC-IDRI | NR | No | Expert readers | No | CT | Nodules |
| Jin et al. 2018 | ResNet 3D | No | 1186 | Nodules | LUNA16 | NR | No | Expert readers | No | CT | Nodules |
| Jung et al. 2018 | 3D DCNN | No | 1186 | Nodules | LUNA16 | NR | No | Expert readers | No | CT | Nodules |
| Kang et al. 2017 | 3D multi view-CNN | No | 776 | Nodules | LIDC-IDRI | NR | No | Expert readers | No | CT | Nodules |
| Kermany et al. 2018 | Inception-v3 | No | 624 | X-rays | Guangzhou Women and Children’s Medical Centre | Random split | No | Expert readers | Yes | X-ray | Pneumonia |
| Kim et al. 2019 | MGI-CNN | No | 1186 | Nodules | LIDC/IDRI | NR | No | Expert readers | No | CT | Nodules |
| Lakhani et al. 201790 | (a) AlexNet; (b) GoogLeNet; (c) Ensemble (AlexNet + GoogLeNet); (d) Radiologist augmented | No | 150 | X-rays | Montgomery County MD, Shenzhen China, Belarus TB public Health Program, Thomas Jefferson University Hospital | Random split | No | Routine clinical reports, expert reader, histopathology | No | X-ray | Tuberculosis |
| Li et al. 2016 | CNN | No | 8937 | Nodules | LIDC-IDRI | Random split | No | Expert readers | No | CT | Nodules |
| Li et al. 201981 | DL-CAD | No | 812 | Nodules | Shenzhen Hospital | NR | No | Expert consensus | Yes | CT | Nodules |
| Li et al. 201980 | CNN | No | 200 | Scans | Massachusetts General Hospital | Random split | No | Routine clinical reports | Yes | CT | Pneumothorax |
| Liang et al. 202068 | CNN | No | 100 | Images | Kaohsiung Veterans General Hospital, Taiwan | NA | Yes | Other imaging | No | X-ray | Nodules |
| Liang et al. 2019 | (a) Custom CNN; (b) VGG-16; (c) DenseNet121; (d) Inception-v3; (e) Xception | No | 624 | X-rays | Guangzhou Women and Children’s Medical Centre | Random split | No | Expert readers | No | X-ray | Pneumonia |
| Liu et al. 2017 | 3D CNN | No | 326 | Nodules | National Lung Cancer Screening Trial and Early Lung Cancer Action Program | Fivefold cross validation | No | Histopathology, follow up | No | CT | Nodules |
| Liu et al. 2019 | CDP-ResNet | No | 539 | Nodules | LIDC-IDRI | Random split | No | Expert readers | No | CT | Nodules |
| Liu H et al. 2019 | Segmentation-based deep fusion network | No | 112,120 | X-rays | Chest X-ray14 | NR | No | Routine clinical reports | No | X-ray | (a) Atelectasis; (b) cardiomegaly; (c) effusion; (d) infiltration; (e) mass; (f) nodule; (g) pneumonia; (h) pneumothorax; (i) consolidation; (j) oedema; (k) emphysema; (l) fibrosis; (m) fibrosis; (n) pleural thickening; (o) hernia |
| Majkowska et al. 201982 | CNN | No | (a–d) 1818; (e–h) 1962 | X-rays | (a–d) Hospital group in India (Bangalore, Bhubaneshwar, Chennai, Hyderabad, New Delhi); (e–h) Chest X-ray14 | Random split | No | Expert consensus | Yes | X-ray | (a) Pneumothorax (b) nodule; (c) opacity; (d) fracture; (e) pneumothorax; (f) nodule; (g) opacity; (h) fracture |
| Monkam et al. 2018 | CNN | No | 2600 | Nodules | LIDC-IDRI | Random split | No | Expert readers | No | CT | Nodules |
| Nam et al. 201869 | CNN | No | (a) 600; (b) 181; (c) 182; (d) 181; (e) 149 | Chest radiographs | (a) Internal validation; (b) Seoul National University Hospital; (c) Boromae Hospital; (d) National Cancer Centre, Korea; (e) University of California an Francisco Medical Centre | Random split | Yes | (a) Routine clinical reports, histopathology; (b–e) histopathology, follow up, other imaging | No | X-ray | Nodules |
| Naqi et al. 2018 | Two-level stacked autoencoder + softmax | No | 777 | Nodules | LIDC-IDRI | NR | No | Expert readers | No | CT | Nodules |
| Nasrullah et al. 2019 | Faster R-CNN | No | 2562 | Nodules | LIDC/IDRI | NR | No | Expert readers | No | CT | Nodules |
| Nibali et al. 2017 | ResNet | No | 166 | Nodules | LIDC-IDRI | Random split | No | Expert readers | No | CT | Nodules |
| Nishio et al. 2018 | VGG-16 | No | 123 | Nodules | Kyoto University Hospital | Random split | No | NR | No | CT | Nodules |
| Onishi et al. 2019 | AlexNet | No | 60 | Nodules | NR | NR | No | Histopathology, follow up | No | CT | Nodules |
| Onishi et al. 2019 | Wasserstein generative adversarial network | No | 60 | Nodules | Fujita Health University Hospital | NR | No | Histopathology, follow up | No | CT | Nodules |
| Park et al. 201989 | YOLO | No | 503 | X-rays | Asan Medical Centre and Seoul National University Bundang Hospital | Hold-out method | No | Expert reader | No | X-ray | Pneumothorax |
| Park et al. 201983 | CNN | No | 200 | Images | Asan Medical Centre and Seoul National University Bundang Hospital | Hold-out method | No | Expert consensus | Yes | X-ray | (a) Nodules; (b) opacity; (c) effusion; (d) pneumothorax; (e) abnormal chest X-ray |
| Pasa et al. 2019 | Custom CNN | No | 220 | X-rays | NIH Tuberculosis Chest X-ray dataset and Belarus Tuberculosis Portal dataset | Random split | No | NR | No | X-ray | Tuberculosis |
| Patel et al. 201984 | CheXMax | No | 50 | X-rays | Stanford University | Hold-out method | No | Expert reader, other imaging, clinical notes | Yes | X-ray | Pneumonia |
| Paul et al. 2018 | VGG-s CNN | No | 237 | Nodules | National Lung Cancer Screening Trial | Hold-out method | No | Expert readers, follow up | No | CT | Nodules |
| Pesce et al. 2019 | Convolution networks with attention feedback (CONAF) | No | 7850 | X-rays | Guy’s and St. Thomas’ NHS Foundation Trust | Random split | No | Routine clinical reports | No | X-ray | Lung lesions |
| Pezeshk et al. 2019 | 3D CNN | No | 128 | Nodules | LUNA16 | Random split | No | Expert readers | No | CT | Nodules |
| Qin et al. 201970 | (a) Lunit; (b) qXR (Qure.ai); (c) CAD4TB | No | 1196 | X-rays | Nepal and Cameroon | NA | Yes | Expert readers | Yes | X-ray | Tuberculosis |
| Rajpurkar et al. 201885 | CNN | No | 420 | X-rays | ChestXray-14 | Random split | No | Routine clinical reports | Yes | X-ray | (a) Atelectasis; (b) cardiomegaly; (c) consolidation; (d) oedema; (e) effusion; (f) emphysema; (g) fibrosis; (h) hernia; (i) infiltration; (j) mass; (k) nodule; (l) pleural thickening; (m) pneumonia; (n) pneumothorax |
| Ren et al. 2019 | Manifold regularized classification deep neural network | No | 98 | Nodules | LIDC-IDRI | Random split | No | Expert readers | No | CT | Nodules |
| Sahu et al. 2019 | Multi-section CNN | No | 130 | Nodules | LIDC-IDRI | Tenfold cross validation | No | Expert readers | No | CT | Nodules |
| Schwyzer et al. 2018 | CNN | No | 100 | Patients | NR | NR | No | NR | No | FDG-PET | Lung cancer |
| Setio et al. 201671 | ConvNet | No | (a) 1186; (b) 50; (c) 898 | (a) Nodules; (b) scans; (c) nodules | LIDC-IDRI | Fivefold cross validation | Yes | (a) Expert readers; (b, c) NR | No | CT | Nodules |
| Shaffie et al. 2018 | Deep autoencoder | No | 727 | Nodules | LIDC-IDRI | NR | No | Expert readers | No | CT | Nodules |
| Shen et al. 2017 | Multiscale CNN | No | 1375 | Nodules | LIDC-IDRI | NR | No | Expert readers | No | CT | Nodules |
| Sim et al. 201972 | ResNet50 | No | 800 | Images | Freiberg University Hospital Freiburg, Massachusetts General Hospital Boston, Samsung Medical Centre Seoul, Severance Hospital Seoul | NA | Yes | Other imaging, histopathology | Yes | X-ray | Nodules |
| Singh et al. 201886 | Qure-AI | No | 724 | Chest radiographs | Chest X-ray8 | Random split | No | Routine clinical reports | Yes | X-ray | (a) Lesions; (b) effusion; (c) hilar prominence; (d) cardiomegaly |
| Song et al. 2017 | (a) CNN; (b) DNN; (c) stacked autoencoder | No | 5024 | Nodules | LIDC-IDRI | Random split | No | Expert readers | No | CT | Nodules |
| Stephen et al. 2019 | CNN | No | 2134 | Images | Guangzhou Women and Children’s Medical Centre | Random split | No | NR | No | X-ray | Pneumonia |
| Sun et al. 2017 | (a) CNN; (b) deep belief network; (c) stacked denoising autoencoder | No | 88,948 | Samples | LIDC-IDRI | Tenfold cross validation | No | Expert readers | No | CT | Nodules |
| Tan et al. 2019 | CNN | No | 280 | Nodules | LIDC-IDRI | Tenfold cross validation | No | NR | No | CT | Nodules |
| Taylor et al. 201873 | (a) Inception-v3; (b) VGG-19; (c) Inception-v3; (d) VGG-19 | No | (a, b) 1990; (c, d) 112,120 | X-rays | (a,b) Internal validation (c,d) Chest X-ray14 | Random split | Yes | Expert consensus | No | X-ray | Pneumothorax |
| Teramoto et al. 2016 | CNN | No | 104 | Scans | Fujita Health University Hospital | NR | No | Expert reader | No | PET/CT | Nodules |
| Togacar et al. 2019 | AlexNet + VGG-16 + VGG-19 | No | 1754 | X-rays | Firat University, Turkey | Random split | No | NR | No | X-ray | Pneumonia |
| Togacar et al. 2020 | (a) LeNet; (b) AlexNet; (c) VGG-16 | No | 100 | Images | Cancer Imaging Archive | NR | No | Expert readers | No | CT | Lung cancer |
| Tran et al. 2019 | LdcNet | No | 1186 | Nodules | LUNA16 | Tenfold cross validation | No | Expert readers | No | CT | Nodules |
| Tu et al. 2017 | CNN | No | 20 | Nodules | LIDC-IDRI | Tenfold cross validation | No | Expert readers | No | CT | (a) Nodules—non-solid; (b) nodules—part-solid; (c) nodules—solid |
| Uthoff et al. 201974 | CNN | No | 100 | Nodules | INHALE STUDY | NA | Yes | Histopathology, follow up | No | CT | Nodules |
| Walsh et al. 201887 | Inception-ResNet-v2 | No | 150 | Scans | La Fondazione Policlinico Universitario A Gemelli IRCCS, Rome, Italy, and University of Parma, Parma, Italy | Random split | No | Expert readers | Yes | CT | Interstitial lung disease |
| Wang et al. 2017 | AlexNet | No | 230 | X-rays | Japanese Society of Radiological Technology (JSRT) database | Tenfold cross validation | No | Other imaging | No | X-ray | Nodules |
| Wang et al. 201888 | 3D CNN | No | 200 | Scans | Fudan University Shanghai Cancer Centre | Random split | No | Expert readers, histopathology | Yes | HRCT | Lung cancer |
| Wang et al. 2018 | VGG-16 | No | 744 | X-rays | JSRT, OpenI, SZCX and MC | Random split | No | Other imaging | No | X-ray | (a) Abnormal chest X-ray; (b) normal chest X-ray |
| Wang et al. 2019 | ChestNet | No | 442 | X-rays | Zhejiang University School of Medicine (ZJU-2) and Chest X-ray14 | Random split | No | Expert readers | No | X-ray | Pneumothorax |
| Wang et al. 2019 | (a) AlexNet; (b) GoogLeNet; (c) ResNet | No | 7580 | Nodules | LIDC-IDRI | Random split | No | Expert readers | No | CT | Nodules |
| Wang et al. 2019 | ResNet152 | No | 25,596 | X-rays | Chest X-ray14 | Random split | No | Routine clinical reports | No | X-ray | (a) Atelectasis; (b) cardiomegaly; (c) effusion; (d) infiltration; (e) mass; (f) nodule; (g) pneumonia; (h) pneumothorax; (i) consolidation; (j) oedema; (k) emphysema; (l) fibrosis; (m) pleural thickening; (n) hernia; (o) abnormal chest X-ray |
| Xie et al. 2018 | LeNet-5 | No | 1972 | Nodules | LIDC-IDRI | Random split | No | Expert readers | No | CT | Nodules |
| Xie et al. 2019 | ResNet50 | No | 1945 | Nodules | LIDC-IDRI | Tenfold cross validation | No | Expert readers | No | CT | Nodules |
| Yates et al. 2018 | Inception-v3 | No | 5505 | X-rays | Chest X-ray14 + Indiana University | Random split | No | Routine clinical reports | No | X-ray | Abnormal chest X-ray |
| Ye et al. 2019 | (a) AlexNet; (b) GoogLeNet; (c) Res-Net150 | No | (a) 321; (b) 321; (c) 593 | (a) Nodules; (b) nodules; (c) regions of interest | (a, b) LIDC-IDRI; (c) private | Random split | No | Expert readers | No | CT | (a, b) Nodules; (c) ground glass opacity |
| Zech et al. 201875 | CNN | No | (a) 30,450; (b) 3807 | X-rays | (a) Mount Sinai and Chest X-ray14; (b) Indiana University Network for Patient Care | Random split | Yes | Expert readers | No | X-ray | Pneumonia |
| Zhang et al. 2018 | 3D DCNN | No | 1186 | Nodules | LUNA16 | NR | No | Expert readers | No | CT | Nodules |
| Zhang et al. 2019 | Voxel-level-1D CNN | No | 67 | Nodules | Stony Brook University Hospital | Twofold cross validation | No | Histopathology | No | CT | Nodules |
| Zhang et al. 2019 | 3D deep dual path network | No | 1004 | Nodules | LIDC/IDRI | Tenfold cross validation | No | Expert readers | No | CT | Nodules |
| Zhang C et al. 2019 | 3D CNN | Yes | 50 | Images | Guangdong Lung Cancer Institute | Random split | Yes | Histopathology, follow up | Yes | CT | Nodules |
| Zhang et al. 201963 | Mask R-CNN | No | 134 | Slices | Shenzhen Hospital | Random split | No | Expert readers | No | CT/PET | Lung cancer |
| Zhang S et al. 2019 | Le-Net5 | No | 762 | Nodules | LIDC/IDRI | Random split | No | Expert readers | No | CT | Nodules |
| Zhang T et al. 2017 | Deep Belief Network | No | 1664 | Nodules | LIDC-IDRI | Random split | No | Expert readers | No | CT | Nodules |
| Zhao X et al. 2018 | Agile CNN | No | 743 | Nodules | LIDC-IDRI | Random split | No | Expert readers | No | CT | Nodules |
| Zhao X et al. 2019 | (a) AlexNet; (b) GoogLeNet; (c) ResNet; (d) VifarNet | No | 2028 | Nodules | LIDC-IDRI | Random split | No | Expert readers | No | CT | Nodules |
| Zheng et al. 2019 | CNN | No | 1186 | Nodules | LIDC-IDRI | Random split | No | Expert readers | No | CT | Nodules |
| Zhou et al. 2019 | Inception-v3 and ResNet50 | No | 600 | Images | Chest X-ray8 | Random split | No | Routine clinical reports | No | X-ray | Cardiomegaly |
Only two studies62,63 used prospectively collected data and 13 (refs. 63–75) studies validated algorithms on external data. No studies provided a prespecified sample size calculation. Twenty-one54,63–67,70,72,76–88 studies compared algorithm performance against healthcare professionals. Reference standards varied greatly as did the method of internal validation used. There was high heterogeneity across all studies (see Table 3).
Lung nodules: Fifty-six studies with 74 separate patient cohorts reported diagnostic accuracy for identifying lung nodules on CT scans on a per lesion basis, compared with nine studies and 14 patient cohorts on CXR. AUC was 0.937 (95% CI 0.924–0.949) for CT versus 0.884 (95% CI 0.842–0.925) for CXR. Seven studies reported on diagnostic accuracy for identifying lung nodules on CT scans on a per scan basis, these were not included in the meta-analysis.
Lung cancer or mass: Six studies with nine patient cohorts reported diagnostic accuracy for identifying mass lesions or lung cancer on CT scans compared with eight studies and ten cohorts on CXR. AUC was 0.887 (95% CI 0.847–0.928) for CT versus 0.864 (95% CI 0.827–0.901) for CXR.
Abnormal Chest X-ray: Twelve studies reported diagnostic accuracy for abnormal CXR with 13 different patient cohorts. AUC was 0.917 (95% CI 0.869–0.966), sensitivity was 0.873 (95% CI 0.762–0.985) and specificity was 0.894 (95% CI 0.860–0.929).
Pneumothorax: Ten studies reported diagnostic accuracy for pneumothorax on CXR with 14 different patient cohorts. AUC was 0.910 (95% CI 0.863–0.957), sensitivity was 0.718 (95% CI 0.433–1.004) and specificity was 0.918 (95% CI 0.870–0.965). Five patient cohorts from two studies73,89 provided contingency tables with raw diagnostic accuracy. When averaging across the cohorts, the pooled sensitivity was 0.70 (95% CI 0.45–0.87) and pooled specificity was 0.94 (95% CI 0.90–0.97). The AUC of the SROC curve was 0.94 (95% CI 0.92–0.96)—see Supplementary Fig. 2.
Pneumonia: Ten studies reported diagnostic accuracy for pneumonia on CXR with 15 different patient cohorts. AUC was 0.845 (95% CI 0.782–0.907), sensitivity was 0.951 (95% CI 0.936–0.965) and specificity was 0.716 (95% CI 0.480–0.953).
Tuberculosis: Six studies reported diagnostic accuracy for tuberculosis on CXR with 17 different patient cohorts. AUC was 0.979 (95% CI 0.978–0.981), sensitivity was 0.998 (95% CI 0.997–0.999) and specificity was 1.000 (95% CI 0.999–1.000). Four patient cohorts from one study90 provided contingency tables with raw diagnostic accuracy. When averaging across the cohorts, the pooled sensitivity was 0.95 (95% CI 0.91–0.97) and pooled specificity was 0.97 (95% CI 0.93–0.99). The AUC of the SROC curve was 0.97 (95% CI 0.96–0.99)—see Supplementary Fig. 3.
X-ray imaging was also used to identify atelectasis, pleural thickening, fibrosis, emphysema, consolidation, hiatus hernia, pulmonary oedema, infiltration, effusion, mass and cardiomegaly. CT imaging was also used to diagnose COPD, ground glass opacity and interstitial lung disease, but these were not included in the meta-analysis.
Breast imaging
Eighty-two studies with 100 separate patient cohorts report on diagnostic accuracy of DL on breast disease (see Table 4 and Supplementary References 3). The four imaging modalities of mammography (MMG), digital breast tomosynthesis (DBT), ultrasound and magnetic resonance imaging (MRI) were used to diagnose breast cancer.
No studies used prospectively collected data and eight91–98 studies validated algorithms on external data. No studies provided a prespecified sample size calculation. Sixteen studies62,91,92,94,97–107 compared algorithm performance against healthcare professionals. Reference standards varied greatly as did the method of internal validation used. There was high heterogeneity across all studies (see Table 4).
Breast cancer: Forty-eight studies with 59 separate patient cohorts reported diagnostic accuracy for identifying breast cancer on MMG (AUC 0.873 [95% CI 0.853–0.894]), 22 studies and 25 patient cohorts on ultrasound (AUC 0.909 [95% CI 0.881–0.936]), and eight studies on MRI (AUC 0.868 [95% CI 0.850–0.886]) and DBT (AUC 0.908 [95% CI 0.880–0.937]).
Other specialities
Our literature search also identified 224 studies in other medical specialities reporting on diagnostic accuracy of DL algorithms to identify disease. These included large numbers of studies in the fields of neurology/neurosurgery (78), gastroenterology/hepatology (24) and urology (25). Out of the 224 studies, only 55 compared algorithm performance against healthcare professionals, although 80% of studies in the field of dermatology did (see Supplementary References 4, Supplementary Table 1 and Supplementary Fig. 4).
Variation of reporting
A key finding of our review was the large degree of variation in methodology, reference standards, terminology and reporting among studies in all specialities. The most common variables amongst DL studies in medical imaging include issues with the quality and size of datasets, metrics used to report performance and methods used for validation (see Table 5). Only eight studies in ophthalmology imaging14,21,32,33,43,55,108,109, ten studies in respiratory imaging64,66,70,72,75,79,82,87,89,110 and six studies in breast imaging62,91,97,104,106,111 mentioned adherence to the STARD-2015 guidelines or had a STARD flow diagram in the manuscript.
Table 5.
Variation in DL imaging studies.
| Data | |
| Image pre-processing, augmentation and preparation |
Are data augmentation techniques such as cropping, padding and flipping used? Is there quality control of the images being used to train the algorithm? I.e., were poor quality images excluded. Were relevant images manually selected? |
| Study design | Retrospective or prospective data collection. |
| Image eligibility |
How are images chosen for inclusion in the study? Were the data from private or open-access repositories? |
| Training, validation, test sets |
Are each of the three sets independent of each other, without overlap? Does data from the same patient appear in multiple datasets? |
| Datasets |
Are the datasets used single or multicentre? Is a public or open-source dataset used? |
| Size of datasets |
Wide variation in size of datasets for training and testing. Is the size of the datasets justified? Are sample size statistical considerations applied for the test set? |
| Use of ‘external’ test sets for final reporting |
Is an independent test set used for ‘external validation’? Is the independent test set constructed using an unenriched representative sample? |
| Multi-vendor images |
Are images from different scanners and vendors included in the datasets to enhance generalisability? Are imaging acquisition parameters described? |
| Algorithm | |
| Index test |
Was sufficient detail given on the algorithm to allow replication and independent validation? What type of algorithm was used? E.g., CNN, Autoencoder, SVM. Was the algorithm made publicly or commercially available? Was the construct or architecture of the algorithm made available? |
| Additional AI algorithmic information | Is the algorithm a static model or is it continuously evolving? |
| Demonstrate how algorithm makes decisions | Is there a specific design for end-user interpretability, e.g., saliency or probability maps |
| Methods | |
| Transfer learning | Was transfer learning used for training and validation? |
| Cross validation | Was k-fold cross validation used during training to reduce the effects of randomness in dataset splits? |
| Reference standard |
Is the reference standard used of high quality and widely accepted in the field? What was the rationale for choosing the reference standard? |
| Additional clinical information | Was additional clinical information given to healthcare professionals to simulate normal clinical process? |
| Performance benchmarking |
What was performance of algorithm benchmarked to? What is expertise level and level of consensus of healthcare professionals if used? |
| Results | |
| Raw diagnostic accuracy data | Are raw diagnostic accuracy data reported in a contingency table demonstrating TP, FP, FN, TN? |
| Metrics for estimating diagnostic accuracy performance | Which diagnostic accuracy metrics reported? Sensitivity, Sensitivity, PPV, NPV, Accuracy, AUROC |
| Unit of assessment | Which unit of assessment reported, e.g., per patient, per scan or per lesion. |
Rows in bold are part of STARD-2015 criteria.
Funnel plots were produced for the diagnostic accuracy outcome measure with the largest number of patient cohorts in each medical speciality, in order to detect bias in the studies included112 (see Supplementary Figs. 5–7). These demonstrate that there is high risk of bias in studies detecting lung nodules on CT scans and detecting DR on RFP, but not for detecting breast cancer on MMG.
Assessment of the validity and applicability of the evidenc
The overall risk of bias and applicability using Quality Assessment of Diagnostic Accuracies Studies 2 (QUADAS-2) led to a majority of studies in all specialities being classified as high risk, particularly with major deficiencies in regard to patient selection, flow and timing and applicability of the reference standard (see Fig. 2). For the patient selection domain, a high or unclear risk of bias was seen in 59/82 (72%) of ophthalmic studies, 89/115 (77%) of respiratory studies and 62/82 (76%) or breast studies. These were mostly related to a case–control study design and sampling issues. For the flow and timing domain, a high or unclear risk of bias was seen in 66/82 (80%) of ophthalmic studies, 93/115 (81%) of respiratory studies and 70/82 (85%) of breast studies. This was largely due to missing information about patients not receiving the index test or whether all patients received the same reference standard. For the reference standard domain, concerns regarding applicability was seen in 60/82 (73%) of ophthalmic studies, 104/115 (90%) of respiratory studies and 78/82 (95%) of breast studies. This was mostly due to reference standard inconsistencies if the index test was validated on external datasets.
Fig. 2. QUADAS-2 summary plots.
Risk of bias and applicability concerns summary about each QUADAS-2 domain presented as percentages across the 82 included studies in ophthalmic imaging (a), 115 in respiratory imaging (b) and 82 in breast imaging (c).
Discussion
This study sought to (1) quantify the diagnostic accuracy of DL algorithms to identify specific pathology across distinct radiological modalities, and (2) appraise the variation in study reporting of DL-based radiological diagnosis. The findings of our speciality-specific meta-analysis suggest that DL algorithms generally have a high and clinically acceptable diagnostic accuracy in identifying disease. High diagnostic accuracy with analogous DL approaches was identified in all specialities despite different workflows, pathology and imaging modalities, suggesting that DL algorithms can be deployed across different areas in radiology. However, due to high heterogeneity and variance between studies, there is considerable uncertainty around estimates of diagnostic accuracy in this meta-analysis.
In ophthalmology, the findings suggest features of diseases, such as DR, AMD and glaucoma can be identified with a high sensitivity, specificity and AUC, using DL on both RFP and OCT scans. In general, we found higher sensitivity, specificity, accuracy and AUC with DL on OCT scans over RFP for DR, AMD and glaucoma. Only sensitivity was higher for DR on RFP over OCT.
In respiratory medicine, our findings suggest that DL has high sensitivity, specificity and AUC to identify chest pathology on CT scans and CXR. DL on CT had higher sensitivity and AUC for detecting lung nodules; however, we found a higher specificity, PPV and F1 score on CXR. For diagnosing cancer or lung mass, DL on CT had a higher sensitivity than CXR.
In breast cancer imaging, our findings suggest that DL generally has a high diagnostic accuracy to identify breast cancer on mammograms, ultrasound and DBT. The performance was found to be very similar for these modalities. In MRI, however, the diagnostic accuracy was lower; this may be due to small datasets and the use of 2D images. The utilisation of larger databases and multiparametric MRI may increase the diagnostic accuracy113.
Extensive variation in the methodology, data interpretability, terminology and outcome measures could be explained by a lack of consensus in how to conduct and report DL studies. The STARD-2015 checklist114, designed for reporting of diagnostic accuracy studies is not fully applicable to clinical DL studies115. The variation in reporting makes it very difficult to formally evaluate the performance of algorithms. Furthermore, differences in reference standards, grader capabilities, disease definitions and thresholds for diagnosis make direct comparison between studies and algorithms very difficult. This can only be improved with well-designed and executed studies that explicitly address questions concerning transparency, reproducibility, ethics and effectiveness116 and specific reporting standards for AI studies115,117.
The QUADAS-2 (ref. 118) assessment tool was used to systematically evaluate the risk of bias and any applicability concerns of the diagnostic accuracy studies. Although this tool was not designed for DL diagnostic accuracy studies, the evaluation allowed us to judge that a majority of studies in this field are at risk of bias or concerning for applicability. Of particular concern was the applicability of reference standards and patient selection.
Despite our results demonstrating that DL algorithms have a high diagnostic accuracy in medical imaging, it is currently difficult to determine if they are clinically acceptable or applicable. This is partially due to the extensive variation and risk of bias identified in the literature to date. Furthermore, the definition of what threshold is acceptable for clinical use and tolerance for errors varies greatly across diseases and clinical scenarios119.
Limitations in the literature
Dataset
There are broad methodological deficiencies among the included studies. Most studies were performed using retrospectively collected data, using reference standards and labels that were not intended for the purposes of DL analysis. Minimal prospective studies and only two randomised studies109,120, evaluating the performance of DL algorithms in clinical settings were identified in the literature. Proper acquisition of test data is essential to interpret model performance in a real-world clinical setting. Poor quality reference standards may result in the decreased model performance due to suboptimal data labelling in the validation set28, which could be a barrier to understanding the true capabilities of the model on the test set. This is symptomatic of the larger issue that there is a paucity of gold-standard, prospectively collected, representative datasets for the purposes of DL model testing. However, as there are many advantages to using retrospectively collected data, the resourceful use of retrospective or synthetic data with the use of labels of varying modality and quality represent important areas of research in DL121.
Study methodology
Many studies did not undertake external validation of the algorithm in a separate test set and relied upon results from the internal validation data; the same dataset used to train the algorithm initially. This may lead to an overestimation of the diagnostic accuracy of the algorithm. The problem of overfitting has been well described in relation to machine learning algorithms122. True demonstration of the performance of these algorithms can only be assumed if they are externally validated on separate test sets with previously unseen data that are representative of the target population.
Surprisingly, few studies compared the diagnostic accuracy of DL algorithms against expert human clinicians for medical imaging. This would provide a more objective standard that would enable better comparison of models across studies. Furthermore, application of the same test dataset for diagnostic performance assessment of DL algorithms versus healthcare professionals was identified in only select studies13. This methodological deficiency limits the ability to gauge the clinical applicability of these algorithms into clinical practice. Similarly, this issue can extend to model-versus-model comparisons. Specific methods of model training or model architecture may not be described well enough to permit emulation for comparison123. Thus, standards for model development and comparison against controls will be needed as DL architectures and techniques continue to develop and are applied in medical contexts.
Reporting
There was varying terminology and a lack of transparency used in DL studies with regards to the validation or test sets used. The term ‘validation’ was identified as being used interchangeably to either describe an external test set for the final algorithm or for an internal dataset that is used to fine tune the model prior to ‘testing’. Furthermore, the inconsistent terminology led to difficulties understanding whether an independent external test set was used to test diagnostic performance13.
Crucially, we found broad variation in the metrics used as outcomes for the performance of the DL algorithms in the literature. Very few studies reported true positives, false positives, true negatives and false negatives in a contingency table as should be the minimum for diagnostic accuracy studies114. Moreover, some studies only reported metrics, such as dice coefficient, F1 score, competition performance metric and Top-1 accuracy that are often used in computer science, but may be unfamiliar to clinicians13. Metrics such as AUC, sensitivity, specificity, PPV and NPV should be reported, as these are more widely understood by healthcare professionals. However, it is noted that NPV and PPV are dependent on the underlying prevalence of disease and as many test sets are artificially constructed or balanced, then reporting the NPV or PPV may not be valid. The wide range of metrics reported also leads to difficulty in comparing the performance of algorithms on similar datasets.
Study strengths and limitations
This systematic review and meta-analysis statistically appraises pooled data collected from 279 studies. It is the largest study to date examining the diagnostic accuracy of DL on medical imaging. However, our findings must be viewed in consideration of several limitations. Firstly, as we believe that many studies have methodological deficiencies or are poorly reported, these studies may not be a reliable source for evaluating diagnostic accuracy. Consequently, the estimates of diagnostic performance provided in our meta-analysis are uncertain and may represent an over-estimation of the true accuracy. Secondly, we did not conduct a quality assessment for the transparency of reporting in this review. This was because current guidelines to assess diagnostic accuracy reporting standards (STARD-2015114) were not designed for DL studies and are not fully applicable to the specifics and nuances of DL research115. Thirdly, due to the nature of DL studies, we were not able to perform classical statistical comparison of measures of diagnostic accuracy between different imaging modalities. Fourthly, we were unable to separate each imaging modality into different subsets, to enable comparison across subsets and allow the heterogeneity and variance to be broken down. This was because our study aimed to provide an overview of the literature in each specific speciality, and it was beyond the scope of this review to examine each modality individually. The inherent differences in imaging technology, patient populations, pathologies and study designs meant that attempting to derive common lessons across the board did not always offer easy comparisons. Finally, our review concentrated on DL for speciality-specific medical imaging, and therefore it may not be appropriate to generalise our findings to other forms of medical imaging or AI studies.
Future work
For the quality of DL research to flourish in the future, we believe that the adoption of the following recommendations are required as a starting point.
Availability of large, open-source, diverse anonymised datasets with annotations
This can be achieved through governmental support and will enable greater reproducibility of DL models124.
Collaboration with academic centres to utilise their expertise in pragmatic trial design and methodology125
Rather than classical trials, novel experimental and quasi-experimental methods to evaluate DL have been proposed and should be evaluated126. This may include ongoing evaluation of algorithms once in clinical practice, as they continue to learn and adapt to the population that they are implemented in.
Creation of AI-specific reporting standards
A major reason for the difficulties encountered in evaluating the performance of DL on medical imaging are largely due to inconsistent and haphazard reporting. Although DL is widely considered as a ‘predictive’ model (where TRIPOD may be applied) the majority of AI interventions close to translation currently published are predominantly in the field of diagnostics (with specifics on index tests, reference standards and true/false positive/negatives and summary diagnostic scores, centred directly in the domain of STARD). Existing reporting guidelines for diagnostic accuracy studies (STARD)114, prediction models (TRIPOD)127, randomised trials (CONSORT)128 and interventional trial protocols (SPIRIT)129 do not fully cover DL research due to specific considerations in methodology, data and interpretation required for these studies. As such, we applaud the recent publication of the CONSORT-AI117 and SPIRIT-AI130 guidelines, and await AI-specific amendments of the TRIPOD-AI131 and STARD-AI115 statements (which we are convening). We trust that when these are published, studies being conducted will have a framework that enables higher quality and more consistent reporting.
Development of specific tools for determining the risk of study bias and applicability
An update to the QUADAS-2 tool taking into account the nuances of DL diagnostic accuracy research should be considered.
Updated specific ethical and legal framework
Outdated policies need to be updated and key questions answered in terms of liability in cases of medical error, doctor and patient understanding, control over algorithms and protection of medical data132. The World Health Organisation133 and others have started to develop guidelines and principles to regulate the use of AI. These regulations will need to be adapted by each country to fit their own political and healthcare context134. Furthermore, these guidelines will need to proactively and objectively evaluate technology to ensure best practices are developed and implemented in an evidence-based manner135.
Conclusion
DL is a rapidly developing field that has great potential in all aspects of healthcare, particularly radiology. This systematic review and meta-analysis appraised the quality of the literature and provided pooled diagnostic accuracy for DL techniques in three medical specialities. While the results demonstrate that DL currently has a high diagnostic accuracy, it is important that these findings are assumed in the presence of poor design, conduct and reporting of studies, which can lead to bias and overestimating the power of these algorithms. The application of DL can only be improved with standardised guidance around study design and reporting, which could help clarify clinical utility in the future. There is an immediate need for the development of AI-specific STARD and TRIPOD statements to provide robust guidance around key issues in this field before the potential of DL in diagnostic healthcare is truly realised in clinical practice.
Methods
This systematic review was conducted in accordance with the guidelines for the ‘Preferred Reporting Items for Systematic Reviews and Meta-Analyses’ extension for diagnostic accuracy studies statement (PRISMA-DTA)136.
Eligibility criteria
Studies that report upon the diagnostic accuracy of DL algorithms to investigate pathology or disease on medical imaging were sought. The primary outcome was various diagnostic accuracy metrics. Secondary outcomes were study design and quality of reporting.
Data sources and searches
Electronic bibliographic searches were conducted in Medline and EMBASE up to 3rd January 2020. MESH terms and all-field search terms were searched for ‘neural networks’ (DL or convolutional or cnn) and ‘imaging’ (magnetic resonance or computed tomography or OCT or ultrasound or X-ray) and ‘diagnostic accuracy metrics’ (sensitivity or specificity or AUC). For the full search strategy, please see Supplementary Methods 1. The search included all study designs. Further studies were identified through manual searches of bibliographies and citations until no further relevant studies were identified. Two investigators (R.A. and V.S.) independently screened titles and abstracts, and selected all relevant citations for full-text review. Disagreement regarding study inclusion was resolved by discussion with a third investigator (H.A.).
Inclusion criteria
Studies that comprised a diagnostic accuracy assessment of a DL algorithm on medical imaging in human populations were eligible. Only studies that stated either diagnostic accuracy raw data, or sensitivity, specificity, AUC, NPV, PPV or accuracy data were included in the meta-analysis. No limitations were placed on the date range and the last search was performed in January 2020.
Exclusion criteria
Articles were excluded if the article was not written in English. Abstracts, conference articles, pre-prints, reviews and meta-analyses were not considered because an aim of this review was to appraise the methodology, reporting standards and quality of primary research studies being published in peer-reviewed journals. Studies that investigated the accuracy of image segmentation or predicting disease rather than identification or classification were excluded.
Data extraction and quality assessment
Two investigators (R.A. and V.S.) independently extracted demographic and diagnostic accuracy data from the studies, using a predefined electronic data extraction spreadsheet. The data fields were chosen subsequent to an initial scoping review and were, in the opinion of the investigators, sufficient to fulfil the aims of this review. Data were extracted on (i) first author, (ii) year of publication, (iii) type of neural network, (iv) population, (v) dataset—split into training, validation and test sets, (vi) imaging modality, (vii) body system/disease, (viii) internal/external validation methods, (ix) reference standard, (x) diagnostic accuracy raw data—true and false positives and negatives, (xi) percentages of AUC, accuracy, sensitivity, specificity, PPV, NPV and other metrics reported.
Three investigators (R.A., V.S. and GM) assessed study methodology using the QUADAS-2 checklist to evaluate the risk of bias and any applicability concerns of the studies118.
Data synthesis and analysis
A bivariate model for diagnostic meta-analysis was used to calculate summary estimates of sensitivity, specificity and AUC data137. Independent proportion and their differences were calculated and pooled through DerSimonian and Laird random-effects modelling138. This considered both between-study and within-study variances that contributed to study weighting. Study-specific estimates and 95% CIs were computed and represented on forest plots. Heterogeneity between studies was assessed using I2 (25–49% was considered to be low heterogeneity, 50–74% was moderate and >75% was high heterogeneity). Where raw diagnostic accuracy data were available, the SROC model was used to evaluate the relationship between sensitivity and specificity139. We utilised Stata version 15 (Stata Corp LP, College Station, TX, USA) for all statistical analyses.
We chose to appraise the performance of DL algorithms to identify individual disease or pathology patterns on different imaging modalities in isolation, e.g., identifying lung nodules on a thoracic CT scan. We felt that combining imaging modalities and diagnoses would add heterogeneity and variation to the analysis. Meta-analysis was only performed where there were greater than or equal to three patient cohorts, reporting for each specific pathology and imaging modality. This study is registered with PROSPERO, CRD42020167503.
Reporting summary
Further information on research design is available in the Nature Research Reporting Summary linked to this article.
Supplementary information
Acknowledgements
Infrastructure support for this research was provided by the NIHR Imperial Biomedical Research Centre (BRC).
Author contributions
H.A. conceptualised the study, R.A., V.S., G.M. and H.A. designed the study, extracted data, conducted the analysis and wrote the manuscript. D.S.W.T., A.K., D.K. and A.D. assisted in writing and editing the manuscript. All authors approved the final version of the manuscript and take accountability for all aspects of the work.
Data availability
The authors declare that all the data included in this study are available within the paper and its Supplementary Information files.
Competing interests
D.K. and A.K. are employees of Google Health. A.D. is an adviser at Google Health. D.S.W.T holds a patent on a deep learning system for the detection of retinal diseases.
Footnotes
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary information
The online version contains supplementary material available at 10.1038/s41746-021-00438-z.
References
- 1.LeCun Y, Bengio Y, Hinton G. Deep learning. Nature. 2015;521:436–444. doi: 10.1038/nature14539. [DOI] [PubMed] [Google Scholar]
- 2.Obermeyer Z, Emanuel EJ. Predicting the future — big data, machine learning, and clinical medicine. N. Engl. J. Med. 2016;375:1216–1219. doi: 10.1056/NEJMp1606181. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Esteva A, et al. A guide to deep learning in healthcare. Nat. Med. 2019;25:24–29. doi: 10.1038/s41591-018-0316-z. [DOI] [PubMed] [Google Scholar]
- 4.Litjens G, et al. A survey on deep learning in medical image analysis. Med. Image Anal. 2017;42:60–88. doi: 10.1016/j.media.2017.07.005. [DOI] [PubMed] [Google Scholar]
- 5.Bluemke DA, et al. Assessing radiology research on artificial intelligence: a brief guide for authors, reviewers, and readers—from the radiology editorial board. Radiology. 2020;294:487–489. doi: 10.1148/radiol.2019192515. [DOI] [PubMed] [Google Scholar]
- 6.Wahl B, Cossy-Gantner A, Germann S, Schwalbe NR. Artificial intelligence (AI) and global health: how can AI contribute to health in resource-poor settings? BMJ Glob. Health. 2018;3:e000798–e000798. doi: 10.1136/bmjgh-2018-000798. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Zhang L, Wang H, Li Q, Zhao M-H, Zhan Q-M. Big data and medical research in China. BMJ. 2018;360:j5910. doi: 10.1136/bmj.j5910. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Nakajima Y, Yamada K, Imamura K, Kobayashi K. Radiologist supply and workload: international comparison. Radiat. Med. 2008;26:455–465. doi: 10.1007/s11604-008-0259-2. [DOI] [PubMed] [Google Scholar]
- 9.Kelly CJ, Karthikesalingam A, Suleyman M, Corrado G, King D. Key challenges for delivering clinical impact with artificial intelligence. BMC Med. 2019;17:195. doi: 10.1186/s12916-019-1426-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Topol EJ. High-performance medicine: the convergence of human and artificial intelligence. Nat. Med. 2019;25:44–56. doi: 10.1038/s41591-018-0300-7. [DOI] [PubMed] [Google Scholar]
- 11.Benjamens S, Dhunnoo P, Meskó B. The state of artificial intelligence-based FDA-approved medical devices and algorithms: an online database. npj Digital Med. 2020;3:118. doi: 10.1038/s41746-020-00324-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Beam AL, Kohane IS. Big data and machine learning in health care. JAMA. 2018;319:1317–1318. doi: 10.1001/jama.2017.18391. [DOI] [PubMed] [Google Scholar]
- 13.Liu X, et al. A comparison of deep learning performance against health-care professionals in detecting diseases from medical imaging: a systematic review and meta-analysis. Lancet Digital Health. 2019;1:e271–e297. doi: 10.1016/S2589-7500(19)30123-2. [DOI] [PubMed] [Google Scholar]
- 14.Abràmoff MD, Lavin PT, Birch M, Shah N, Folk JC. Pivotal trial of an autonomous AI-based diagnostic system for detection of diabetic retinopathy in primary care offices. npj Digital Med. 2018;1:39. doi: 10.1038/s41746-018-0040-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Bellemo V, et al. Artificial intelligence using deep learning to screen for referable and vision-threatening diabetic retinopathy in Africa: a clinical validation study. Lancet Digital Health. 2019;1:e35–e44. doi: 10.1016/S2589-7500(19)30004-4. [DOI] [PubMed] [Google Scholar]
- 16.Christopher M, et al. Performance of deep learning architectures and transfer learning for detecting glaucomatous optic neuropathy in fundus photographs. Sci. Rep. 2018;8:16685. doi: 10.1038/s41598-018-35044-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Gulshan V, et al. Performance of a deep-learning algorithm vs manual grading for detecting diabetic retinopathy in India. JAMA Ophthalmol. 2019;137:987–993. doi: 10.1001/jamaophthalmol.2019.2004. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Keel S, Wu J, Lee PY, Scheetz J, He M. Visualizing deep learning models for the detection of referable diabetic retinopathy and glaucoma. JAMA Ophthalmol. 2019;137:288–292. doi: 10.1001/jamaophthalmol.2018.6035. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.Sandhu HS, et al. Automated diagnosis and grading of diabetic retinopathy using optical coherence tomography. Investig. Ophthalmol. Vis. Sci. 2018;59:3155–3160. doi: 10.1167/iovs.17-23677. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Zheng C, et al. Detecting glaucoma based on spectral domain optical coherence tomography imaging of peripapillary retinal nerve fiber layer: a comparison study between hand-crafted features and deep learning model. Graefes Arch. Clin. Exp. Ophthalmol. 2020;258:577–585. doi: 10.1007/s00417-019-04543-4. [DOI] [PubMed] [Google Scholar]
- 21.Kanagasingam Y, et al. Evaluation of artificial intelligence-based grading of diabetic retinopathy in primary care. JAMA Netw. Open. 2018;1:e182665–e182665. doi: 10.1001/jamanetworkopen.2018.2665. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22.Alqudah AM. AOCT-NET: a convolutional network automated classification of multiclass retinal diseases using spectral-domain optical coherence tomography images. Med. Biol. Eng. Comput. 2020;58:41–53. doi: 10.1007/s11517-019-02066-y. [DOI] [PubMed] [Google Scholar]
- 23.Asaoka R, et al. Validation of a deep learning model to screen for glaucoma using images from different fundus cameras and data augmentation. Ophthalmol. Glaucoma. 2019;2:224–231. doi: 10.1016/j.ogla.2019.03.008. [DOI] [PubMed] [Google Scholar]
- 24.Bhatia KK, et al. Disease classification of macular optical coherence tomography scans using deep learning software: validation on independent, multicenter data. Retina. 2020;40:1549–1557. doi: 10.1097/IAE.0000000000002640. [DOI] [PubMed] [Google Scholar]
- 25.Chan, G. C. Y. et al. Fusing results of several deep learning architectures for automatic classification of normal and diabetic macular edema in optical coherence tomography. In Conference proceedings: Annual International Conference of the IEEE Engineering in Medicine and Biology Society. IEEE Engineering in Medicine and Biology Society. Annual Conference, Vol. 2018, 670–673 (IEEE, 2018). [DOI] [PubMed]
- 26.Gargeya R, Leng T. Automated identification of diabetic retinopathy using deep learning. Ophthalmology. 2017;124:962–969. doi: 10.1016/j.ophtha.2017.02.008. [DOI] [PubMed] [Google Scholar]
- 27.Grassmann F, et al. A deep learning algorithm for prediction of age-related eye disease study severity scale for age-related macular degeneration from color fundus photography. Ophthalmology. 2018;125:1410–1420. doi: 10.1016/j.ophtha.2018.02.037. [DOI] [PubMed] [Google Scholar]
- 28.Gulshan V, et al. Development and validation of a deep learning algorithm for detection of diabetic retinopathy in retinal fundus photographs. JAMA. 2016;316:2402–2410. doi: 10.1001/jama.2016.17216. [DOI] [PubMed] [Google Scholar]
- 29.Hwang DK, et al. Artificial intelligence-based decision-making for age-related macular degeneration. Theranostics. 2019;9:232–245. doi: 10.7150/thno.28447. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30.Keel S, et al. Development and validation of a deep-learning algorithm for the detection of neovascular age-related macular degeneration from colour fundus photographs. Clin. Exp. Ophthalmol. 2019;47:1009–1018. doi: 10.1111/ceo.13575. [DOI] [PubMed] [Google Scholar]
- 31.Krause J, et al. Grader variability and the importance of reference standards for evaluating machine learning models for diabetic retinopathy. Ophthalmology. 2018;125:1264–1272. doi: 10.1016/j.ophtha.2018.01.034. [DOI] [PubMed] [Google Scholar]
- 32.Li F, et al. Automatic detection of diabetic retinopathy in retinal fundus photographs based on deep learning algorithm. Transl. Vis. Sci. Technol. 2019;8:4. doi: 10.1167/tvst.8.6.4. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33.Li Z, et al. An automated grading system for detection of vision-threatening referable diabetic retinopathy on the basis of color fundus photographs. Diabetes Care. 2018;41:2509–2516. doi: 10.2337/dc18-0147. [DOI] [PubMed] [Google Scholar]
- 34.Liu H, et al. Development and validation of a deep learning system to detect glaucomatous optic neuropathy using fundus photographs. JAMA Ophthalmol. 2019;137:1353–1360. doi: 10.1001/jamaophthalmol.2019.3501. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35.Liu S, et al. A deep learning-based algorithm identifies glaucomatous discs using monoscopic fundus photographs. Ophthalmol. Glaucoma. 2018;1:15–22. doi: 10.1016/j.ogla.2018.04.002. [DOI] [PubMed] [Google Scholar]
- 36.MacCormick IJC, et al. Accurate, fast, data efficient and interpretable glaucoma diagnosis with automated spatial analysis of the whole cup to disc profile. PLoS ONE. 2019;14:e0209409. doi: 10.1371/journal.pone.0209409. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37.Phene S, et al. Deep learning and glaucoma specialists: the relative importance of optic disc features to predict glaucoma referral in fundus photographs. Ophthalmology. 2019;126:1627–1639. doi: 10.1016/j.ophtha.2019.07.024. [DOI] [PubMed] [Google Scholar]
- 38.Ramachandran N, Hong SC, Sime MJ, Wilson GA. Diabetic retinopathy screening using deep neural network. Clin. Exp. Ophthalmol. 2018;46:412–416. doi: 10.1111/ceo.13056. [DOI] [PubMed] [Google Scholar]
- 39.Raumviboonsuk P, et al. Deep learning versus human graders for classifying diabetic retinopathy severity in a nationwide screening program. npj Digital Med. 2019;2:25. doi: 10.1038/s41746-019-0099-8. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40.Sayres R, et al. Using a deep learning algorithm and integrated gradients explanation to assist grading for diabetic retinopathy. Ophthalmology. 2019;126:552–564. doi: 10.1016/j.ophtha.2018.11.016. [DOI] [PubMed] [Google Scholar]
- 41.Ting DSW, et al. Development and validation of a deep learning system for diabetic retinopathy and related eye diseases using retinal images from multiethnic populations with diabetes. JAMA. 2017;318:2211–2223. doi: 10.1001/jama.2017.18152. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 42.Ting DSW, et al. Deep learning in estimating prevalence and systemic risk factors for diabetic retinopathy: a multi-ethnic study. npj Digital Med. 2019;2:24. doi: 10.1038/s41746-019-0097-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43.Verbraak FD, et al. Diagnostic accuracy of a device for the automated detection of diabetic retinopathy in a primary care setting. Diabetes Care. 2019;42:651. doi: 10.2337/dc18-0148. [DOI] [PubMed] [Google Scholar]
- 44.Van Grinsven MJ, van Ginneken B, Hoyng CB, Theelen T, Sánchez CI. Fast convolutional neural network training using selective data sampling: application to hemorrhage detection in color fundus images. IEEE Trans. Med. Imaging. 2016;35:1273–1284. doi: 10.1109/TMI.2016.2526689. [DOI] [PubMed] [Google Scholar]
- 45.Rogers TW, et al. Evaluation of an AI system for the automated detection of glaucoma from stereoscopic optic disc photographs: the European Optic Disc Assessment Study. Eye. 2019;33:1791–1797. doi: 10.1038/s41433-019-0510-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 46.Al-Aswad LA, et al. Evaluation of a deep learning system for identifying glaucomatous optic neuropathy based on color fundus photographs. J. Glaucoma. 2019;28:1029–1034. doi: 10.1097/IJG.0000000000001319. [DOI] [PubMed] [Google Scholar]
- 47.Brown JM, et al. Automated diagnosis of plus disease in retinopathy of prematurity using deep convolutional neural networks. JAMA Ophthalmol. 2018;136:803–810. doi: 10.1001/jamaophthalmol.2018.1934. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 48.Burlina P, et al. Utility of deep learning methods for referability classification of age-related macular degeneration. JAMA Ophthalmol. 2018;136:1305–1307. doi: 10.1001/jamaophthalmol.2018.3799. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 49.Burlina PM, et al. Automated grading of age-related macular degeneration from color fundus images using deep convolutional neural networks. JAMA Ophthalmol. 2017;135:1170–1176. doi: 10.1001/jamaophthalmol.2017.3782. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 50.Burlina P, Pacheco KD, Joshi N, Freund DE, Bressler NM. Comparing humans and deep learning performance for grading AMD: a study in using universal deep features and transfer learning for automated AMD analysis. Computers Biol. Med. 2017;82:80–86. doi: 10.1016/j.compbiomed.2017.01.018. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 51.De Fauw J, et al. Clinically applicable deep learning for diagnosis and referral in retinal disease. Nat. Med. 2018;24:1342–1350. doi: 10.1038/s41591-018-0107-6. [DOI] [PubMed] [Google Scholar]
- 52.Gómez-Valverde JJ, et al. Automatic glaucoma classification using color fundus images based on convolutional neural networks and transfer learning. Biomed. Opt. Express. 2019;10:892–913. doi: 10.1364/BOE.10.000892. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 53.Jammal AA, et al. Human versus machine: comparing a deep learning algorithm to human gradings for detecting glaucoma on fundus photographs. Am. J. Ophthalmol. 2019;211:123–131. doi: 10.1016/j.ajo.2019.11.006. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 54.Kermany DS, et al. Identifying medical diagnoses and treatable diseases by image-based deep learning. Cell. 2018;172:1122–1131.e1129. doi: 10.1016/j.cell.2018.02.010. [DOI] [PubMed] [Google Scholar]
- 55.Li F, et al. Deep learning-based automated detection of retinal diseases using optical coherence tomography images. Biomed. Opt. Express. 2019;10:6204–6226. doi: 10.1364/BOE.10.006204. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 56.Long E, et al. An artificial intelligence platform for the multihospital collaborative management of congenital cataracts. Nat. Biomed. Eng. 2017;1:0024. doi: 10.1038/s41551-016-0024. [DOI] [Google Scholar]
- 57.Matsuba S, et al. Accuracy of ultra-wide-field fundus ophthalmoscopy-assisted deep learning, a machine-learning technology, for detecting age-related macular degeneration. Int. Ophthalmol. 2019;39:1269–1275. doi: 10.1007/s10792-018-0940-0. [DOI] [PubMed] [Google Scholar]
- 58.Nagasato D, et al. Automated detection of a nonperfusion area caused by retinal vein occlusion in optical coherence tomography angiography images using deep learning. PLoS ONE. 2019;14:e0223965. doi: 10.1371/journal.pone.0223965. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 59.Peng Y, et al. DeepSeeNet: a deep learning model for automated classification of patient-based age-related macular degeneration severity from color fundus photographs. Ophthalmology. 2019;126:565–575. doi: 10.1016/j.ophtha.2018.11.015. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 60.Shibata N, et al. Development of a deep residual learning algorithm to screen for glaucoma from fundus photography. Sci. Rep. 2018;8:14665. doi: 10.1038/s41598-018-33013-w. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 61.Zhang Y, et al. Development of an automated screening system for retinopathy of prematurity using a deep neural network for wide-angle retinal images. IEEE Access. 2019;7:10232–10241. doi: 10.1109/ACCESS.2018.2881042. [DOI] [Google Scholar]
- 62.Becker AS, et al. Classification of breast cancer in ultrasound imaging using a generic deep learning analysis software: a pilot study. Br. J. Radio. 2018;91:20170576. doi: 10.1259/bjr.20170576. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 63.Zhang C, et al. Toward an expert level of lung cancer detection and classification using a deep convolutional neural network. Oncologist. 2019;24:1159–1165. doi: 10.1634/theoncologist.2018-0908. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 64.Ardila D, et al. End-to-end lung cancer screening with three-dimensional deep learning on low-dose chest computed tomography. Nat. Med. 2019;25:954–961. doi: 10.1038/s41591-019-0447-x. [DOI] [PubMed] [Google Scholar]
- 65.Hwang EJ, et al. Deep learning for chest radiograph diagnosis in the emergency department. Radiology. 2019;293:573–580. doi: 10.1148/radiol.2019191225. [DOI] [PubMed] [Google Scholar]
- 66.Hwang EJ, et al. Development and validation of a deep learning–based automated detection algorithm for major thoracic diseases on chest radiographs. JAMA Netw. Open. 2019;2:e191095–e191095. doi: 10.1001/jamanetworkopen.2019.1095. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 67.Hwang EJ, et al. Development and validation of a deep learning–based automatic detection algorithm for active pulmonary tuberculosis on chest radiographs. Clin. Infect. Dis. 2018 doi: 10.1093/cid/ciy967. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 68.Liang CH, et al. Identifying pulmonary nodules or masses on chest radiography using deep learning: external validation and strategies to improve clinical practice. Clin. Radiol. 2020;75:38–45. doi: 10.1016/j.crad.2019.08.005. [DOI] [PubMed] [Google Scholar]
- 69.Nam JG, et al. Development and validation of deep learning–based automatic detection algorithm for malignant pulmonary nodules on chest radiographs. Radiology. 2018;290:218–228. doi: 10.1148/radiol.2018180237. [DOI] [PubMed] [Google Scholar]
- 70.Qin ZZ, et al. Using artificial intelligence to read chest radiographs for tuberculosis detection: A multi-site evaluation of the diagnostic accuracy of three deep learning systems. Sci. Rep. 2019;9:15000. doi: 10.1038/s41598-019-51503-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 71.Setio AAA, et al. Pulmonary nodule detection in CT images: false positive reduction using multi-view convolutional networks. IEEE Trans. Med. Imaging. 2016;35:1160–1169. doi: 10.1109/TMI.2016.2536809. [DOI] [PubMed] [Google Scholar]
- 72.Sim Y, et al. Deep convolutional neural network–based software improves radiologist detection of malignant lung nodules on chest radiographs. Radiology. 2020;294:199–209. doi: 10.1148/radiol.2019182465. [DOI] [PubMed] [Google Scholar]
- 73.Taylor AG, Mielke C, Mongan J. Automated detection of moderate and large pneumothorax on frontal chest X-rays using deep convolutional neural networks: a retrospective study. PLOS Med. 2018;15:e1002697. doi: 10.1371/journal.pmed.1002697. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 74.Uthoff, J. et al. Machine learning approach for distinguishing malignant and benign lung nodules utilizing standardized perinodular parenchymal features from CT. Med. Phys.46, 3207–3216 (2019). [DOI] [PMC free article] [PubMed]
- 75.Zech JR, et al. Variable generalization performance of a deep learning model to detect pneumonia in chest radiographs: a cross-sectional study. PLOS Med. 2018;15:e1002683. doi: 10.1371/journal.pmed.1002683. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 76.Cha MJ, Chung MJ, Lee JH, Lee KS. Performance of deep learning model in detecting operable lung cancer with chest radiographs. J. Thorac. Imaging. 2019;34:86–91. doi: 10.1097/RTI.0000000000000388. [DOI] [PubMed] [Google Scholar]
- 77.Chae KJ, et al. Deep learning for the classification of small (≤2 cm) pulmonary nodules on ct imaging: a preliminary study. Acad. Radiol. 2020;27:E55–E63. doi: 10.1016/j.acra.2019.05.018. [DOI] [PubMed] [Google Scholar]
- 78.Ciompi F, et al. Towards automatic pulmonary nodule management in lung cancer screening with deep learning. Sci. Rep. 2017;7:46479. doi: 10.1038/srep46479. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 79.Dunnmon JA, et al. Assessment of convolutional neural networks for automated classification of chest radiographs. Radiology. 2018;290:537–544. doi: 10.1148/radiol.2018181422. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 80.Li X, et al. Deep learning-enabled system for rapid pneumothorax screening on chest CT. Eur. J. Radiol. 2019;120:108692. doi: 10.1016/j.ejrad.2019.108692. [DOI] [PubMed] [Google Scholar]
- 81.Li L, Liu Z, Huang H, Lin M, Luo D. Evaluating the performance of a deep learning-based computer-aided diagnosis (DL-CAD) system for detecting and characterizing lung nodules: comparison with the performance of double reading by radiologists. Thorac. Cancer. 2019;10:183–192. doi: 10.1111/1759-7714.12931. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 82.Majkowska A, et al. Chest radiograph interpretation with deep learning models: assessment with radiologist-adjudicated reference standards and population-adjusted evaluation. Radiology. 2019;294:421–431. doi: 10.1148/radiol.2019191293. [DOI] [PubMed] [Google Scholar]
- 83.Park S, et al. Deep learning-based detection system for multiclass lesions on chest radiographs: comparison with observer readings. Eur. Radiol. 2019;30:1359–1368. doi: 10.1007/s00330-019-06532-x. [DOI] [PubMed] [Google Scholar]
- 84.Patel BN, et al. Human–machine partnership with artificial intelligence for chest radiograph diagnosis. npj Digital Med. 2019;2:111. doi: 10.1038/s41746-019-0189-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 85.Rajpurkar P, et al. Deep learning for chest radiograph diagnosis: a retrospective comparison of the CheXNeXt algorithm to practicing radiologists. PLOS Med. 2018;15:e1002686. doi: 10.1371/journal.pmed.1002686. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 86.Singh R, et al. Deep learning in chest radiography: detection of findings and presence of change. PLoS ONE. 2018;13:e0204155. doi: 10.1371/journal.pone.0204155. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 87.Walsh SLF, Calandriello L, Silva M, Sverzellati N. Deep learning for classifying fibrotic lung disease on high-resolution computed tomography: a case-cohort study. Lancet Respir. Med. 2018;6:837–845. doi: 10.1016/S2213-2600(18)30286-8. [DOI] [PubMed] [Google Scholar]
- 88.Wang S, et al. 3D convolutional neural network for differentiating pre-invasive lesions from invasive adenocarcinomas appearing as ground-glass nodules with diameters ≤3 cm using HRCT. Quant. Imaging Med. Surg. 2018;8:491–499. doi: 10.21037/qims.2018.06.03. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 89.Park S, et al. Application of deep learning-based computer-aided detection system: detecting pneumothorax on chest radiograph after biopsy. Eur. Radio. 2019;29:5341–5348. doi: 10.1007/s00330-019-06130-x. [DOI] [PubMed] [Google Scholar]
- 90.Lakhani P, Sundaram B. Deep learning at chest radiography: automated classification of pulmonary tuberculosis by using convolutional neural networks. Radiology. 2017;284:574–582. doi: 10.1148/radiol.2017162326. [DOI] [PubMed] [Google Scholar]
- 91.Becker AS, et al. Deep learning in mammography: diagnostic accuracy of a multipurpose image analysis software in the detection of breast cancer. Investig. Radio. 2017;52:434–440. doi: 10.1097/RLI.0000000000000358. [DOI] [PubMed] [Google Scholar]
- 92.Ciritsis A, et al. Automatic classification of ultrasound breast lesions using a deep convolutional neural network mimicking human decision-making. Eur. Radio. 2019;29:5458–5468. doi: 10.1007/s00330-019-06118-7. [DOI] [PubMed] [Google Scholar]
- 93.Cogan T, Cogan M, Tamil L. RAMS: remote and automatic mammogram screening. Comput. Biol. Med. 2019;107:18–29. doi: 10.1016/j.compbiomed.2019.01.024. [DOI] [PubMed] [Google Scholar]
- 94.McKinney SM, et al. International evaluation of an AI system for breast cancer screening. Nature. 2020;577:89–94. doi: 10.1038/s41586-019-1799-6. [DOI] [PubMed] [Google Scholar]
- 95.Peng W, Mayorga RV, Hussein EMA. An automated confirmatory system for analysis of mammograms. Comput. Methods Prog. Biomed. 2016;125:134–144. doi: 10.1016/j.cmpb.2015.09.019. [DOI] [PubMed] [Google Scholar]
- 96.Ribli D, Horváth A, Unger Z, Pollner P, Csabai I. Detecting and classifying lesions in mammograms with deep learning. Sci. Rep. 2018;8:4165. doi: 10.1038/s41598-018-22437-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 97.Rodríguez-Ruiz A, et al. Detection of Breast cancer with mammography: effect of an artificial intelligence support system. Radiology. 2018;290:305–314. doi: 10.1148/radiol.2018181371. [DOI] [PubMed] [Google Scholar]
- 98.Rodriguez-Ruiz A, et al. Stand-alone artificial intelligence for breast cancer detection in mammography: comparison with 101 radiologists. J. Natl Cancer Inst. 2019;111:916–922. doi: 10.1093/jnci/djy222. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 99.Byra M, et al. Breast mass classification in sonography with transfer learning using a deep convolutional neural network and color conversion. Med. Phys. 2019;46:746–755. doi: 10.1002/mp.13361. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 100.Choi JS, et al. Effect of a deep learning framework-based computer-aided diagnosis system on the diagnostic performance of radiologists in differentiating between malignant and benign masses on breast ultrasonography. Korean J. Radio. 2019;20:749–758. doi: 10.3348/kjr.2018.0530. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 101.Dalmis MU, et al. Artificial intelligence–based classification of breast lesions imaged with a multiparametric breast mri protocol with ultrafast DCE-MRI, T2, and DWI. Investig. Radiol. 2019;54:325–332. doi: 10.1097/RLI.0000000000000544. [DOI] [PubMed] [Google Scholar]
- 102.Fujioka T, et al. Distinction between benign and malignant breast masses at breast ultrasound using deep learning method with convolutional neural network. Jpn J. Radio. 2019;37:466–472. doi: 10.1007/s11604-019-00831-5. [DOI] [PubMed] [Google Scholar]
- 103.Kim SM, et al. A comparison of logistic regression analysis and an artificial neural network using the BI-RADS Lexicon for ultrasonography in conjunction with introbserver variability. J. Digital Imaging. 2012;25:599–606. doi: 10.1007/s10278-012-9457-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 104.Truhn D, et al. Radiomic versus convolutional neural networks analysis for classification of contrast-enhancing lesions at multiparametric breast MRI. Radiology. 2019;290:290–297. doi: 10.1148/radiol.2018181352. [DOI] [PubMed] [Google Scholar]
- 105.Wu N, et al. Deep neural networks improve radiologists’ performance in breast cancer screening. IEEE Trans. Med. Imaging. 2020;39:1184–1194. doi: 10.1109/TMI.2019.2945514. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 106.Yala A, Schuster T, Miles R, Barzilay R, Lehman C. A deep learning model to triage screening mammograms: a simulation study. Radiology. 2019;293:38–46. doi: 10.1148/radiol.2019182908. [DOI] [PubMed] [Google Scholar]
- 107.Zhou J, et al. Weakly supervised 3D deep learning for breast cancer classification and localization of the lesions in MR images. J. Magn. Reson. Imaging. 2019;50:1144–1151. doi: 10.1002/jmri.26721. [DOI] [PubMed] [Google Scholar]
- 108.Li Z, et al. Efficacy of a deep learning system for detecting glaucomatous optic neuropathy based on color fundus photographs. Ophthalmology. 2018;125:1199–1206. doi: 10.1016/j.ophtha.2018.01.023. [DOI] [PubMed] [Google Scholar]
- 109.Lin H, et al. Diagnostic efficacy and therapeutic decision-making capacity of an artificial intelligence platform for childhood cataracts in eye clinics: a multicentre randomized controlled trial. EClinicalMedicine. 2019;9:52–59. doi: 10.1016/j.eclinm.2019.03.001. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 110.Annarumma M, et al. Automated triaging of adult chest radiographs with deep artificial neural networks. Radiology. 2019;291:196–202. doi: 10.1148/radiol.2018180921. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 111.Yala A, Lehman C, Schuster T, Portnoi T, Barzilay R. A deep learning mammography-based model for improved breast cancer risk prediction. Radiology. 2019;292:60–66. doi: 10.1148/radiol.2019182716. [DOI] [PubMed] [Google Scholar]
- 112.Sedgwick P. Meta-analyses: how to read a funnel plot. BMJ. 2013;346:f1342. doi: 10.1136/bmj.f1342. [DOI] [PubMed] [Google Scholar]
- 113.Herent P, et al. Detection and characterization of MRI breast lesions using deep learning. Diagn. Inter. Imaging. 2019;100:219–225. doi: 10.1016/j.diii.2019.02.008. [DOI] [PubMed] [Google Scholar]
- 114.Bossuyt PM, et al. STARD 2015: an updated list of essential items for reporting diagnostic accuracy studies. BMJ. 2015;351:h5527. doi: 10.1136/bmj.h5527. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 115.Sounderajah V, et al. Developing specific reporting guidelines for diagnostic accuracy studies assessing AI interventions: the STARD-AI Steering Group. Nat. Med. 2020;26:807–808. doi: 10.1038/s41591-020-0941-1. [DOI] [PubMed] [Google Scholar]
- 116.Vollmer S, et al. Machine learning and artificial intelligence research for patient benefit: 20 critical questions on transparency, replicability, ethics, and effectiveness. BMJ. 2020;368:l6927. doi: 10.1136/bmj.l6927. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 117.Liu X, et al. Reporting guidelines for clinical trial reports for interventions involving artificial intelligence: the CONSORT-AI extension. Nat. Med. 2020;26:1364–1374. doi: 10.1038/s41591-020-1034-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 118.Whiting PF, et al. QUADAS-2: a revised tool for the quality assessment of diagnostic accuracy studies. Ann. Intern. Med. 2011;155:529–536. doi: 10.7326/0003-4819-155-8-201110180-00009. [DOI] [PubMed] [Google Scholar]
- 119.Food, U. & Administration, D. Artificial Intelligence and Machine Learning in Software as a Medical Device (US Food and Drug Administratio, 2019).
- 120.Titano JJ, et al. Automated deep-neural-network surveillance of cranial images for acute neurologic events. Nat. Med. 2018;24:1337–1341. doi: 10.1038/s41591-018-0147-y. [DOI] [PubMed] [Google Scholar]
- 121.Rankin D, et al. Reliability of supervised machine learning using synthetic data in health care: Model to preserve privacy for data sharing. JMIR Med. Inform. 2020;8:e18910. doi: 10.2196/18910. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 122.Cawley GC, Talbot NL. On over-fitting in model selection and subsequent selection bias in performance evaluation. J. Mach. Learn. Res. 2010;11:2079–2107. [Google Scholar]
- 123.Blalock, D., Ortiz, J., Frankle, J. & Guttag, J. What is the state of neural network pruning? Preprint at https://arxiv.org/abs/2003.03033 (2020).
- 124.Beam AL, Manrai AK, Ghassemi M. Challenges to the reproducibility of machine learning models in health care. JAMA. 2020;323:305–306. doi: 10.1001/jama.2019.20866. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 125.Celi LA, et al. Bridging the health data divide. J. Med. Internet Res. 2016;18:e325. doi: 10.2196/jmir.6400. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 126.Shah P, et al. Artificial intelligence and machine learning in clinical development: a translational perspective. npj Digital Med. 2019;2:69. doi: 10.1038/s41746-019-0148-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 127.Collins GS, Reitsma JB, Altman DG, Moons KG. Transparent reporting of a multivariable prediction model for individual prognosis or diagnosis (TRIPOD): the TRIPOD statement. BMJ. 2015;350:g7594. doi: 10.1136/bmj.g7594. [DOI] [PubMed] [Google Scholar]
- 128.Schulz KF, Altman DG, Moher D. CONSORT 2010 Statement: updated guidelines for reporting parallel group randomised trials. BMJ. 2010;340:c332. doi: 10.1136/bmj.c332. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 129.Chan A-W, et al. SPIRIT 2013 statement: defining standard protocol items for clinical trials. Ann. Intern. Med. 2013;158:200–207. doi: 10.7326/0003-4819-158-3-201302050-00583. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 130.Cruz Rivera S, et al. Guidelines for clinical trial protocols for interventions involving artificial intelligence: the SPIRIT-AI extension. Nat. Med. 2020;26:1351–1363. doi: 10.1038/s41591-020-1037-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 131.Collins GS, Moons KG. Reporting of artificial intelligence prediction models. Lancet. 2019;393:1577–1579. doi: 10.1016/S0140-6736(19)30037-6. [DOI] [PubMed] [Google Scholar]
- 132.Ngiam KY, Khor IW. Big data and machine learning algorithms for health-care delivery. Lancet Oncol. 2019;20:e262–e273. doi: 10.1016/S1470-2045(19)30149-4. [DOI] [PubMed] [Google Scholar]
- 133.World Health Organization. Big Data and Artificial Intelligence for Achieving Universal Health Coverage: an International Consultation on Ethics: Meeting Report, 12–13 October 2017 (World Health Organization, 2018).
- 134.Cath C, Wachter S, Mittelstadt B, Taddeo M, Floridi L. Artificial Intelligence and the ‘Good Society’: the US, EU, and UK approach. Sci. Eng. Ethics. 2018;24:505–528. doi: 10.1007/s11948-017-9901-7. [DOI] [PubMed] [Google Scholar]
- 135.Mittelstadt B. The ethics of biomedical ‘Big Data’ analytics. Philos. Technol. 2019;32:17–21. doi: 10.1007/s13347-019-00344-z. [DOI] [Google Scholar]
- 136.McInnes MDF, et al. Preferred reporting items for a systematic review and meta-analysis of diagnostic test accuracy studies: the PRISMA-DTA statement. JAMA. 2018;319:388–396. doi: 10.1001/jama.2017.19163. [DOI] [PubMed] [Google Scholar]
- 137.Reitsma JB, et al. Bivariate analysis of sensitivity and specificity produces informative summary measures in diagnostic reviews. J. Clin. Epidemiol. 2005;58:982–990. doi: 10.1016/j.jclinepi.2005.02.022. [DOI] [PubMed] [Google Scholar]
- 138.DerSimonian R, Laird N. Meta-analysis in clinical trials. Controlled Clin. Trials. 1986;7:177–188. doi: 10.1016/0197-2456(86)90046-2. [DOI] [PubMed] [Google Scholar]
- 139.Jones CM, Ashrafian H, Darzi A, Athanasiou T. Guidelines for diagnostic tests and diagnostic accuracy in surgical research. J. Investig. Surg. 2010;23:57–65. doi: 10.3109/08941930903469508. [DOI] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
The authors declare that all the data included in this study are available within the paper and its Supplementary Information files.

