Abstract
Background: Opportunistic prostate cancer (PCa) screening is a controversial topic. Magnetic resonance imaging (MRI) has proven to detect prostate cancer with a high sensitivity and specificity, leading to the idea to perform an image-guided prostate cancer (PCa) screening; Methods: We evaluated a prospectively enrolled cohort of 49 healthy men participating in a dedicated image-guided PCa screening trial employing a biparametric MRI (bpMRI) protocol consisting of T2-weighted (T2w) and diffusion weighted imaging (DWI) sequences. Datasets were analyzed both by human readers and by a fully automated artificial intelligence (AI) software using deep learning (DL). Agreement between the algorithm and the reports—serving as the ground truth—was compared on a per-case and per-lesion level using metrics of diagnostic accuracy and k statistics; Results: The DL method yielded an 87% sensitivity (33/38) and 50% specificity (5/10) with a k of 0.42. 12/28 (43%) Prostate Imaging Reporting and Data System (PI-RADS) 3, 16/22 (73%) PI-RADS 4, and 5/5 (100%) PI-RADS 5 lesions were detected compared to the ground truth. Targeted biopsy revealed PCa in six participants, all correctly diagnosed by both the human readers and AI. Conclusions: The results of our study show that in our AI-assisted, image-guided prostate cancer screening the software solution was able to identify highly suspicious lesions and has the potential to effectively guide the targeted-biopsy workflow.
Keywords: prostatic neoplasms, early detection of cancer, magnetic resonance imaging, deep learning
1. Introduction
Opportunistic prostate cancer (PCa) screening is a controversial topic in the urological literature. Large prostate-specific antigen (PSA)-based screening programs in Europe (European Randomised Study of Screening for Prostate Cancer (ERSPC)) and the U.S. (Prostate Cancer Screening in the Randomized Prostate, Lung, Colorectal, and Ovarian Cancer Screening Trial (PLCO)) were able demonstrate that early diagnosis and early treatment can help to reduce prostate-cancer-specific mortality [1,2]. However, screening programs are associated with relevant rates of overdiagnosis of 27–56% [3] and overtreatment of clinically insignificant cancers [4]. Furthermore, between 15% and 44% of biopsy-proven cancer occurs in patients with PSA levels below 4 ng/mL, which represents the accepted cutoff value to perform prostate biopsy [5]. Omitting biopsy due to low PSA level can ultimately lead to missing clinically relevant cancer.
For more than a decade, magnetic resonance imaging (MRI) has been established as a powerful tool for prostate cancer diagnosis. The PROMIS study has demonstrated that prostate MRI is a suitable triage tool for biopsy-naïve men, reducing the number of unnecessary biopsies by a quarter while improving the detection of clinically significant cancer [6]. The PRECISION study randomized patients to either systematic biopsies or MRI with no biopsy if MRI was negative, and targeted biopsy if MRI was positive. Targeted biopsies guided by MRI detected significantly more clinically significant cancers while reducing the number of clinically insignificant cancers [7]. Because of these findings, MRI for prostate cancer diagnosis has been integrated into established guidelines [8].
In order to overcome the outlined weaknesses of PSA-based prostate cancer screening programs, the incorporation of the diagnostic strengths of imaging techniques into an image-guided prostate cancer screening, analogous to breast cancer screening, has already been discussed in the literature [9,10].
The dilemma of an increased workload due to such an image-guided, opportunistic prostate cancer screening may be addressed with technological advances: (i) shortening scan protocols and (ii) automatization of the image acquisition and reporting processes. There is growing evidence [11,12,13] that biparametric MRI (bpMRI) protocols consisting of T2w and diffusion weighted imaging yield a similar diagnostic performance compared to the conventional multiparametric MRI (mpMRI) approach while reducing scan times in selected patients to as low as 5 min is theoretically possible [14]. Embedding these standardized imaging sequences in automated acquisition and processing environments represents an essential step to make prostate MR imaging a best clinical practice tool for either screening or diagnostic procedures. Concerning the automatization of reporting, artificial intelligence (AI)-supported workflows have been shown to achieve similar performances in detecting suspicious lesions in prostate MRI examinations compared to human readers [15] and to provide valuable assistance if used as a concurrent reader [16].
This work focuses on specific cohort: a true MRI screening population consisting of healthy, biopsy-naïve men enrolled in a prospective trial. The purpose of this trial was to evaluate the efforts and resources required to implement a solely bpMRI-based prostate cancer screening program and to employ a state-of-the-art deep learning for detection and classification purposes in order to test the capabilities of this technology for automatization.
Therefore, we investigate whether a deep-learning-based algorithm in combination with biparametric imaging can be used for detection and classification of Prostate Imaging Reporting and Data System (PI-RADS) lesions in asymptomatic men enrolled in this prospective, MRI-based prostate cancer screening trial. We hypothesize that a deep-learning-based algorithm would provide a high-accuracy solution, potentially allowing to integrate this technology in an image-guided screening workflow.
2. Materials and Methods
2.1. Screening Program and Prospective Trial Information
The study was approved by the local ethics committees (ethics committee Northwest and Central Switzerland; EKNZ 2018-01965, approved: 26 November 2018) and all patients gave informed consent. Participants were prospectively enrolled in a national registered trial (NCT03749993). The primary purpose of this trial is to evaluate the efforts and resources required to implement a solely bpMRI-based prostate cancer screening program. The presented results are part of a post hoc analysis. Participants were included when the following inclusion criteria were met: biopsy-naïve men >45 years with no history of or suspicion for prostate cancer and a life expectancy >10 years. Exclusion criteria comprised the following: acute urinary tract infection, clinical suspicion of severe voiding disorders and/or chronic inflammation of the prostate, and contraindications for MRI examinations. For the present proof-of-concept investigation, participants were eligible for inclusion during a 6-month observation period ranging from December 2018 to June 2019. Within this time period, 49 participants were recruited and enrolled into the study population.
2.2. Artificial Intelligence Software Solution
A deep-learning-based, not commercially available, prototype AI solution (ProstateAI, Siemens Healthineers, Erlangen, Germany, termed in the following text: algorithm) was used for fully automatic prostate lesion detection and classification. A detailed visualization of the network architecture can be found in Figure 1. As illustrated in Figure 2, ProstateAI contains two parts: a preprocessing pipeline and a deep-learning-based lesion detection and classification component. The preprocessing pipeline directly takes the acquired bpMRI sequences and generates the required well-formatted and transformed data volumes. In particular, the preprocessing pipeline first parses and filters the acquired Digital Imaging and Communications in Medicine (DICOM) files loading only the T2w and diffusion weighted imaging (DWI) series. From DWI series, a logarithmic extrapolation method [17] is adopted to compute a new DWI volume with b-value of 2000 s/mm2. This step can simultaneously eliminate the b-value variances among the datasets and also improve lesion detection performance [18]. Moreover, apparent diffusion coefficient (ADC) maps are computed. Next, whole-organ gland segmentation is performed on T2w volumes using a learning-based method as presented in [19]. After segmentation, a rigid registration [20] is conducted to align all other sequences (DWI-2000 and ADC) to T2WI.
ProstateAI then automatically detects clinically relevant lesions and classifies each detected lesion according to PI-RADS categories. This is achieved by a sequence of coupled deep neural networks that were trained separately. First, a fully convolutional localization net (Candidate Localization Network in Figure 1) is able to generate a semantic lesion candidate heatmap (see examples in D1 and D2 of Figure 3); then, a sub-volume-based, false-positive-reduction net (Candidate Qualification Network in Figure 1) further improves the detection accuracy by removing false positives; last, another sub-volume-based PI-RADS scoring net (Classification Network in Figure 1) stages the level of malignancy for each detection by assigning them to the corresponding PI-RADS categories. The detailed description of the architectures of Candidate Localization and Candidate Qualification Networks can be found in [21].
For this study, the algorithm was trained using 2170 bpMRI prostate examinations consisting of 944 lesion-free cases and 1226 positive cases; all of which had lesion-based PI-RADS information and pixel-based annotations of the lesion boundaries. The anonymized datasets were acquired and labeled at eight different institutions; each institutions’ review board (IRB) provided either exemptions from further review—due to the anonymized nature of the data sets—or full-board IRB approval after review. Using a multicenter approach with standardized reporting, data inhomogeneity for efficient AI training was ensured. Furthermore, we obtained the central processing unit (CPU)-based computational time per case, including the preprocessing and deep-learning component. The CPU-based approach has been chosen in order to most accurately simulate a clinical environment. In this study, we used an Intel® Core™ i7-8850H CPU@2.60 GHz.
2.3. MRI Examination
All MRI examinations were performed on a single 3T scanner (MAGNETOM Prisma, Siemens Healthineers, Erlangen, Germany, see Table 1 for a detailed study sequence description). The sequences were embedded into a day optimizing throughput (DOT) workflow, which automatically centers the prostate in the field of view, adapts the size of the field of view, and performs a three-dimensional correction of spatial axes. After coil placement, the DOT workflow does not require further adaptations by technicians while at any time allowing interruptions of the scan process. The total scan time per patient from the start of the first sequence to the end of the last sequence was ~9 min and 30 s, while the scan time for the workflow-relevant series (T2w turbo-spin echo (TSE) tra and DWI) was ~6 min and 20 s. A detailed workflow visualization is outlined in Figure 2.
Table 1.
Parameter | Localizer | T2w tra (TSE) | DWI | T1w tra (TSE) |
---|---|---|---|---|
TA (in min) | 0.13 | 3:37 | 2:43 | 2:45 |
TR (in ms) | 3.51 | 7000 | 3500 | 700 |
TE (in ms) | 1.53 | 104 | 66 | 12 |
ST (in mm) | 6 | 3 | 3 | 2 |
Voxel size (in mm) | 1.6 × 1.6 × 6.0 | 0.3 × 0.3 × 3.0 | 0.9 × 0.9 × 3.0 | 0.4 × 0.4 × 3.0 |
AF | / | 3 | / | 3 |
b-values (in s/mm2) | / | / | 0, 800 | / |
Note that T2-weighted coronal and sagittal orientations were re-sliced from the transversal orientation. TA = acquisition time, TR = repetition time, TE = echo time, ST = slice thickness, AF = acceleration factor, TSE = turbo-spin echo, DWI = diffusion-weighted imaging.
All participants underwent a routine clinical reading process in an academic institution with two board-certified radiologists with at least 5 years of experience in prostate imaging reading the cases; at our institution, all prostate imaging studies are read by two independent radiologists as consensus read and the resultant report is highly structured—as suggested by the PI-RADS v2.0 guidelines [22]—including: number of suspicious lesions separated for the peripheral zone (PZ) and transition zone (TZ) with exact locations using series and image number descriptions and reference to the PI-RADS sector map, and PI-RADS score/lesion based on the PI-RADS assessment per zone; especially, the index lesion with the highest PI-RADS score specifically was highlighted per zone.
2.4. Histopatholgical Analysis
The biopsies were prepared in the following standardized manner by a uropathologist with more than 3 years subspecialty experience: After fixation in 10% buffered formalin, the biopsy probes were embedded in paraffin wax and sectioned and stained with hematoxylin and eosin according to the pathology committee of the European Randomized Study of Screening for Prostate Cancer. Every biopsy with pathologic prostate parenchyma was attributed a specific Gleason grade on the basis of the underlying glandular pattern. Biopsies with benign prostatic tissue were graded as “normal” if anatomically adjacent cores were tumor free and, additionally, did not show any significant signs of inflammation.
2.5. Comparison between Ground Truth and AI
For all lesions with a PI-RADS lesion score of ≥3, targeted transrectal MRI-TRUS Fusion biopsies were performed by a board-certified urologist. Systematic biopsies were not performed. A minimum of 3 cores per lesion were obtained (median number of cores: 3, range: 2–5).
In order to evaluate the agreement between the ground truth extracted from the written reports with histopathologic correlation and the automatically computed output of the AI software solution, every dataset was manually annotated by a radiology fellow (D.J.W) using a proprietary software (Annotator Tool, V03_B41). The written report contained detailed information about the reported lesions, especially mentioning series numbers on either T2w or ADC with accompanying image numbers and were further visualized as lesions in a PI-RADS sector map. Each lesion was then carefully identified on the DWI series and corresponding ADC maps, using the T2-weighted images as morphological reference. All confirmed lesions were subsequently segmented three-dimensionally on the T2-weighted images in a slice-by-slice fashion and labeled according to the PI-RADS assessment score from 3 to 5 on the ADC map for PZ lesions. In order to evaluate the agreement between the radiologists who identified the lesions in the academic reading process and the annotating fellow, all annotations were reviewed by a senior radiologist (D.T.B.).
The criterion for a true positive labeling of a lesion, defined by the PI-RADS assessment score, was that the detection point of the deep learning software was localized less than 10 mm away from the centroid of the lesion in the annotation in 3D, analogously to (9). False negatives (FN) were defined as lesions annotated by the human reader but not detected by the AI algorithm.
2.6. Statistical Testing
Sensitivity, specificity, positive predictive value (PPV), and negative predictive value (NPV) on a case level were calculated based on the algorithm’s results compared to the histologically proven ground truth defined by the radiologists’ reports. A case was classified as positive when a PI-RADS lesions ≥3 was mentioned in the imaging report. Accordingly, a case was evaluated as negative when only PI-RADS 1 and 2 scores were reported. On a per-lesion and index level, sensitivities were calculated. Kappa statistics were applied to compare the agreement concerning the PI-RADS classification. The k coefficients were assessed as follows: 0.01–020, slight agreement; 0.21–0.40, fair agreement; 0.41–0.60, moderate agreement; 0.61–0.80, substantial agreement; and 0.81–0.99, almost perfect agreement. p values <0.05 were considered significant. All statistical evaluations were performed using Python (version 3.5, Python Software Foundation; https://www.python.org/).
3. Results
From the 49 participants, 1 participant was excluded due to distortion artifacts from the gas-filled rectum and consecutive failure of the image registration between the T2w sequence and ADC map. In total, 48 screening cases were included in the analysis (see Figure 4). The mean age ± standard deviation was 58 ± 8 years (range: 45–75) and the mean PSA value was 2.68 ± 5.48 µg/mL (median: 1.07). The demographic and clinical information is summarized in Table 2. Detailed information concerning the metrics of diagnostic performance can be found in Table 3. The mean CPU-based computational time including image preprocessing, lesion detection, and classification was 14 s per case. All 3D lesion annotations done by the fellow were confirmed by another senior radiologist.
Table 2.
Variable | Participants (n = 48) |
---|---|
Demographic and clinical information | |
Mean age (mean ± SD, in years) | 58 ± 8 (range: 45–75) |
PSA in µg/mL (median, mean ± SD) | 1.07, 2.68 ± 5.48 |
Algorithm detection rate * | |
PI-RADS 1 and 2 | 5/10 (50) |
PI-RADS 3 | 14/18 (78) |
PI-RADS 4 | 14/15 (93) |
PI-RADS 5 | 5/5 (100) |
Positive histopathology results per PI-RADS score | |
PI-RADS 3 | 1/28 (3) |
PI-RADS 4 | 2/22 (9) |
PI-RADS 5 | 3/5 (60) |
* Data represent results on a case level with counts for PI-RADS 3, 4, and 5 categories representing the index lesion score per patient and nested PI-RADS 1 and 2 categories if no or clearly benign lesions were reported. Values in parentheses are percentages.
Table 3.
Parameter | Sensitivity (%) | Specificity (%) | PPV (%) | NPV (%) | Kappa k |
---|---|---|---|---|---|
Case level | 87 | 50 | 87 | 50 | 0.42 |
Lesion level | |||||
PI-RADS category 3 | 43 | / | / | / | / |
PI-RADS category 4 | 73 | / | / | / | / |
PI-RADS category 5 | 100 | / | / | / | / |
Combined PI-RADS category 4 and 5 | 78 | / | / | / | / |
PI-RADS = Prostate Imaging Reporting and Data System, PPV = positive predictive value, NPV = negative predictive value.
3.1. Case-Level Performance
All detected lesions were peripheral zone lesions. With regard to the ground truth, 38/48 (80%) cases had a ≥3 PIRADS score and 10/48 (20%) cases were defined as lesion free based on the radiologists’ reports. The AI solution achieved a case-level sensitivity of 87% (33/38) with a PPV of 87%. The case-level specificity was 50% (5/10) with an NPV of 50%. The case-level k was 0.42.
3.2. Lesion-Level Performance
In total, 28 PI-RADS 3, 22 PI-RADS 4, and 5 PI-RADS 5 lesions were assigned in the written reports. Moreover, 22 participants had 1 lesion, 15 participants had 2 lesions, and 1 participant had 3 lesions. The AI solution detected all the PI-RADS 5 lesions (100%), 16/22 PI-RADS 4 lesions (73%), and 12/28 PI-RADS 3 lesions (43%). Overall, 33 of 55 lesions were detected (60%). The mean false-positive rate per patient was 0.875.
3.3. Index-Level Performance
In total, 38 index lesions were assigned by the human readers, according to the total number of positive cases. Of these, 18 index-lesions were classified as PI-RADS 3, 15 as PI-RADS 4, and 5 as PI-RADS 5. The algorithm detected 14/18 PI-RADS 3 (78%), 14/15 PI-RADS 4 (93%), and 5/5 PI-RADS 5 (100%) lesions, resulting in an overall 87% sensitivity on an index-lesion level.
3.4. Biopsy Results
Transrectal biopsies revealed a positive result in histopathology in a total of 6 patients (see Figure 4). Three participants had a Gleason grade group (GGG) of 1 (corresponding PI-RADS scores in these locations were 3, 4, and 5), two had a GGG of 2 (PI-RADS scores: 4 and 5) and one participant had a GGG of 3 (PI-RADS score: 5), see Table 4. All these lesions were detected by the human readers and the AI software solution.
Table 4.
Participants | GGG | PI-RADS | PSA (µg/mL) | Maximum Diameter (mm) |
---|---|---|---|---|
1 | 1 | 5 | 3.99 | 15.7 |
2 | 1 | 3 | 2.18 | 6.3 |
3 | 2 | 4 | 6.47 | 12.1 |
4 | 1 | 4 | 0.81 | 9.05 |
5 | 2 | 5 | 5.52 | 21.1 |
6 | 3 | 5 | 39.70 | 35.0 |
The PI-RADS score was derived from the written reports. The PSA value was determined before the biopsy. The maximum diameter of the lesions was derived from the axial slice of the ADC map with the largest tumor diameter. PI-RADS = Prostate Imaging Reporting and Data System, PSA = prostate-specific antigen, GGG = Gleason grade group.
4. Discussion
With the ongoing success of deep learning techniques in medical image analysis, those techniques have been used for the detection of prostate cancer. A recent study of Schelb et al. [15], investigating a deep learning system for the detection of suspicious lesions in prostate MRI examinations in men suspected of having clinically significant prostate cancer, showed sensitivity and specificity values of 96% and 31% at a U-Net probability cut-off ≥0.22, and 92% and 47% using a cut-off ≥0.33 in their test set, respectively. The performance metrics in our MRI screening population are very similar, however, our setting is defined by the screening protocol, corresponding to different subject statistics. The total computational time per case in our study was 14 s. It is a known statistical problem that higher sensitivity comes at a price of higher false-positive (FP) rates. A study conducted by Vos et al. [23] showed that at a sensitivity of 74% the FP level was at 5 per patient while the sensitivity dropped to 41% at a FP level of 1. In our study, we opted for a balanced sensitivity and specificity using a FP-reduction (FPR) strategy as the deep learning algorithm tended to overestimate lesions. This approach resulted in a very low false-positive rate per patient of 0.875. Experiments without the FPR strategy yielded sensitivity and specificity values of 97% and 16% on a case level.
The ultimate implementation of such AI-based software solutions in has two prerequisites: (i) agreement between the human reader and the software solution should be non-inferior to human interobserver metrics and (ii) accurate guidance of the biopsy workflow. Concerning the first point, Muller et al. [24] evaluated the interobserver variability of the PI-RADS v2 lexicon in a five-reader study with varying reader experience. The investigators found a k of 0.46 concerning the overall suspicion score. Our results show a similar performance comparing human readers and AI with a k score of 0.42. With regard to the second point, we were able to show that both human readers and the AI solution were able to identify all biopsy-verified prostate cancer lesions. Interestingly, from a screening point of view, only 3/6 participants with a positive histopathology demonstrated PSA values ≥4 µg/mL. In fact, the PSA values demonstrated good capabilities to identify a GGG of ≥2. However, all patients with a GGG = 1 would not have been detected. Therefore, our results may provide new insights in the sense that an MRI-based screening is better suitable as an early warning system.
We found clinically significant cancers in 3/38 participants (8%). Consequently, in the remaining 35 participants the biopsy did not reveal a clinically significant cancer despite the presence of a PI-RADS lesion ≥3. This may be due to the low tumor yield in PI-RADS 3 lesions in our cohort (3%), potentially warranting a change in the biopsy decision workflow in the future. The PI-RADS score does not equal a cancer identification score but merely is an ordinal, probability score for the presence of cancer. With regard to PI-RADS 3, 4, and 5 lesions, cancer has been detected in the following ranges 12–33%, 22–70%, and 72–91%, respectively [25,26], in studies using extended biopsy sampling schemes. Our values range below these reported values not only because we used targeted biopsies but also due to our distinct study objects, representing healthy individuals with extremely low PSA values and not patients with clinical suspicion for prostate cancer; thus, resulting in a much lower pre-test probability for prostate cancer as compared to a patient cohort with a high level of suspicion for prostate cancer. With regard to the spectrum of pre-test probabilities as published by Lavelle et al. [27], participants in our screening cohort would have been found rather on the “exclusion threshold” for prostate cancer. This fact may explain the slightly lower k values in our screening study compared to data in the literature: as PI-RADS category 3 lesions represented 51% of the reported lesions in our screening cohort (versus a reported incidence of 32% to 22% in patient cohorts [28]) and due to the inherent uncertainty of this PI-RADS category in the guidelines, the detection and classification task presented here is generally more difficult than findings obvious tumor in patients with suspicion for prostate cancer. Larger studies will have to reveal whether the PI-RADS classification scheme can be applied on such a cohort with a certain validity and what management strategies can be developed for those men.
Our study has several limitations. The training data of the algorithm contained more cases with PI-RADS ≥3 lesions than lesion-free cases. Due to this discrepant distribution, the algorithm is somehow prone to a certain over-detection, reducing the overall specificity. However, our specificity ranged between experienced and intermediate/low-experienced readers [29]. Second, all of the lesions under investigation in the current study were peripheral zone lesions, and thus, the results are not valid for transition zone lesions. While reported detection rates of peripheral zone lesions—especially PI-RADS ≥4—are sufficiently high, the detection of transition zone lesions is hampered due to a difficult differentiation of benign and malignant processes [30,31] due to common image features. Here, the capability of artificial-intelligence-based software solutions to detect patterns, which potentially remain invisible to the human eye, hold some promises. Future studies need to investigate the performance of the present or different algorithm(s) in the transition zone. Third, in this cohort we performed targeted biopsies of suspicious lesions only, and no random biopsies were taken. Therefore, it cannot be ruled out completely that lesions not detected in bpMRI and not detected by the algorithm were missed. Fourth, the sample size in our study is rather small. Reasons for this fact are the difficulty to establish a screening population and that the study was designed as a first evaluation of the approach in a proof-of-concept setting.
5. Conclusions
In conclusion, this study demonstrates that a deep-learning-based software solution can autonomously detect and classify PI-RADS lesions with a high sensitivity on both a lesion and case level with a moderate classification performance, potentially allowing to use this technology in a screening setting. Furthermore, the AI was able to detect and correctly classify all lesions that contained histopathologically proven cancer, allowing to use that technology in a consecutive targeted biopsy workflow. In an outlook, our approach should be tested in a larger, prospective cohort and the predicted lesions of the AI and the human PI-RADS assessment scores should be compared with the histopathology yield per case and core in a prospective manner.
Acknowledgments
We would like to thank the following collaboration partners for supporting this study with datasets, without which the study would not have been possible: Henkjan Huisman (Department of Radiology Radboud University Medical Center Nijmegen, The Netherlands), Dieter Szolar (Diagnostikum Graz Graz, Austria), Evan Johnson (Montclair Radiology Clifton, NJ, USA), Ivan Shabunin (Patero Clinic, Moscow, Russia), Moon Hyung Choi (Eunpyeong St. Mary’s Hospital, Catholic University of Korea, Seoul, Republic of Korea), Qingsong Yang (Radiology Department, Changhai Hospital of Shanghai, Shanghai, China), Tobias Penzkofer (Charité, Universitätsmedizin Berlin, Berlin, Germany).
Author Contributions
D.J.W.: conception, design, analysis, and interpretation; C.W.: conception, design, analysis, and interpretation; M.O.M.: acquisition; B.L.: analysis and creation of new software; A.K.: analysis and creation of new software; B.S.: analysis and creation of new software; D.C.: analysis and creation of new software; H.-H.S.: conception, design, acquisition, and interpretation; C.A.R.: conception, design, acquisition, and interpretation; D.T.B.: conception, design, analysis, and interpretation. All authors have read and agreed to the published version of the manuscript.
Funding
This research received no external funding.
Conflicts of Interest
D.J.W. was enrolled in a one-year, paid fellowship program in medical image analysis/artificial intelligence at Siemens Medical Imaging Technologies, Princeton, NJ. A.K., B.S., B.L., and D.C. are full-time employees of Siemens Medical Imaging Technologies. However, there are no financial arrangements or other relationships that could be construed as a conflict of interest concerning this manuscript. Furthermore, we report no subject overlap.
Footnotes
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.
References
- 1.Schröder F.H., Hugosson J., Roobol M.J., Tammela T.L.J., Zappa M., Nelen V., Kwiatkowski M., Lujan M., Määttänen L., Lilja H., et al. Screening and Prostate Cancer Mortality: Results of the European Randomised Study of Screening for Prostate Cancer (ERSPC) at 13 Years of Follow-Up. Lancet. 2014;384:2027–2035. doi: 10.1016/S0140-6736(14)60525-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Andriole G.L., Crawford E.D., Grubb R.L., 3rd, Buys S.S., Chia D., Church T.R., Fouad M.N., Isaacs C., Kvale P.A., Reding D.J., et al. Prostate Cancer Screening in the Randomized Prostate, Lung, Colorectal, and Ovarian Cancer Screening Trial: Mortality Results after 13 Years of Follow-Up. J. Natl. Cancer Inst. 2012;104:125–132. doi: 10.1093/jnci/djr500. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Draisma G., Etzioni R., Tsodikov A., Mariotto A., Wever E., Gulati R., Feuer E., de Koning H. Lead Time and Overdiagnosis in Prostate-Specific Antigen Screening: Importance of Methods and Context. J. Natl. Cancer Inst. 2009;101:374–383. doi: 10.1093/jnci/djp001. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.Tsodikov A., Gulati R., Heijnsdijk E.A.M., Pinsky P.F., Moss S.M., Qiu S., de Carvalho T.M., Hugosson J., Berg C.D., Auvinen A., et al. Reconciling the Effects of Screening on Prostate Cancer Mortality in the ERSPC and PLCO Trials. Ann. Intern. Med. 2017;167:449–455. doi: 10.7326/M16-2586. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.Thompson I.M., Pauler D.K., Goodman P.J., Tangen C.M., Lucia M.S., Parnes H.L., Minasian L.M., Ford L.G., Lippman S.M., Crawford E.D., et al. Prevalence of Prostate Cancer among Men with a Prostate-Specific Antigen Level ≤4.0 Ng per Milliliter. N. Engl. J. Med. 2004;350:2239–2246. doi: 10.1056/NEJMoa031918. [DOI] [PubMed] [Google Scholar]
- 6.Ahmed H.U., El-Shater Bosaily A., Brown L.C., Gabe R., Kaplan R., Parmar M.K., Collaco-Moraes Y., Ward K., Hindley R.G., Freeman A., et al. Diagnostic Accuracy of Multi-Parametric MRI and TRUS Biopsy in Prostate Cancer (PROMIS): A Paired Validating Confirmatory Study. Lancet. 2017;389:815–822. doi: 10.1016/S0140-6736(16)32401-1. [DOI] [PubMed] [Google Scholar]
- 7.Kasivisvanathan V., Rannikko A.S., Borghi M., Panebianco V., Mynderse L.A., Vaarala M.H., Briganti A., Budäus L., Hellawell G., Hindley R.G., et al. MRI-Targeted or Standard Biopsy for Prostate-Cancer Diagnosis. N. Engl. J. Med. 2018;378:1767–1777. doi: 10.1056/NEJMoa1801993. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.EAU Guidelines Edn; Presented at the EAU Annual Congress Barcelona 2019; Barcelona, Spain. 15–19 March 2019; [Google Scholar]
- 9.Litjens G., Debats O., Barentsz J., Karssemeijer N., Huisman H. Computer-Aided Detection of Prostate Cancer in MRI. IEEE Trans. Med. Imaging. 2014;33:1083–1092. doi: 10.1109/TMI.2014.2303821. [DOI] [PubMed] [Google Scholar]
- 10.Kim S.J., Vickers A.J., Hu J.C. Challenges in Adopting Level 1 Evidence for Multiparametric Magnetic Resonance Imaging as a Biomarker for Prostate Cancer ScreeningEvidence for Multiparametric Magnetic Resonance Imaging as a Biomarker for Prostate Cancer ScreeningEvidence for Multiparame. JAMA Oncol. 2018;4:1663–1664. doi: 10.1001/jamaoncol.2018.4160. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Barth B.K., De Visschere P.J., Cornelius A. Detection of Clinically Significant Prostate Cancer: Short Dual-Pulse Sequence versus Standard Multi-Parametric MR Imaging—A Multireader Study. Radiology. 2017;284:725. doi: 10.1148/radiol.2017162020. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Kuhl C.K., Bruhn R., Krämer N., Nebelung S., Heidenreich A., Schrading S. Abbreviated Biparametric Prostate MR Imaging in Men with Elevated Prostate-Specific Antigen. Radiology. 2017;282:493–505. doi: 10.1148/radiol.2017170129. [DOI] [PubMed] [Google Scholar]
- 13.Woo S., Suh C.H., Kim S.Y., Cho J.Y., Kim S.H., Moon M.H. Head-to-Head Comparison between Biparametric and Multiparametric MRI for the Diagnosis of Prostate Cancer: A Systematic Review and Meta-Analysis. Am. J. Roentgenol. 2018;211:W226–W241. doi: 10.2214/AJR.18.19880. [DOI] [PubMed] [Google Scholar]
- 14.Weiss J., Martirosian P., Notohamiprodjo M., Kaufmann S., Othman A.E., Grosse U., Nikolaou K., Gatidis S. Implementation of a 5-Minute Magnetic Resonance Imaging Screening Protocol for Prostate Cancer in Men with Elevated Prostate-Specific Antigen before Biopsy. Investig. Radiol. 2018;53:186–190. doi: 10.1097/RLI.0000000000000427. [DOI] [PubMed] [Google Scholar]
- 15.Schelb P., Kohl S., Radtke J.P., Wiesenfarth M., Kickingereder P., Bickelhaupt S., Kuder T.A., Stenzinger A., Hohenfellner M., Schlemmer H.-P., et al. Classification of Cancer at Prostate MRI: Deep Learning versus Clinical PI-RADS Assessment. Radiology. 2019:190938. doi: 10.1148/radiol.2019190938. [DOI] [PubMed] [Google Scholar]
- 16.Giannini V., Mazzetti S., Armando E., Carabalona S., Russo F., Giacobbe A., Muto G., Regge D. Multiparametric Magnetic Resonance Imaging of the Prostate with Computer-Aided Detection: Experienced Observer Performance Study. Eur. Radiol. 2017;27:4200–4208. doi: 10.1007/s00330-017-4805-0. [DOI] [PubMed] [Google Scholar]
- 17.Maas M.C., Fütterer J.J., Scheenen T.W. Quantitative Evaluation of Computed High b Value Diffusion-Weighted Magnetic Resonance Imaging of the Prostate. Investig. Radiol. 2013;48:779. doi: 10.1097/RLI.0b013e31829705bb. [DOI] [PubMed] [Google Scholar]
- 18.Rosenkrantz A.B., Parikh N., Kierans A.S., Kong M.X., Babb J.S., Taneja S.S., Ream J.M. Prostate Cancer Detection Using Computed Very High B-Value Diffusion-Weighted Imaging: How High Should We Go? Acad. Radiol. 2016;23:704–711. doi: 10.1016/j.acra.2016.02.003. [DOI] [PubMed] [Google Scholar]
- 19.Yang D., Xu D., Zhou S.K., Georgescu B., Chen M., Grbic S., Metaxas D., Comaniciu D. Automatic Liver Segmentation Using an Adversarial Image-to-Image Network. In: Descoteaux M., Maier-Hein L., Franz A., Jannin P., Collins D.L., Duchesne S., editors. Med Image Comput Comput Assist Interv—MICCAI 2017. Springer International Publishing; Cham, Switzerland: 2017. pp. 507–515. [DOI] [Google Scholar]
- 20.Khamene A., Chisu R., Wein W., Navab N., Sauer F. A Novel Projection Based Approach for Medical Image Registration; Proceedings of the Third International Conference on Biomedical Image Registration (WBIR’06); Utrecht, The Netherlands. 9–11 July 2006; Berlin/Heidelberg, Germany: Springer; 2006. pp. 247–256. [DOI] [Google Scholar]
- 21.Yu X., Lou B., Bibo S., Winkel D.J., Arrahmane N., Diallo M., Tongbai M., von Busch H., Grimm R., Kiefer B., et al. False Positive Reduction Using Multiscale Contextual Features for Prostate Cancer Detection in Multi-Parametric MRI Scans; Proceedings of the IEEE International Symposium on Biomedical Imaging (ISBI); Iowa City, IA, USA. 3–7 April 2020. [Google Scholar]
- 22.Weinreb J.C., Barentsz J.O., Choyke P.L., Cornud F., Haider M.A., Macura K.J., Margolis D., Schnall M.D., Shtern F., Tempany C.M., et al. PI-RADS Prostate Imaging—Reporting and Data System: 2015, Version 2. Eur. Urol. 2016 doi: 10.1016/j.eururo.2015.08.052. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23.Vos P.C., Barentsz J.O., Karssemeijer N., Huisman H.J. Automatic Computer-Aided Detection of Prostate Cancer Based on Multiparametric Magnetic Resonance Image Analysis. Phys. Med. Biol. 2012;57:1527. doi: 10.1088/0031-9155/57/6/1527. [DOI] [PubMed] [Google Scholar]
- 24.Muller B.G., Shih J.H., Sankineni S., Marko J., Rais-Bahrami S., George A.K., de la Rosette J.J.M.C.H., Merino M.J., Wood B.J., Pinto P., et al. Prostate Cancer: Interobserver Agreement and Accuracy with the Revised Prostate Imaging Reporting and Data System at Multiparametric MR Imaging. Radiology. 2015;277:741–750. doi: 10.1148/radiol.2015142818. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.Mehralivand S., Bednarova S., Shih J.H., Mertan F.V., Gaur S., Merino M.J., Wood B.J., Pinto P.A., Choyke P.L., Turkbey B. Prospective Evaluation of PI-RADSTM Version 2 Using the International Society of Urological Pathology Prostate Cancer Grade Group System. J. Urol. 2017;198:583–590. doi: 10.1016/j.juro.2017.03.131. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Greer M.D., Shih J.H., Lay N., Barrett T., Bittencourt L.K., Borofsky S., Kabakus I.M., Law Y.M., Marko J., Shebel H., et al. Validation of the Dominant Sequence Paradigm and Role of Dynamic Contrast-Enhanced Imaging in Pi-RADS Version 2. Radiology. 2017;285:859–869. doi: 10.1148/radiol.2017161316. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Lavelle L.P., Dunne R.M., Carroll A.G., Malone D.E. Evidence-Based Practice of Radiology. RadioGraphics. 2015;35:1802–1813. doi: 10.1148/rg.2015150027. [DOI] [PubMed] [Google Scholar]
- 28.Schoots I.G. MRI in Early Prostate Cancer Detection: How to Manage Indeterminate or Equivocal PI-RADS 3 Lesions? Transl. Androl. Urol. 2018;7:70–82. doi: 10.21037/tau.2017.12.31. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Greer M.D., Shih J.H., Lay N., Barrett T., Bittencourt L., Borofsky S., Kabakus I., Law Y.M., Marko J., Shebel H., et al. Interreader Variability of Prostate Imaging Reporting and Data System Version 2 in Detecting and Assessing Prostate Cancer Lesions at Prostate MRI. Am. J. Roentgenol. 2019;212:1197–1205. doi: 10.2214/AJR.18.20536. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30.Hoeks C.M.A., Hambrock T., Yakar D., Hulsbergen–van de Kaa C.A., Feuth T., Witjes J.A., Fütterer J.J., Barentsz J.O. Transition Zone Prostate Cancer: Detection and Localization with 3-T Multiparametric MR Imaging. Radiology. 2013;266:207–217. doi: 10.1148/radiol.12120281. [DOI] [PubMed] [Google Scholar]
- 31.Weinreb J.C. Organized Chaos: Does PI-RADS Version 2 Work in the Transition Zone? Radiology. 2018;288:492–494. doi: 10.1148/radiol.2018180123. [DOI] [PubMed] [Google Scholar]