Abstract
Purpose
To develop a cascaded deep learning model trained with apparent diffusion coefficient (ADC) and T2-weighted imaging (T2WI) for fully automated detection and localization of clinically significant prostate cancer (csPCa).
Methods
This retrospective study included 347 consecutive patients (235 csPCa, 112 non-csPCa) with high-quality prostate MRI data, which were randomly selected for training, validation, and testing. The ground truth was obtained using manual csPCa lesion segmentation, according to pathological results. The proposed cascaded model based on Res-UNet takes prostate MR images (T2WI+ADC or only ADC) as inputs and automatically segments the whole prostate gland, the anatomic zones, and the csPCa region step by step. The performance of the models was evaluated and compared with PI-RADS (version 2.1) assessment using sensitivity, specificity, accuracy, and Dice similarity coefficient (DSC) in the held-out test set.
Results
In the test set, the per-lesion sensitivity of the biparametric (ADC + T2WI) model, ADC model, and PI-RADS assessment were 95.5% (84/88), 94.3% (83/88), and 94.3% (83/88) respectively (all p > 0.05). Additionally, the mean DSC based on the csPCa lesions were 0.64 ± 0.24 and 0.66 ± 0.23 for the biparametric model and ADC model, respectively. The sensitivity, specificity, and accuracy of the biparametric model were 95.6% (108/113), 91.5% (665/727), and 92.0% (773/840) based on sextant, and were 98.6% (68/69), 64.8% (46/71), and 81.4% (114/140) based on patients. The biparametric model had a similar performance to PI-RADS assessment (p > 0.05) and had higher specificity than the ADC model (86.8% [631/727], p< 0.001) based on sextant.
Conclusion
The cascaded deep learning model trained with ADC and T2WI achieves good performance for automated csPCa detection and localization.
Keywords: deep learning, prostatic neoplasms, magnetic resonance imaging, detection, localization
Introduction
Prostate cancer (PCa) is one of the most common malignant tumors in men worldwide. The clinical behavior of PCa ranges from low-grade indolent that is generally considered to be non-life-threatening to high-grade aggressive tumors with a Gleason of Score 7–10, i.e. clinically significant PCa (csPCa), that may progress rapidly to metastatic disease and may be life-threatening (1). Multiparametric magnetic resonance imaging (mpMRI) has adopted an increasingly significant role in the detection and localization of csPCa, as well as in guiding targeted biopsy (2). Recent large-scale clinical trials have demonstrated that the use of mpMRI before biopsy increases the detection of csPCa, while reducing the detection of those deemed clinically insignificant (3, 4). Furthermore, using mpMRI to triage male patients may enable a quarter to half of them to avoid the need for biopsy (3, 5). To standardize and improve the interpretation of prostate mpMRI, the use of the Prostate Imaging Reporting and Data System (PI-RADS) has been recommended and updated (2, 6). However, the interobserver agreement for subjective evaluation using PI-RADS (version 2) is moderate and influenced by the readers’ experience (7, 8). Additionally, PI-RADS (version 2.1) has shown no significant improvements in overall diagnostic performance compared to PI-RADS (version 2.0) (9, 10). As there is arguably a trend in more people with clinically suspected csPCa undergoing prostate mpMRI, it is clinically desirable to develop more accurate and automated methods for prostate mpMRI interpretation.
In the recent years, artificial intelligence (AI) methods, particularly deep learning, have achieved promising results in automated csPCa diagnosis of mpMRI (11–14). A range of proposed deep learning algorithms were trained based on prior annotated regions of interest (ROIs) to classify them as cancerous or noncancerous lesions (11, 15), or slices, in which each individual image was classified as cancerous or not (16, 17). These methods were unable to precisely detect and locate csPCa, and such predicted results may not be directly applied to clinical practice. Some computer-aided diagnosis (CAD) systems developed for csPCa were based on the manual or semi-automatic segmentation of the prostate gland (13, 18), which also limits their direct clinical use. With the development of the deep convolutional neural network (CNN), some approaches for csPCa detection have been fully automatic with an area under the receiver operator characteristics curve of 0.75–0.86 (12, 19, 20). More studies are needed to improve and optimize these models. Although many generalized AI models have been developed, few studies have reported on how to integrate AI-based prediction into the clinical workflow. More explorations are demanded to move the prostate AI systems from the laboratory to the clinic with perfect output.
Our study aimed to develop a fully automated cascaded deep learning model for the detection and localization of csPCa using apparent diffusion coefficient (ADC) maps and T2-weighted imaging (T2WI), as well as to seamlessly integrate these AI predictions into the radiological workflow using the structured report.
Materials and methods
Study subjects
Our institute’s review board approved this retrospective study and waived the need for informed consent. The inclusion criteria for the study were mpMRI scans performed on a GE 750 3.0T MRI scanner at Peking University First Hospital from March 2017 to February 2020 on consecutive patients who underwent mpMRI before a biopsy, with a clinical suspicion of PCa due to an elevated serum prostate-specific antigen (PSA) level, abnormal digital rectal examination (DRE), and/or abnormal transrectal ultrasound (TRUS) results. Exclusion criteria were patients without a subsequent biopsy performed within 3 months after mpMRI examination, a negative biopsy for csPCa without clinical follow-up >1 year, or showing potential csPCa during the clinical follow-up (progression of PSA or MR findings), as well as images with severe artifacts or incomplete pathology results which could not be matched with MR images. In total, 347 patients were included. Figure 1 displays the flow diagram for the inclusion of patients in the study. In this study, csPCa was defined as the International Society of Urological Pathology Gleason grade group ≥2, i.e., Gleason Score ≥7. The characteristics of the 235 patients with csPCa included are shown in Table 1 . The other 112 patients without csPCa (labeled non-csPCa with a mean age of 64.1 ± 7.5 years) had a median PSA level of 8.0 ng/ml, with an interquartile range of 6.6–13.1 ng/ml. The patients were randomly selected to populate the datasets for training (145 csPCa, 35 non-csPCa), validation (21 csPCa, 6 non-csPCa), and testing (69 csPCa, 71 non-csPCa).
Table 1.
Characteristics | Patients with csPCa (n = 235) |
---|---|
Mean age (years) | 70.2 ± 8.6 |
Median PSA (ng/mL) | 16.3 (9.7–32.6) |
Per-patient maximum Gleason score | |
3 + 4 | 76 |
4 + 3 | 59 |
3 + 5, 5 + 3, 4 + 4 | 46 |
4 + 5, 5 + 4 | 54 |
No. of csPCa lesions per patient | |
One lesion | 173 |
Two lesions | 49 |
Three lesions | 10 |
Four lesions | 3 |
Zone distribution of csPCa lesions | |
Peripheral zone | 212 |
Transition zone | 101 |
csPCa, clinically significant prostate cancer; PSA, serum prostate-specific antigen.
MRI sequences
All of the mpMRI examinations were performed on a 3-T MR machine (Discovery MR750, GE Medical Systems). A 32-channel abdominal phased array coil was used as the receiving coil. All patients were scanned following the unified prostate mpMRI protocol. The main sequence parameters are summarized in Table 2 . The ADC map was automatically generated by the MR vendor software based on diffusion-weighted imaging (DWI) data with different b values. Concerning the patients included in this study, anonymized images were exported in the Digital Imaging and Communication in Medicine (DICOM) format.
Table 2.
T2WI | DWI | DCE | |
---|---|---|---|
Field of view (mm) | 240 × 240 | 240 × 240 | 260 × 260 |
Acquisition Matrix | 320 × 256 | 96 × 96 | 320 × 192 |
Repetition time (ms) | 3200 | 3000 | 4 |
Echo time (ms) | 85 | 60 | 1.3 |
Flip angle (degrees) | 111 | 90 | 13 |
Slice thickness (mm) | 4 | 4 | 3 |
Additional information | … | b values: 0–1400 s/mm2 | Temporal resolution = 13 s;18 phases |
T2WI, T2-weighted imaging; DWI, diffusion-weighted imaging; DCE, dynamic contrast material enhancement.
PI-RADS assessment
All of the mpMRI cases in the dataset of the test were retrospectively interpreted according to PI-RADS (version 2.1) by a urogenital radiologist with 10 years of experience in prostate MRI diagnosis. The radiologist was informed of the clinical information of the patients, such as age, biopsy history, PSA, etc., but was blinded to the pathology results and the previous MRI reports. The lesions detected were delineated on a prostate sector map.
Reference standard and annotation
All of the patients in this study underwent TRUS-guided systematic and targeted biopsy using 12- or six-core needles. For cognitive targeting, lesions suspected of malignancy on mpMRI had been marked on a prostate sector map (6) using structured reports by five dedicated urogenital radiologists during the clinical routine. Before the biopsy, MR images would be reviewed by at least one urogenital radiologist and one urologist at a multidisciplinary meeting to ensure the accuracy of suspicious lesions localization. The urologists obtained additional needle cores (two- to five-core needles) for each of the suspicious lesions during the TRUS-guided biopsy. Histopathology analysis of each specimen was performed by a urogenital pathologist with 11 years of experience.
Two experienced urogenital radiologists (X and Y with 7 and 13 years of experience in prostate MRI diagnosis, respectively) retrospectively reviewed all csPCa cases and mapped the detailed pathology results of the csPCa foci to the MR images with consensus. The ground truth of the csPCa lesion segmentation was obtained using manual delineation by the urogenital radiologist (X), in consensus with and under the supervision of the other urogenital radiologist (Y), using the open-source segmentation software ITK-SNAP (version 3.6 2015; available at www.itksnap.org) (21). The format of ADC and T2WI was converted from DICOM to NIFTI. Three-dimensional volumes of interest (VOIs) were manually drawn along the boundaries of the csPCa lesions on consecutive axial sections of ADC images.
Image preprocessing
After collecting the mpMRI data, the first step of image preprocessing is T2WI and DWI/ADC image registration. Patient motion is minimal and the two sequences are temporally close to each other during the scanning. T2WI and DWI/ADC images were registered via rigid transformation using the coordinate information stored in the DICOM image headers. B-spline interpolation to the third order was employed for all MR image interpolation tasks, while Gaussian label interpolation was used for the csPCa and prostate segmentation masks. Following this, a coarse segmentation of the prostate was obtained by K-means clustering in DWI high b value images to localize the prostate region. Once the prostate region was identified, the images were cropped to a patch of size 128 × 128. The prostate region of interest was then normalized into the range of [0, 1]. We augmented the data in the training set by mirroring, random rotation (rotation angle within 10°), and adding noise (within 0.001, which means each pixel value randomly fluctuates within one thousandth).
Deep learning framework
The base architecture for the deep learning framework used in this study is a CNN inspired by the 2D U-Net (22) and Res-Net (23) architectures and is termed Res-UNet. U-Net is one of the end-to-end methods of deep learning, which is also a pixel-to-pixel method and, with long skip connections, considers feature maps of the encoder path to obtain good segmentation performance in medical images. Res-Net (23) proposed a residual connection architecture to make the network deeper and avoid gradient vanishing and take advantage of strengths from both architectures by modifying the original U-Net architecture and adding residual blocks into the contracting and symmetric expanding paths of the U-Net architecture. In building the Res-UNet, we define a basic convolution operator by a 3 × 3 convolution (Conv) followed by a batch normalization (BN) and a rectified linear unit (ReLu). The residual block was designed by using a 1 × 1 Conv layer, plus an addition operation and ReLU function. Figure 2 depicts the Res-UNet architecture.
Following McNeal’s criterion (24), the prostate is typically partitioned into two distinct zones: the Central Gland (CG, including both the transition zone and the central zone) and the Peripheral Zone (PZ). PCa lesions vary in frequency and malignancy depending on the zone, and there are different evaluation criteria for different regions in the PI-RADS. Therefore, just like a radiologist, a model for automated PCa detection and classification will invariably benefit from having both CG and PZ mask priors provided as inputs, in addition to the mpMRI. Accordingly, a cascading system of a three-segmentation Res-UNet which was previously developed in our institution (25) was developed for automatic prostate CG and PZ segmentation and PCa lesion segmentation. The cascade is designed to decompose the multi-task segmentation problem into a sequence of three smaller binary segmentation problems according to the subregion hierarchy. As can be seen in Figure 2 , the first Res-UNet model takes prostate MR images (T2WI + ADC or only ADC) as inputs and produces a prostate segmentation mask as an output. Then, the second model takes the MR images and the prostate segmentation mask which were obtained in the previous step as inputs and produces a PZ segmentation mask. The CG segmentation mask can be computed by subtracting the PZ mask from the prostate mask. Finally, the last model takes the MR images, the PZ mask, and the CG mask as inputs and segments to the csPCa.
Training setup
All of the training steps were performed using a GPU NVIDIA Tesla P100 16G at Peking University First Hospital, using the operating system Ubuntu 16.04. The software and packages used included Python 3.6, Pytorch 0.4.1, Opencv 3.4.0.12, Numpy 1.16.2, and SimpleITK 1.2.0. The input data were 128 × 128 images of ADC alone and ADC combined with T2WI, respectively, with an annotation of the csPCa lesions. The automated segmentation of the whole prostate gland and its different zones was completed using the previously developed and described method (25). For training the architectures for csPCa segmentation, the batch size was set as 20 with a learning rate of 0.0001. The networks were trained for 120 epochs. The pixel classification threshold was 0.5, while Adam was used as a training optimizer. The Dice similarity coefficient (DSC) was used to evaluate the performance of the networks in the segmentation of the csPCa, which is calculated as
here X is the pixel set of csPCa segmented manually as the ground truth and Y is the pixel set of csPCa prediction by the model.
Prediction results integrated into the structured report
Initially, the prediction results were “csPCa” or “non-csPCa” depending on the patient concerned. When “csPCa” was the output, the three-dimensional diameter of the suspicious csPCa lesions and the whole prostate gland would be filled into the structured report, as well as the key image of the suspicious csPCa lesions ( Supplementary Figure S1 ). Furthermore, sextant localization of the suspicious csPCa lesions would be labeled in the prostate sector map. The prostate sextant is defined according to the standard sextant biopsy (26), based on the automatic segmentation of the prostate gland.
Statistical analysis
Statistical analysis was carried out using SPSS 20.0 and MedCalc 15.8. We evaluated the performance of the biparametric (ADC + T2WI) model, ADC model, and PI-RADS assessment using the testing set. For PI-RADS assessment, PI-RADS ≥3 was considered positive for csPCa. For per-lesion analysis, to limit the influence of very small overlap regions, only the predicted lesions of the model overlapping ≥50% of the manual csPCa segmentation lesions in at least one slice were considered as true positive. Otherwise, the predicted lesions were considered to be false positives. For sextants analysis, only sextants overlapping at least 50% of an MRI lesion, or being occupied at least 50% by an MRI lesion, were considered to contain the MRI lesion (13). For per-patient analysis, if a patient had one or more than one csPCa lesions, the prediction of the model or PI-RADS assessment, was considered as true positive when at least one csPCa lesion was detected. On the other hand, for a patient without csPCa, the prediction was considered as false positive as long as one lesion was predicted. The performance of the models and PI-RADS assessment for csPCa detection and localization were evaluated based on the lesions, sextants, and patients, respectively. The sensitivity, specificity, and accuracy of the models and PI-RADS were evaluated and compared using the McNemar test. A p value of less than 0.05 was considered statistically significant.
Results
Based on lesions
In the test set, 88 csPCa lesions were included and the mean greatest dimension was 1.6 ± 0.7 cm. Table 3 depicts the sensitivity of the models and PI-RADS assessment on the per-lesion analysis. The per-lesion sensitivity of the biparametric model, ADC model, and PI-RADS assessment was 95.5% (84/88), 94.3% (83/88), and 94.3% (83/88), respectively, with all p > 0.05. For the csPCa lesions with the greatest dimension ≥1.5cm, the sensitivity of the biparametric model was 100%, and the sensitivity of the ADC model and PI-RADS assessment was 97.6% (40/41, with p > 0.05). The sensitivity showed no significant difference between the models and PI-RADS regardless of whether the csPCa lesions were in the PZ or the TZ.
Table 3.
Biparametric model (%) | ADC model (%) | PI-RADS (%) | p | |
---|---|---|---|---|
Total | 95.5 (84/88) [88.8, 98.8] | 94.3 (83/88) [87.2, 98.1] | 94.3 (83/88) [87.2, 98.1] | >0.05 |
Peripheral zone | 95.0 (57/60) [86.1, 99.0] | 93.3 (56/60) [83.8, 98.2] | 96.7 (58/60) [88.5, 99.6] | >0.05 |
Transition zone | 96.4 (27/28) [81.7, 99.9] | 96.4 (27/28) [81.7, 99.9] | 89.3 (25/28) [71.8, 97.7] | >0.05 |
Dimension 0.4-1.5cm | 91.5 (43/47) [79.6, 97.6] | 91.5 (43/47) [79.6, 97.6] | 91.5 (43/47) [79.6, 97.6] | >0.05 |
Dimension ≥1.5cm | 100.0 (41/41) [96.4, 100] | 97.6 (40/41) [87.1, 99.9] | 97.6 (40/41) [87.1, 99.9] | >0.05 |
PI-RADS, Prostate Imaging Reporting and Data System.
In addition, the mean DSC based on csPCa lesions in the test were 0.64 ± 0.24 and 0.66 ± 0.23 for the biparametric model and the ADC model, respectively. Figure 3 demonstrates examples of the csPCa segmentation of the biparametric model.
Based on sextants
A total of 840 sextants from the test set were analyzed, including 113 sextants of csPCa and 727 sextants of non-csPCa. The diagnostic efficacy and comparisons of the models and the PI-RADS assessment based on sextants are summarized in Tables 4 and 5 . The biparametric model and PI-RADS assessment both had relatively high sensitivity, specificity and accuracy, i.e., 95.6% (108/113) vs. 92.9% (105/113), 91.5% (665/727) vs. 92.2% (670/727), and 92.0% (773/840) vs. 92.3% (775/840), respectively, all with p > 0.05. The ADC model demonstrated a comparable sensitivity of 91.2% (103/113) when compared with the biparametric model where p = 0.125, while the ADC model had a specificity of 86.8% (631/727) and accuracy of 87.4% (734/840), which were both lower than the biparametric model (all p< 0.001). Considering all the mpMRI sequences and detailed pathological results, 61.5% (59/96) of the false-positive sextants from the ADC model were hyperplastic nodules and asymmetric central zone, which is shown in Figure 4 , while the ratio was 45.2% (28/62) for the biparametric model.
Table 4.
Sensitivity (%) | Specificity (%) | Accuracy (%) | |
---|---|---|---|
Based on sextant | |||
Biparametric model | 95.6 (108/113) [90.0, 98.6] | 91.5 (665/727) [89.2, 93.4] | 92.0 (773/840) |
ADC model | 91.2 (103/113) [84.3, 95.7] | 86.8 (631/727) [84.1, 89.2] | 87.4 (734/840) |
PI-RADS | 92.9 (105/113) [86.5, 96.9] | 92.2 (670/727) [90.0, 94.0] | 92.3 (775/840) |
Based on patient | |||
Biparametric model | 98.6 (68/69) [92.2, 99.9] | 64.8 (46/71) [52.5, 75.8] | 81.4 (114/140) |
ADC model | 97.1 (67/69) [89.9, 99.7] | 54.9 (39/71) [42.7, 66.8] | 75.7 (106/140) |
PI-RADS | 98.6 (68/69) [92.2, 99.9] | 66.2 (47/71) [54.0, 77.0] | 82.1 (115/140) |
PI-RADS, Prostate Imaging Reporting and Data System.
Data in brackets are 95% CIs.
Table 5.
Sensitivity | Specificity | Accuracy | |
---|---|---|---|
Based on sextants | |||
Biparametric model vs. ADC model | 0.125 | <0.001 | <0.001 |
Biparametric model vs. PI-RADS | 0.508 | 0.630 | 0.910 |
ADC model vs. PI-RADS | 0.754 | <0.001 | <0.001 |
Based on patients | |||
Biparametric model vs. ADC model | 1.000 | 0.118 | 0.077 |
Biparametric model vs. PI-RADS | 1.000 | 1.000 | 1.000 |
ADC model vs. PI-RADS | 1.000 | 0.077 | 0.064 |
PI-RADS, Prostate Imaging Reporting and Data System.
Bold characters indicate that the difference was statistically significant (p< 0.05).
Based on patients
For the 140 patients (csPCa, 69; non-csPCa, 71) in the test set, the performance and comparisons of the models and the PI-RADS assessment based on patients are shown in Tables 4 and 5 . Biparametric model and PI-RADS assessment had comparable per-patient sensitivity, specificity, and accuracy, i.e. 98.6% (68/69) vs. 98.6% (68/69), 64.8% (46/71) vs. 66.2% (47/71) and 81.4% (114/140) vs. 82.1% (115/140), respectively (all p > 0.05). ADC model had a similar sensitivity of 97.1% (67/69) compared with the combined model and PI-RADS assessment. The specificity and accuracy of the biparametric model were slightly higher than those of the ADC model (specificity, 54.9%; accuracy, 75.7%); however, the statistical significance was not reached (p = 0.118, 0.077).
Discussion
Our approach using cascaded CNNs could automatically detect and segment the suspicious csPCa lesions on MR images without any human intervention. The whole prediction process could be completed within a few seconds per case, which was much faster than human interpretation using PI-RADS, which normally takes several minutes. There are several benefits of using a cascaded framework for the segmentation of hierarchical structures. First, many proposed methods try to solve the segmentation problem using a single neural network. Considering the great variability in the shape, size, texture, and appearance of the prostate gland and PCa, we suggest using cascaded CNNs for the segmentation task to improve the segmentation accuracy, and each network can focus on one segmentation problem. Thus, they are easier to train and can reduce over-fitting. Second, in consideration of the PCa lesions, these can vary in frequency and malignancy depending on the zone; the hierarchical pipeline follows the anatomical structures of the prostate and uses them as spatial constraints. Thus, the model for automated csPCa detection and classification will likely benefit.
The results demonstrated that the biparametric model had high sensitivity (95.5%, 95.6%, and 98.6% based on lesions, sextants, and patients respectively) and acceptable specificity (64.8%, 91.5%; based on patients and sextants) and had comparable performance to PI-RADS evaluation by an experienced radiologist, which is consistent with Schelb’s findings (13). The preliminary results of our study add to the evidence that fully automated deep learning models for csPCa detection have now even reached the level of an experienced radiologist (13, 27). Further prospective studies based on large consecutive data are needed for clinical validation. Furthermore, our model could also determine the boundary of csPCa precisely. The DSC based on csPCa lesions was 0.64 and 0.66 for the biparametric model and ADC model, respectively, which was higher than that in similar studies on csPCa detection that reported 0.35–0.58 (13, 28, 29). The good segmentation performance would facilitate the 3D prostate MRI-TRUS fusion targeted biopsy. Additionally, a 3D model for the visualization of csPCa and the adjacent vital structures, based on accurate segmentation, may be helpful for the urologist in the surgery, as well as for patient education (30). However, DSC as a voxel-level metric remains limited for lesion-level PCa detection and can misrepresent the accuracy for evaluating the localization of multifocal PCa (31, 32). Therefore, our study based on the actual clinical practice mainly used sensitivity and specificity at the lesion, sextant, and patient levels to comprehensively evaluate the performance of the model.
T2WI and ADC derived from DWI are recommended by PI-RADS as the most important sequences for the evaluation of TZ and PZ lesions, respectively. Many studies demonstrated that the diagnostic performance of biparametric MRI without DCE was similar to those of mpMRI (33, 34). Therefore, this study mainly used those two parameters to develop the model. Additionally, the ADC of DWI is considered to be the current best monoparametric sequence of prostate MRI assessment, which is reported to have a strong relationship with the Gleason scores (GS) of PCa and is even superior to TRUS-guided prostate biopsy for the assessment of PCa aggressiveness (35–37). For this reason, the proposed model trained with monoparametric ADC was evaluated specifically. The results of our study demonstrated that the monoparameter ADC model had a high sensitivity for csPCa detection (94.3% [83/88] and 97.1% [67/69], based on lesions and patients, respectively), which showed no significant difference (p > 0.05) with the biparametric model and PI-RADS assessment, regardless of whether the csPCa lesions were located on the PZ or TZ. Zabihollahy et al.’s study also showed that deep learning using only ADC was highly sensitive and could even reach a 100% sensitivity at the level of dominant PZ tumor detection (12), which is slightly higher than ours. The reason may be that their study only considered the most dominant lesion on PZ, which was easily identifiable, while our study detected all the csPCa on the MR images. Further research with larger volumes of testing data is needed to verify the performance of the ADC model. The specificity of the biparametric model in our study was higher than that of the ADC model based on sextants (p< 0.001). When compared with the ADC model, the biparametric model outputted fewer false-positive lesions such as hyperplastic nodules and the central zone. Nevertheless, the high sensitivity of the ADC model using a single parameter instead of time-consuming mpMRI may facilitate the promotion of prostate MRI screening.
In contrast with studies that were training models using public data (15, 38), the data in our study were collected consecutively based on real-world clinical scenarios, which would allow the model to be more easily integrated into a clinical setting. The amount of data in the training and validation sets (145 and 21 csPCa cases, including 1204 slices and 173 slices, respectively, after automatic prostate segmentation) was comparable to that used in some studies (11, 12), and was larger than that in other studies (19, 20). Therefore, we put more data in the testing set (69 csPCa, 71 non-csPCa cases) to better evaluate and verify the generalization ability of the model. Our proposed biparametric model yielded high sensitivity of 98.6% for csPCa detection based on patients, as well as other studies with the sensitivity ranging from 82.9% to 97% (13, 14, 39–41). However, it is worth noting that the specificity was not as high as expected in our study (64.8%) and in other studies (47%–76%) (12–14, 39). That is to say, the success of the AI came at the cost of a high false-positive rate of even 50% (42). Yu et al. (43)proposed a cascaded approach to reduce the false positive for PCa detection, where the second-stage classifier was able to reduce false positives at the expense of nearly an 8% decrease in detection sensitivity. Saha et al. (29) present a multi-stage 3D CAD model for csPCa localization in biparametric MRI with the addition of a residual patch-wise 3D classifier to improve the model specificity. The results demonstrated that up to 12.89% less false positives were generated per patient, while retaining the same sensitivity (92.29%) as before. Min et al. (44) explored the feasibility of controlling the false positives/negatives during training by incorporating the cost-sensitive classification losses. More studies are needed to further explore how to improve the specificity of the prostate CAD.
An optimal AI model should not only have good performance for csPCa diagnosis but should also have a perfect output form facilitating clinical practice. One advantage of the proposed approach is that it had a perfect output and added AI into the radiological workflow seamlessly by automatically integrating the prediction results into structured reports, which makes this approach more convenient for clinical application. At present, our model can output whether a patient has csPCa or not and the size of the prostate gland. If csPCa lesions were found, it would further output the size of csPCa lesions and also mark the area of csPCa on MR images. These prediction results could be automatically transferred into a structured report before radiologists open the reporting system. However, a complete structured report includes many other contents, such as whether the csPCa lesions invade the prostate capsule and adjacent structures, as well as lymph node condition, bone metastasis, etc. Our institution is now exploring each of the above, and some of them have achieved good performance (45). Our ultimate goal was to develop a fully automatic intelligent structured report, thereby freeing radiologists from heavy clinical paperwork. When radiologists open structured reports, they just need to check the accuracy of each item.
Several limitations of our study were as follows. First, all of the images were from a single MR machine in a single institution. Multi-center and multi-machine data functionality should be added to improve the generalization ability of the model in further studies. Secondly, this U-Net model was trained only using ADC and T2WI. Future research involving the addition of more MRI sequences and/or clinical information may be investigated to improve the performance of the model. On the other hand, the model in this study achieved good results with biparametric MRI, so streamlining MRI sequences with an advanced algorithm may be another possible research direction. Thirdly, even though the reference standard using TRUS-guided systematic and targeted biopsy had high sensitivity for csPCa, it still has a false-negative rate when compared with radical prostatectomy. Nevertheless, our cohorts may be optimal, for radical prostatectomy cohorts would exclude many patients who only had a prostate biopsy and could lead to bias. Finally, this model was only applicable to the detection and localization of csPCa instead of staging and active surveillance.
Conclusion
In conclusion, our study demonstrated that a cascaded deep learning model trained with ADC and T2WI achieved good performance for the fully automated detection and segmentation of csPCa and demonstrated comparable performance with an experienced radiologist using PI-RADS (version 2.1). The proposed approach can automatically integrate prediction results into the radiological workflow seamlessly by using a structured report. As a preliminary exploration, this study provided a reference for future AI clinical implementation. Further studies are needed to explore the optimal paradigm of AI clinical integration.
Data availability statement
The datasets presented in this article are not readily available because the datasets are privately owned by Peking University First Hospital and are not made public. Requests to access the datasets should be directed to XYW, wangxiaoying@bjmu.edu.cn.
Ethics statement
The studies involving human participants were reviewed and approved by Peking University First Hospital. Written informed consent for participation was not required for this study in accordance with the national legislation and the institutional requirements.
Author contributions
LZ and XYW designed the study. CH, XL, and DL contributed to acquisition of data. LZ and XYW annotated the images data. YZ, WL, XPW, and JZ designed the model and implemented the main algorithm. LZ and GG analyzed the data. LZ wrote the paper. XYW and XZ reviewed the paper. All authors contributed to the article and approved the submitted version.
Acknowledgments
The authors would like to acknowledge Xin Yue and Suisui Zhang from the Beijing Smart Tree Medical Technology Co., Ltd., for their help in constructing the structured report.
Conflict of interest
Authors WL, XPW and JZ are employed by Beijing Smart Tree Medical Technology Co. Ltd.
The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Supplementary material
The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fonc.2022.958065/full#supplementary-material
References
- 1. Johnson LM, Turkbey B, Figg WD, Choyke PL. Multiparametric MRI in prostate cancer management. Nat Rev Clin Oncol (2014) 11:346–53. doi: 10.1038/nrclinonc.2014.69 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2. Turkbey B, Rosenkrantz AB, Haider MA, Padhani AR, Villeirs G, Macura KJ, et al. Prostate imaging reporting and data system version 2.1: 2019 update of prostate imaging reporting and data system version 2. Eur Urol (2019) 76:340–51. doi: 10.1016/j.eururo.2019.02.033 [DOI] [PubMed] [Google Scholar]
- 3. Ahmed HU, El-Shater Bosaily A, Brown LC, Gabe R, Kaplan R, Parmar MK, et al. Diagnostic accuracy of multi-parametric MRI and TRUS biopsy in prostate cancer (PROMIS): a paired validating confirmatory study. Lancet (2017) 389:815–22. doi: 10.1016/S0140-6736(16)32401-1 [DOI] [PubMed] [Google Scholar]
- 4. Kasivisvanathan V, Rannikko AS, Borghi M, Panebianco V, Mynderse LA, Vaarala MH, et al. MRI-Targeted or standard biopsy for prostate-cancer diagnosis. N Engl J Med (2018) 378:1767–77. doi: 10.1056/NEJMoa1801993 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5. van der Leest M, Cornel E, Israel B, Hendriks R, Padhani AR, Hoogenboom M, et al. Head-to-head comparison of transrectal ultrasound-guided prostate biopsy versus multiparametric prostate resonance imaging with subsequent magnetic resonance-guided biopsy in biopsy-naive men with elevated prostate-specific antigen: A Large prospective multicenter clinical study. Eur Urol (2019) 75:570–8. doi: 10.1016/j.eururo.2018.11.023 [DOI] [PubMed] [Google Scholar]
- 6. Weinreb JC, Barentsz JO, Choyke PL, Cornud F, Haider MA, Macura KJ, et al. PI-RADS prostate imaging - reporting and data system: 2015, version 2. Eur Urol (2016) 69:16–40. doi: 10.1016/j.eururo.2015.08.052 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7. Rosenkrantz AB, Ginocchio LA, Cornfeld D, Froemming AT, Gupta RT, Turkbey B, et al. Interobserver reproducibility of the PI-RADS version 2 lexicon: A multicenter study of six experienced prostate radiologists. Radiology (2016) 280:793–804. doi: 10.1148/radiol.2016152542 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8. Byun J, Park KJ, Kim MH, Kim JK. Direct comparison of PI-RADS version 2 and 2.1 in transition zone lesions for detection of prostate cancer: Preliminary experience. J Magn Reson Imaging (2020) 52:577–86. doi: 10.1002/jmri.27080 [DOI] [PubMed] [Google Scholar]
- 9. Rudolph MM, Baur ADJ, Cash H, Haas M, Mahjoub S, Hartenstein A, et al. Diagnostic performance of PI-RADS version 2.1 compared to version 2.0 for detection of peripheral and transition zone prostate cancer. Sci Rep (2020) 10:15982. doi: 10.1038/s41598-020-72544-z [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10. Linhares Moreira AS, De Visschere P, Van Praet C, Villeirs G. How does PI-RADS v2.1 impact patient classification? a head-to-head comparison between PI-RADS v2.0 and v2.1. Acta Radiol (2021) 62:839–47. doi: 10.1177/0284185120941831 [DOI] [PubMed] [Google Scholar]
- 11. Aldoj N, Lukas S, Dewey M, Penzkofer T. Semi-automatic classification of prostate cancer on multi-parametric MR imaging using a multi-channel 3D convolutional neural network. Eur Radiol (2020) 30:1243–53. doi: 10.1007/s00330-019-06417-z [DOI] [PubMed] [Google Scholar]
- 12. Zabihollahy F, Ukwatta E, Krishna S, Schieda N. Fully automated localization of prostate peripheral zone tumors on apparent diffusion coefficient map MR images using an ensemble learning method. J Magn Reson Imaging (2020) 51:1223–34. doi: 10.1002/jmri.26913 [DOI] [PubMed] [Google Scholar]
- 13. Schelb P, Kohl S, Radtke JP, Wiesenfarth M, Kickingereder P, Bickelhaupt S, et al. Classification of cancer at prostate MRI: Deep learning versus clinical PI-RADS assessment. Radiology (2019) 293:607–17. doi: 10.1148/radiol.2019190938 [DOI] [PubMed] [Google Scholar]
- 14. Cao R, Mohammadian Bajgiran A, Afshari Mirak S, Shakeri S, Zhong X, Enzmann D, et al. Joint prostate cancer detection and Gleason score prediction in mp-MRI via FocalNet. IEEE Trans Med Imaging (2019) 38:2496–506. doi: 10.1109/TMI.2019.2901928 [DOI] [PubMed] [Google Scholar]
- 15. Song Y, Zhang YD, Yan X, Liu H, Zhou M, Hu B, et al. Computer-aided diagnosis of prostate cancer using a deep convolutional neural network from multiparametric MRI. J Magn Reson Imaging (2018) 48:1570–7. doi: 10.1002/jmri.26047 [DOI] [PubMed] [Google Scholar]
- 16. Yoo S, Gujrathi I, Haider MA, Khalvati F. Prostate cancer detection using deep convolutional neural networks. Sci Rep (2019) 9:19518. doi: 10.1038/s41598-019-55972-4 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17. Wang Z, Liu C, Cheng D, Wang L, Yang X, Cheng KT. Automated detection of clinically significant prostate cancer in mp-MRI images based on an end-to-End deep neural network. IEEE Trans Med Imaging (2018) 37:1127–39. doi: 10.1109/TMI.2017.2789181 [DOI] [PubMed] [Google Scholar]
- 18. Lemaitre G, Marti R, Rastgoo M, Meriaudeau F. Computer-aided detection for prostate cancer detection based on multi-parametric magnetic resonance imaging. Annu Int Conf IEEE Eng Med Biol Soc (2017) 2017:3138–41. doi: 10.1109/EMBC.2017.8037522 [DOI] [PubMed] [Google Scholar]
- 19. Seetharaman A, Bhattacharya I, Chen LC, Kunder CA, Shao W, Soerensen SJC, et al. Automated detection of aggressive and indolent prostate cancer on magnetic resonance imaging. Med Phys (2021) 48:2960–72. doi: 10.1002/mp.14855 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20. Sanyal J, Banerjee I, Hahn L, Rubin D. An automated two-step pipeline for aggressive prostate lesion detection from multi-parametric MR sequence. AMIA Jt Summits Transl Sci Proc (2020) 2020:552–60. [PMC free article] [PubMed] [Google Scholar]
- 21. Yushkevich PA, Piven J, Hazlett HC, Smith RG, Ho S, Gee JC, et al. User-guided 3D active contour segmentation of anatomical structures: significantly improved efficiency and reliability. NeuroImage (2006) 31:1116–28. doi: 10.1016/j.neuroimage.2006.01.015 [DOI] [PubMed] [Google Scholar]
- 22. Ronneberger O, Fischer P, Brox T. U-Net: Convolutional networks for biomedical image segmentation. In: Navab N, Hornegger J, Wells, W, Frangi A. (eds). in: Medical image computing and computer-assisted intervention – MICCAI 2015. MICCAI 2015. Lecture notes in computer science. Cham: Springer; (2015) 9351:234–41. doi: 10.1007/978-3-319-24574-4_28 [DOI] [Google Scholar]
- 23. Nah S, Kim TH, Lee KM. (2017). Deep multi-scale convolutional neural network for dynamic scene deblurring, in: Proceedings of the IEEE conference on computer vision and pattern recognition. pp. 3883–91. doi: 10.48550/arXiv.1612.02177 [DOI] [Google Scholar]
- 24. McNeal JE. Anatomy of the prostate: an historical survey of divergent views. Prostate (1980) 1:3–13. doi: 10.1002/pros.2990010103 [DOI] [PubMed] [Google Scholar]
- 25. Zhu Y, Wei R, Gao G, Ding L, Zhang X, Wang X, et al. Fully automatic segmentation on prostate MR images based on cascaded fully convolution network. J Magn Reson Imaging (2019) 49:1149–56. doi: 10.1002/jmri.26337 [DOI] [PubMed] [Google Scholar]
- 26. Presti JC. Prostate biopsy: current status and limitations. Rev Urol (2007) 9:93–8. [PMC free article] [PubMed] [Google Scholar]
- 27. Hosseinzadeh M, Saha A, Brand P, Slootweg I, de Rooij M, Huisman H. Deep learning-assisted prostate cancer detection on bi-parametric MRI: minimum training data size requirements and effect of prior knowledge. Eur Radiol (2022) 32:2224–34. doi: 10.1007/s00330-021-08320-y [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28. Vente C, Vos P, Hosseinzadeh M, Pluim J, Veta M. Deep learning regression for prostate cancer detection and grading in bi-parametric MRI. IEEE Trans BioMed Eng (2021) 68:374–83. doi: 10.1109/TBME.2020.2993528 [DOI] [PubMed] [Google Scholar]
- 29. Saha A, Hosseinzadeh M, Huisman H. End-to-end prostate cancer detection in bpMRI via 3D CNNs: Effects of attention mechanisms, clinical priori and decoupled false positive reduction. Med Image Anal (2021) 73:102155. doi: 10.1016/j.media.2021.102155 [DOI] [PubMed] [Google Scholar]
- 30. Wake N, Rosenkrantz AB, Huang R, Park KU, Wysock JS, Taneja SS, et al. Patient-specific 3D printed and augmented reality kidney and prostate cancer models: impact on patient education. 3D Print Med (2019) 5:4. doi: 10.1186/s41205-019-0041-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 31. Carass A, Roy S, Gherman A, Reinhold JC, Jesson A, Arbel T, et al. Evaluating white matter lesion segmentations with refined sørensen-dice analysis. Sci Rep (2020) 10:8242. doi: 10.1038/s41598-020-64803-w [DOI] [PMC free article] [PubMed] [Google Scholar]
- 32. Yan W, Yang Q, Syer T, Min Z, Punwani S, Emberton M, et al. The impact of using voxel-level segmentation metrics on evaluating multifocal prostate cancer localisation. arXiv preprint arXiv (2022) 2203:16415. doi: 10.48550/arXiv.2203.16415 [DOI] [Google Scholar]
- 33. Choi MH, Kim CK, Lee YJ, Jung SE. Prebiopsy biparametric MRI for clinically significant prostate cancer detection with PI-RADS version 2: A multicenter study. AJR Am J roentgenol (2019) 212:839–46. doi: 10.2214/AJR.18.20498 [DOI] [PubMed] [Google Scholar]
- 34. Tamada T, Kido A, Yamamoto A, Takeuchi M, Miyaji Y, Moriya T, et al. Comparison of biparametric and multiparametric MRI for clinically significant prostate cancer detection with PI-RADS version 2.1. J Magn Reson Imaging (2021) 53:283–91. doi: 10.1002/jmri.27283 [DOI] [PubMed] [Google Scholar]
- 35. Bonekamp D, Kohl S, Wiesenfarth M, Schelb P, Radtke JP, Gotz M, et al. Radiomic machine learning for characterization of prostate lesions with MRI: Comparison to ADC values. Radiology (2018) 289:128–37. doi: 10.1148/radiol.2018173064 [DOI] [PubMed] [Google Scholar]
- 36. Kim TH, Kim CK, Park BK, Jeon HG, Jeong BC, Seo SI, et al. Relationship between Gleason score and apparent diffusion coefficients of diffusion-weighted magnetic resonance imaging in prostate cancer patients. Can Urol Assoc J Nov-Dec (2016) 10:E377–82. doi: 10.5489/cuaj.3896 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37. Li C, Chen M, Wang J, Wang X, Zhang W, Zhang C. Apparent diffusion coefficient values are superior to transrectal ultrasound-guided prostate biopsy for the assessment of prostate cancer aggressiveness. Acta Radiol (2017) 58:232–9. doi: 10.1177/0284185116639764 [DOI] [PubMed] [Google Scholar]
- 38. Yuan Y, Qin W, Buyyounouski M, Ibragimov B, Hancock S, Han B, et al. Prostate cancer classification with multiparametric MRI transfer learning model. Med Phys (2019) 46:756–65. doi: 10.1002/mp.13367 [DOI] [PubMed] [Google Scholar]
- 39. Yang X, Liu C, Wang Z, Yang J, Min HL, Wang L, et al. Co-Trained convolutional neural networks for automated detection of prostate cancer in multi-parametric MRI. Med Image Anal (2017) 42:212–27. doi: 10.1016/j.media.2017.08.006 [DOI] [PubMed] [Google Scholar]
- 40. Wong T, Schieda N, Sathiadoss P, Haroon M, Abreu-Gomez J, Ukwatta E. Fully automated detection of prostate transition zone tumors on T2-weighted and apparent diffusion coefficient (ADC) map MR images using U-net ensemble. Med Phys (2021) 48:6889–900. doi: 10.1002/mp.15181 [DOI] [PubMed] [Google Scholar]
- 41. Netzer N, Weißer C, Schelb P, Wang X, Qin X, Görtz M, et al. Fully automatic deep learning in bi-institutional prostate magnetic resonance imaging: Effects of cohort size and heterogeneity. Invest Radiol (2021) 56:799–808. doi: 10.1097/RLI.0000000000000791 [DOI] [PubMed] [Google Scholar]
- 42. Padhani AR, Turkbey B. Detecting prostate cancer with deep learning for MRI: A small step forward. Radiology (2019) 293:618–9. doi: 10.1148/radiol.2019192012 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43. Yu X, Lou B, Shi B, Winkel D, Szolar D. (2020). False positive reduction using multiscale contextual features for prostate cancer detection in multi-parametric MRI scans, in: 2020 IEEE 17th International Symposium on Biomedical Imaging (ISBI). Iowa City, IA, USA: IEEE. pp. 1355–9. doi: 10.1109/ISBI45749.2020.9098338 [DOI] [Google Scholar]
- 44. Min Z, Bianco FJ, Yang Q, Rodell R, Yan W, Barratt D, et al. (2021). Controlling false Positive/Negative rates for deep-Learning-Based prostate cancer detection on multiparametric MR images, in: 25th UK Conference on Medical Image Understanding and Analysis. Switzerland: Springer Nature. pp. 56–70. doi: 10.48550/arXiv.2106.02385 [DOI] [Google Scholar]
- 45. Liu X, Han C, Cui Y, Xie T, Zhang X, Wang X. Detection and segmentation of pelvic bones metastases in MRI images for patients with prostate cancer based on deep learning. Front Oncol (2021) 11:773299. doi: 10.3389/fonc.2021.773299 [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
The datasets presented in this article are not readily available because the datasets are privately owned by Peking University First Hospital and are not made public. Requests to access the datasets should be directed to XYW, wangxiaoying@bjmu.edu.cn.