Abstract
Background
To develop and evaluate the feasibility of a data-driven deep learning approach (deepAC) for positron-emission tomography (PET) image attenuation correction without anatomical imaging. A PET attenuation correction pipeline was developed utilizing deep learning to generate continuously valued pseudo-computed tomography (CT) images from uncorrected 18F-fluorodeoxyglucose (18F-FDG) PET images. A deep convolutional encoder-decoder network was trained to identify tissue contrast in volumetric uncorrected PET images co-registered to CT data. A set of 100 retrospective 3D FDG PET head images was used to train the model. The model was evaluated in another 28 patients by comparing the generated pseudo-CT to the acquired CT using Dice coefficient and mean absolute error (MAE) and finally by comparing reconstructed PET images using the pseudo-CT and acquired CT for attenuation correction. Paired-sample t tests were used for statistical analysis to compare PET reconstruction error using deepAC with CT-based attenuation correction.
Results
deepAC produced pseudo-CTs with Dice coefficients of 0.80 ± 0.02 for air, 0.94 ± 0.01 for soft tissue, and 0.75 ± 0.03 for bone and MAE of 111 ± 16 HU relative to the PET/CT dataset. deepAC provides quantitatively accurate 18F-FDG PET results with average errors of less than 1% in most brain regions.
Conclusions
We have developed an automated approach (deepAC) that allows generation of a continuously valued pseudo-CT from a single 18F-FDG non-attenuation-corrected (NAC) PET image and evaluated it in PET/CT brain imaging.
Keywords: Deep learning, PET, CT, MRI, PET/MR, PET/CT, Attenuation correction
Background
Positron-emission tomography (PET) is a non-invasive imaging modality that provides direct imaging of biomarkers for physiology through the use of radiolabeled molecules, such as 18F-fluorodeoxyglucose (FDG) to assess glucose metabolism. PET activity is observed by detecting pairs of coincident gamma rays emitted from the PET tracer, sorted into sinograms, and reconstructed into a volumetric image. Knowledge of the tissue-dependent attenuation (typically Compton scatter) that gamma rays undergo is crucial to achieve quantitatively accurate PET reconstruction. In a typical PET scanner, the attenuation map (or μ-map) is obtained by performing additional anatomic imaging, which typically increases patient ionizing radiation dose. In a combined PET/CT system, an acquired CT image is used to generate an attenuation map for 511 keV photons by simple piecewise scaling of a CT image [1]. However, the attenuation map is estimated using a single snapshot in time which does not reflect misregistration due to patient motion between the acquisition of the CT and PET scans. In simultaneous PET/MR systems, estimation of the required attenuation map is based on MR images (which does not increase patient ionizing radiation dose) and is particularly challenging because the bone, the tissue with the largest attenuation coefficient, is not visible with positive contrast under typical MR acquisitions. Consequently, the bone is often ignored or estimated using atlas registration methods [2]. Specialized MRI acquisitions using a short echo time (STE), ultrashort echo time (UTE), or zero echo time (ZTE) can be implemented to allow the measurement of the rapidly decaying MR signal in the bone tissue to estimate the bone [3–6]. Unfortunately, UTE and ZTE acquisitions provide little clinical value compared to the conventional diagnostic imaging sequences. Additionally, even with advanced acquisitions, bony structure and air often remain difficult to distinguish, and errors remain in attenuation calculation [3, 7, 8].
Machine learning approaches have previously been proposed to estimate attenuation maps. A few pilot studies utilizing neural network methods have shown promising results in PET/MR imaging [9]. These studies typically rely on the inputs of T1-weighted MR images [10, 11], ultrashort-echo time MR images and transmission images [12] to estimate pseudo-CTs which can be applied to PET attenuation correction. More recently, deep learning approaches using Convolutional Neural Networks (CNN) have been applied to medical imaging with successful implementations across a diverse range of applications [13]. Deep learning methods are more advanced forms of neural networks and utilizing many levels of network structures capable of learning image features by a series of image convolution processes [14]. One recent study used deep learning to generate discrete pseudo-CTs for PET/MR attenuation correction using a single T1-weighted head image, which significantly reduced PET error in an evaluation dataset of ten brain images in comparison with vendor-provided segmentation- and atlas-based methods [15]. However, in all previous studies, besides PET data, additional acquisitions of either MR images or transmission images were required inputs. In this study, we propose a novel automated PET image attenuation correction approach, deepAC, that allows generation of a continuously valued pseudo-CT using only a non-attenuation-corrected (NAC) PET image as input. To the best of our knowledge, this proposed method is the first pilot study performing PET attenuation correction utilizing a self-regularized approach. The feasibility of this new approach is demonstrated in the human brain utilizing 18F-FDG PET/CT datasets and is compared to reconstruction utilizing the acquired CT. Given the favorable results of deepAC, we expect deep learning-based approaches to have a substantial impact to maintain accurate PET reconstruction while reducing the ionizing radiation dose and increasing the resilience to subject motion.
Methods
Convolutional encoder-decoder architecture
The key component of our proposed method is a deep convolutional encoder-decoder (CED) network, which is capable of mapping the NAC PET image into a pixel-wise continuously valued pseudo-CT image. The CED framework was modified based on the network structure used in a previous study for generating discrete three-class pseudo-CT in PET/MR attenuation correction [15]. A schematic demonstration of this deep learning network is shown in Fig. 1. The network features a connected encoder network and a decoder network. The encoder is designed to compress input image data while detecting robust and spatially invariant features. The Visual Geometry Group 16 (VGG16) network [16] was used as an encoder because this network has been proven to be efficient in capturing image features in object recognition and to be effective in CED-based medical image applications [7, 15, 17, 18]. Each unit layer in the VGG16 encoder consists of 2D convolution layer with a set of 2D filters, batch normalization (BN) [19], rectified linear unit (ReLU) activation [20], and followed by a max-pooling for the reduction of data dimensions. An increased number of 2D filters from 64 to 512 were generated by convolution layers from the first unit layer to the last in the encoder network to efficiently extract the input image features. Batch normalization was used to reduce the internal image covariate shift during network training so as to increase the training efficiency [19]. The ReLU activation function was used to perform non-linear transformation of the input image features so as to increase the network capability for learning a complex information. Max-pooling used a 2 × 2 window and stride 2 which leads to sub-sampled feature maps with a size reduction factor of 2. This unit layer was repeated 13 times in the VGG16 configuration to achieve sufficient data compression.
To reconstruct pixel-wise continuously valued CT images, a decoder network is applied directly after the encoder network. This decoder network is the reverse process of the encoder and consists of mirrored layers from the VGG16 network. The max-pooling layer in the decoder is replaced by an un-pooling layer (i.e., upsampling) where the image features were upsampled using bilinear interpolation. At the end of the decoder network, an additional 2D convolutional layer with one 2D filter was added to synthesize the output pseudo-CT images based on the decoded image features from the upstream decoder network.
In contrast to the network used in pseudo-CT generation in Reference [15], the proposed network also features shortcut connections (SC) which are used to forwardly pass the image features from an encoder network to a decoder network. As shown in the current study, synthesis of the continuously valued CT images requires preservation of additional rich image features compared to the discretely valued CT images in Reference [15], for which the added shortcut connections are valuable [21, 22]. Since it is well known that training a deep network is challenging due to the vanishing gradient problem [21], shortcut connections are helpful for enabling efficient training of networks with a large number of layers and complex structure. As shown in Fig. 1, the shortcut connections occur symmetrically in multiple layers in the networks and link feature maps by adding ones from the encoder network to the ones in the decoder network element-wise. A total of four shortcut connections were created between the network layers, and one additional shortcut connection was also generated from the input image directly to the output image. For each shortcut connection, the layer insertion (Fig. 1) to transfer feature maps occurred prior to the BN and ReLU activation in the unit layer of the decoder. This strategy was also described as the full pre-activation version of deep residual network design [22].
Proposed PET processing pipeline
The proposed processing pipeline consists of two independent phases for training retrospective data and reconstructing new data, respectively. In the training phase (Fig. 2), the training data for the proposed CED network consists of NAC PET images as input and reference non-contrast-enhanced CT data. For each training dataset, NAC PET and co-registered CT images were first offset to positive values and then scaled by pixel intensity of 6000 Bq/ml and 2000 Hounsfield unit (HU), respectively, to ensure a similar dynamic range. 3D NAC PET and CT images were cropped to enclose the image object with a minimally sized bounding box to remove redundant background prior to deep learning training. 2D axial slices from the 3D volumetric NAC and CT images were used as inputs to the deep learning network. All 2D input images were first transformed pixel-wise using a Softsign activation function in order to maintain a compact dynamic range and resampled to a matrix size of 200 × 180 using bilinear interpolation before being used as input to the CED. The encoder and decoder network weights were initialized using a normal distribution centered on 0 with a standard deviation varying according to the number of input weights. This initialization scheme was described in Reference [23]. Network updating used a gradient-based optimization algorithm (ADAM) [24] with a fixed and empirically selected learning rate of 0.001. The CED network iteratively estimated continuous CT images and compared them to the reference real CT data. The data consistency between estimated and real CT image was ensured by using mean squared error (MSE) as an image loss objective function where the loss was calculated in a mini-batch of 12 images in each iteration given our current hardware setup. The network training was performed for 20,000 iterations which corresponded to 50 epochs for our training data to achieve training loss convergence.
Once the training phase was complete, the CED network at the epoch with the least image loss was fixed and was used for generating continuous pseudo-CT for new PET data, which were subsequently used for PET reconstruction (Fig. 2).
In this study, the proposed framework was implemented in a hybrid computing environment involving Python (version 2.7, Python Software Foundation, Wilmington, DE, USA) and MATLAB (version 2013a, MathWorks, Natick, MA, USA). The image processing and analysis steps were performed in MATLAB. The CED network was coded using the Keras package with Tensorflow deep learning libraries as the computing backend [25].
Image datasets for training and evaluation
Our study was performed in compliance with the Health Insurance Portability and Accountability Act (HIPAA) regulations and with approval from our Institutional Review Board (IRB). Data collection for training and evaluation of the proposed deepAC method was performed utilizing an IRB-approved protocol for retrospective analysis. Subject eligibility criteria included any subjects who underwent a whole-body 18F-FDG PET and a non-contrast CT scan on the Discovery PET/CT 710 scanner (GE Healthcare, Waukesha, WI, USA) at our institution in 2016. Images from the first 140 consecutive patients meeting our eligibility criteria were retrieved from PACS in chronological order. Subjects were excluded from the training and evaluation if visual inspection revealed a spatial mismatch between NAC PET and CT images due to subject motion, which resulted in the removal of 12 subjects from the dataset. The dataset included both arms-up and arms-down positioning, with no discrimination upon either type of positioning. Subjects used for training and testing had a median age of 65 (range, 19–92) with 52 males and 76 females. Per our institution’s FDG PET protocol, subjects fasted for at least 6 h before the exam, had blood glucose less than 200 mg/dl, and were injected with 0.14 mCi/kg of tracer. Scanning began 56.5 ± 2.5 min after injection of FDG, with 3 min PET acquisitions per bed position. CT images were obtained with the following acquisition/reconstruction settings: 1.37 mm transaxial voxel dimensions, 5 mm slice thicknesses with 3.27 mm interslice spacing, 140 kVp, automatic exposure control with GE noise index of 25, and 0.52 helical pitch. NAC PET images were reconstructed offline (PET Toolbox, GE Healthcare) without attenuation correction and without time-of-flight, using the following parameters: 256 × 256 matrix, 700 × 700 mm2 field of view, OSEM reconstruction algorithm, 24 iterations, 3 subsets, SharpIR, and 4-mm post filter.
For training and evaluation, image data from the head was selectively used from each whole-body PET/CT dataset. The CED network was trained using 100 randomly selected subjects and evaluated in the remaining 28 subjects. All training and testing were performed on a desktop computer running a 64-bit Linux operating system with an Intel Xeon W3520 quad-core CPU, 12 GB DDR3 RAM, and a Nvidia Quadro K4200 graphic card (1344 CUDA cores, 4 GB GDDR5 RAM).
Pseudo CT generation and evaluation
After a NAC PET image is fed into the network, the output of the CED is scaled back to HUs. Analysis of pseudo-CT accuracy was performed on 28 subjects not included in the training phase of the CED network. Pixel-wise Dice coefficient, a similarity measure ranging from 0 to 1 that describes the overlap between two labels, was calculated for air, soft tissue, and bone. For calculation of Dice coefficients, the continuous pseudo-CT and real CT images were discretized by thresholding as follows: bone if HU > 300, air if HU < − 400, otherwise soft tissue. Additionally, the mean absolute error (MAE) between the generated pseudo-CT and real CT images was evaluated in each subject. Statistical analysis was performed using MATLAB.
PET image reconstruction
Offline PET reconstruction (PET Toolbox, GE Healthcare) was performed with a pseudo-CT image generated by deepAC and the acquired CT image for attenuation correction. PET reconstruction parameters were 256 × 256 matrix, 700 × 700 mm2 field of view, TOF-OSEM reconstruction algorithm, 24 iterations, 3 subsets, SharpIR, model-based scatter correction, and 4-mm post filter.
Evaluation of PET quantification
Evaluation of reconstructed PET image quality in deepAC was performed with the 28 testing subjects. PET images reconstructed using the deepAC-based attenuation correction were compared to those reconstructed using the acquired CT-based attenuation correction (CTAC). Pixel-wise error maps were obtained using the percentage error:
1 |
where IdeepAC and ICTAC are the PET image intensity (Bq/ml) from deepAC and CTAC, respectively. We also calculated the absolute value of Err as the absolute percentage error. Region of interest (ROI) analysis was performed using the IBASPM parcellation software with a PET brain atlas to compute ROI-level errors in 21 brain regions [26]. The data normality assumption was checked, and paired-sample t tests were used to perform the pairwise comparison for the PET activity in these brain regions using deepAC and CTAC. Statistical analysis was performed using MATLAB and R (R Development Core Team) with statistical significance defined as a p < 0.05 with Bonferroni correction for minimizing type-I error (equivalent to uncorrected p < 0.0024).
Results
An example of an acquired NAC PET image, deepAC pseudo-CT and real CT for a 48-year-old female subject is shown in Fig. 3. As shown in this figure, deepAC was able to identify the air, skull, and soft tissue in the NAC PET images and synthesize continuous CT values for distinct tissue types. The total training stage required approximately 30 h, whereas generating a single pseudo-CT output using the trained model required approximately 0.5 min. The training loss curve is shown in Fig. 3. The curve indicated a successful training convergence where the training loss gradually decreased prior to 40 epochs and reached a plateau after 40 epochs. Dice coefficients for the evaluation subset (n = 28) comparing the pseudo-CT to the real CT were high for air, 0.80 ± 0.02; soft tissue, 0.94 ± 0.01; and bone, 0.75 ± 0.03. The MAE between pseudo-CT and real CT was 111 ± 16 HU for all subjects with a range between 88 and 157 HU.
Another example of an acquired NAC PET image, deepAC pseudo-CT and real CT are shown in Fig. 4 for a non-compliant patient who exhibited significant movement between the PET and CT scans. This 52-year-old male subject is one of the cases excluded from training and PET evaluation. The misregistration can be clearly identified between NAC PET and real CT images in the axial and sagittal slices (red arrow) at the same image location. In contrast, the pseudo-CT generated from deepAC is free from misregistration with respect to the NAC PET data.
Figure 5 shows the reconstructed PET image for the subject of Fig. 3, utilizing the pseudo- and real CT for attenuation correction, respectively, as well as pixel-wise relative difference images between these. As seen in Fig. 5, deepAC results in PET error of less than 1% in most of the brain regions.
A challenging case for deepAC is shown in Fig. 6 for an 80-year-old female with a significant right and frontal skull abnormality. The generated pseudo-CT was able to predict the parts of missing skull in the forehead, indicated by the red arrows in the real CT image. The average reconstructed PET error for this subject in all brain pixels is 0.87%. Despite significant skull abnormalities relative to typical patients, PET reconstruction error was maintained at a low level utilizing deepAC.
Another case for deepAC is shown in Fig. 7 for a 59-year-old male with a brain tumor, indicated by the red arrow in the reconstructed PET image using CTAC. The average reconstructed PET error for this subject in all brain pixels is 0.52%. Despite the presence of brain tumor which was rare in our dataset, deepAC PET reconstruction was able to correctly identify tumor region and the global error was maintained at a low level.
Table 1 provides an average error, standard deviation, and p values from paired t tests within different brain regions across all 28 subjects. Reconstructed PET error on the ROI level showed that deepAC provided average PET errors below 1% for 15 out of 21 regions and below 2% for all regions tested. Paired t tests showed that for 14 out of 21 ROIs, mean PET activity was not significantly different (p > 0.05) between the deepAC approach and the reference PET (CTAC) image. For the ROIs that were significantly different, the average signed differences in PET values were small, which included left parietal lobe (p = 0.01, average difference = − 1.7%), right parietal lobe (p = 0.005, average difference = − 1.92%), left occipital lobe (p = 0.01, average difference = − 1.78%), right occipital lobe (p = 0.004, average difference = − 1.92%), right putamen (p = 0.009, difference = − 0.74%), right globus pallidus (p = 0.013, difference = − 0.56%), and left cingulate region (p = 0.049, difference = − 0.5%). Note that when corrected for multiple comparisons using a Bonferroni corrected significance level (p < 0.0024), none of the region-wise differences remained significant.
Table 1.
Brain regions | deepAC error (%) | deepAC absolute error (%) | p value (deepAC vs CTAC) |
---|---|---|---|
Frontal lobe left | − 1.04 ± 2.35 (− 6.67, 3.32) | 2.43 ± 1.57 (0.65, 6.67) | 0.18 |
Frontal lobe right | − 1.15 ± 2.56 (− 6.61, 4.06) | 2.59 ± 1.69 (0.82, 6.62) | 0.15 |
Temporal lobe left | − 0.79 ± 1.70 (− 4.45, 2.22) | 2.11 ± 0.94 (1.07, 4.67) | 0.18 |
Temporal lobe right | − 0.73 ± 1.99 (− 4.77, 2.68) | 2.32 ± 0.96 (1.10, 4.88) | 0.055 |
Parietal lobe left | − 1.70 ± 2.25 (− 5.56, 2.22) | 2.52 ± 1.63 (0.51, 5.56) | 0.01 |
Parietal lobe right | − 1.92 ± 2.38 (− 5.60, 2.28) | 2.79 ± 1.61 (0.69, 5.60) | 0.005 |
Occipital lobe left | − 1.78 ± 1.95 (− 6.24, 1.40) | 2.38 ± 1.43 (0.69, 6.26) | 0.01 |
Occipital lobe right | − 1.92 ± 2.15 (− 6.12, 2.81) | 2.75 ± 1.25 (0.89, 6.12) | 0.004 |
Cerebellum left | − 0.22 ± 1.62 (− 3.86, 2.68) | 1.70 ± 0.76 (0.67, 3.94) | 0.154 |
Cerebellum right | − 0.27 ± 1.78 (− 3.79, 2.83) | 1.78 ± 0.85 (0.54, 3.79) | 0.146 |
Brainstem | 0.69 ± 1.79 (− 3.20, 3.81) | 1.77 ± 0.88 (0.74, 3.81) | 0.354 |
Caudate nucleus left | 0.37 ± 1.71 (− 3.38, 3.69) | 1.50 ± 0.85 (0.43, 3.69) | 0.613 |
Caudate nucleus right | 0.32 ± 1.64 (− 3.47, 3.21) | 1.33 ± 0.86 (0.33, 3.47) | 0.451 |
Putamen left | − 0.67 ± 1.58 (− 3.80, 2.31) | 1.41 ± 1.00 (0.24, 3.80) | 0.115 |
Putamen right | − 0.74 ± 1.53 (− 3.94, 2.08) | 1.40 ± 1.02 (0.27, 3.94) | 0.009 |
Thalamus left | − 0.07 ± 1.59 (− 3.90, 3.05) | 1.40 ± 0.88 (0.31, 3.91) | 0.172 |
Thalamus right | 0.00 ± 1.56 (− 4.04, 3.27) | 1.34 ± 0.91 (0.20, 4.04) | 0.283 |
Globus pallidus left | − 0.39 ± 1.51 (− 3.71, 3.11) | 1.25 ± 0.91 (0.12, 3.71) | 0.056 |
Globus pallidus right | − 0.56 ± 1.41 (− 3.83, 2.35) | 1.24 ± 0.92 (0.29, 3.83) | 0.013 |
Cingulate region left | − 0.50 ± 1.68 (− 4.02, 2.70) | 1.61 ± 0.95 (0.35 4.02) | 0.049 |
Cingulate region right | − 0.45 ± 1.63 (− 3.81, 2.28) | 1.58 ± 0.87 (0.35, 3.81) | 0.057 |
All regions | − 0.64 ± 1.99 (− 4.18, 2.22) | 1.74 ± 0.94 (0.29, 4.20) |
Discussion
In this study, we demonstrated the feasibility of a new approach utilizing only 18F-FDG NAC PET images input into a deep learning model trained with PET/CT data to generate pseudo-CT images that can be used to perform quantitatively accurate PET-only imaging. For all ROIs tested, the percentage error in PET reconstruction was less than 2% on average, with the vast majority less than 1% on average. For individual subject data, the maximal percent error in any single region was 6.7%. When compared with CTAC reconstruction, this provided a statistically significant different quantitative PET result in only seven of the regions studied (yet still with an error of less than 2%). After correcting for multiple comparisons (Bonferroni correction), these differences were no longer statistically significant. Given the large training size of 100 subjects, these findings suggest that deep learning approaches can be leveraged to provide quantitatively accurate PET without the acquisition of a CT image. Given the inclusion of anatomically abnormal subjects into the evaluation dataset, e.g. as shown in Figs. 6 and 7, it can be expected that state-of-the-art deep learning approaches, such as the demonstrated herein, can indeed be robust to individual patient variations. With an increased number of training datasets, particularly datasets that encompass the range of variability and abnormalities present, this robustness would be expected to increase. Finally, the extension of deepAC to other regions outside the brain is expected to be valuable. However, we expect to require body-region-specific models such that models can be trained to best match the features of specific anatomical regions.
Reconstruction-based approaches have been previously proposed to synthesize μ-maps directly from PET data [27]. Recent embodiments of these approaches have demonstrated quantitatively accurate PET images with approximately 2–7% error [28, 29]. While these techniques use complicated reconstruction algorithms, they are also likely to benefit from the application of a deep learning approach such as deepAC, to augment or constrain data.
The applications of deepAC are expected to be numerous. In particular, methods that do not require the acquisition of a CT could be used to reduce the ionizing radiation exposure for patients undergoing repeated PET imaging studies. Depending on the protocol used, the patient exposure from CT is typically comparable to the whole-body equivalent PET exposure [30]. For pediatric and pregnant patients, this would significantly reduce ionizing radiation exposure. Other potential applications include management of misregistration between the PET and CT, where if significant motion occurs between the CT and PET acquisitions (e.g., Fig. 4), deepAC would be able to accurately compensate for this movement. The current implementation of deepAC does not necessarily help for fully non-compliant patients with significant motion during PET acquisition. One possible solution for this approach is to use deepAC to generate multi-phase dynamic pseudo-CTs to capture more intensive and dynamic motion. However, this would require additional prospective experimental design to obtain both dynamic PET and dynamic CT data. Additional applications of deepAC include utilization in research-only PET studies where CT adds minimal value beyond attenuation correction. Finally, the most impactful application is expected to be with PET/MR imaging, where a pseudo-CT could be generated directly for PET NAC data, requiring no additional time to acquire an attenuation correction scan, improving PET/MR workflow.
This study has several limitations. First, the training group was selected from a clinical database. While the included patients accurately reflect a population of subjects undergoing PET, the group was not selected to be optimal for healthy volunteers with no known pathology. However, the data is representative of a realistic patient population. With the addition of an increased number of datasets in the training data, it is expected that even greater resilience to this potential source of bias (which was not observed herein) could be achieved. A second limitation is that the trained model is only applicable to 18F-FDG PET data. Development of models for other radiotracers would require training sets of PET/CT data utilizing that specific tracer. This could be potentially limiting in tracers that are highly specific and do not have a sufficiently global distribution, as it would be expected that the deep learning model would require patterns of global physiological uptake to fully train the model. Furthermore, the model was trained on 18F-FDG PET data that was acquired 60 min after the administration. Utilization of deepAC with agents that have significant changes in regional uptake based and rapid pharmacokinetics may also be challenging. However, in the context of clinically routine 18F-FDG PET imaging as demonstrated herein, deepAC performs considerably well. Third, for the ROI approach used herein to assess PET reconstruction error, the computed error is averaged within the entire volume of the ROI. In certain regions of the brain, the error may not be distributed uniformly, and may actually be larger near the interfaces of air, bone, and soft tissue. For example, the cortical regions of the brain that are near the skull may have a larger error relative to the regions further inside the skull. However, these errors are perhaps better depicted visually, as can be seen in Figs. 5 and 6. Fourth, the current study implemented 2D CNN for pseudo-CT generation from NAC images. Potential improvement for the generation quality might be achieved using methods considering 3D image context, such as using 3D CNNs or augmented 2D CNNs with 3D spatial refinement [7, 17, 18, 31]. Finally, the pseudo-CT generated by deepAC is not a replacement for a diagnostic CT. While it may provide an anatomical reference, it does not necessarily reflect the true underlying tissue contrast. For studies that require a true anatomical reference, additional imaging must be performed. In particular, we expect extensions of deepAC for PET/MR applications, where anatomic imaging is readily available, to be particularly valuable.
Conclusions
We have demonstrated a deep learning approach to produce accurate quantitative PET imaging by using only NAC 18F-FDG PET images. Such approaches will likely have a substantial impact on future work in PET, PET/CT, and PET/MR studies to reduce ionizing radiation dose and increase resilience to subject misregistration between the PET acquisition and attenuation map acquisition.
Availability of data and materials
All results are provided in the manuscript. The datasets supporting the conclusions of this article are available upon request.
Abbreviations
- 18F-FDG
18F-fluorodeoxyglucose
- BN
Batch normalization
- CED
Convolutional encoder-decoder
- CNN
Convolutional Neural Network
- CT
Computed tomography
- deepAC
Deep attenuation correction
- MAE
Mean absolute error
- MR
Magnetic resonance
- NAC
Non-attenuation-corrected
- PET
Positron-emission tomography
- ReLU
Rectified linear unit
- VGG16
Visual Geometry Group 16
Authors’ contributions
FL, HJ, TB, and ABM are the guarantors of integrity of the entire study. All authors contributed to the study concepts/study design, data acquisition, data analysis/interpretation, manuscript drafting, manuscript revision for important intellectual content, literature research, and manuscript editing. All authors approved the final version of the manuscript. All authors agree to ensure that any questions related to the work are appropriately resolved. FL, HJ, and TB contributed to the clinical studies. FL and HJ contributed to the experimental studies and statistical analysis.
Ethics approval and consent to participate
Our study was performed in compliance with the Health Insurance Portability and Accountability Act (HIPAA) regulations and with approval from our Institutional Review Board (IRB). Data collection for training and evaluation of the proposed method was performed utilizing an IRB-approved protocol for retrospective analysis.
Consent for publication
Not applicable.
Competing interests
The authors declare that they have no competing interests.
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
References
- 1.Burger C, Goerres G, Schoenes S, Buck A, Lonn A, von Schulthess G. PET attenuation coefficients from CT images: experimental evaluation of the transformation of CT into PET 511-keV attenuation coefficients. Eur J Nucl Med Mol Imaging. 2002;29:922–927. doi: 10.1007/s00259-002-0796-3. [DOI] [PubMed] [Google Scholar]
- 2.Ladefoged CN, Law I, Anazodo U, Lawrence KS, Izquierdo-Garcia D, Catana C, et al. A multi-centre evaluation of eleven clinically feasible brain PET/MRI attenuation correction techniques using a large cohort of patients. NeuroImage. 2016;147:346–359. doi: 10.1016/j.neuroimage.2016.12.010. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Jang H, Liu F, Bradshaw T, AB MM. Rapid dual echo ramped hybrid encoding MR-based attenuation correction (dRHE-MRAC) for PET/MR. Magn Reson Med Off J Soc Magn Reson Med Magn Reson Med. 2017; In press. Available from: http://doi.wiley.com/10.1002/mrm.26953. Cited 6 Oct 2017. [DOI] [PMC free article] [PubMed]
- 4.Leynes AP, Yang J, Shanbhag DD, Kaushik SS, Seo Y, Hope TA, et al. Hybrid ZTE/Dixon MR-based attenuation correction for quantitative uptake estimation of pelvic lesions in PET/MRI. Med Phys. 2017;44:902–913. doi: 10.1002/mp.12122. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.Wiesinger F, Bylund M, Yang J, Kaushik S, Shanbhag D, Ahn S, et al. Zero TE-based pseudo-CT image conversion in the head and its application in PET/MR attenuation correction and MR-guided radiation therapy planning. Magn Reson Med. 2018; Available from: http://doi.wiley.com/10.1002/mrm.27134. Cited 27 Mar 2018. [DOI] [PubMed]
- 6.Fathi Kazerooni A, Ay MR, Arfaie S, Khateri P, Saligheh Rad H. Single STE-MR acquisition in MR-based attenuation correction of brain PET imaging employing a fully automated and reproducible level-set segmentation approach. Mol Imaging Biol. 2017;19:143–152. doi: 10.1007/s11307-016-0990-5. [DOI] [PubMed] [Google Scholar]
- 7.Jang H, Liu F, Zhao G, Bradshaw T, McMillan AB. Technical note: deep learning based MRAC using rapid ultra-short echo time imaging. Med Phys. 2018; In-press. Available from: http://www.ncbi.nlm.nih.gov/pubmed/29763997. Cited 11 June 2018. [DOI] [PMC free article] [PubMed]
- 8.Leynes AP, Yang J, Wiesinger F, Kaushik SS, Shanbhag DD, Seo Y, et al. Direct pseudoCT generation for pelvis PET/MRI attenuation correction using deep convolutional neural networks with multi-parametric MRI: zero echo-time and dixon deep pseudoCT (ZeDD-CT). J Nucl Med. 2017:jnumed.117.198051. Available from: http://www.ncbi.nlm.nih.gov/pubmed/29084824. Cited 3 Nov 2017. [DOI] [PMC free article] [PubMed]
- 9.Mehranian A, Arabi H, Zaidi H. Vision 20/20: magnetic resonance imaging-guided attenuation correction in PET/MRI: challenges, solutions, and opportunities. Med Phys. 2016;43:1130–1155. doi: 10.1118/1.4941014. [DOI] [PubMed] [Google Scholar]
- 10.Wagenknecht G, Kops ER, Kaffanke J, Tellmann L, Mottaghy F, Piroth MD, et al. IEEE Nuclear Science Symposium Conference Record. 2010. CT-based evaluation of segmented head regions for attenuation correction in MR-PET systems; pp. 2793–2797. [Google Scholar]
- 11.Wagenknecht G, Kops ER, Tellmann L, Herzog H. IEEE Nuclear Science Symposium Conference Record. 2009. Knowledge-based segmentation of attenuation-relevant regions of the head in T1-weighted MR images for attenuation correction in MR/PET systems; pp. 3338–3343. [Google Scholar]
- 12.Santos Ribeiro A, Rota Kops E, Herzog H, Almeida P. Hybrid approach for attenuation correction in PET/MR scanners. Nucl Inst Methods Phys Res A. 2014;734:166–170. doi: 10.1016/j.nima.2013.09.034. [DOI] [Google Scholar]
- 13.Litjens G, Kooi T, Bejnordi BE, Setio AAA, Ciompi F, Ghafoorian M, et al. A survey on deep learning in medical image analysis. Med Image Anal. 2017;42:60–88. doi: 10.1016/j.media.2017.07.005. [DOI] [PubMed] [Google Scholar]
- 14.LeCun Y, Bengio Y, Hinton G. Deep learning. Nature. 2015;521:436–444. doi: 10.1038/nature14539. [DOI] [PubMed] [Google Scholar]
- 15.Liu F, Jang H, Kijowski R, Bradshaw T, McMillan AB. Deep learning MR imaging–based attenuation correction for PET/MR imaging. Radiology. 2017:170700. Available from: http://pubs.rsna.org/doi/10.1148/radiol.2017170700. Cited 26 Sept 2017. [DOI] [PMC free article] [PubMed]
- 16.Simonyan K, Zisserman A. Very deep convolutional networks for large-scale image recognition. ArXiv e-prints [Internet]. 2014;1409.1556. Available from: http://arxiv.org/abs/1409.1556. Cited 27 Mar 2018.
- 17.Liu F, Zhou Z, Jang H, Samsonov A, Zhao G, Kijowski R. Deep convolutional neural network and 3D deformable approach for tissue segmentation in musculoskeletal magnetic resonance imaging. Magn Reson Med. 2017; 10.1002/mrm.26841. Available from: http://www.ncbi.nlm.nih.gov/pubmed/28733975. Cited 29 July 2017. [DOI] [PMC free article] [PubMed]
- 18.Zhou Z, Zhao G, Kijowski R, Liu F. Deep convolutional neural network for segmentation of knee joint anatomy. Magn Reson Med. 2018; 10.1002/mrm.27229. [DOI] [PMC free article] [PubMed]
- 19.Ioffe S, Szegedy C. Batch normalization: accelerating deep network training by reducing internal covariate shift [Internet]. ArXiv e-prints. 2015. Available from: http://adsabs.harvard.edu/abs/2015arXiv150203167I. Accessed 11 Mar 2017.
- 20.Nair V, Hinton GE. Proceedings 27th International Conference on Machine Learning. 2010. Rectified linear units improve restricted boltzmann machines; pp. 807–814. [Google Scholar]
- 21.He K, Zhang X, Ren S, Sun J. Deep residual learning for image recognition. ArXiv e-prints [Internet]. 2015. Available from: http://arxiv.org/abs/1512.03385. Cited 14 May 2017.
- 22.He K, Zhang X, Ren S, Sun J. Identity mappings in deep residual networks. ArXiv e-prints [Internet]. 2016. Available from: http://arxiv.org/abs/1603.05027. Cited 14 May 2017.
- 23.He K, Zhang X, Ren S, Sun J. Delving deep into rectifiers: surpassing human-level performance on ImageNet classification [Internet]. ArXiv e-prints. 2015. Available from: http://adsabs.harvard.edu/abs/2015arXiv150201852H. Accessed 5 June 2017.
- 24.Kingma DP, Ba J. Adam: a method for stochastic optimization. ArXiv e-prints [Internet]. 2014. Available from: http://arxiv.org/abs/1412.6980. Cited 5 July 2017.
- 25.Abadi M, Agarwal A, Barham P, Brevdo E, Chen Z, Citro C, et al. TensorFlow: large-scale machine learning on heterogeneous distributed systems. ArXiv e-prints [Internet]. 2016; Available from: http://arxiv.org/abs/1603.04467. Cited 30 June 2017.
- 26.Alemán-Gómez Y, Melie-García L, Valdés-Hernandez P. 12th Annual Meeting of the Organization for Human Brain Mapping Florence, Italy. 2006. IBASPM: toolbox for automatic parcellation of brain structures. [Google Scholar]
- 27.Censor Y, Gustafson DE, Lent A, Tuy H. A new approach to the emission computerized tomography problem: simultaneous calculation of attenuation and activity coefficients. IEEE Trans Nucl Sci. 1979;26:2775–2779. doi: 10.1109/TNS.1979.4330535. [DOI] [Google Scholar]
- 28.Mehranian A, Zaidi H. Joint estimation of activity and attenuation in whole-body TOF PET/MRI using constrained Gaussian mixture models. IEEE Trans Med Imaging. 2015;34:1808–1821. doi: 10.1109/TMI.2015.2409157. [DOI] [PubMed] [Google Scholar]
- 29.Benoit D, Ladefoged CN, Rezaei A, Keller SH, Andersen FL, Højgaard L, et al. Optimized MLAA for quantitative non-TOF PET/MR of the brain. Phys Med Biol. 2016;61:8854–8874. doi: 10.1088/1361-6560/61/24/8854. [DOI] [PubMed] [Google Scholar]
- 30.Huang B, Law MW-M, Khong P-L. Whole-body PET/CT scanning: estimation of radiation dose and cancer risk. Radiology. 2009;251:166–174. doi: 10.1148/radiol.2511081300. [DOI] [PubMed] [Google Scholar]
- 31.Xiang L, Wang Q, Nie D, Zhang L, Jin X, Qiao Y, et al. Deep embedding convolutional neural network for synthesizing CT image from T1-weighted MR image. Med Image Anal. 2017;47:31–44. doi: 10.1016/j.media.2018.03.011. [DOI] [PMC free article] [PubMed] [Google Scholar]