Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2022 Sep 1.
Published in final edited form as: Med Phys. 2021 Jul 27;48(9):5115–5129. doi: 10.1002/mp.15073

Generation of synthetic PET images of synaptic density and amyloid from 18F-FDG images using deep learning

Rui Wang 1,2,3, Hui Liu 1,2,3, Takuya Toyonaga 1, Luyao Shi 1, Jing Wu 1, John Aaron Onofrey 1, Yu-Jung Tsai 1, Mika Naganawa 1, Tianyu Ma 2,3, Yaqiang Liu 2,3, Ming-Kai Chen 1, Adam P Mecca 4,5, Ryan S O’Dell 4,5, Christopher H van Dyck 4,5, Richard E Carson 1, Chi Liu 1
PMCID: PMC8455448  NIHMSID: NIHMS1737581  PMID: 34224153

Abstract

Purpose:

Positron emission tomography (PET) imaging with various tracers is increasingly used in Alzheimer’s disease (AD) studies. However, access to PET scans using new or less-available tracers with sophisticated synthesis and short half-life isotopes may be very limited. Therefore, it is of great significance and interest in AD research to assess the feasibility of generating synthetic PET images of less-available tracers from the PET image of another common tracer, in particular 18F-FDG.

Methods:

We implemented advanced deep learning methods using the U-Net model to predict 11C-UCB-J PET images of synaptic vesicle protein 2A (SV2A), a surrogate of synaptic density, from 18F-FDG PET data. Dynamic 18F-FDG and 11C-UCB-J scans were performed in 21 participants with normal cognition (CN) and 33 participants with Alzheimer’s disease (AD). Cerebellum was used as the reference region for both tracers. For 11C-UCB-J image prediction, four network models were trained and tested, which included 1) 18F-FDG SUV ratio (SUVR) to 11C-UCB-J SUVR, 2) 18F-FDG Ki ratio to 11C-UCB-J SUVR, 3) 18F-FDG SUVR to 11C-UCB-J distribution volume ratio (DVR), and 4) 18F-FDG Ki ratio to 11C-UCB-J DVR. The normalized root mean square error (NRMSE), structure similarity index (SSIM), and Pearson’s correlation coefficient were calculated for evaluating the overall image prediction accuracy. Mean bias of various ROIs in the brain and correlation plots between predicted images and true images were calculated for ROI-based prediction accuracy. Following a similar training and evaluation strategy, 18F-FDG SUVR to 11C-PiB SUVR network was also trained and tested for 11C-PiB static image prediction.

Results:

The results showed that all four network models obtained satisfactory 11C-UCB-J static and parametric images. For 11C-UCB-J SUVR prediction, the mean ROI bias was −0.3% ± 7.4% for the AD group and −0.5% ± 7.3% for the CN group with 18F-FDG SUVR as the input, −0.7% ± 8.1% for the AD group, and −1.3% ± 7.0% for the CN group with 18F-FDG Ki ratio as the input. For 11C-UCB-J DVR prediction, the mean ROI bias was −1.3% ± 7.5% for the AD group and −2.0% ± 6.9% for the CN group with 18F-FDG SUVR as the input, −0.7% ± 9.0% for the AD group, and −1.7% ± 7.8% for the CN group with 18F-FDG Ki ratio as the input. For 11C-PiB SUVR image prediction, which appears to be a more challenging task, the incorporation of additional diagnostic information into the network is needed to control the bias below 5% for most ROIs.

Conclusions:

It is feasible to use 3D U-Net-based methods to generate synthetic 11C-UCB-J PET images from 18F-FDG images with reasonable prediction accuracy. It is also possible to predict 11C-PiB SUVR images from 18F-FDG images, though the incorporation of additional non-imaging information is needed.

Keywords: brain PET, deep learning, image processing, multi-tracer, parametric image

1 |. INTRODUCTION

Alzheimer’s disease (AD) is an insidious neurodegenerative disorder resulting in significant progressive cognitive and functional decline, and is the most common cause of dementia, accounting for approximately 60%–80% of cases. AD pathology is traditionally characterized by a distinct accumulation of extracellular β-amyloid plaques and intracellular neurofibrillary tangles (NFTs) of tau,1,2 but also demonstrates significant spatiotemporal alterations in synaptic density and glucose metabolism.3 Accordingly, the clinical symptoms of AD are related to a distinct pathology, including β-amyloid plaques and neurofibrillary tangles, and are well-correlated with synaptic loss and abnormal glucose metabolism in the brain. Positron emission tomography (PET) imaging has been increasingly employed in AD studies to detect this accumulation of β-amyloid4 and NFTs and alterations in glucose metabolism5 and synaptic density3,6 with specific tracers in both cross-sectional and longitudinal models.

Traditionally, 18F-FDG-PET has been used as a surrogate marker of disease progression, measuring synaptic activity via glucose metabolism.7 However, glucose metabolism cannot be interpreted as a direct measure of synaptic density, and 18F-FDG can easily be confounded by medications, sensory stimulation, and glucose levels.8,9 Alternatively, to assess synaptic loss more directly, one kind of suitable target is the synaptic vesicle glycoprotein 2 (SV2). One of its isoforms, SV2A, is ubiquitously expressed in neurons and is involved in the regulation of neurotransmitter release.6 A recently developed tracer, 11C-UCB-J, has been demonstrated to detect SV2A quantitatively in the brain, giving the possibility for detecting synaptic density in the brain directly.10 Furthermore, recent brain PET imaging studies have demonstrated that SV2A is a promising biomarker to quantify the loss of synaptic density in the AD.11,12 Besides, β-amyloid plaques and abnormal tau protein deposition in the form of neurofibrillary tangles are also two defining pathological indicators and hallmarks of AD.13 Pittsburgh Compound-B (PiB) is a widely used PET tracer to provide quantitative information on fibrillar amyloid deposits.2 Among these tracers, 11C-UCB-J and 11C-PiB are labeled with 11C, which has a 20-min half-life. The generation of those tracers requires an onsite cyclotron, which is not easily accessible in most research centers. Besides, it can often be burdensome for patients to have multiple PET scans. Therefore, it is of potential significance and interest in AD research to assess whether synthetic 11C-UCB-J or 11C-PiB images can be conveniently created from a single scan of the most clinically available tracer, 18F-FDG.

Deep learning has been in a wide variety of fields, including computer vision, speech recognition, and object detection.14 Two networks, the encoder-decoder networks15 and generative adversarial networks (GANs)16 accelerated the use of deep learning in the image-to-image translation domain17 and image restoration.18 Specifically, in the nuclear medicine imaging field, the convolutional network has been used for image denoising, attenuation correction, and image reconstruction. Recently, deep learning methods have been demonstrated to be effective in cross-modality image-to-image translation, including the synthesis of CT images from MRI using deep convolutional neural networks,19 MR to CT translation with deep convolution adversarial networks,20 β-amyloid PET to MRI image translation with GANs,21 MRI images to higher SNR PET images with the U-Net model,22 and generation of attenuation maps from SPECT emission data using deep learning models.23

Inspired by cross-modality image-to-image translation, we proposed cross-tracer PET image translation methods using deep learning. Specifically, we observed that the level of glucose metabolism in brain regions is positively correlated with the amount of the SV2A,24,25 which provides the possibility of predicting the PET image of 11C-UCB-J from the PET image with 18F-FDG. Conversely, if some of the 11C-UCB-J signals cannot be reproduced from 18F-FDG by this methodology, that may provide information about aspects of the 11C-UCB-J and 18F-FDG signals that are independent.

In this work, we investigated a fully 3D U-Net architecture to predict the brain PET image of 11C-UCB-J based on the brain PET image of 18F-FDG. In addition to investigating deep learning networks for static PET studies, we also trained the network for dynamic parametric PET images. We comprehensively optimized parameters and systematically investigated the quantification accuracy of ROIs in the brain using U-Net-based image prediction approaches. In addition, we explored a similar network training strategy to generate synthetic 11C-PiB static images from 18F-FDG images.

2 |. MATERIALS AND METHODS

2.1 |. Human study dataset and brain imaging

As shown in Table 1, we included 54 subjects (25 males and 29 females; 33 Alzheimer’s disease [AD] and 21 cognitively normal [CN]; age 50–84). Similar datasets, as well as screening procedures and eligibility criteria, have been previously described.12 Subjects underwent three brain dynamic PET scans of 18F-FDG, 11C-UCB-J, and 11C-PiB acquired on an HRRT scanner. The mean injected activity was ~5 mCi for 18F-FDG, ~16 mCi for 11C-UCB-J, and ~13 mCi for 11C-PiB. All the data were reconstructed with a motion-compensation list-mode reconstruction method,26 in which OSEM was used with 2 iterations and 30 subsets. The original reconstructed images included 27 frames in 90 min after injection. The size of each image was 256 × 256 × 207 with a voxel size of 1.219 × 1.219 × 1.231 mm3.

TABLE 1.

Demographics and clinical diagnosis of dataset

Characteristic Dataset (n = 54)
Age, mean (SD) [range], years 71.12 (8.07) [50.29–84.49]
Sex (Male: Female) 25:29
Weight, mean (SD) [range], kg 76.13 (16.24) [45–108.3]
18F-FDG Injection Activity, mean (SD) [range], mCi 4.85 (0.27) [3.74–5.15]
11C-UCB-J Injection Activity, mean (SD) [range], mCi 16.24 (4.71) [4.08–20.59]
11C-PiB-J Injection Activity, mean (SD) [range], mCi 12.58 (2.64) [4.78–14.82]
Diagnosis(CN:AD) 21:33

Abbreviations: AD, Alzheimer’s Disease; CN, cognitively normal.

For the static PET dataset, standard uptake value (SUV) images were generated from 60 to 90 min for 18F-FDG, 40 to 60 min for 11C-UCB-J, and 50 to 70 min for 11C-PiB. Then the SUVR image with the whole cerebellum as the reference was computed for all the SUV images. For the parametric PET dataset, 18F-FDG net uptake parameter Ki images were estimated with Patlak analysis using a population-based input function (PBIF). The time window for Patlak analysis was 30 to 90 min post-injection. Ki ratio images with the cerebellum as the reference region were calculated as (Ki/Ki[cerebellum]). Images of 11C-UCB-J non-displaceable binding potential (BPND) were estimated with the Simplified Reference Tissue Model 2 (SRTM2) method by 90 min of data with centrum semiovale as the reference tissue.27 Then, the distribution volume ratio with the cerebellum as the reference region (DVRCb) was calculated as (BPND + 1)/(BPND[cerebellum] + 1).

For each subject, a T1-weighted magnetic resonance imaging (MRI) scan was obtained at the Yale MRI Center for the registration between PET and MR images as well as the definition for regions of interest (ROI). Registration to MRI was implemented by the BioImage Suite software for the PET images.28 Specifically, both the 18F-FDG images of SUVR and Ki Ratio and the 11C-UCB-J images of SUVR and DVR of the same subject were registered with a rigid transformation to each individual’s MR image. The image size after registration was resliced to 176 × 240 × 256 with a voxel size of 1.2 × 1.055 × 1.055 mm3 in MR space for both 18F-FDG and 11C-UCB-J images. Then, a brain mask derived from the aligned MR image was used to crop the images. ROIs were derived from the MR images using FreeSurfer version 6 (Massachusetts General Hospital).29 Similarly, we registered 18F-FDG datasets (SUVR and Ki Ratio) and 11C-PiB datasets (SUVR and Ki Ratio) to the MR space and generated ROIs from the MR image of each subject.

2.2 |. Deep learning network architecture

U-Net is a convolutional network architecture that has been widely applied to biomedical image processing. Unlike the fully convolutional network,30 a large number of filters are also applied in the upsampling step in order to propagate context information to higher resolution layers.15

Since training such a 3D network based on entire images is computationally intensive, we extracted patches randomly from the images within the brain mask as the input and output. Moreover, the training data in terms of the number of patches are much larger than the number of images. The structure of the patch-based U-Net is shown in Figure 1. It consists of a contracting path (4 layers) and an expansive path (4 layers). U-net has concatenating skip connections between the contracting path and the corresponding expansive path, which makes up a U-shape. The contracting path contains a layer of a 3 × 3 × 3 convolution kernel followed by a rectified linear unit (ReLU) and a layer of 2 × 2 × 2 max pooling operation. The expansive path consists of a 2 × 2 × 2 up-convolution kernel and a 3 × 3 × 3 convolution kernel, followed by a ReLU. Symmetric padding was applied to maintain the same patch size after convolution. A fully convolutional layer was used to map the feature map to the label at the final stage. Thirty-two features were extracted in the first layer and were doubled at each downsampling step, then halved at each upsampling step as shown in Figure 1.

FIGURE 1.

FIGURE 1

The Architecture of The U-Net 3D model

2.3 |. Experimental setup

For 11C-UCB-J image generation, two groups of studies were conducted to investigate the prediction accuracy of both SUVR and DVR images of 11C-UCB-J using four networks, as summarized in Table 2. Two networks aimed to predict static 11C-UCB-J SUVR images, where 18F-FDG SUVR images were used as the input data for Network #1 and 18F-FDG Ki Ratio images were used as the input data for Network #2. Another two networks were trained to predict parametric 11C-UCB-J DVR images, where 18F-FDG SUVR images were used as the input data for Network #3 and 18F-FDG Ki Ratio images were used as the input data for Network #4. In 11C-PiB PET image generation, 18F-FDG SUVR images were used as input, and 11C-PiB SUVR images were used as output for Network #5.

TABLE 2.

Five networks used for 11C-UCB-J and 11C-PiB image generation

Network ID Input image
Output image
18F-FDG 11C-UCB-J 11C-PiB
#1 SUVR SUVR /
#2 K i Ratio SUVR /
#3 SUVR DVR /
#4 K i Ratio DVR /
#5 SUVR / SUVR

During the training, 240 epochs each containing 16000 patches were chosen for the convergence of the cost function. Each epoch consisted of 500 small batches. The patch size of 32 × 32 × 32 was applied in the 11C-UCB-J image generation, while 64 × 64 × 64 was applied in the 11C-PiB SUVR image generation. We chose the patch size based on a pilot study31,32 (methods in Supplemental Materials), where we found the performances of 32 × 32 × 32 and 64 × 64 × 64 patches were similar for 11C-UCB-J, while 64 × 64 × 64 patches outperformed 32 × 32 × 32 patches for 11C-PiB image generation(Figure S1). The Adam optimizer was applied with an initial learning rate of 0.001 and the learning decay rate of 0.96 by each epoch. L1 norm was used to calculate the loss between the generated patch and the target patch for the update since it showed better performance and faster convergence speed than L2 during the training in this case. For testing, the patch size was set to 240 × 240 × 160 which is larger than the size of the input image to cover the whole image of each subject without overlap. Since we had 54 subjects, a “leave-11-out” cross-validation method was used, which means 43 subjects were used for training and the other 11 subjects were used for testing. Then we repeated this process four times until all the subjects had been used for testing. Subjects for both training and testing were selected randomly from AD and CN groups separately to form a balanced group with approximately the same ratio between the number of AD and CN subjects. All the experiments were carried out on the server at Yale PET Center with two GPUs (NVIDIA Quadra RTX 8000).

The proposed deep learning-based image synthesis approach was also compared to a simple fitting approach, based on the potential correlation between glucose metabolism and synaptic density. 11C-UCB-J image prediction from 18F-FDG images using the polynomial fitting with the same dataset was applied. A population-averaged fitting curve was established for all brain voxels in 18F-FDG images as the input variables and the matching voxels in 11C-UCB-J images as the target variables based on the same cross-validation strategy, which means the fitting curve was established based on 43 subjects, then the prediction was applied to 11 other testing subjects. This process was repeated four times until all the subjects were used for testing.

2.4 |. Statistical analysis

The normalized root mean square error (NRMSE) inside the brain between the predicted image and the true image was calculated as follows:

NRMSE=IPredITrue2/ITrue2,

where IPred is the predicted image from the 18F-FDG image and ITrue is the true image of 11C-UCB-J or 11C-PiB. Structural similarity (SSIM) between the predicted image and the true image was calculated as follows:

SSIM=(2μxμy+c1)(2σxy+c2)(μx2+μy2+c1)(σx2+σy2+c2),

where μx and μy are the mean values across all the voxels within the brain mask of the predicted and true images, σx and σy are the standard deviations of the predicted and true images across all voxels, and σxy is the covariance of the two images. c1 and c2 are two variables to stabilize the division with a weak denominator. By default, c1 = (0.01 × L)2, c2 = (0.03 × L)2, where L is the specified dynamic range value. Pearson’s correlation coefficient between the predicted image and the true image was calculated as follows:

ρX,Y=cov(X,Y)σXσY,

where X and Y are the vector representations of the predicted image and true image. For ROI-based quantification, the percent bias between the predicted image and the ground truth in a number of typical ROIs were calculated as follows:

Bias=μx,Rμy,Rμy,R×100%,

where R is the specific ROI as shown in Table 3.

TABLE 3.

Brain regions-of-interest

Index ROI Index ROI
1 Hippocampus 10 Insula
2 Entorhinal 11 Thalamus
3 Precuneus 12 Caudate
4 Posterior Cingulate 13 Putamen
5 Parietal 14 Pallidum
6 Pons 15 Amygdala
7 Frontal 16 Cerebral white matter
8 Occipital 17 Anterior Cingulate
9 Temporal

Independent samples t-tests were applied between the CN and AD groups to test if there is a group difference in ROI percentage bias. A one-sample t-test was applied to check if the mean of ROI percentage bias is significantly different from zero for each network to compare the difference between predicted images and the true images. In addition, scatter plots with linear fitting were drawn to evaluate the consistency between the true images and the predicted images.

3 |. RESULTS

As shown in Figure 2, both using 18F-FDG SUVR and Ki Ratio images as the input can provide a satisfactory prediction of 11C-UCB-J SUVR images. Visually, the predicted images appear to have less noise than the true images, and the predicted images are smoother than the true images. The smooth effect was also observed in other image denoising studies using the U-Net model.32 The U-Net model minimized the squared error between the predicted image and the true image, which may cause a smooth effect during prediction.33 The mean NRMSE, SSIM, and Pearson coefficient across all AD and CN subjects for the 18F-FDG SUVR11C-UCB-J SUVR network and 18F-FDG Ki Ratio11C-UCB-J SUVR network are listed in Table 4. The SSIM and Pearson coefficient between the predicted 11C-UCB-J SUVR image and the true image were statistically larger than that between the 18F-FDG input image and true 11C-UCB-J SUVR for Network #1 and Network #2 according to paired-sample t-tests. The NRMSE of 18F-FDG SUVR11C-UCB-J SUVR network is 16.2% ± 2.9% for the AD group and 15.6% ± 2.6% for the CN group, which outperformed the prediction results of 18F-FDG Ki Ratio11C-UCB-J SUVR network. The SSIM and Pearson coefficient of 18F-FDG SUVR11C-UCB-J SUVR network outperformed evaluation results of 18F-FDG Ki Ratio11C-UCB-J SUVR network as well. Besides, using both 18F-FDG SUVR images and 18F-FDG Ki Ratio images as input could provide a reasonable prediction of 11C-UCB-J DVR images. Using 18F-FDG SUVR image as the input provided superior NRMSE, SSIM, and Pearson coefficient across all the subjects as compared to using 18F-FDG Ki Ratio images as input as shown in Table 4.

FIGURE 2.

FIGURE 2

Representative predicted results of four networks including #1: 18F-FDG SUVR11C-UCB-J SUVR, #2: 18F-FDG Ki Ratio11C-UCB-J SUVR, #3: 18F-FDG SUVR11C-UCB-J DVR, and #4: 18F-FDG Ki Ratio11C-UCB-J DVR for (a) one Alzheimer’s disease subject and (b) one cognitively normal subject

TABLE 4.

NRMSE, SSIM, and Pearson correlation coefficients of predicted results compared to true 11C-UCB-J images from four networks

#1: SUVR to SUVR
#2: Ki Ratio to SUVR
#3: SUVR to DVR
#4: Ki Ratio to DVR
Pred-True Input-True Pred-True Input-True Pred-True Input-True Pred-True Input-True
AD
 NRMSE 16.2% (2.9%) / 17.8% (3.3%) / 18.1% (3.3%) / 20.0% (3.5%) /
 SSIM 0.903 (0.033) 0.866 (0.037) 0.880 (0.032) 0.737 (0.037) 0.890 (0.029) 0.773 (0.033) 0.875 (0.026) 0.795 (0.040)
ρ 0.963 (0.012) 0.932 (0.014) 0.955 (0.014) 0.884 (0.042) 0.937 (0.020) 0.863 (0.024) 0.924 (0.023) 0.799 (0.057)
CN
 NRMSE 15.6% (2.6%) / 16.8% (2.7%) / 17.4% (2.4%) / 19.0% (2.5%) /
 SSIM 0.910 (0.030) 0.874 (0.037) 0.890 (0.029) 0.751 (0.021) 0.902 (0.019) 0.781 (0.028) 0.885 (0.024) 0.807 (0.024)
ρ 0.967 (0.011) 0.944 (0.012) 0.961 (0.012) 0.908 (0.022) 0.945 (0.014) 0.883 (0.016) 0.935 (0.016) 0.835 (0.029)
All
 NRMSE 16.0% (2.8%) / 17.4% (3.1%) / 17.9% (3.0%) / 19.7% (3.2%) /
 SSIM 0.906 (0.032) 0.869 (0.037) 0.884 (0.031) 0.743 (0.033) 0.895 (0.026) 0.776 (0.032) 0.879 (0.026) 0.799 (0.035)
ρ 0.965 (0.011) 0.937 (0.014) 0.957 (0.013) 0.894 (0.037) 0.940 (0.018) 0.870 (0.023) 0.928 (0.021) 0.813 (0.051)

All the results were displayed by mean (SD)

Abbreviations: AD, Alzheimer’s Disease; All, All the Subjects; CN, Cognitively Normal.

Consistency analysis of histogram distribution was performed between the true 11C-UCB-J image in relation to both the predicted 11C-UCB-J image and the input 18F-FDG image across all four networks. As shown in Figure 3, for all the networks, the distribution of values in the predicted image is more consistent with that of the true image in terms of correlation as compared to the distribution of values in the input 18F-FDG image in both AD and CN subjects.

FIGURE 3.

FIGURE 3

Log-scaled joint histograms between predicted images and the ground truth, as well as between the input images and the ground truth for all the networks. The color represents the log-base 10 of the number of voxels with a given set of SUVR, DVR, or Ki Ratio values. AD refers to one Alzheimer’s disease subject and CN refers to one cognitively normal subject (same subjects as in Figure 2)

The predicted images based on the polynomial fitting method are shown in Figure 4. We observed that the texture of the predicted images is similar to the input 18F-FDG images in all the four image synthesis pairs, indicating that polynomial fitting could not help obtain additional structural features of 11C-UCB-J images. We compared the overall quantification results of both deep learning network and polynomial fitting method as shown in Figure 5. For all the four image synthesis pairs, deep learning-based methods clearly showed smaller NRMSE and higher SSIM and Pearson coefficient indicating that the network could learn more structural information and outperforms the simple curve fitting method.

FIGURE 4.

FIGURE 4

Representative predicted results using polynomial fitting of four synthesis pairs including #1:18F-FDG SUVR11C-UCB-J SUVR, #2: 18F-FDG Ki Ratio11C-UCB-J SUVR, #3: 18F-FDG SUVR11C-UCB-J DVR, and #4: 18F-FDG Ki Ratio11C-UCB-J DVR for (a) one Alzheimer’s disease subject and (b) one cognitively normal subject

FIGURE 5.

FIGURE 5

Comparison between deep learning method (DL) and polynomial fitting (PF) curve method based on NRMSE(a, b), SSIM(c, d), and ρ(e, f) in both AD and CN groups. AD: Alzheimer’s Disease, CN: cognitively normal

The mean percentage bias and standard deviation of all the selected ROIs are shown in Table 5 and Table 6. As shown in Table 5, the mean percentage bias across all ROIs is −0.33% ± 7.43% in the AD group and −0.47% ± 7.28% in the CN group for Network #1 (18F-FDG SUVR11C-UCB-J SUVR). For Network #2 (18F-FDG Ki Ratio11C-UCB-J SUVR), the mean percentage bias is −0.68% ± 8.10% in the AD group, and −1.31% ± 6.99% in the CN group. In addition, the ROI bias of gray matter was calculated in both the AD and CN groups for global error estimation. The mean percentage bias is −1.61% ± 4.99% (p = 0.07) in the AD group and −1.33% ± 5.61% (p = 0.29) in the CN group for Network #1, while for Network #2, the mean percentage bias is −2.76% ± 4.96% (p < 0.05) in the AD group and −3.05% ± 5.33% (p < 0.05) in the CN group. The observed high standard deviation of ROI bias is secondary to tracer uptake variability across subjects.

TABLE 5.

ROI percent bias across AD and CN Groups for 18F-FDG SUVR to 11C-UCB-J SUVR Network and 18F-FDG Ki Ratio to 11C-UCB-J SUVR Network

ROI 18F-FDG SUVR11C-UCB-J SUVR
18F-FDG KiRatio11C-UCB-J SUVR
AD CN AD CN
Hippocampus −0.29% (8.01%) −1.50% (7.71%) 0.25% (10.3%) −2.11% (7.17%)
Entorhinal −2.84% (9.61%) −0.89% (7.24%) −2.90% (8.45%) −3.02% (7.70%)
Precuneus −2.17% (7.26%) −1.38% (7.14%) −4.32% (7.36%)* −3.64% (6.65%)*
Cingulate Post −1.50% (6.98%) −1.15% (8.49%) −3.05% (7.27%)* −2.95% (6.86%)
Parietal −2.13% (6.54%) −1.17% (6.71%) −3.78% (6.60%)* −3.23% (6.37%)*
Pons 0.49% (7.34%) −3.56% (6.15%)* 2.31% (9.17%) −0.86% (7.71%)
Frontal −1.60% (5.81%) −2.00% (7.09%) −3.15% (5.75%)* −4.39% (6.97%)*
Occipital −2.49% (6.75%)* −1.88% (5.94%) −3.15% (6.95%)* −3.51% (5.64%)*
Temporal −1.75% (6.73%) −0.08% (6.42%) −2.94% (6.40%)* −2.03% (5.91%)
Insula 0.05% (6.63%) −0.53% (6.66%) −0.94% (7.34%) −2.67% (5.93%)
Thalamus −0.17% (6.73%) −2.18% (6.32%) 1.06% (6.33%) −1.77% (5.00%)
Caudate 1.10% (7.12%) 1.38% (6.77%) −0.14% (7.71%) 0.20% (5.15%)
Putamen 2.33% (8.19%) 2.09% (7.71%) 2.22% (7.47%) 2.61% (6.93%)
Pallidum 3.04% (6.23%)* −0.16% (6.63%) 3.83% (7.57%)* 2.20% (6.96%)
Amygdala 1.29% (9.59%) 2.96% (9.94%) 3.32% (11.0%) 3.50% (8.11%)
Cerebral White Matter 1.13% (6.32%) 1.37% (6.24%) 1.61% (7.15%) 0.87% (6.34%)
Cingulate Ant −0.08% (7.60%) 0.66% (8.83%) −1.79% (7.66%) −1.51% (8.25%)
Gray Matter −1.61% (4.99%) −1.33% (5.61%) −2.76% (4.96%)* −3.05% (5.33%)*

All values are expressed as mean% ± SD%.

p values are obtained by a two-tailed one-sample t-test compared with the mean of zero.

*

p < 0.05.

TABLE 6.

ROI percent bias across AD and CN Groups for 18F-FDG SUVR to 11C-UCB-J DVR Network and 18F-FDG Ki Ratio to 11C-UCB-J DVR Network

ROI 18F-FDG SUVR11C-UCB-J DVR
18F-FDG Ki Ratio11C-UCB-J DVR
AD CN AD CN
Hippocampus −0.37% (7.89%) −3.55% (6.91%)* −0.85% (9.39%) −2.73% (6.95%)
Entorhinal −1.95% (10.1%) −1.43% (7.58%) −0.99% (9.92%) −2.70% (8.73%)
Precuneus −4.01% (7.64%)* −3.12% (6.57%)* −5.34% (8.62%)* −4.30% (7.56%)*
Cingulate Post −3.17% (7.15%)* −2.93% (7.61%) −5.07% (8.60%)* −3.94% (8.43%)*
Parietal −2.87% (6.43%)* −2.49% (6.75%) −4.99% (7.44%)* −3.82% (7.24%)*
Pons −0.91% (9.07%) −0.71% (7.71%) 10.12% (11.8%)* 5.88% (9.38%)*
Frontal −1.90% (5.48%) −3.61% (6.54%)* −3.85% (6.34%)* −4.59% (7.30%)*
Occipital −2.81% (7.62%)* −3.01% (6.27%)* −4.14% (7.75%)* −5.22% (6.91%)*
Temporal −2.27% (6.69%) −1.89% (5.87%) −2.83% (6.50%)* −2.84% (6.15%)*
Insula −1.16% (6.66%) −3.67% (6.14%)* −1.15% (6.88%) −4.37% (6.50%)*
Thalamus −0.63% (6.12%) −3.16% (7.10%) −0.04% (6.56%) −1.36% (7.22%)
Caudate −0.43% (7.24%) −1.39% (5.70%) 0.14% (7.73%) −1.96% (5.50%)
Putamen −0.30% (7.36%) −1.05% (6.49%) 1.48% (7.77%) −0.84% (6.39%)
Pallidum 0.33% (6.53%) −2.03% (7.68%) 1.54% (7.23%) −0.31% (7.21%)
Amygdala 0.03% (10.2%) 1.26% (8.62%) 3.64% (12.0%) 4.00% (8.69%)*
Cerebral White Matter 0.69% (6.26%) 0.91% (5.78%) 0.60% (7.76%) 1.15% (6.82%)
Cingulate Ant −0.96% (6.61%) −2.22% (6.96%) −0.62% (7.06%) −2.19% (7.38%)
Gray Matter −2.31%(4.99%)* −2.83%(5.27%)* −3.56%(5.33%)* −3.74%(5.76%)*

All values are expressed as mean% ± SD%.

p values are obtained by a two-tailed one-sample t-test compared with the mean of zero.

*

p < 0.05.

Two statistical comparisons were made. First, there were no statistically significant differences in bias between the AD and CN groups for any ROI. Within the hippocampus, for example, a region with significant 11C-UCB-J group differences, p-values were 0.58, 0.32, 0.13, and 0.40 for Networks #1–4, respectively. Second, statistical analyses were performed to determine if mean biases across any group were significantly different from 0. As shown in Table 5, for Network #1, biases were significantly different from 0 within the occipital and pallidum ROIs in the AD group, and within the pons in the CN group, although the mean bias was <4% (in absolute value). The number of biased results is slightly larger in Network #2, indicating that the 11C-UCB-J SUVR prediction network with 18F-FDG SUVR as the input leads to better evaluation performance. As shown in Table 6, the mean bias of all ROIs is −1.34% ± 7.47% in the AD group and −2.00% ± 6.88% in the CN group for Network #3 (18F-FDG SUVR11C-UCB-J DVR), while −0.73% ±9.04% in the AD group and −1.70% ± 7.80% in the CN group for Network #4 (18F-FDG Ki Ratio11C-UCB-J DVR). As with Networks #1 and #2, Networks #3 and #4 demonstrated no significant differences in ROI percentage bias between CN and AD groups. Regarding ROI mean biases significantly different from 0, Network #3 led to superior results in the pons, frontal, and temporal ROIs (one-sample t-tests, p > 0.05) in the AD group as compared to Network #4. Within the CN group, Network #3 network led to superior results in the posterior cingulate, parietal, pons, temporal, and amygdala ROIs (one-sample t-tests, p > 0.05) as compared to Network #4. The mean percentage bias of gray matter is −2.31% ± 4.99% (p < 0.05) in the AD group and −2.83% ± 5.27% (p < 0.05) in the CN group for Network #3, while for Network #4 is −3.56% ± 5.33% (p < 0.05) in the AD group and −3.74% ± 5.76% (p < 0.05) in the CN group.

A comparison between the deep learning method (DL) and polynomial curve fitting (PF) method on ROI bias in AD and CN groups is depicted in Figure 6. For polynomial results, the mean bias of all the ROIs is −3.78% ± 19.41% in the AD group and −1.34% ± 17.13% in the CN group for Network #1, while for Network #2 is −3.79% ± 20.17% in the AD group and −2.43% ± 18.05% in the CN group. Similarly, the mean bias of all ROIs is −5.29% ± 24.46% in the AD group and −2.47% ± 23.39% in the CN group for Network #3, while for Network #4 is −4.96% ± 25.29% in the AD group and −3.28% ± 24.55% in the CN group. For all four image synthesis pairs, the deep learning method shows smaller bias across most ROIs, indicating that the deep learning method is more applicable for cross-tracer image synthesis than the simple polynomial fitting method.

FIGURE 6.

FIGURE 6

Comparison between deep learning method (DL) and polynomial fitting curve (PF) method based on ROI percentage bias for 11C-UCB-J SUVR prediction in the AD group (a) and the CN group (c) as well as 11C-UCB-J DVR prediction in the AD group (b) and the CN group (d). AD, Alzheimer’s disease subjects; CN, cognitively normal subjects

Figure 7 depicted the mean correlation between the predicted image and true image across all subjects in three sample ROIs (hippocampus, precuneus, caudate) for all four networks. The results suggest that the predicted 11C-UCB-J results are more consistent with those of the true 11C-UCB-J results in the hippocampus and caudate. Across all networks, the r2 values ranged from ~0.8 in the caudate, 0.5 to 0.6 in the hippocampus, and 0.4 to 0.6 in the precuneus, indicating that the prediction accuracy could be ROI dependent. This may reflect the biological differences between regions for SV2A and glucose metabolism in AD.

FIGURE 7.

FIGURE 7

Correlation of the mean values of three representative ROIs (hippocampus, precuneus, and caudate) across all the subjects for all the networks between 18F-FDG and 11C-UCB-J. Blue triangles represent Alzheimer’s disease subjects, red dots represent cognitively normal subjects. The black line is the linear fitting of mean values across all the subjects of one specific ROI. The green line is the line of identity

4 |. DISCUSSION

In the 11C-UCB-J image prediction study, we demonstrated the feasibility of generating synthetic 11C-UCB-J PET images of synaptic density from 18F-FDG images. Since the two tracers have intrinsically different distributions and quantitation, we focused on generating the SUVR and DVR images, instead of absolute SUV images. Our results showed that using both 18F-FDG SUVR and Ki Ratio images can lead to a robust prediction of synthetic SUVR and DVR images of 11C-UCB-J. The quantification bias of various ROIs is smaller than 8% for the 11C-UCB-J SUVR image prediction and smaller than 10% for the 11C-UCB-J DVR image prediction. Regarding input images, both SUVR and Ki Ratio images provided a similar prediction of 11C-UCB-J static and parametric images. Using SUVR images as the input is slightly better in ROI analysis than using Ki Ratio images as input, potentially due to the lower noise in SUVR images. Since generating SUVR images required a much shorter acquisition time, it is more practical to use SUVR images as input as compared to using Ki Ratio images.

While the primary analyses focused on the generation of synthetic 11C-UCB-J images from 18F-FDG data, an exploratory analysis investigated the possibility that 11C-PiB images could be generated from 18F-FDG images with proper training using paired data. The predicted images of amyloid deposition (11C-PiB SUVR) from 18F-FDG SUVR for a single subject with AD and normal cognition are depicted in Figure S2. Visual inspection of 11C-PiB SUVR predicted images (Figure S2, row 3) demonstrates little overlap with the true 11C-PiB SUVR image (Figure S2, row 2), in both AD and CN groups due to the very different spatial pattern of tracer uptake in both CN and AD subjects. Therefore, we incorporated an additional binary channel with information about the subject as CN or AD (health vs. disease) into the network. With the addition of this new channel (Figure S2, row 4), the predicted images are more consistent with the true 11C-PiB SUVR images after incorporating diagnosis information. The distribution of the predicted 11C-PiB SUVR image is more consistent with that of the true 11C-PiB SUVR image for both AD and CN subjects according to Figure S3. The mean ROI bias is 0.34% ± 15.72% for AD subjects and −3.58% ± 14.44% for CN subjects according to Table S1. As significant amyloid deposition (as assessed by 11C-PiB SUV) has been observed in the frontal, parietal, temporal, and occipital of patients with Alzheimer’s disease,2 we compared the SUVR of both the predicted images and the true images for these typical ROIs to evaluate the physiological characteristics prediction accuracy as shown in Table S2. Overall, the mean SUVR of predicted 11C-PiB images is consistent with the mean SUVR of true 11C-PiB images in selected ROIs.

We have demonstrated the feasibility to predict 11C-UCB-J or 11C-PiB images from 18F-FDG images using the U-Net model. Both static and parametric images of 11C-UCB-J generation have been achieved with reasonable accuracy. We also implemented 11C-PiB static image prediction with the additional patient diagnosis information since the image distribution is quite different between CN and AD subjects. Importantly, as such diagnostic information may not be readily available during the initial evaluation of new patients, 11C-PiB image prediction from 18F-FDG may be more useful in monitoring disease progression and evaluating response to therapy for research use.

Although we demonstrated the possibility of using deep learning methods to generate 11C-UCB-J or 11C-PiB images from 18F-FDG images, the synthetic resultant images cannot replace true images. The 11C-UCB-J is not FDA approved for clinical use yet. However, these generated images can still potentially be of use in an AD research setting as an additional piece of diagnostic information, especially for research institutions where 11C-UCB-J are not immediately readily accessible. Related to amyloid tracers, we do have FDA-approved 18F amyloid tracers for clinical use that could be used in locations without on-site cyclotrons, so there is a demand for generating 11C-PiB images from 18F-FDG PET. Similar to 11C-UCB-J, we expect the synthesized 11C-PiB images could be used as a piece of additional information in the AD research study. Besides, the use of a single 18F-FDG scan will reduce total imaging sessions, radiation exposure, and costs for participants who need multiple scans.

This study has some limitations. First, the different noise levels of 11C-UCB-J and 11C-PiB images due to the differences in tracer uptake among the subjects after injection could increase the evaluation bias. Second, the training results could be improved via the inclusion of a larger dataset. For future work, we will attempt to generate other tracer images such as tau using 18F-FDG images. In addition, images from two tracers instead of only one could be used as the input to generate a third tracer image.

5 |. CONCLUSIONS

It is feasible to use fully 3D U-Net architecture to predict both 11C-UCB-J and 11C-PiB PET images from 18F-FDG images. Both 18F-FDG SUVR images and Ki Ratio images could be used as the input for the prediction of11C-UCB-J static and parametric images. Using the SUVR image as input is slightly better in terms of evaluation accuracy as well as its accessibility. 18F-FDG SUVR images along with diagnostic information could be used as the input for the prediction of11C-PiB SUVR images.

Supplementary Material

supplemental fig 1
supplemental fig 2
supplemental fig 3
supplemental tables
supplemental methods

Funding information

This work was supported by NIH grants R01EB025468, R01AG052560, R01AG062276, P30AG066508, R01NS094253, and China Scholar Council.

Footnotes

CONFLICT OF INTEREST

The authors have no relevant conflict of interest to disclose.

SUPPORTING INFORMATION

Additional supporting information may be found online in the Supporting Information section.

DATA AVAILABILITY STATEMENT

The data that support the findings of this study are available from the corresponding author upon reasonable request.

REFERENCES

  • 1.Goedert M Tau protein and the neurofibrillary pathology of Alzheimer’s disease. Trends Neurosci. 1993;16(11):460–465. [DOI] [PubMed] [Google Scholar]
  • 2.Klunk WE, Engler H, Nordberg A, et al. Imaging brain amyloid in Alzheimer’s disease with Pittsburgh compound-B. Ann Neurol. 2004;55(3):306–319. [DOI] [PubMed] [Google Scholar]
  • 3.Hamos JE, DeGennaro LJ, Drachman DA. Synaptic loss in Alzheimer’s disease and other dementias. Neurology. 1989;39(3):355. [DOI] [PubMed] [Google Scholar]
  • 4.Nordberg A PET imaging of amyloid in Alzheimer’s disease. Lancet Neurol. 2004;3(9):519–527. [DOI] [PubMed] [Google Scholar]
  • 5.Marcus C, Mena E, Subramaniam RM. Brain PET in the diagnosis of Alzheimer’s disease. Clin Nucl Med. 2014;39(10):e413. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Finnema SJ, Nabulsi NB, Eid T, et al. Imaging synaptic density in the living human brain. Sci Transl Med. 2016;8(348):348ra96. [DOI] [PubMed] [Google Scholar]
  • 7.Landau SM, Harvey D, Madison CM, et al. Associations between cognitive, functional, and FDG-PET measures of decline in AD and MCI. Neurobiol Aging. 2011;32(7):1207–1218. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Ishibashi K, Onishi A, Fujiwara Y, Ishiwata K, Ishii K. Relationship between Alzheimer disease-like pattern of 18F-FDG and fasting plasma glucose levels in cognitively normal volunteers. J Nucl Med. 2015;56(2):229–233. [DOI] [PubMed] [Google Scholar]
  • 9.Teipel SJ, Drzezga A, Bartenstein P, Möller H-J, Schwaiger M, Hampel H. Effects of donepezil on cortical metabolic response to activation during 18FDG-PET in Alzheimer’s disease: a double-blind cross-over trial. Psychopharmacology. 2006;187(1):86–94. [DOI] [PubMed] [Google Scholar]
  • 10.Nabulsi NB, Mercier J, Holden D, et al. Synthesis and pre-clinical evaluation of 11C-UCB-J as a PET tracer for imaging the synaptic vesicle glycoprotein 2A in the brain. J Nucl Med. 2016;57(5):777–784. [DOI] [PubMed] [Google Scholar]
  • 11.Chen M-K, Mecca AP, Naganawa M, et al. Assessing synaptic density in alzheimer disease with synaptic vesicle glycoprotein 2A positron emission tomographic imaging. JAMA Neurol. 2018;75(10):1215–1224. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Mecca AP, Chen M-K, O’Dell RS, et al. In vivo measurement of widespread synaptic loss in Alzheimer’s disease with SV2A PET. Alzheimer’s Dementia. 2020;16(7):974–982. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Sisodia S, Koo E, Beyreuther K, Unterbeck A, Price D. Evidence that beta-amyloid protein in Alzheimer’s disease is not derived by normal processing. Science. 1990;248(4954):492–495. [DOI] [PubMed] [Google Scholar]
  • 14.LeCun Y, Bengio Y, Hinton G. Deep learning. Nature. 2015;521(7553):436–444. [DOI] [PubMed] [Google Scholar]
  • 15.Ronneberger O, Fischer P. U-net BT: convolutional networks for biomedical image segmentation. Paper presented at: International Conference on Medical image computing and computer-assisted intervention. 2015. [Google Scholar]
  • 16.Goodfellow I, Pouget-Abadie J, Mirza M, et al. Generative adversarial nets. Paper presented at: Advances in neural information processing systems. 2014. [Google Scholar]
  • 17.Kaji S, Kida S. Overview of image-to-image translation by use of deep neural networks: denoising, super-resolution, modality conversion, and reconstruction in medical imaging. Radiol Phys Technol. 2019;12(3):235–248. [DOI] [PubMed] [Google Scholar]
  • 18.Mao X-J, Shen C, Yang Y-B. Image restoration using convolutional auto-encoders with symmetric skip connections. arXiv preprint arXiv:160608921. 2016. [Google Scholar]
  • 19.Han X MR-based synthetic CT generation using a deep convolutional neural network method. Med Phys. 2017;44(4):1408–1419. [DOI] [PubMed] [Google Scholar]
  • 20.Nie D, Trullo R, Lian J, et al. Medical image synthesis with deep convolutional adversarial networks. IEEE Trans Biomed Eng. 2018;65(12):2720–2730. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Choi H, Lee DS. Alzheimer’s disease neuroimaging I. Generation of structural MR images from amyloid PET: application to MR-less quantification. J Nucl Med. 2018;59(7):1111–1117. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Liu C-C, Qi J. Higher SNR PET image prediction using a deep learning model and MRI image. Phys Med Biol. 2019;64(11):115004. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Shi L, Onofrey JA, Liu H, Liu Y-H, Liu C. Deep learning-based attenuation map generation for myocardial perfusion SPECT. Eur J Nucl Med Mol Imaging. 2020;47(10):2383–2395. 10.1007/s00259-020-04746-6. [DOI] [PubMed] [Google Scholar]
  • 24.Chen M-K, Mecca A, Toyonaga T, et al. Correlation between FDG PET for neuronal function and 11C-UCB-J PET for synaptic density using SUV ratios with cerebellum reference in Alzheimer’s disease. J Nucl Med. 2019;60(supplement 1):422. [Google Scholar]
  • 25.van Aalst J, Ceccarini J, Sunaert S, Dupont P, Koole M. Van Laere K. In vivo synaptic density relates to glucose metabolism at rest in healthy subjects, but is strongly modulated by regional differences [published online ahead of print 2021/01/15]. J Cereb Blood Flow Metab. 2021. 10.1177/0271678x20981502:271678x20981502. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Carson RE, Barker WC, Jeih-San L, Johnson CA. Design of a motion-compensation OSEM list-mode algorithm for resolution-recovery reconstruction for the HRRT. Paper presented at: 2003 IEEE Nuclear Science Symposium. Conference Record (IEEE Cat. No.03CH37515; ); 19–25October. 2003, 2003. [Google Scholar]
  • 27.Wu Y, Carson RE. Noise reduction in the simplified reference tissue model for neuroreceptor functional imaging. J Cereb Blood Flow Metab. 2002;22(12):1440–1452. [DOI] [PubMed] [Google Scholar]
  • 28.Duncan JS, Papademetris X, Yang J, Jackowski M, Zeng X, Staib LH. Geometric strategies for neuroanatomic analysis from MRI. NeuroImage. 2004;23:S34–S45. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.Dale AM, Fischl B, Sereno MI. Cortical surface-based analysis: I. Segmentation and surface reconstruction. NeuroImage. 1999;9(2):179–194. [DOI] [PubMed] [Google Scholar]
  • 30.Long J, Shelhamer E, Darrell T. Fully convolutional networks for semantic segmentation. Paper presented at: Proceedings of the IEEE conference on computer vision and pattern recognition. 2015. [DOI] [PubMed] [Google Scholar]
  • 31.Lu W, Onofrey JA, Lu Y, et al. An investigation of quantitative accuracy for deep learning based denoising in oncological PET. Phys Med Biol. 2019;64(16):165019. [DOI] [PubMed] [Google Scholar]
  • 32.Liu H, Wu J, Lu W, Onofrey JA, Liu Y-H, Liu C. Noise reduction with cross-tracer and cross-protocol deep transfer learning for low-dose PET. Phys Med Biol. 2020;65(18):185006. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Wolterink JM, Leiner T, Viergever MA, Išgum I. Generative adversarial networks for noise reduction in low-dose CT. IEEE Trans Med Imaging. 2017;36(12):2536–2545. [DOI] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

supplemental fig 1
supplemental fig 2
supplemental fig 3
supplemental tables
supplemental methods

Data Availability Statement

The data that support the findings of this study are available from the corresponding author upon reasonable request.

RESOURCES