Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2025 Jan 1.
Published in final edited form as: Eur J Nucl Med Mol Imaging. 2023 Oct 3;51(2):358–368. doi: 10.1007/s00259-023-06417-8

PET image denoising based on denoising diffusion probabilistic model

Kuang Gong 1,2,3, Keith A Johnson 2, Georges El Fakhri 2, Quanzheng Li 3, Tinsu Pan 4
PMCID: PMC10958486  NIHMSID: NIHMS1972295  PMID: 37787849

Abstract

Purpose:

Due to various physical degradation factors and limited counts received, PET image quality needs further improvements. The denoising diffusion probabilistic model (DDPM) was a distribution learning-based model, which tried to transform a normal distribution into a specific data distribution based on iterative refinements. In this work, we proposed and evaluated different DDPM-based methods for PET image denoising.

Methods:

Under the DDPM framework, one way to perform PET image denoising was to provide the PET image and/or the prior image as the input. Another way was to supply the prior image as the network input with the PET image included in the refinement steps, which could fit for scenarios of different noise levels. 150 brain [18F]FDG datasets and 140 brain [18F]MK-6240 (imaging neurofibrillary tangles deposition) datasets were utilized to evaluate the proposed DDPM-based methods.

Results:

Quantification showed that the DDPM-based frameworks with PET information included generated better results than the nonlocal mean, Unet and generative adversarial network (GAN)-based denoising methods. Adding additional MR prior in the model helped achieved better performance and further reduced the uncertainty during image denoising. Solely relying on MR prior while ignoring the PET information resulted in large bias. Regional and surface quantification showed that employing MR prior as the network input while embedding PET image as a data-consistency constraint during inference achieved the best performance.

Conclusions:

DDPM-based PET image denoising is a flexible framework, which can efficiently utilize prior information and achieve better performance than the nonlocal mean, Unet and GAN-based denoising methods.

Keywords: PET image denoising, Denoising diffusion probabilistic model, Low-dose PET, Generative models

Introduction

Positron Emission Tomography (PET) was widely used in oncology, cardiology and neurology studies because of its high sensitivity and quantitative merits. Due to various physical degradation factors and limited counts received, the signal-to-noise ratio (SNR) and resolution of PET was low, which comprised its clinical values in diagnosis, prognosis, staging and treatment monitoring. Additionally, to improve the hospital’s throughput or reduce the radiation exposures to patients, faster or low-dose PET imaging was desirable, where the counts received during the scan was even less. This further challenged our ability to attain high-quality PET images from limited counts.

For the past decades, various approaches were explored for PET image denoising, e.g., wavelet [1], highly constrained backprojection (HYPR) [2], nonlocal mean [3, 4], guided image filtering [5], block matching [6], and dictionary learning [7]. These methods were mainly based on specifically designed similarity calculation or small-scale feature extraction/learning. With the availability of enormous training data and strong computational power, deep learning-based image processing were extensively studied recently. More specifically, convolutional neural network (CNN)-based PET image denoising achieved superior performance than other state-of-the-art methods [816]. However, over-smoothness of the network output was one pitfall of most CNN-based approaches. Conditional generative adversarial networks (cGANs)-based PET image denoising were developed to address this issue [1722], where the learned discriminative loss could further push the image appearance closer to the ground truth. One issue of the CNN and cGAN-based methods was that if there were data-distribution mismatches between training and testing data, the network performance could be compromised. Transfer learning was investigated as one approach to address the mismatch issue and was demonstrated effective in various PET studies [2327].

The deep learning approaches described above were deterministic models, where only one denoised image was obtained based on the pre-specified network input. The denoising diffusion probabilistic model (DDPM) and score function-based model [2830], denoted as diffusion models here, were a category of distribution learning-based models, which tried to transform a normal distribution into a specific data distribution based on iterative refinements. During the forward pass, noise was injected in the image gradually. During the reverse pass, data samples were generated by removing noise from the image after multiple refinement steps. Fig. 1 showed the diagram of the forward and reverse pass. Additional prior image could be added to the model to conduct conditional image generation. By running the reverse pass multiple times, an uncertainty map could also be calculated from the generated data samples, which could be used to further guide diagnosis and treatment monitoring [3133]. The diffusion models have shown encouraging results in various computer-vision tasks, e.g., image generation [2830, 34, 35], super resolution [34, 36], image in-painting [37], and MR image reconstruction [38, 39].

Figure 1:

Figure 1:

Diagram of the DDPM framework.

As the dynamic range of different organs’ uptake was high and the noise levels of different scans varied, the data distribution of PET images was more complicated than natural images. Additionally, apart from PET image itself, there was patients’ anatomical prior information (e.g., CT or MR images) available from the same or separate scans. Thus, further evaluating the diffusion models and exploring approaches to embed the patient’s prior information for PET image denoising deserved further investigations. In this work, we evaluated DDPM for PET image denoising and further explored different approaches of embedding the PET and the prior MR information in DDPM. Nonlocal mean, Unet and Pix2Pix [40] methods were employed as the reference methods. Unet and Pix2Pix were widely used CNN and cGAN-based methods, respectively. 150 brain [18F]FDG datasets and 140 [18F]MK-6240 (imaging neurofibrillary tangles deposition) datasets were utilized in two evaluation studies to evaluate the performance of different methods based on global and regional quantitative metrics.

Materials and methods

DDPM model

Suppose the target data x0qx0. During the forward process, we could define a Markov diffusion process q that added Gaussian noise to x0 in each step, which could be written as [28]

q(x1,..,xTx0)=t=1Tq(xtxt1),q(xtxt1)=𝒩(xt;1βtxt1,βtI). (1)

One property of the forward process q was that it allowed us to sample at any arbitrary step directly conditioned on x0. Denoting αt=1βt and αt=l=1tαl, we could have

q(xtx0)=𝒩(xt;α¯tx0,(1α¯t)I),xt=α¯tx0+(1α¯t)ϵ, (2)

where ϵ𝒩(0,I). Based on Bayes theorem, the posterior qxt1xt also followed Gaussian distribution,

qxt1xt=𝒩(xt1;μ˜txt,x0,1αt11αtβtI), (3)
withμ˜txt,x0=αt1βt1αtx0+αt1αt11αtxt. (4)

If we aimed to sample from the target data distribution qx0, we could first sample from qxT, which was an isotropic Gaussian distribution given a large enough T. Then based on the posterior distribution qxt1xt, we could obtain samples of x0. However, qxt1xt was not computable as the distribution of x0 was unknown. The DDPM framework tried to approximate qxt1xt by pθxt1xt through a network with parameter θ, where

pθxt1xt=𝒩xt1;μ˜θxt,t,σt2I. (5)

Instead of directly approximating μ˜θxt,t by a neural network, Ho et al [28] proposed to approximate the noise ϵ in equation (2) by a network as ϵθxt,t, which could also be interpreted as the gradient of the data log-likelihood, known as the score function. Based on equations (2) and (4), μ˜θxt,t could expressed as

μ˜θxt,t=1αtxtβt1αtϵθxt,t. (6)

Based on the trained score function ϵθˆxt,t, and equations (5) and (6), each refinement step during inference was

xt1=1αtxtβt1αtϵθˆxt,t+σtz,wherez𝒩(0,I). (7)

Conditional PET image denoising based on DDPM

The original DDPM stated above was for unconditional image generation. For PET image denoising, the noisy PET image xnoisy existed and the patient’s prior image xprior might also be available. We could directly supply xnoisy and xprior (if available) as the additional network input of the score function, as was done in image super-resolution [36], and the score function ϵθxt,t changed to ϵθxt,t,xnoisy,xprior. This framework required specific low- and high-quality training pairs during the score function training. Another approach was to only supply xprior as the additional input to the score function, while including xnoisy during inference. Compared to the previous approach, this approach did not need low- and high-quality pairs during training and could thus work for noisy PET images of various noise levels during inference. In this approach, we could approximate qxt1xt by pθxt1xt,xnoisy,xprior instead of pθxt1xt in the original DDPM framework, which could be rewritten as

pθxt1xt,xnoisy,xpriorpθxt1xt,xpriorpxnoisyxt1,xt,xprior. (8)

Suppose the PET image noise followed Gaussian distribution and was independent of xprior, we could have

pxnoisyxt1,xt,xprior𝒩xt,σd2I, (9)

where σd indicated the noise level of xnoisy. As for pθxt1xt,xprior, similar to equations (5) and (6), ϵ was approximated by ϵθxt,t,xprior, where xprior was supplied as an additional input to the score function. Based on equation (10) from Ref. 34 and equation (9), the refinement step during inference changed to

xt1=1αtxtβt1αtϵθˆxt,t,xpriorσt2σd2xnoisyxt+σtz. (10)

Datasets

Two experiments based on [18F]FDG and [18F]MK-6240 datasets were conducted to evaluate the performance of DDPM for PET image denoising and also explore different approaches of embedding prior information. Fig. 1 showed a general description of the datasets used for the two experiments.

Firstly, 150 brain [18F]FDG datasets acquired from the GE DMI PET-CT scanner at MD Anderson Cancer Center were utilized. The scans were performed 1 hour post injection with 2 min per bed position. The injected doses were 359.82 ± 41.49 MBq. No head and neck position immobilization nor PET/CT alignment were performed. Of all the datasets, 85 datasets were used for training, 5 datasets for validation and the remaining 60 datasets for testing. 87 data pairs were generated from each patient dataset and there were in total 7395 training data pairs. Normal-dose datasets were used as the ground truth and low-dose datasets were generated by extracting 1/4 events from the normal-dose listmode datasets. All the datasets were reconstructed into a matrix of 256 × 256 × 89 with a voxel size of 1.17 × 1.17 × 2.79 mm3 based on the ordered subset expectation maximization (OSEM) algorithm (3 iterations and 17 subsets) including point spread function (PSF) and time of flight (TOF) modeling. The images were later scaled to the standardized uptake value (SUV) unit before being further processed using different methods. For the proposed DDPM-based method, low-dose PET images were supplied as the network input and this method was denoted as DDPM-PET.

The other experiment was based on [18F]MK-6240 datasets with corresponding T1-weighted MR images. In total 140 dynamic [18F]MK-6240 datasets acquired on the GE DMI PET/CT scanner (injected doses around 185 MBq) were utilized in this experiment. The datasets were acquired from 66 subjects. Some of the subjects had longitudinal scans. The testing data did not include datasets scanned from subjects that were included in the training/validation population. The demographics of the subjects were shown in Table 1. For each dataset, the events from 0 min to 10 min post injection, were extracted and combined to construct the normal-dose dataset. These early-time-frame datasets mainly contained brain perfusion information, which was actively investigated as a potential biomarker for neurodegenerative diseases [4143]. The image reconstruction algorithm, matrix size and pixel size were the same as the above-mentioned [18F]FDG datasets. Additional T1-weighted MR images were acquired from the Siemens 3T MR scanner and were registered to the PET images through rigid registration by ANTS [44]. To simulate low-dose scenerios, we further down sampled the listmode events to generate 1/4 low-dose datasets. Of all the 140 data pairs, 116 were employed for training, 4 for validation and the remaining 20 for testing. 87 data pairs were generated from each patient dataset and there were in total 10092 training data pairs. To further test the robustness of different methods to mismatches between training and testing datasets, we also downsampled the normal-dose test datasets to 1/16 low-dose test datasets. In this experiment, we have different DDPM-based methods by varying embedding approaches of the low-dose PET and MR prior images. The method using PET image only as the network input was denoted as DDPM-PET (the same as in the [18F]FDG experiment). The method using MR image only as the input was denoted as DDPM-MR. The method using both PET and MR images as the input was denoted as DDPM-PETMR. Finally, the method using the MR image as the network input while using the PET image in the data consistency item, as shown in equation (10), was denoted as DDPM-MR-PETCon.

Table 1:

Demographics of the subjects in the [18F]MK-6240 study. CN: cognitive normal. MCI: mild cognitive impairment. AD: Alzheimer’s disease. MMSE: Mini-Mental State Examination.

CN (N = 57) MCI (N = 6) AD (N = 3)

Age 68.0 ±11.5 69.9 ±10.0 69.1±12.9
Females 47.4% 50.0% 67.7%
MMSE 28.6±1.3 21.7 ±5.0 21.7 ±6.1

Training details and reference methods

The network used to generate the score function ϵθ was a Unet structure with attention and residual blocks as described in [34]. Two neighboring axial slices were additionally supplied as the network input, to avoid the axial artifacts. During network training, the batch size was set to 20 for all methods. The network input was 256 × 256 ×3, 256 × 256 ×3, 256 × 256 ×3, and 256 × 256 ×6, for DDPM-PET, DDPM-MR, DDPM-MR-PETCon, and DDPM-PETMR, respectively. For the score function training, the loss function used was the same as in [28],

L(θ)=Et[1,T],x0qx0,ϵ𝒩(0,I)(ϵϵθxt,t22). (11)

where E denoted expectation, the number of time steps T was set to 1000, and t was uniformly chosen from 1 to T.βt was set to increase linearly from β1=0.0001 to βT=0.02 as in [28]. For each DDPM-based method, the training took around 10 days using 4 NVIDIA RTX 8000 GPUs. The inference took around 50 mins based on 1 GPU.

For both [18F]FDG and [18F]MK-6240 experiments, the nonlocal mean (NLM), Unet, and Pix2Pix-based GAN were adopted as the reference methods. For the Unet and GAN methods, the training datasets and the number of training parameters are the same as the DDPM-based methods. For the [18F]FDG experiment, the similarity in the NLM method was calculated based on the low-dose PET image. The three-channel input of the Unet and GAN is also the low-dose PET image. In the [18F]MK-6240 experiment, the similarity in the NLM method was calculated based on the MR prior image. The six-channel input of the Unet and GAN contains both the low-dose PET and MR prior images.

Data analysis

For [18F]FDG and [18F]MK-6240 test datasets, the peak signal-to-noise ratio (PSNR) and and structural similarity index measure (SSIM) were calculated for different methods with the normal-dose PET images as the ground truth. PSNR was calculated as

PSNR=10log10MAXI2MSE (12)

where MAXI was the maximum possible pixel value of the image. MSE=1ni=1nPETpostiPETnormali2, where PETpost and PETnormal were the low-dose image after post-processing and the normal-dose image, respectively. Given two image vectors x and y, the general definition of SSIM was

SSIM(x,y)=[l(x,y)]α[c(x,y)]β[s(x,y)]γ, (13)

where l(x,y)=2μxμy+C1μx2+μy2+C1,c(x,y)=2σxσy+C2σx2+σy2+C2 and s(x,y)=σxy+C3σxσy+C3 were the luminance, contrast and similarity measures, respectively. Here μx,μy,σx2,σy2 were the means and variances of x and y, respectively, and σxy was the covariance between x and y.C1,C2 and C3 were small constants used to stabilize the results. All parameters were based on the default setting in MATLAB 2021a (Mathworks, MA) The Wilcoxon signed-rank tests of PSNR and SSIM were conducted to compare different methods. In the [18F]MK-6240 experiment, as the patient’s MR image was available, Freesurfer [45] was used for cortical parcellation based on the MR image to obtain the region of interests (ROIs). Further quantification of the cortical ROIs based on PSNR was performed. The Braak stage-related cortices—hippocampus, entorhinal, parahippocampal, amygdala, inferior temporal, fusiform, posterior cingulate, lingual, precuneus, insula, pericalcarine, cuneus and precentral—were chosen as the evaluation ROIs.

We have further performed surface analysis to comprehensively evaluate different methods’ performance on generating high-quality PET images in the [18F]MK-6240 experiment. Firstly, the relative PET error, PETerror, was calculated as

PETerror=PETpostPETnormalPETnormal. (14)

Afterwards, the surface map of PETerror was generated for each [18F]MK-6240 test dataset and was registered to the FSAverage template in Freesurfer to construct the averaged surface map of all the 20 test datasets.

One characteristic of the diffusion models is the ability to generate the uncertainty map based on multiple forward passes during inference. We forwarded pass the same test dataset through the network 20 times. The variance of the 20 network outputs were used to represent the uncertainty map of the network output. Note that the uncertainty of the network output is not the same as the error of the model prediction. To understand the relationship between the uncertainty of the network output and the model prediction error, we calculated the correlation between the output uncertainty at different brain regions and the mean squared error (MSE) calculated at the same brain regions. The MSE was calculated by utilizing the normal-dose data as the ground truth.

Results

Fig. 3 showed three views of one [18F]FDG dataset processed with different methods. The error images generated by different methods compared to the normal-dose image were also shown. It could be observed that both the Unet and the DDPM-PET methods could generate results with better cortical details and smaller noise in the while matter region compared to the NLM method. From the error images, we observed that the DDPM-PET method could generate a smaller error compared with other methods. Fig. 4 presented the PSNR and SSIM values of different methods based on 60 test datasets, which showed that the DDPM-PET method had better quantification results than other reference methods.

Figure 3:

Figure 3:

Three views of one [18F]FDG test dataset processed using different methods along with the 1/4 low-dose image (1st column) and the normal-dose image (last column). For each view, the second row shows the difference images of difference methods when compared with the normal-dose image.

Figure 4:

Figure 4:

The PSNR (left) and SSIM (right) values using different methods calculated based on 60 [18F]FDG test datasets. ***, **, ns, located at the top of the bar plot represents p-value< 0.001, p-value< 0.01 and p-value≥ 0.05, respectively.

Fig. 5 showed three views of one [18F]MK-6240 dataset processed using different methods along with the co-registered MR prior image for the 1/4 and 1/16 low-dose datasets. Compared to DDPM-PET, DDPM-PETMR had higher image resolution, which demonstrated the benefit of including the additional MR prior image as the network input. It was observed that the DDPM-MR method could generate results with the highest resolution and the lowest noise. However, as shown in the error images, large bias existed at some regions, e.g., the caudate and putamen regions. Based on the additional PET data-consistency constraint, results of the DDPM-MR-PETCon method shared more similarity with the normal-dose results. Compared to Unet, DDPM-MR-PETCon had a higher image contrast. Fig. 6 showed the PSNR and SSIM values of different methods based on 20 test datasets. These global quantification values showed that DDPM-PETMR had the best performance and was more robust to the noise levels during testing, followed by DDPM-MR-PETCon. The regional PSNR plot at 14 cortical regions for different methods was presented at Fig. 7. Fig. 8 further showed the surface plots of the left hemisphere regarding the PET relative error for different methods. The regional and the surface results both showed that DDPM-MR-PETCon had the lowest error and was also more robust to the noise levels during testing, followed by DDPM-PETMR.

Figure 5:

Figure 5:

Three views of one [18F]MK-6240 test dataset processed using different methods along with the MR prior image (1st column), the 1/4 low-dose image (2nd column) and the normal-dose image (last column). For each view, the second row shows the difference images of difference methods when compared with the normal-dose image.

Figure 6:

Figure 6:

The PSNR (left column) and SSIM (right column) values using different methods calculated based on 20 [18F]MK-6240 test datasets for 1/4 low-dose (top row) and 1/16 low-dose datasets. ***, **, *, ns, located at the top of the bar plot represents p-value< 0.001, p-value< 0.01, p-value< 0.05, and p-value≥ 0.05, respectively.

Figure 7:

Figure 7:

PSNR values of the 14 cortical regions for different methods calculated based on 20 [18F]MK-6240 test datasets. The top figure is for the case of 1/4 low-dose datasets and the bottom figure is for the case of 1/16 low-dose datasets.

Figure 8:

Figure 8:

Surface maps of the left hemisphere showing mean PET relative errors for different methods calculated based on 20 [18F]MK-6240 test datasets regarding the (a) 1/4 low-dose datasets and (b) 1/16 low-dose datasets. For (a), the color-bar range is [0 0.13]. For (b), the color-bar range is [0 0.18].

Fig. 9 shows the uncertainty maps of the DDPM-MR, DDPM-PET, DDPM-PETMR and DDPM-MR-PETCon methods based on 20 realizations from the same dataset as in Fig. 5. Comparing the results of DDPM-PET and DDPM-PETMR, we observed that adding additional MR prior image could reduce the uncertainty. Comparing the results of DDPM-MR and DDPM-MR-PETCon, we observed that adding PET data-consistency constraint could further reduce the uncertainty. The uncertainty values of DDPM-MR-PETCon were smaller than that of DDPM-PETMR, which showed that including the PET image during the refinement steps could further reduce the uncertainty compared to supplying it as the network input. We also calculated the correlation between the MSE and the generated predictive uncertainty for different brain regions, which was 0.6048, 0.947, 0.9209, and 0.9246 for DDPM-MR, DDPM-PET, DDPM-PETMR, and DDPM-MR-PETCons respectively. This showed that the DDPM generated results with PET info had high correlation with the MSE, and the correlation of DDPM-MR with MSE was low.

Figure 9:

Figure 9:

Three views of the uncertainty maps of one [18F]MK-6240 test dataset calculated from 20 realizations for different DDPM-based methods.

Discussion

In this work we developed DDPM-based approaches for PET image denoising and evaluated their performance on datasets from two different tracers. Quantitative results showed that the DDPM-based frameworks with PET information included had better performance than the NLM, Unet, and GAN-based denoising methods. One benefit of the DDPM-based framework compared to Unet and GAN is the ability to calculate the uncertainty map by generating multiple realizations, which has the potential to be used in progression tracking and longitudinal studies.

Regarding different DDPM-based methods, by comparing DDPM-PET and DDPM-PETMR, our evaluations showed that further adding the MR prior images as the additional network input helped improve the model’s performance both globally (as shown in Fig. 6) and regionally (as shown in Figs. 7 and 8). It also further reduced the uncertainty values as shown in Fig. 9. The results of DDPM-MR presented in Figs. 5 and 6 showed that only replying on the MR prior image was difficult to synthesize a pseudo-PET image with high accuracy as large bias existed in the results due to the high dynamic range of PET uptakes at different regions. Further adding PET information as a data-consistency constraint in the inference step helped improve the results as shown by the performance of DDPM-MR-PETCon.

Fig. 6 showed that DDPM-PETMR had better performance than DDPM-MR-PETCon regarding global quantification. Figs. 7 and 8 revealed that DDPM-MR-PETCon was better than DDPM-PETMR with respect to local quantification. One advantage of DDPM-MR-PETCon compared to DDPM-PETMR and Unet was that it was more robust regarding difference PET image noise levels as demonstrated in Figs. 6, 7 and 8. This was because during the training phase of the score function ϵθ, PET information was not included. Another advantage of DDPM-MR-PETCon was that it could be further extended to PET image reconstruction. This could be achieved by changing the Gaussian distribution assumption in equation (9) to Poisson distribution with PET sinograms included [46]. Considering the performance, the invariability to nosie levels, and the potential extension to PET image reconstruction, the DDPM-MR-PETCon framework was preferable over DDPM-PETMR.

For the two experiments with [18F]FDG and [18F]MK-6240, the ground truth utilized in the quantification was based on the normal-dose datasets, whose quality was not good enough and this was one limitation of the quantification. In our work, a 2D network was used in the DDPM-based methods and the neighboring two axial slices were supplied as the network input to reduce axial artifacts. We also tested the 3D network and the training was 4 times slower than the 2D network, which was not realistic based on our current computing resources. Additionally, the inference step took around 50 mins for one 3D dataset. Compared to CNN and GAN-based methods, though the diffusion models could be adapted to PET images of different noise levels and also provide additional predictive uncertainty map, the high computational cost during inference was one issue of the diffusion models. A lot of on-going works are trying to reduce the inference time, e.g., finding a better initialization instead of starting from the random noise [47]. Further extending the framework to enable efficient 3D network training and reducing the inference time while preserving the sampling quality were our future research directions. In our current study, the experiment based on [18F]MK-6240 datasets was based on 20 test datasets due to the limitation of datasets available for the experiment, which limited its statistical power. Additionally, we believe results of the human reader studies will be the best metric to evaluate deep learning-based image enhancement methods. Further evaluating the performance of the DDPM-based methods on more clinical datasets of different tracers and performing human reader studies are also our future research directions.

Conclusion

In this work, DDPM-based PET imaging denoising frameworks were proposed and evaluated based on [18F]FDG and [18F]MK-6240 datasets. Results showed that DDPM-based frameworks including PET information had better performance than the nonlocal mean and the Unet-based denoising methods. Adding additional MR prior in the model achieved better performance and further reduce the uncertainty during image denoising. Future work will focus on extending the framework to 3D network, further reducing the inference time, and performing more evaluations based on different tracers.

Figure 2:

Figure 2:

Diagram of the datasets used in the training/validation/test scenerios for the two experiments.

Funding

This work was supported by the National Institutes of Health under grants R21AG067422, R03EB030280, R01AG078250, P41EB022544 and P01AG036694.

Footnotes

Statements and Declarations

Competing Interests The authors have no relevant financial or non-financial interests to disclose.

Ethics approval All procedures performed in studies involving human participants were in accordance with the ethical standards of the institutional and/or national research committee and with the 1964 Helsinki declaration and its later amendments or comparable ethical standards.

Informed consent For the [18F]FDG datasets, informed consent was waived due to the retrospective merits of the datasets. For the [18F]MK-6240 datasets, informed consent was obtained from all the participants.

References

  • 1.Lin J-W, Laine AF, and Bergmann SR, “Improving pet-based physiological quantification through methods of wavelet denoising,” IEEE Transactions on Biomedical Engineering, vol. 48, no. 2, pp. 202–212, 2001. [DOI] [PubMed] [Google Scholar]
  • 2.Christian BT, Vandehey NT, Floberg JM, and Mistretta CA, “Dynamic pet denoising with hypr processing,” Journal of Nuclear Medicine, vol. 51, no. 7, pp. 1147–1154, 2010. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Chan C, Fulton R, Barnett R, Feng DD, and Meikle S, “Postreconstruction nonlocal means filtering of whole-body pet with an anatomical prior,” IEEE Transactions on medical imaging, vol. 33, no. 3, pp. 636–650, 2014. [DOI] [PubMed] [Google Scholar]
  • 4.Dutta J, Leahy RM, and Li Q, “Non-local means denoising of dynamic pet images,” PloS one, vol. 8, no. 12, p. e81390, 2013. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Yan J, Lim JC-S, and Townsend DW, “Mri-guided brain pet image filtering and partial volume correction,” Physics in medicine and biology, vol. 60, no. 3, p. 961, 2015. [DOI] [PubMed] [Google Scholar]
  • 6.Ote K, Hashimoto F, Kakimoto A, Isobe T, Inubushi T, Ota R, Tokui A, Saito A, Moriya T, Omura T, et al. , “Kinetics-induced block matching and 5-d transform domain filtering for dynamic pet image denoising,” IEEE Transactions on Radiation and Plasma Medical Sciences, vol. 4, no. 6, pp. 720–728, 2020. [Google Scholar]
  • 7.Wang Y, Ma G, An L, Shi F, Zhang P, Lalush DS, Wu X, Pu Y, Zhou J, and Shen D, “Semisupervised tripled dictionary learning for standard-dose pet image prediction using low-dose pet and multimodal mri,” IEEE transactions on biomedical engineering, vol. 64, no. 3, pp. 569–579, 2016. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Xiang L, Qiao Y, Nie D, An L, Lin W, Wang Q, and Shen D, “Deep auto-context convolutional neural networks for standard-dose pet image estimation from low-dose pet/mri,” Neurocomputing, vol. 267, pp. 406–416, 2017. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Kaplan S and Zhu Y-M, “Full-dose pet image estimation from low-dose pet image using deep learning: a pilot study,” Journal of digital imaging, vol. 32, no. 5, pp. 773–778, 2019. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Cui J, Gong K, Guo N, Wu C, Meng X, Kim K, Zheng K, Wu Z, Fu L, Xu B, et al. , “Pet image denoising using unsupervised deep learning,” European journal of nuclear medicine and molecular imaging, vol. 46, no. 13, pp. 2780–2789, 2019. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Chen KT, Gong E, de Carvalho Macruz FB, Xu J, Boumis A, Khalighi M, Poston KL, Sha SJ, Greicius MD, Mormino E, et al. , “Ultra–low-dose 18f-florbetaben amyloid pet imaging using deep learning with multi-contrast mri inputs,” Radiology, vol. 290, no. 3, pp. 649–656, 2019. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Hashimoto F, Ohba H, Ote K, Teramoto A, and Tsukada H, “Dynamic pet image denoising using deep convolutional neural networks without prior training datasets,” IEEE access, vol. 7, pp. 96594–96603, 2019. [Google Scholar]
  • 13.da Costa-Luis CO and Reader AJ, “Micro-networks for robust mr-guided low count pet imaging,” IEEE transactions on radiation and plasma medical sciences, vol. 5, no. 2, pp. 202–212, 2020. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Schramm G, Rigie D, Vahle T, Rezaei A, Van Laere K, Shepherd T, Nuyts J, and Boada F, “Approximating anatomically-guided pet reconstruction in image space using a convolutional neural network,” Neuroimage, vol. 224, p. 117399, 2021. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Mehranian A, Wollenweber SD, Walker MD, Bradley KM, Fielding PA, Su K-H, Johnsen R, Kotasidis F, Jansen FP, and McGowan DR, “Image enhancement of whole-body oncology [18f]-fdg pet scans using deep neural networks to reduce noise,” European journal of nuclear medicine and molecular imaging, vol. 49, no. 2, pp. 539–549, 2022. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Daveau RS, Law I, Henriksen OM, Hasselbalch SG, Andersen UB, Anderberg L, Højgaard L, Andersen FL, and Ladefoged CN, “Deep learning based low-activity pet reconstruction of [11c] pib and [18f] fe-pe2i in neurodegenerative disorders,” NeuroImage, vol. 259, p. 119412, 2022. [DOI] [PubMed] [Google Scholar]
  • 17.Ouyang J, Chen KT, Gong E, Pauly J, and Zaharchuk G, “Ultra-low-dose pet reconstruction using generative adversarial network with feature matching and task-specific perceptual loss,” Medical physics, vol. 46, no. 8, pp. 3555–3564, 2019. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Lei Y, Dong X, Wang T, Higgins K, Liu T, Curran WJ, Mao H, Nye JA, and Yang X, “Whole-body pet estimation from low count statistics using cycle-consistent generative adversarial networks,” Physics in Medicine & Biology, vol. 64, no. 21, p. 215017, 2019. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Zhou L, Schaefferkoetter JD, Tham IW, Huang G, and Yan J, “Supervised learning with cyclegan for low-dose fdg pet image denoising,” Medical image analysis, vol. 65, p. 101770, 2020. [DOI] [PubMed] [Google Scholar]
  • 20.Song T-A, Chowdhury SR, Yang F, and Dutta J, “Pet image super-resolution using generative adversarial networks,” Neural Networks, vol. 125, pp. 83–91, 2020. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Sanaat A, Shiri I, Arabi H, Mainta I, Nkoulou R, and Zaidi H, “Deep learning-assisted ultra-fast/low-dose whole-body pet/ct imaging,” European journal of nuclear medicine and molecular imaging, vol. 48, no. 8, pp. 2405–2415, 2021. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Xue S, Guo R, Bohn KP, Matzke J, Viscione M, Alberts I, Meng H, Sun C, Zhang M, Zhang M, et al. , “A cross-scanner and cross-tracer deep learning method for the recovery of standard-dose imaging quality from low-dose pet,” European journal of nuclear medicine and molecular imaging, vol. 49, no. 6, pp. 1843–1856, 2022. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Gong K, Guan J, Liu C-C, and Qi J, “Pet image denoising using a deep neural network through fine tuning,” IEEE Transactions on Radiation and Plasma Medical Sciences, vol. 3, no. 2, pp. 153–161, 2018. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.Liu H, Wu J, Lu W, Onofrey JA, Liu Y-H, and Liu C, “Noise reduction with cross-tracer and cross-protocol deep transfer learning for low-dose pet,” Physics in Medicine & Biology, vol. 65, no. 18, p. 185006, 2020. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Chen KT, Schürer M, Ouyang J, Koran MEI, Davidzon G, Mormino E, Tiepolt S, Hoffmann K-T, Sabri O, Zaharchuk G, et al. , “Generalization of deep learning models for ultra-low-count amyloid pet/mri using transfer learning,” European journal of nuclear medicine and molecular imaging, vol. 47, no. 13, pp. 2998–3007, 2020. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Cui J, Gong K, Guo N, Wu C, Kim K, Liu H, and Li Q, “Populational and individual information based pet image denoising using conditional unsupervised learning,” Physics in Medicine & Biology, vol. 66, no. 15, p. 155001, 2021. [DOI] [PubMed] [Google Scholar]
  • 27.Zhou B, Miao T, Mirian N, Chen X, Xie H, Feng Z, Guo X, Li X, Zhou SK, Duncan JS, et al. , “Federated transfer learning for low-dose pet denoising: A pilot study with simulated heterogeneous data,” IEEE Transactions on Radiation and Plasma Medical Sciences, 2022. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Ho J, Jain A, and Abbeel P, “Denoising diffusion probabilistic models,” Advances in Neural Information Processing Systems, vol. 33, pp. 6840–6851, 2020. [Google Scholar]
  • 29.Song Y and Ermon S, “Generative modeling by estimating gradients of the data distribution,” Advances in Neural Information Processing Systems, vol. 32, 2019. [Google Scholar]
  • 30.Song Y, Sohl-Dickstein J, Kingma DP, Kumar A, Ermon S, and Poole B, “Score-based generative modeling through stochastic differential equations,” arXiv preprint arXiv:2011.13456, 2020. [Google Scholar]
  • 31.Edupuganti V, Mardani M, Vasanawala S, and Pauly J, “Uncertainty quantification in deep MRI reconstruction,” IEEE Transactions on Medical Imaging, vol. 40, no. 1, pp. 239–250, 2020. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Wu P, Sisniega A, Uneri A, Han R, Jones C, Vagdargi P, Zhang X, Luciano M, Anderson W, and Siewerdsen J, “Using uncertainty in deep learning reconstruction for cone-beam CT of the brain,” arXiv preprint arXiv:2108.09229, 2021. [Google Scholar]
  • 33.Sudarshan VP, Upadhyay U, Egan GF, Chen Z, and Awate SP, “Towards lower-dose PET using physics-based uncertainty-aware multimodal learning with robustness to out-of-distribution data,” Medical Image Analysis, vol. 73, p. 102187, 2021. [DOI] [PubMed] [Google Scholar]
  • 34.Dhariwal P and Nichol A, “Diffusion models beat gans on image synthesis,” Advances in Neural Information Processing Systems, vol. 34, pp. 8780–8794, 2021. [Google Scholar]
  • 35.Rombach R, Blattmann A, Lorenz D, Esser P, and Ommer B, “High-resolution image synthesis with latent diffusion models,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 10684–10695, 2022. [Google Scholar]
  • 36.Saharia C, Ho J, Chan W, Salimans T, Fleet DJ, and Norouzi M, “Image super-resolution via iterative refinement,” arXiv preprint arXiv:2104.07636, 2021. [DOI] [PubMed] [Google Scholar]
  • 37.Lugmayr A, Danelljan M, Romero A, Yu F, Timofte R, and Van Gool L, “Repaint: Inpainting using denoising diffusion probabilistic models,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 11461–11471, 2022. [Google Scholar]
  • 38.Jalal A, Arvinte M, Daras G, Price E, Dimakis AG, and Tamir J, “Robust compressed sensing mri with deep generative priors,” Advances in Neural Information Processing Systems, vol. 34, pp. 14938–14954, 2021. [Google Scholar]
  • 39.Chung H and Ye JC, “Score-based diffusion models for accelerated mri,” Medical Image Analysis, p. 102479, 2022. [DOI] [PubMed] [Google Scholar]
  • 40.Isola P, Zhu J-Y, Zhou T, and Efros AA, “Image-to-image translation with conditional adversarial networks,” in Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 1125–1134, 2017. [Google Scholar]
  • 41.Tiepolt S, Hesse S, Patt M, Luthardt J, Schroeter ML, Hoffmann K-T, Weise D, Gertz H-J, Sabri O, and Barthel H, “Early [18f] florbetaben and [11c] pib pet images are a surrogate biomarker of neuronal injury in alzheimer’s disease,” European journal of nuclear medicine and molecular imaging, vol. 43, no. 9, pp. 1700–1709, 2016. [DOI] [PubMed] [Google Scholar]
  • 42.Hammes J, Leuwer I, Bischof GN, Drzezga A, and van Eimeren T, “Multimodal correlation of dynamic [18f]-av-1451 perfusion pet and neuronal hypometabolism in [18f]-fdg pet,” European journal of nuclear medicine and molecular imaging, vol. 44, no. 13, pp. 2249–2256, 2017. [DOI] [PubMed] [Google Scholar]
  • 43.Visser D, Wolters EE, Verfaillie SC, Coomans EM, Timmers T, Tuncel H, Reimand J, Boellaard R, Windhorst AD, Scheltens P, et al. , “Tau pathology and relative cerebral blood flow are independently associated with cognition in alzheimer’s disease,” European journal of nuclear medicine and molecular imaging, vol. 47, no. 13, pp. 3165–3175, 2020. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 44.Avants BB, Tustison N, Song G, et al. , “Advanced normalization tools (ants),” Insight j, vol. 2, no. 365, pp. 1–35, 2009. [Google Scholar]
  • 45.Fischl B, “Freesurfer,” Neuroimage, vol. 62, no. 2, pp. 774–781, 2012. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 46.Qi J and Leahy RM, “Iterative reconstruction techniques in emission computed tomography,” Physics in Medicine & Biology, vol. 51, no. 15, p. R541, 2006. [DOI] [PubMed] [Google Scholar]
  • 47.Chung H, Sim B, and Ye JC, “Come-closer-diffuse-faster: Accelerating conditional diffusion models for inverse problems through stochastic contraction,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 12413–12422, 2022. [Google Scholar]

RESOURCES