Abstract
Low-dose CT image reconstruction has been a popular research topic in recent years. A typical reconstruction method based on post-log measurements is called penalized weighted-least squares (PWLS). Due to the underlying limitations of the post-log statistical model, the PWLS reconstruction quality is often degraded in low-dose scans. This paper investigates a shifted-Poisson (SP) model based likelihood function that uses the pre-log raw measurements that better represents the measurement statistics, together with a data-driven regularizer exploiting a Union of Learned TRAnsforms (SPULTRA). Both the SP induced data-fidelity term and the regularizer in the proposed framework are nonconvex. The proposed SPULTRA algorithm uses quadratic surrogate functions for the SP induced data-fidelity term. Each iteration involves a quadratic subproblem for updating the image, and a sparse coding and clustering subproblem that has a closed-form solution. The SPULTRA algorithm has a similar computational cost per iteration as its recent counterpart PWLS-ULTRA that uses post-log measurements, and it provides better image reconstruction quality than PWLS-ULTRA, especially in low-dose scans.
Index Terms—: Inverse problems, sparse representation, transform learning, shifted-Poisson model, nonconvex optimization, efficient algorithms, machine learning
I. Introduction
Recent years have witnessed the growing deployment of X-ray computed tomography (CT) in medical applications. Simultaneously there has been great concern to reduce the potential risks caused by exposure to X-ray radiation. Strategies for reducing the X-ray radiation in CT include reducing the photon intensity at the X-ray source, i.e., low-dose CT (LDCT), or lowering the number of projection views obtained by the CT machine, i.e., sparse-view CT. In the case where the X-ray radiation is extremely low, the CT image may not be suitable for medical diagnosis, but it is still quite helpful for non-diagnostic applications such as attenuation correction for PET/CT imaging [1]–[3] and virtual CT colonoscopy screening [4]. Reconstructing CT images with reduced radiation is challenging, and many reconstruction methods have been proposed for this setting. Model-based iterative reconstruction (MBIR) is widely used [5] among these approaches. Based on maximum a posteriori (MAP) estimation, MBIR approaches form a cost function that incorporates the statistical model for the acquired measurements and the prior knowledge (model) of the images. This section first reviews some of the statistical models for CT measurements along with recent works on extracting prior knowledge about images for LDCT image reconstruction, and then presents our contributions.
A. Background
Accurate statistical modeling of the measurements in CT scanners is challenging, especially in low-dose imaging, when the electronic noise in the data acquisition system (DAS) becomes significant [6]–[12]. Approximations of the measurement statistics can be categorized into post-log and pre-log models [13], which are detailed next.
The post-log models work on data obtained from the logarithmic transformation of the raw measurements, which is often assumed Gaussian distributed. Since the logarithmic transformation approximately linearizes the raw measurements, methods based on post-log data can readily exploit various optimization approaches and regularization designs with efficiency and convergence guarantees for this reconstruction problem [14]–[16]. The post-log methods however have a major drawback: the raw measurements may contain non-positive values on which the logarithmic transformation cannot be taken (or near-zero positive measurements whose logarithm can be very negative), particularly when the electronic noise becomes significant as compared to the photon statistical noise in low-dose cases.
There are many pre-correction approaches to deal with non-positive raw measurements for post-log methods. Examples of such approaches include using a statistical weight of zero for such measurements [17], replacing the non-positive measurements with a small positive value [18] and filtering neighboring measurements [6]. Thibault et al. [19] proposed a recursive filter which preserves the local mean to pre-process noisy measurements, but still used a non-linear function to map all noisy measurements to strictly positive values. Chang et al. [20] applied the local linear minimum mean-square error (LLMMSE) filter to pre-process the raw measurements, but the LLMMSE filter does not guarantee positivity in its output sinograms and introduces correlations among neighbouring channels. This correlation violates the assumption of independence of sinogram data on which MAP reconstruction formulations rely. Chang et al. [20] also proposed a pointwise Bayesian restoration (PBR) approach, which better preserves the independence of sinogram data while reducing bias for photon-starved CT data. When pre-processing a large percentage of non-positive values for LDCT measurements, these pre-correction methods may still introduce bias in the reconstructed image and can degrade image quality [13], [19]. The logarithmic transformation itself causes a positive bias in the line integrals from which the image is reconstructed [13], [21]. A typical method for reconstructing images from the post-log data is penalized weighted least squares (PWLS) [18] that optimizes an objective consisting of a weighted least squares data fidelity term and a regularization penalty. However, the pre-correction process and non-linear logarithmic operation create challenges in estimating the statistical weights for the PWLS methods [19], [22].
Contrary to the post-log methods, the pre-log methods work directly with the raw measurements. A robust statistical model for the pre-log raw CT measurements is the shifted-Poisson (SP) model. This model shifts the measurements by the variance of the electric readout noise. The shifted measurement has its variance equal to its mean, so that it could be approximated to be Poisson distributed. Since the shifted-Poisson model is a better approximation for CT measurement statistics compared to the Gaussian model [13], [23]–[26], and no pre-correction of the data is needed for most LDCT dose levels [13], this paper uses this SP model for LDCT image reconstruction.
There has been growing interest in improving CT image reconstruction by extracting prior knowledge from previous patient scans. Many methods have been proposed in this regard, such as prior image constrained compressed sensing methods (PICCS) [27]–[29], or the previous normal-dose scan induced nonlocal means method [30], [31]. More recently, inspired by the success of learning-based methods in image processing and computer vision, researchers have incorporated data-driven approaches along with statistical models for LDCT image reconstruction. One such approach proposed by Xu et al. [32] combined dictionary learning techniques with the PWLS method for LDCT image reconstruction. The dictionary they used was either pre-learned from a training image set (consisting of 2D images) and fixed during reconstruction, or adaptively learned while reconstructing the image. The 2D dictionary model for image patches was later extended to a 2.5D dictionary (where different dictionaries were trained from 2D image patches extracted from axial, sagittal, and coronal planes of 3D data) [33], and then to a 3D dictionary trained from 3D image patches [34]. These dictionary learning and reconstruction methods are typically computationally expensive, because they involve repeatedly optimizing NP-hard problems [35] for estimating the sparse coefficients of patches. The learning of sparsifying transforms (ST) was proposed in recent works [36], [37] as a generalized analysis dictionary learning method, where the sparse coefficients are estimated directly by simple and efficient thresholding. Pre-learned square sparsifying transforms have been recently incorporated into 2D LDCT image reconstruction with both post-log Gaussian statistics [38] and pre-log SP measurement models [39]. Especially, Zheng et al. [38] showed promise for PWLS with a union of pre-learned sparsifying transforms [40] regularization that generalizes the square sparsifying transform approach.
In addition to the dictionary learning-based approaches, some works have incorporated neural networks in CT image reconstruction. Adler and Öktem proposed a learned primal-dual reconstruction method [41], that uses convolutional neural networks (CNNs) to learn parameterized proximal operators. This method was applied to relatively simple 2D phantoms. Wu et. al [42] proposed a k-sparse autoencoder (KSAE) based regularizer for LDCT image reconstruction, where they trained three independent KSAEs from axial, sagittal and coronal slices for 3D reconstruction via artificial neural networks. Chen et al. [43] proposed to unfold the classical iterative reconstruction procedure into a CNN-based recurrent residual network so that the original fixed regularizers and the balancing parameters within the iterative scheme can vary for each layer. The reconstruction with this network was only performed slice by slice. He et al. proposed a parameterized plug-and-play alternating direction method (3pADMM) for PWLS model based low-dose CT image reconstruction [44]. By regarding the ADMM optimization steps as network modules, this method can optimize the 3p prior and the related parameters simultaneously. These methods are fully supervised learning methods requiring large datasets consisting of both undersampled images or measurements and the corresponding high-quality images. Some post-processing approaches involving neural networks such as a U-Net or a residual net also improve CT image quality [45], [46], but such post-processing methods usually construct an image-to-image mapping without fully incorporating the physics of the imaging process. Additionally, the generalization of supervised learning methods may be limited in the sense that the trained model may only work well on the data that is similar to the training set.
B. Contributions
Considering the robustness and accuracy offered by the SP statistics, and inspired by the data-driven image modeling methods not requiring paired training data or previous registered normal-dose images, here we propose a new LDCT image reconstruction method named SPULTRA that combines robust SP measurement modeling with a union of learned sparsifying transforms (ULTRA) based regularizer. Since the SP model leads to a nonconvex data-fidelity term, we design a series of quadratic surrogate functions for this term in our optimization. For each surrogate function combined with the ULTRA regularizer (a majorizer of the SPULTRA objective), we optimize it by alternating between an image update step and a sparse coding and clustering step. The proposed SPULTRA scheme is proved to converge to the critical points of the overall nonconvex problem. In the experiments, we compare SPULTRA with the recent PWLS-ULTRA scheme [38] under different incident photon intensity levels for 3D XCAT phantom simulations. The results demonstrate that the proposed method avoids bias in image regions caused by the PWLS-ULTRA method, especially for low X-ray doses. At the same time, SPULTRA achieves better image reconstruction quality than PWLS-ULTRA given the same number of iterations, or alternatively, SPULTRA achieves a desired image reconstruction quality much faster than the competing PWLS-ULTRA scheme, especially for low X-ray doses. We verify the bias avoidance property of SPULTRA on a synthesized 3D clinical chest scan, and an ultra low-dose 2D shoulder phantom scan simulated from standard-dose raw measurements that also involve beam-hardening effects. We compared SPULTRA with a recent deep-learning based denoising framework [46] on the 2D data demonstrating the better reconstruction and generalization ability of SPULTRA.
This paper significantly extends our previous conference work [39] by incorporating the ULTRA regularizer and proposing a faster optimization procedure with convergence guarantees. We performed extensive numerical evaluations compared to the 2D LDCT XCAT phantom results in [39].
C. Organization
The rest of this paper is organized as follows. Section II presents the proposed problem formulation for low-dose CT image reconstruction. Section III briefly reviews the ULTRA learning method and describes the proposed SPULTRA image reconstruction algorithm. Section IV discusses the convergence properties of the SPULTRA methodology. Section V presents detailed experimental results and comparisons. Section VI presents conclusions.
II. Problem Formulation for SPULTRA
The goal in LDCT image reconstruction is to estimate the linear attenuation coefficients from CT measurements . We propose to obtain the reconstructed image by solving a SP model-based penalized-likelihood problem:
(P0) |
where is a large constant. The objective function G(x) is composed of a negative log-likelihood function L(x) based on the SP model for the measurements, and a penalty term R(x) that is based on the ULTRA model [38], [47]. The SP model can be described as , where Yi is the shifted quantity of the ith measurement for i = 1,…,Nd, σ2 is the variance of the electronic noise, I0 is the incident photon count per ray from the source, fi(·) models the beam-hardening effect, and is the CT system matrix. Denoting li(x) ≜ [Ax]i (or li in short), the data-fidelity term L(x) can be written as
(1) |
where
(2) |
The beam-hardening model fi(·) is usually approximated as a polynomial [13]. For simplicity, we use a second order polynomial, i.e., , where and are coefficients of the polynomial for the ith measurement.
The ULTRA regularizer R(x) has the following form [38]:
(3) |
where denotes the set consisting of all possible partitions of {1,2,…,Np} into K disjoint subsets, K is the number of classes and Ck denotes the set of indices of patches belonging to the kth class. The operator is the patch extraction operator that extracts the jth patch of υ voxels for j = 1,…,Ñ, from x, where Ñ is the number of extracted patches. The learned transform corresponding to the kth class maps the patches to the transform domain. Vector denotes the sparse approximation of the transformed jth patch, with the parameter controlling its sparsity level. We use the ℓ0 “norm” (that counts the number of nonzero elements in zj) to enforce sparsity. The patch-based weight τj is defined as ‖Pjκ‖1/υ [38], [48], where is defined to help encourage resolution uniformity as [49, eq(39)], with aij denoting the entries of A, and is approximated as [13, eq(10)], where is the beam-hardening corrected, post-log sinogram data. To balance the data-fidelity term and the regularizer in the formulation, R(x) is scaled by a positive parameter β.
III. Algorithm
The proposed SPULTRA algorithm is based on a pre-learned union of sparsifying transforms. The process of learning such a union of transforms from a dataset of image patches has been detailed in [38]. The learning problem in [38] simultaneously groups the training patches into K classes and learns a transform in each class along with the sparse coefficients (in the transform domain) of the patches. This learning is accomplished by an alternating algorithm (see [38]). This section focuses on describing the algorithm in the reconstruction stage for SPULTRA, i.e., for (P0).
The data-fidelity term L(x) in (P0) is nonconvex when the electronic noise variance σ2 is nonzero. It is challenging to directly optimize such a logarithmic nonconvex function. We propose to iteratively design quadratic surrogate functions for this data-fidelity term L(x). In each iteration, we optimize the surrogate function that is a quadratic data-fidelity term together with the ULTRA regularizer using alternating minimization that alternates between an image update step and a sparse coding and clustering step that has closed-form solution [40]. We use the relaxed OS-LALM algorithm for the image update step [50]. We perform only one alternation between the two steps for each designed surrogate function, which saves runtime and works well in practice.
A. Surrogate function design
We design a series of quadratic surrogate functions for L(x) as follows:
(4) |
where (·)n denotes values at the nth iteration and is a row vector capturing the gradient information and is defined as . The curvatures of the nth updated parabola (surrogate) are described by . In this paper, we use the optimum curvatures [51] that are defined as follows:
(5) |
where is the second-order derivative, and operator [·]+ sets the non-positive values to zero. In practice, we replace negative values with a small positive number so that the diagonal matrix Wn is invertible. Due to numerical precision, (5) might become extremely large when is nonzero but small. To avoid this problem, we use an upper bound of the maximum second derivative for the curvature when .
By ignoring the terms irrelevant to x in (4), we get the following equivalent form of φ(x;xn):
(6) |
where “≡” means equal to within irrelevant constants of x, and . The overall surrogate function at the nth iteration for the penalized-likelihood objective function G(x) in (P0) is then
(7) |
We descend the surrogate function Φ(x;xn) in (7) by alternating once between an image update step, and a sparse coding and clustering step.
B. Image Update Step
In the image update step, we update the image x with fixed sparse codes {zj} and class assignments {Ck}. The relevant part of the majorizer for this step is
(8) |
Although we have a box constraint on x, i.e., , in practice, the upper bound xmax can be set high such that it will not be active. We applied the relaxed OS-LALM algorithm [50] to minimize (8) with the constraint. This algorithm is shown in Algorithm 1 (steps 7–10). The OS-LALM method uses majorizing matrices. In particular, the matrix ATWnA is majorized by DA ≜ diag{ATWnA1}, where 1 denotes a vector of ones. Denoting the regularization term in (8) as R2(x), its gradient is
(9) |
The Hessian of R2(x) is majorized by the following diagonal matrix:
(10) |
The (over-)relaxation parameter α ∈ [1,2) and the parameter ρt > 0 decreases with iterations t in OS-LALM according to the following equation [50]:
(11) |
C. Sparse Coding and Clustering Step
Here, with x fixed, we jointly update the sparse codes and the class memberships of patches. The relevant part of the cost function for the sparse coding and clustering step is
(12) |
Problem (12) is separable in terms of the patches, so each patch is clustered and sparse coded independently in parallel. The optimal sparse code zj in (12) is obtained by hard-thresholding, i.e., , where represents a vector hard-thresholding operator that zeros out elements whose magnitudes are smaller than γc, and leaves other entries unchanged. Then the optimized class for the jth patch is computed as follows [38]:
(13) |
We compute the cost values on the right hand side of (13) for each k = 1,···, K, and determine the that gives the minimal cost value, i.e., patch Pjx is grouped with the transform that provides the smallest value of the cost in (13). Then, the corresponding optimal sparse code is .
Algorithm 1 illustrates the proposed optimization algorithm for Problem (P0).
D. Computational Cost
The SPULTRA algorithm has a similar structure in each iteration as the recent PWLS-ULTRA [38], except for several initializations in the image update step. Since forward and backward projections are used to compute DA and during initialization, the image update step of SPULTRA is slightly slower than PWLS-ULTRA. In our experiments, we observed that the initializations took around 20% of the runtime in each outer iteration. However, in practice, especially for low doses, SPULTRA reconstructs images better than PWLS-ULTRA for a given number of outer iterations. Or alternatively, SPULTRA takes much fewer outer iterations (and runtime) to achieve the same image reconstruction quality as PWLS-ULTRA. These results are detailed in Sec. V.
Algorithm 1.
SPULTRA Algorithm
![]() |
IV. Convergence Analysis
The objective function (P0) of SPULTRA is highly nonconvex due to the non-convexity of the data-fidelity term and the regularizer. The proposed algorithm efficiently optimizes it by using surrogate functions and alternating minimization. This section provides a convergence analysis for the general optimization approach. While a recent work [47] analyzed the convergence of a related optimization method, it did not involve the use of surrogate functions and involved adaptive learning of transforms.
In the proposed method, the sparse coding and clustering step is solved exactly. For the image update step, where the cost function is quadratic as in (8), many approaches may be used to optimize it, e.g., [16], [50], [52]. Our convergence proof in the supplement assumes for simplicity that the image update step is solved exactly.
The convergence result uses the following notation. We use Z for the sparse code matrix concatenated by column vectors zj, and use a vector , whose elements represent the classes indices for the patches, i.e., Γj ∈ {1,···, K}. For an initial (x0,Z0,Γ0), we let {xn,Zn,Γn} denote the sequence of iterates generated by alternating algorithm. The objective function in (P0) is denoted as G(x,Z,Γ) and includes the constraint on x as an added barrier penalty (which takes the value +∞ when the constraint is violated and is zero otherwise). The convergence result is as follows.
Theorem 1. Assume the image update step is solved exactly. For an initial (x0,Z0,Γ0), iterative sequence (xn,Zn,Γn) generated by the SPULTRA algorithm is bounded, and the corresponding objective sequence {G(xn,Zn,Γn)} decreases monotonically and converges to G* ≜ G*(x0,Z0,Γ0). Moreover, all the accumulation points of the iterate sequence are equivalent and achieve the same value G* of the objective. Each accumulation point (x*,Z*,Γ*) also satisfies the following partial optimality conditions:
(14) |
where ∂x denotes the sub-differential operator for the function G with respect to x [53]–[55]. Finally, ‖xn+1 − xn‖2 → 0 as n → ∞.
The above theorem implies that for each initial (x0,Z0,Γ0), the objective sequence converges (although the limit may depend on initialization) and the iterate sequence in the optimization framework converges to an equivalence class of accumulation points (i.e., all accumulation points have the same objective value G*) that are also partial optimizers satisfying (14). Moreover, the image sequence satisfies ‖xn+1 − xn‖2 → 0.
When K = 1, (14) readily implies that the iterate sequence in the algorithm converges to an equivalence class of critical points [53] (that are generalized stationary points) of the nonconvex cost G(x,Z,Γ).
A detailed proof is included in the supplement1.
V. Experimental Results
Here we present numerical experiments demonstrating the behavior of SPULTRA. We evaluated the proposed SPULTRA method on the 3D XCAT phantom [56] and synthesized clinical data at multiply X-ray doses, as well as an ultra low-dose 2D shoulder phantom scan simulated from real raw data, and compared its performance with that of the state-of-the-art PWLS-ULTRA [38]. We computed the root mean square error (RMSE) and structural similarity index (SSIM) [32], [57] of XCAT images reconstructed by various methods in a region of interest (ROI). The RMSE is defined as , where Np,ROI is the number of pixels in the ROI, is the reconstructed image, and x* is the ground-truth image. We also compared to PWLS reconstruction with an edge-preserving regularizer (PWLS-EP) , where Nj represents the neighborhood of the jth pixel, κj and κk are elements of κ that encourages resolution uniformity [49]. The potential function for 3D reconstruction was ϕ(t) = δ2(|t/δ| − log(1 + |t/δ|)) with2 δ = 10 HU, and that for 2D shoulder phantom simulations was with δ = 100 HU. The results obtained by PWLS-EP were taken as initial images for methods we compared with in this section.
The SPULTRA method shifts uncorrected pre-log data by the variance of electronic noise. Such un-preprocessed pre-log data and the variance of the electronic noise on a CT scanner are proprietary to CT vendors, especially for LDCT. In our experiments of XCAT phantom simulations and the synthesized clinical data, we generated pre-log data ŷ from the XCAT phantom as well as from a clinical image reconstructed by the PWLS-ULTRA method as follows:
(15) |
where denotes a Gaussian distribution with mean μ and variance σ2. We refer to the image used for generating the synthesized clinical data as the “true” clinical image. We also simulated an ultra low-dose scan from raw (pre-log) measurements of a standard-dose scan of a 2D shoulder phantom as:
(16) |
where α is a scale factor we used to lower the dose from standard-dose measurements, and denotes the raw standard-dose measurements. We set σ = 5 for all the simulations, as suggested in prior works [13], [38]. We implemented the system model A via the separable footprint projector methods [58]. MATLAB code to reproduce the results in this work is released at http://web.eecs.umich.edu/~fessler/. Some additional results are included in the supplement.
A. XCAT phantom results
1). Framework:
We pre-learned a union of 15 square transforms from 8 × 8 × 8 overlapping patches extracted from a 420×420×54 XCAT phantom with a patch stride 2×2×2. These transforms were initialized during training [38] with 3D DCT, and the clusters were initialized randomly. We simulated 3D axial cone-beam scans using a 840 × 840 × 96 XCAT phantom with Δx = Δy = 0.4883 mm and Δz = 0.625 mm. We generated sinograms of size 888 × 64 × 984 using GE LightSpeed cone-beam geometry corresponding to a mono-energetic source with I0 = 1×104, 5 × 103, 3 × 103, and 2 × 103 incident photons per ray and no scatter, respectively. Tab. I shows percentages of non-positive measurements under different dose levels. We set these non-positive measurements to 1×10−5 for generating the post-log sinogram that PWLS-based methods rely on [13]. We reconstructed the 3D volume with a size of 420 × 420 × 96 at a coarser resolution of Δx = Δy = 0.9766 mm and Δz = 0.625 mm. The patch size during reconstruction was 8 × 8 × 8 and the stride was 3 × 3 × 3. For evaluating reconstruction performance, we chose an ROI that was composed of the central 64 out of 96 axial slices, and refer to it as the reconstruction targeted ROI. Fig. 1 shows the central slices of the true XCAT phantom inside this ROI along three directions. In the reconstruction stage of PWLS-ULTRA and SPULTRA, we used 4 iterations for the image update step, i.e., P = 4, for a good tradeoff between algorithms’ convergence and computational costs. We used 12 ordered subsets, i.e., M = 12, to speed up the algorithm. The initial image for the ULTRA methods was reconstructed by PWLS-EP, whose regularization parameter was set empirically to ensure good reconstruction quality as βep = 213 for all the experimented dose cases. We used an analytical filtered back-projection (FBP) method FDK [59] to initialize PWLS-EP. The FDK images of XCAT phantom for all the dose levels are shown in the supplement. Due to the fact that SPULTRA has a similar cost function as PWLS-ULTRA in each outer iteration, we used the same parameter settings for both methods: β = 4×104 and γc = 4×10−4, which we observed worked well for all the dose levels we tested.
TABLE I:
Percentages of non-positive measurements under different dose levels for XCAT phantom simulations.
I0 | 1 × 104 | 5 × 103 | 3 × 103 | 2 × 103 |
---|---|---|---|---|
Non-positive Percentage (%) | 0.06 | 0.20 | 0.48 | 0.96 |
Fig. 1:
Reconstruction targeted ROI of the true XCAT phantom displayed with central slices along the axial, sagittal and coronal directions. The display window is [800, 1200] HU.
2). Behavior of the learned ULTRA Models:
The learned union of transforms contributes to the clustering and sparsification of image patches. To illustrate the behavior of the learned transforms, we selected 3 of the 15 transforms that capture important structures/features of the reconstructed image (with I0 = 1 × 104) in their classes. Fig. 2 (first column) shows three voxel-level classes (voxels are clustered by majority vote among patches overlapping them) for the reconstructed central axial slice. The top image only contains soft tissues, whereas the middle image shows some edges and bones in the vertical direction, and the bottom image captures some high-contrast structures. Fig. 2 (second column) shows the transforms for the corresponding classes. Each learned transform has 512 8 × 8 × 8 filters, and we show the first 8 × 8 slice of 256 of these filters that show gradient-like and directional features. Fig. 2 also shows the central axial slice of the sparse coefficient maps (volumes) for different filters of the transforms in the third, fourth and fifth columns. Each voxel value in a sparse coefficient map is obtained by applying the specific 3D filter to a 3D patch (whose front top left corner is at that voxel) and hard-thresholding the result. Coefficients for patches not belonging to the specific class are set to zero (masked out). The sparse code maps capture different types of image features (e.g., edges at different orientations or contrasts) depending on the filters and classes.
Fig. 2:
The three rows correspond to the 1st, 13th, and 14th classes respectively. The first column displays three voxel-level clustered images of the central axial slice. Each of them is formed by image patches lie in the corresponding class. The second column displays part of the transforms for the corresponding classes. The third, fourth and fifth columns show the central axial slice of the sparse coefficient maps obtained by applying specific filters (shown in the top left corner) to patches belonging to the corresponding classes. The patch stride for plotting these figures was 1 × 1 × 1.
3). Numerical Results:
We compare the RMSE and the SSIM for SPULTRA with those for PWLS-EP and PWL-SULTRA. Tab. II lists the two metrics for the reconstruction targeted ROI after sufficient iterations (800 iterations) for convergence of PWLS-EP, PWLS-ULTRA, and SPULTRA, for various dose levels. The results show that SPULTRA achieves significant improvements in RMSE and SSIM in low-dose situations. Notably, compared to PWLS-ULTRA, SPULTRA further decreases the RMSE by up to 1.3 HU when I0 = 3 × 103, and by around 3.3 HU when I0 = 2 × 103.
TABLE II:
RMSE (HU) and SSIM of the reconstruction targeted ROI at various dose levels (I0) using the PWLS-EP, PWLS-ULTRA, and SPULTRA methods for the XCAT phantom simulations.
(a) RMSE (HU) | (b) SSIM | |||||||
---|---|---|---|---|---|---|---|---|
I0 | PWLS-EP | PWLS-ULTRA | SPULTRA | I0 | PWLS-EP | PWLS-ULTRA | SPULTRA | |
1 × 104 | 45.3 | 29.1 | 28.9 | 1 × 104 | 0.941 | 0.974 | 0.974 | |
5 × 103 | 47.1 | 33.3 | 32.8 | 5 × 103 | 0.937 | 0.969 | 0.970 | |
3 × 103 | 49.7 | 37.7 | 36.4 | 3 × 103 | 0.927 | 0.961 | 0.963 | |
2 × 103 | 53.5 | 43.2 | 39.9 | 2 × 103 | 0.911 | 0.948 | 0.956 |
The RMSE improvement of SPULTRA over PWLS-ULTRA can be more clearly observed from Fig. 3 that shows the RMSE evolution with the number of outer iterations under different dose levels. At low-doses, SPULTRA decreases the RMSE more quickly (from the same initial value) and to much lower levels than PWLS-ULTRA. Fig. 3 shows that to achieve the same RMSE as PWLS-ULTRA at 600 outer iterations, SPULTRA takes 487, 365, 251 and 133 outer iterations under I0 = 1×104, 5 × 103, 3 × 103, and 2 × 103, respectively.
Fig. 3:
RMSE comparison of SPULTRA and PWLS-ULTRA. The cursors indicate the RMSEs (Y) at specific number of outer iterations (X).
4). Computational Costs:
As discussed in Sec. III-D, SPULTRA has a similar computational cost per iteration as PWL-SULTRA, except for computing some initializations for image update. Fig. 3 shows that the SPULTRA method requires much fewer number of outer iterations than PWLS-ULTRA to achieve the same RMSE for the reconstruction, especially at low doses. When the dose is very low, e.g., when I0 = 2×103, SPULTRA takes only a quarter the number of outer iterations as PWLS-ULTRA to achieve the same RMSE. Thus, the total runtime to achieve a specific reconstruction quality at low doses is typically much lower for SPULTRA than for PWL-SULTRA. When the dose is not very low, for example when I0 = 1×104, the SPULTRA and the PWLS-ULTRA methods have similar computational costs and runtimes. To achieve RMSE of 29.26 HU (see Fig. 3a), PWLS-ULTRA requires 600 outer iterations, while SPULTRA requires 487×120% ≈ 584 effective outer iterations where the additional 20% runtime is associated with initializations in each SPULTRA outer iteration.
5). Visual Results and Image Profiles:
Fig. 4 shows the reconstructed images and the corresponding error images for PWLS-EP, PWLS-ULTRA, and SPULTRA, at I0 = 3 × 103 and I0 = 2 × 103. Compared to the PWLS-EP result, both PWLS-ULTRA and SPULTRA achieved significant improvements in image quality in terms of sharper reconstructions of anatomical structures such as bones and soft tissues, and suppressing the noise. However, the PWLS-ULTRA method introduces bias in the reconstructions, which leads to larger reconstruction errors compared to the proposed SPULTRA method. In Fig. 4, we marked three 3D ROIs in the axial plane, i.e., ROI 1, ROI 2, and ROI 3. Fig. 5 shows the zoom-in images of a 3D plot of ROI 1, and those of ROI 2 and ROI 3 are shown in the supplement. We also plot the evolution of RMSE through the axial slices of the three 3D ROIs in Fig. 6. The figures demonstrate that SPULTRA clearly outperforms the competing PWLS-EP and PWLS-ULTRA schemes.
Fig. 4:
Comparison of reconstructions and reconstruction errors at (a) I0 = 3×103 and (b) I0 = 2×103 dose levels. The 3D images are displayed with the central slices along the axial, sagittal, and coronal directions. The unit of the display windows is HU.
Fig. 5:
3D displays of reconstructions of ROI 1 defined in Fig. 4. The display windows are [900, 1200] HU.
Fig. 6:
RMSE (HU) for each axial slice of the 3D ROIs (ROI 1, ROI 2, and ROI 3). The X-axis shows slice indices of the central 64 out of 96 axial slices. Left plot: I0 = 3 × 103. Right plot: I0 = 2 × 103.
The above advantages of SPULTRA can be seen more clearly when observing the image profiles. Fig. 7 plots the image profiles for the three methods together with that of the ground-truth image. Fig. 4 shows the horizontal green solid line and the vertical red dashed line, whose intensities are plotted in Fig. 7. It is obvious that the profiles for SPULTRA are closest to the ground-truth among the three compared methods. The gap between the profiles of the PWLS-based methods and the ground-truth shows the bias caused by the compared PWLS methods.
Fig. 7:
Image profiles along the horizontal and vertical lines indicated in Fig. 4. Left plot: I0 = 3 × 103. Right plot: I0 = 2 × 103.
B. Synthesized Clinical Data
1). Framework:
We used the pre-learned union of 15 square transforms from the XCAT phantom simulations to reconstruct the synthesized helical chest scan volume of size 420 × 420 × 222 with Δx = Δy = 1.1667 mm and Δz = 0.625 mm. The sinograms were of size 888×64×3611. Since the clinical data is synthesized via the PWLS-ULTRA reconstruction, the noise model for this synthesized data is obscure, making it difficult to determine appropriate low-dose levels for such data. We tested the radiation dose of I0 = 1 × 104 with an electronic noise variance the same as the XCAT phantom simulation, i.e., σ2 = 25. The percentage of non-positive pre-log measurements for the synthesized clinical data in this case was around 0.14%. Such non-positive values were replaced by 1 × 10−5 for PWLS-based methods. Fig. 8a shows the “true” clinical image that was reconstructed from real clinical regular-dose sinogram using the PWL-SULTRA method. Similar to the XCAT phantom simulation, the initial image for both SPULTRA and PWLS-ULTRA was a reconstruction obtained using PWLS-EP. We set the regularizer parameter βep for PWLS-EP to 215 to generate a smoother (with less noise) initial image, which led to good visual image equality for the SPULTRA and PWLS-ULTRA reconstructions. Since the optimization problem for PWLS-EP is strictly convex, we simply initialized PWLS-EP with a zero image. Fig. 8b shows the PWLS-EP reconstructed image for I0 = 1 × 104. We set the regularizer parameters for both PWLS-ULTRA and SPULTRA as γc = 5 × 10−4, and β = 1.5 × 104.
Fig. 8:
(a) “true” clinical image (HU), (b) the reconstruction (HU) of the synthesized data with PWLS-EP for I0 = 1×104 with βep = 215. The central axial, sagittal, and coronal slices of the volume are shown.
2). Reconstruction results for the synthesized clinical data:
Fig. 9 shows three axial slices from the 3D reconstructions with SPULTRA and PWLS-ULTRA at I0 = 1 × 104: the middle slice (No. 67) and two slices located farther away from the center (No. 90 and No. 120). The image profiles along a horizontal line (shown in green) in the displayed slices are also shown in Fig. 9. The reconstructed slices using PWLS-ULTRA appear darker around the center compared to the “true” clinical image and the reconstructions with SPULTRA. This means PWLS-ULTRA produces a strong bias in the reconstruction. The bias can be observed more clearly in the profile plots: the pixel intensities for the SPULTRA reconstruction better follow those of the “true” clinical image, while those for the PWL-SULTRA reconstruction are much worse than the “true” values. Moreover, SPULTRA achieves sharper rising and failing edges compared to PWLS-ULTRA. In other words, SPULTRA also achieves better resolution than PWLS-ULTRA. Fig. 9 also shows a zoomed-in ROI for each of the chosen slices, and highlights some small details with arrows. It is clear that in addition to reducing the bias, SPULTRA reconstructs image details better than PWLS-ULTRA.
Fig. 9:
Reconstructed images (columns 1 to 3) and the image profiles (the 4th column) along the green line in the “true” clinical image for the synthesized clinical data with I0 = 1×104 and σ2 = 25. (a) Results for axial slice No. 67, (b) results for slice No. 90, and (c) results for slice No. 120. We selected one ROI for each of these three slices and the arrows point out some small structures in the image. The display windows for reconstructed images are [800, 1200] HU, and those for the zoomed-in ROIs are [950, 1200] HU.
C. Ultra Low-dose Experiments with Raw Data
1). Framework:
We obtained from GE a 2D fan-beam raw (pre-log) scan of a shoulder phantom, which included the beam-hardening effect. The provided 200 mA with 1 second scan can be viewed as a standard-dose scan and all the raw measurements are positive. Based on this standard-dose scan, we simulated an ultra low-dose scan as shown in (15) with α = 200, and added Poisson and Gaussian noise (σ = 5) to the measurements. The simulated measurements have about 0.4% non-positive values. The sinograms were of size 888 × 984, and reconstructed images were of size 512 × 512 with Δx = Δy = 0.9766 mm.
For PWLS-ULTRA and SPULTRA, we pre-learned a union of five square transforms using 8 × 8 overlapping image patches with stride 1 × 1 from five 512 × 512 XCAT phantom slices [38]. Here, we also compared SPULTRA with a recent deep-learning based low-dose CT denoising framework “WavResNet” combined with an RNN architecture [46]. The iterative RNN version of WavResNet was pre-trained based on the 2016 Low-Dose CT Grand Challenge data set [46]. During reconstruction, WavResNet, PWLS-ULTRA, and SPULTRA were initialized with the image reconstructed by PWLS-EP with βep = 0.1. The parameters (β,γc) for both PWLS-ULTRA and SPULTRA were set as (0.05,80). These values worked well in our experiment. In the supplement, we discuss in detail the parameter selection procedure of (β, γc) for both PWLS-ULTRA and SPULTRA. Parameters for testing WavResNet were set according to [46], and the pixel values of the input to WavResNet were converted to match the network required scalings. Since the WavResNet was trained with images reconstructed with the filtered backprojection (FBP) method [46], we also tested on this shoulder phantom that initialized WavResNet with an FBP reconstructed image. Although initializing WavResNet with an FBP reconstructed image better matches the trained model than the PWLS-EP reconstructed image does, the latter still provided better results. We included in the supplement the denoised image initialized with the FBP reconstruction.
2). Results:
Fig. 10 shows the reconstructions for the 200 mA scan (reference image) along with the reconstructions for the simulated ultra low-dose scan obtained with PWLS-EP, WavResNet, PWLS-ULTRA, and SPULTRA. Visually, WavResNet fails to reconstruct the image but improves over the initial PWLS-EP reconstruction, while PWLS-ULTRA and SPULTRA provide better image quality. This indicates that the ULTRA-based methods may have a better generalization property than WavResNet, since they learn more fundamental features of CT images (also see [38]). We selected three smooth ROIs, where the pixel values are approximately constant. Tab. III shows the mean and the standard deviation of pixel values for these ROIs for various methods and the standard-dose reference. Since the iterative RNN version of WavResNet only has small improvements over PWLS-EP, the pixel values do not change much compared with PWLS-EP. PWLS-ULTRA however reduces the bias in the central region of the image (ROI 2), but fails to correct the bias in the regions near the bones (ROI 1 and ROI 3). SPULTRA reduces the bias in the central region of the image, and also significantly corrects the bias near the bone regions. The standard deviations of the ROIs reconstructed by SPULTRA are comparable to those reconstructed by PWLS-ULTRA, and are close to those of the reference ROIs. Additionally, SPULTRA reconstructs the bone (indicated by the magenta arrow in the last two subfigures of Fig. 10) better than PWLS-ULTRA.
Fig. 10:
Reconstructions for ultra low-dose 2D scan simulated from raw measurements. The leftmost image is the PWLS-EP reconstructed image for the 200 mA scan. The second image is the PWLS-EP reconstruction for the simulated ultra low-dose scan, and it is the initial image for WavResNet [46], PWLS-ULTRA [38], and SPULTRA. The display windows are [800, 1400] HU.
TABLE III:
Mean (HU) and standard deviation (STD) (HU) of the ROIs for ultra low-dose shoulder phantom simulations
(a) Mean (HU) | (b) STD (HU) | |||||||
---|---|---|---|---|---|---|---|---|
Methods | ROI1 | ROI 2 | ROI 3 | Methods | ROI 1 | ROI 2 | ROI 3 | |
Reference | 1052.1 | 1060.1 | 1053.4 | Reference | 8.12 | 8.81 | 6.98 | |
PWLS-EP | 1032.7 | 977.5 | 1026.3 | PWLS-EP | 19.45 | 19.45 | 30.46 | |
WavResNet [46] | 1037.6 | 981.1 | 1031.2 | WavResNet [46] | 18.91 | 18.91 | 30.16 | |
PWLS-ULTRA [38] | 1031.1 | 1043.0 | 1024.2 | PWLS-ULTRA [38] | 14.82 | 10.92 | 19.29 | |
SPULTRA | 1054.7 | 1044.0 | 1049.6 | SPULTRA | 16.34 | 11.42 | 11.60 |
VI. Conclusions
This paper proposes a new LDCT reconstruction method dubbed SPULTRA that combines the shifted-Poisson statistical model with the union of learned transforms or ULTRA regularizer. To deal with the nonconvex data-fidelity term arising from the shifted-Poisson model, we iteratively designed quadratic surrogate functions for this term in the proposed algorithm. In each surrogate function update iteration, the overall cost function (i.e., majorizer) has a similar structure as in the very recent PWLS-ULTRA method, and is optimized by performing an image update step with a quadratic cost and a sparse coding and clustering step with an efficient closed-form update. We evaluated the proposed SPULTRA scheme with numerical experiments on the XCAT phantom, synthesized clinical data, and beam-hardened ultra low-dose raw measurement simulations. SPULTRA outperformed prior methods in terms of eliminating bias and noise in the reconstructed image while maintaining the resolution of the reconstruction under very low X-ray doses. SPULTRA was also much faster than PWLS-ULTRA in achieving a desired reconstruction quality for low-doses, and it had better generalization property than the WavResNet based denoising scheme. Moreover, we investigated the convergence guarantees of the proposed surrogate function based alternating minimization scheme. We will investigate variations or generalizations of the SPULTRA model such as exploiting unions of overcomplete or tall transforms, or rotationally invariant transforms in future work.
Supplementary Material
Acknowledgments
This work was supported in part by the SJTU-UM Collaborative Research Program, NSFC (61501292), NIH grant U01 EB018753, ONR grant N00014-15-1-2141, DARPA Young Faculty Award D14AP00086, and ARO MURI grants W911NF-11-1-0391 and 2015-05174-05.
Footnotes
Supplementary material is available in the multimedia tab.
“HU” used in this paper is the shifted Hounsfield unit, where air is 0 HU and water is 1000 HU.
Contributor Information
Siqi Ye, University of Michigan - Shanghai Jiao Tong University Joint Institute, Shanghai Jiao Tong University, Shanghai 200240, China..
Saiprasad Ravishankar, Department of Computational Mathematics, Science and Engineering, and the Department of Biomedical Engineering, Michigan State University, East Lansing, MI, 48824 USA..
Yong Long, University of Michigan - Shanghai Jiao Tong University Joint Institute, Shanghai Jiao Tong University, Shanghai 200240, China..
Jeffrey A. Fessler, Department of Electrical Engineering and Computer Science, University of Michigan, Ann Arbor, MI, 48109 USA..
REFERENCES
- [1].E Kinahan P, W Townsend D, Beyer T, and Sashin D, “Attenuation correction for a combined 3D PET/CT scanner,” Med. Phys, vol. 25, no. 10, pp. 2046–2053, 1998. [DOI] [PubMed] [Google Scholar]
- [2].Xia T, Alessio AM, De Man B, Manjeshwar R, Asma E, and Kinahan PE, “Ultra-low dose CT attenuation correction for PET/CT,” Phys. Med. Biol, vol. 57, no. 2, pp. 309, 2011. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [3].Rui X, Cheng L, Long Y, Fu L, Alessio AM, Asma E, E Kinahan P, and De Man B, “Ultra-low dose CT attenuation correction for PET/CT: analysis of sparse view data acquisition and reconstruction algorithms,” Phys. Med. Biol, vol. 60, no. 19, pp. 7437, 2015. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [4].Wang J, Wang S, Li L, Fan Y, Lu H, and Liang Z, “Virtual colonoscopy screening with ultra low-dose CT and less-stressful bowel preparation: A computer simulation study,” IEEE Trans. Nucl. Sci, vol. 55, no. 5, pp. 2566–2575, 2008. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [5].Fessler JA, “Statistical image reconstruction methods for transmission tomography,” in Handbook of Medical Imaging, Volume 2. Medical Image Processing and Analysis, Sonka M and Fitzpatrick J. Michael, Eds., pp. 1–70. SPIE, Bellingham, 2000. [Google Scholar]
- [6].Hsieh J, “Adaptive streak artifact reduction in computed tomography resulting from excessive X-ray photon noise,” Med. Phys, vol. 25, no. 11, pp. 2139–2147, 1998. [DOI] [PubMed] [Google Scholar]
- [7].R Whiting B, “Signal statistics in X-ray computed tomography,” in Medical Imaging 2002: Physics of Medical Imaging. International Society for Optics and Photonics, 2002, vol. 4682, pp. 53–61. [Google Scholar]
- [8].Elbakri IA and Fessler JA, “Efficient and accurate likelihood for iterative image reconstruction in X-ray computed tomography,” in Medical Imaging 2003: Image Processing. International Society for Optics and Photonics, 2003, vol. 5032, pp. 1839–1851. [Google Scholar]
- [9].Yu L, Shiung M, Jondal D, and H McCollough C, “Development and validation of a practical lower-dose-simulation tool for optimizing computed tomography scan protocols,” J. Comput. Assist. Tomogr, vol. 36, no. 4, pp. 477–487, 2012. [DOI] [PubMed] [Google Scholar]
- [10].Ma J, Liang Z, Fan Y, Liu Y, Huang J, Chen W, and Lu H, “Variance analysis of X-ray CT sinograms in the presence of electronic noise background,” Med. Phys, vol. 39, no. 7 Part1, pp. 4051–4065, 2012. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [11].Ding Q, Long Y, Zhang X, and Fessler JA, “Modeling mixed Poisson-Gaussian noise in statistical image reconstruction for X-ray CT,” in Proc. 4th Intl. Mtg. on Image Formation in X-ray CT, 2016, pp. 399–402. [Google Scholar]
- [12].Nuyts J, De Man B, Fessler JA, Zbijewski W, and Beekman FJ, “Modelling the physics in the iterative reconstruction for transmission computed tomography,” Phys. Med. Biol, vol. 58, no. 12, pp. R63, 2013. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [13].Fu L, Lee TC, Kim SM, Alessio AM, Kinahan PE, Chang ZQ, Sauer K, Kalra MK, and De Man B, “Comparison between pre-log and post-log statistical models in ultra-low-dose CT reconstruction,” IEEE Trans. Med. Imag, vol. 36, no. 3, pp. 707–720, 2017. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [14].Beister M, Kolditz D, and A Kalender W, “Iterative reconstruction methods in X-ray CT,” Physica Medica: European Journal of Medical Physics, vol. 28, no. 2, pp. 94–108, 2012. [DOI] [PubMed] [Google Scholar]
- [15].Thibault J-B, Sauer K, Bouman C, and Hsieh J, “A three-dimensional statistical approach to improved image quality for multi-slice helical CT,” Med. Phys, vol. 34, no. 11, pp. 4526–44, November 2007. [DOI] [PubMed] [Google Scholar]
- [16].Kim D, Ramani S, and Fessler JA, “Combining ordered subsets and momentum for accelerated X-ray CT image reconstruction,” IEEE Trans. Med. Imag, vol. 34, no. 1, pp. 167–178, January 2015. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [17].Elbakri IA and Fessler JA, “Statistical image reconstruction for polyenergetic X-ray computed tomography,” IEEE Trans. Med. Imag, vol. 21, no. 2, pp. 89–99, 2002. [DOI] [PubMed] [Google Scholar]
- [18].Wang J, Li T, Lu HB, and Liang ZR, “Penalized weighted least-squares approach to sinogram noise reduction and image reconstruction for low-dose X-ray computed tomography,” IEEE Trans. Med. Imag, vol. 25, no. 10, pp. 1272–1283, 2006. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [19].Thibault J-B, Bouman CA, Sauer KD, and Hsieh J, “A recursive filter for noise reduction in statistical iterative tomographic imaging,” in Proc. SPIE 6065, Computational Imaging IV, 2006, vol. 6065, p. 60650X. [Google Scholar]
- [20].Chang Z, Zhang R, Thibault J-B, Pal D, Fu L, Sauer K, and Bouman C, “Modeling and pre-treatment of photon-starved CT data for iterative reconstruction,” IEEE Trans. Med. Imag, vol. 36, no. 1, pp. 277–287, 2016. [DOI] [PubMed] [Google Scholar]
- [21].Fessler JA, “Hybrid poisson/polynomial objective functions for tomographic image reconstruction from transmission scans,” IEEE Trans. Image Process, vol. 4, no. 10, pp. 1439–1450, 1995. [DOI] [PubMed] [Google Scholar]
- [22].Hayes J, Zhang R, Zhang C, Gomez-Cardona D, and Chen GH, “Unbiased statistical image reconstruction in low-dose ct,” in Medical Imaging 2019: Physics of Medical Imaging. International Society for Optics and Photonics, 2019, vol. 10948, p. 1094859. [Google Scholar]
- [23].La Riviere PJ, “Penalized-likelihood sinogram smoothing for low-dose CT,” Med. Phys, vol. 32, no. 6, pp. 1676–1683, 2005. [DOI] [PubMed] [Google Scholar]
- [24].La Riviere PJ, Bian JG, and Vargas PA, “Penalized-likelihood sinogram restoration for computed tomography,” IEEE Trans. Med. Imag, vol. 25, no. 8, pp. 1022–1036, 2006. [DOI] [PubMed] [Google Scholar]
- [25].Xu JY and Tsui BMW, “Electronic noise modeling in statistical iterative reconstruction,” IEEE Trans. Image Process, vol. 18, no. 6, pp. 1228–1238, 2009. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [26].Tilley S, Jacobson M, Cao Q, Brehler M, Sisniega A, Zbijewski W, and Stayman JW, “Penalized-likelihood reconstruction with high-fidelity measurement models for high-resolution cone-beam imaging,” IEEE Trans. Med. Imag, vol. 37, no. 4, pp. 988–999, 2017. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [27].Chen G-H, Tang J, and Leng S, “Prior image constrained compressed sensing (piccs): a method to accurately reconstruct dynamic CT images from highly undersampled projection data sets,” Med. Phys, vol. 35, no. 2, pp. 660–663, 2008. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [28].Ramirez-Giraldo JC, Trzasko J, Leng S, Yu L, Manduca A, and McCollough CH, “Nonconvex prior image constrained compressed sensing (NCPICCS): Theory and simulations on perfusion CT,” Med. Phys, vol. 38, no. 4, pp. 2157–2167, 2011. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [29].Chen G-H, Thériault-Lauzier J, and Tang P, Nett B, Leng S, Zambelli J, Qi Z, Bevins N, Raval A, Reeder S, et al. , “Time-resolved interventional cardiac c-arm cone-beam CT: An application of the PICCS algorithm,” IEEE Trans. Med. Imag, vol. 31, no. 4, pp. 907–923, 2012. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [30].Ma J, Huang J, Feng Q, Zhang H, Lu H, Liang Z, and Chen W, “Low-dose computed tomography image restoration using previous normal-dose scan,” Med. Phys, vol. 38, no. 10, pp. 5713–5731, 2011. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [31].Zhang H, Zeng D, Zhang H, Wang J, Liang Z, and Ma J, “Applications of nonlocal means algorithm in low-dose X-ray CT image processing and reconstruction: A review,” Med. Phys, vol. 44, no. 3, pp. 1168–1185, 2017. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [32].Xu Q, Yu H, Mou X, Zhang L, Hsieh J, and Wang G, “Low-dose X-ray CT reconstruction via dictionary learning,” IEEE Trans. Med. Imag, vol. 31, no. 9, pp. 1682–97, September 2012. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [33].Luo J, Eri H, Can A, Ramani S, Fu L, and De Man B, “2.5 D dictionary learning based computed tomography reconstruction,” in SPIE Defense+ Security. pp. 98470L–98470L–12, International Society for Optics and Photonics. [Google Scholar]
- [34].Liu J, Hu Y, Yang J, Chen Y, Shu H, Luo L, Feng Q, Gui Z, and Coatrieux G, “3D feature constrained reconstruction for low-dose CT imaging,” IEEE Trans. Circuits Syst. Video Technol, vol. 28, no. 5, pp. 1232–1247, May 2018. [Google Scholar]
- [35].R Garey M and S Johnson D, Computers and intractability, vol. 29, WH Freeman; New York, 2002. [Google Scholar]
- [36].Ravishankar S and Bresler Y, “Learning sparsifying transforms,” IEEE Trans. Signal Process, vol. 61, no. 5, pp. 1072–1086, 2013. [Google Scholar]
- [37].Ravishankar S and Bresler Y, “ℓ0 sparsifying transform learning with efficient optimal updates and convergence guarantees,” IEEE Trans. Signal Process, vol. 63, no. 9, pp. 2389–2404, 2015. [Google Scholar]
- [38].Zheng X, Ravishankar S, Long Y, and Fessler JA, “PWLS-ULTRA: An efficient clustering and learning-based approach for low-dose 3D CT image reconstruction,” IEEE Trans. Med. Imag, vol. 37, no. 6, pp. 1498–1510, 2018. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [39].Ye S, Ravishankar S, Long Y, and Fessler JA, “Adaptive sparse modeling and shifted-Poisson likelihood based approach for low-dose CT image reconstruction,” in Proc. IEEE Wkshp. Machine Learning for Signal Proc, 2017, pp. 1–6. [Google Scholar]
- [40].Wen B, Ravishankar S, and Bresler Y, “Structured overcomplete sparsifying transform learning with convergence guarantees and applications,” Intl. J. Comp. Vision, vol. 114, no. 2–3, pp. 137–167, 2015. [Google Scholar]
- [41].Adler J and Oktem O, “Learned primal-dual reconstruction,”¨IEEE Trans. Med. Imag, vol. 37, no. 6, pp. 1322–1332, 2018. [DOI] [PubMed] [Google Scholar]
- [42].Wu D, Kim K, El Fakhri G, and Li Q, “Iterative low-dose CT reconstruction with priors trained by artificial neural network,” IEEE Trans. Med. Imag, vol. 36, no. 12, pp. 2479–2486, 2017. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [43].Chen H, Zhang Y, Chen Y, Zhang J, Zhang W, Sun H, Lv Y, Liao P, Zhou J, and Wang G, “LEARN: Learned experts assessment-based reconstruction network for sparse-data CT,” IEEE Trans. Med. Imag, vol. 37, no. 6, pp. 1333–1347, June 2018. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [44].He J, Yang Y, Wang Y, Zeng D, Bian Z, Zhang H, Sun J, Xu Z, and Ma J, “Optimizing a parameterized plug-and-play admm for iterative low-dose ct reconstruction,” IEEE Trans. Med. Imag, vol. 38, no. 2, pp. 371–382, 2018. [DOI] [PubMed] [Google Scholar]
- [45].Han YS, Yoo J, and Ye JC, “Deep residual learning for compressed sensing CT reconstruction via persistent homology analysis,” arXiv preprint arXiv:1611.06391, 2016.
- [46].Kang E, Chang W, Yoo J, and Ye JC, “Deep convolutional framelet denoising for low-dose CT via wavelet residual network,” IEEE Trans. Med. Imag, vol. 37, no. 6, pp. 1358–1369, 2018. [DOI] [PubMed] [Google Scholar]
- [47].Ravishankar S and Bresler Y, “Data-driven learning of a union of sparsifying transforms model for blind compressed sensing,” IEEE Trans. Comput. Imag, vol. 2, no. 3, pp. 294–309, 2016. [Google Scholar]
- [48].Chun IY, Zheng X, Long Y, and Fessler JA, “Sparse-view X-ray CT reconstruction using ℓ1 regularization with learned sparsifying transform,” in Proc. Intl. Mtg. on Fully 3D Image Recon. in Rad. and Nuc. Med, 2017, pp. 115–9. [Google Scholar]
- [49].Cho JH and Fessler JA, “Regularization designs for uniform spatial resolution and noise properties in statistical image reconstruction for 3-D X-ray CT,” IEEE Trans. Med. Imag, vol. 34, no. 2, pp. 678–689, 2015. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [50].Nien H and Fessler JA, “Relaxed linearized algorithms for faster X-ray CT image reconstruction,” IEEE Trans. Med. Imag, vol. 35, no. 4, pp. 1090–1098, April 2016. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [51].Erdogan H and Fessler JA, “Monotonic algorithms for transmission tomography,” IEEE Trans. Med. Imag, vol. 18, no. 9, pp. 801–814, 1999. [DOI] [PubMed] [Google Scholar]
- [52].Kim D and Fessler JA, “Generalizing the optimized gradient method for smooth convex minimization,” SIAM J. Optim, vol. 28, no. 2, pp. 1920–1950, 2018. [Google Scholar]
- [53].Rockafellar RT and Wets RJ-B, Variational analysis, vol. 317, Springer Science & Business Media, 2009. [Google Scholar]
- [54].S Mordukhovich B, Variational analysis and generalized differentiation I: Basic theory, vol. 330, Springer Science & Business Media, 2006. [Google Scholar]
- [55].Ravishankar S and Bresler Y, “Efficient blind compressed sensing using sparsifying transforms with convergence guarantees and application to magnetic resonance imaging,” SIAM J. Imaging Sci, vol. 8, no. 4, pp. 2519–2557, 2015. [Google Scholar]
- [56].Segars WP, Mahesh M, Beck TJ, Frey EC, and Tsui BMW, “Realistic CT simulation using the 4D XCAT phantom,” Med. Phys, vol. 35, no. 8, pp. 3800–8, August 2008. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [57].Zhang Y, Mou X, Wang G, and Yu H, “Tensor-based dictionary learning for spectral CT reconstruction,” IEEE Trans. Med. Imag, vol. 36, no. 1, pp. 142–154, 2017. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [58].Long Y, Fessler JA, and Balter JM, “3D forward and back-projection for X-ray CT using separable footprints,” IEEE Trans. Med. Imag, vol. 29, no. 11, pp. 1839–1850, 2010. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [59].Feldkamp LC, Davis LA and Kress JW, “Practical cone-beam algorithm,” J. Opt. Soc. Amer. A, Opt. Image Sci, vol. 1, no. 6, pp. 612–619, 1984. [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.