Abstract
Spectral computed tomography (CT) has recently emerged as an advanced version of medical CT and significantly improves conventional (single-energy) CT. Spectral CT has two main forms: dual-energy computed tomography (DECT) and photon-counting computed tomography (PCCT), which offer image improvement, material decomposition, and feature quantification relative to conventional CT. However, the inherent challenges of spectral CT, evidenced by data and image artifacts, remain a bottleneck for clinical applications. To address these problems, machine learning techniques have been widely applied to spectral CT. In this review, we present the state-of-the-art data-driven techniques for spectral CT.
Keywords: Photon-counting CT (PCCT), Dual-energy CT (DECT), Artificial Intelligence (AI), Machine Learning, Deep Learning
I. Introduction
SINCE Cormack and Hounsfield’s Nobel prize-winning breakthrough, X-ray CT is extensively used in medical applications and produces a huge number of gray-scale CT images. However, these images are often insufficient to distinguish crucial differences between biological tissues and contrast agents. From the perspective of physics, the X-ray spectrum from a medical device is polychromatic, and interactions between X-rays and biological tissues depend on the X-ray energy, which suggests the feasibility to obtain spectral, multi-energy, or true-color, CT images.
Over the past decade, spectral CT has been rapidly developed as a new generation of CT technology. DECT and PCCT are the two main forms of spectral CT. DECT is a method of acquiring two projection datasets at different energy levels. PCCT, on the other hand, uses detectors that measure individual photons and their energy, promising significantly better performance with major improvements in energy resolution, spatial resolution and dose efficiency [1], [2]. Despite the intrinsic merits of spectral CT, there are technical challenges already, being or yet to be addressed [3], [4]. To meet these challenges, the solutions can be hardware-oriented, software-oriented, or hybrid.
Traditionally, CT algorithms are grouped into two categories, which are analytic and iterative reconstruction respectively. A new category of CT algorithms has recently emerged: artificial intelligence (AI)-inspired, learning-based or data-driven reconstruction. These algorithms are commonly implemented as deep neural networks (NNs), which are iteratively trained for image reconstruction and post-processing, and then used for inference in the feed-forward fashion just like a closed-form solution.
Several reviews have been dedicated to machine learning and deep learning in CT. These papers cover a wide range of topics, including image reconstruction, segmentation, classification, and more. For example, Litjens et al. [5] and Sahiner et al. [6] comprehensively surveyed deep learning applications in medical imaging. Domingues et al. [7] proposed a review on deep learning in CT and positron emission tomography (PET). However, few have specifically focused on spectral CT.
This review paper provides a technical overview of the current state-of-the-art of machine learning techniques for spectral CT, especially deep learning ones. The paper is divided into the following sections: DECT and PCCT systems, image reconstruction, material decomposition, pre- and post-processing, hybrid imaging, perspectives and conclusion. Section II describes DECT and PCCT systems. Section III discusses the application of learning-based techniques for multienergy CT reconstruction from energy-binned data, which use shallow or deep network architectures, from dictionary learning (DL) to much deeper contemporary network models. Reconstruction of multi-energy CT images will face the problem of beam hardening. Section IV covers different approaches to material decomposition: image-based techniques, which use as input multi-energy CT, and alternative solutions to beam hardening, projections-based and one-step decompositions. Section V is dedicated to various pre-processing and post-processing aspects, which are based on sinogram data and spectral CT images respectively, including data calibration, image denoising and artifacts correction, as well as image generation. Finally, Section VI covers key issues and future directions of learning-based spectral CT. The structure of this paper is outlined in Fig. 1.
Fig. 1.
Structure of this review paper, with the sections keyed to the main steps in the spectral CT imaging process.
Notations
Vectors (resp. matrices) are represented with bold lowercase (resp. uppercase characters). Images are represented as J-dimensional real-valued vectors which can be reshaped in 2-dimensional (2-D) or 3-dimensional (3-D) objects, where is the number of image pixels or voxels. is the number of rays per energy bin. ‘’ is the matrix transposition symbol. A NN is represented by a bold calligraphic upper case character with a subscript representing the weights to be trained, e.g., . is the semi-norm defined for all as , where denotes the cardinal of set , and , the -norm. For a positive-definite matrix , is the weighted-norm defined for all as , and denotes the Frobenius norm. is the horizontal concatenation of two column vectors and with the same length. denotes an ordered collection of vectors where the number of elements depends on the context. denotes a loss function that evaluates the adequation between 2 vectors, e.g, (negative Poisson log-likelihood), or . is a regularisation functional.
II. DECT and PCCT Systems
The first attempt to differentiate materials using CT with multiple X-ray energy spectra was made in the 1970s [8]. Since then, technologies in spectral CT have been continuously evolving. Traditional DECT and spectrally-resolving PCCT are the two specific forms of spectral CT that are both commercially available. The former uses a minimum of two separate X-ray energy spectra to differentiate two basis materials with different attenuation properties at various energy levels, while the latter usually involves the advanced detector technology known as energy resolving photon-counting detectors (PCD), which resolves spectral information of X-ray photons in two or more energy bins emitted from a polychromatic X-ray source. DECT overcomes several limitations of single energy spectrum CT and has achieved clinical acceptance and widespread applications. In the following, several types of DECT are briefly described. We will not cover all technologies, but we will focus on those that are currently representative. The interested readers may refer to [9]–[18] for more details and comparisons.
Sequential acquisition is perhaps the most straightforward DECT imaging approach. It performs two consecutive or subsequent scans of the same anatomy using an X-ray source operated at a low-peak kilovoltage (kVp) setting and then a high-kVp setting. The approach requires no hardware modification, but may suffer from image mis-registration due to motion artifacts from the delay between low- and high-kVp scans. Advanced DECT technologies all utilize specific hardware to mitigate the misregistration problem and shorten the data acquisition time.
The dual-source DECT scanner was first introduced in 2005 [19], which is featured by two source-detector systems orthogonally arranged in the same gantry to acquire the low- and the high-energy scan simultaneously. Although the 90-degree phase shift between the two scans creates a slight temporal offset, the two X-ray sources can select independent X-ray energy spectra to optimize the spectral separation for material differentiation in the data and/or image domains.
A dual-layer detector or a combination of two detector layers of scintillation material is also a good solution for DECT [20]–[23]. In this approach, low- and high-energy datasets are collected simultaneously by the two detector layers with perfect spatial alignment and excellent synchronicity. This advantage simplifies direct data-domain material decomposition.
Fast kVp-switching DECT is yet another technology that uses a highly specialized X-ray generator that can rapidly witch the tube voltage between low- and high-kVp settings during data acquisition. The first commercially available fast kVp-switching DECT scanner (GE Discovery CT750 HD) is capable of changing the tube voltage for each projection angle, so that each low- and high-kVp projection can be obtained almost simultaneously. The material decomposition can then be performed in the data domain. A similar design has been reported in [24] where the authors have utilized a linear accelerator as X-ray source to generate rapid switching electron pulses of 6 MeV and 9 MeV respectively. This has resulted in an experimental MeV DECT system that has been developed to perform cargo container inspection. Another type of fast kVp-switching DECT scanner has recently been introduced (Canon Aquilion ONE/PRISM) [25] that switches the tube voltage less frequently, allowing it to acquire the same energy from multiple successive projection angles. This design simplifies tube current modulation, making dose balancing at the two energy levels less complex. Along with the fast kVp-switching process, there is also a grating-based method that can help improve data acquisition [26]. In this method, an X-ray filter that combines absorption and filtering gratings is placed between the source and the patient. The gratings move relative to each other and are synchronized with the tube switching process to avoid spectral correlation. Simulation studies have shown improved spectral information with reduced motion-induced artifacts.
PCD technology plays an important role in PCCT imaging. PCDs requires a single layer of semiconductor sensor that converts X-ray photons directly into electrical signals. The main converter materials at present are cadmium zinc telluride (CZT) and Si. CZT is a material with a higher atomic number Z than Si and has a relatively high X-ray stopping power. Thus, the CZT-based PCD can have thin sensor layers of only a few millimeters, whereas Si-based detector lengths must be long enough to ensure good X-ray absorption. In one example of Si-based detector, the Si wafers are mounted sideways or edge-on against incoming X-rays to form a deep Si strip detector [27]. Therefore, building a full-area Si detector system can be more challenging. For imaging performance, both types of PCD have advantages and disadvantages in terms of signal quality as well as detection efficiency. More detailed comparisons can be found in [28], [29].
The innovation of PCD makes PCCT more attractive and offers unique advantages over conventional CT or DECT. These include improved dose efficiency by elimination of electronic noise, improved contrast-to-noise ratio (CNR) ratio through energy weighting [29]–[31], higher spatial resolution due to the small sub-millimeter PCD fabricated without any septa [29], [32], and most importantly, unprecedented material decomposition capabilities potentially for multi-tracer studies. Although PCCT is potentially more advantageous, it has to deal with technical challenges, including charge sharing and pile-up effects together with the need for substantial hardware and system research and development. Currently, the accessibility of PCCT for clinical applications is still limited.
III. Multi-Energy Image Reconstruction
Spectral CT, i.e., DECT and PCCT, offer the possibility to perform separate measurements, each measurement corresponding to an energy spectrum. One possibility is to reconstruct several attenuation CT images at different energies from these binned raw data. These images can then be used, e.g., for image-based material decomposition [33], [34] as illustrated in the top path of Fig. 1; more sophisticated method, in particular the one-step reconstruction of material images, will be discussed in Section IV.
The acquired projections usually suffer from low signal-to-noise ratio due to limited photons in each energy bin [35]. Moreover, practical constraints such as a reduced scanning time restrict CT systems to have a limited number of views. Therefore, the development of specific multi-energy reconstruction algorithms is of major importance.
This section reviews existing reconstruction algorithms for multi-energy CT reconstruction from energy-binned projection data, starting from conventional CT reconstruction algorithms to synergistic multi-energy CT reconstruction, with the incorporation of DL techniques and deep learning architectures. The methods presented here are only a subset of the literature in multichannel image reconstruction and we refer the readers to Arridge et al. [36] for an exhaustive review.
A. Forward and Inverse Problems
In this section, we briefly introduce a forward model that can be equally used for PCCT and DECT. We consider a standard discrete model used in model-based iterative reconstruction (MBIR).
The linear attenuation image takes the form of a spatially- and energy-dependent function , such that for all and for all , is the linear attenuation at position and energy . Standard CT systems perform measurements along a collection of rays where denotes the ray, , with , and being respectively the number of detector pixels and the number of source positions. For all , the expected signal (e.g. the number of photons in PCCT) is given by the Beer-Lambert law as
| (1) |
where ‘’ denotes the line integral along , is the corresponding X-ray photon flux which accounts for the source spectrum and the detector sensitivity (times the energy with energy integrating detectors) and is the background term (e.g., scatter, dark current).
In multi-energy CT (e.g., PCCT and DECT), the measurements are regrouped into energy bins ( for DECT and more for PCCT). For each bin , the expected number of detected X-ray photons is
| (2) |
where is the ray for bin , is the photon flux X-ray intensity for bin and is the background term. In PCCT each bin corresponds to an interval with , although may spillover the neighboring intervals. We assume that the number of detector pixels is equal to for each energy bin .
The forward model (2) applies to both PCCT and DECT. In PCCT, the detector records the deposited energy in each interaction and the energy binning is performed the same way for each ray so that is independent of the bin . In contrast, DECT systems (except dual-layer detectors) perform 2 independent acquisitions with 2 different photon flux X-ray intensity and , possibly at different source locations (i.e., via rapid kVp switching) so that the rays generally depend on .
One of the possible tasks in PCCT and DECT is to estimate a collection of attenuation CT images, i.e., one image per each of the binned measurements , . The energy-dependent image to reconstruct is sampled on a grid of pixels, assuming that can be decomposed on a basis of “pixel-functions” such that
| (3) |
where is the energy-dependent attenuation at pixel . The line integrals in Eq. (1) and Eq. (2) can be therefore rewritten as
| (4) |
with defined as and is the discretized energy-dependent attenuation, and we consider the following model which is an approximate version of Eq. (2)
| (5) |
where and for each the image is an “average” attenuation image corresponding to energy bin .
The reconstruction of each is achieved by “fitting” the expectation to the measurement , for example by solving the inverse problem
| (6) |
with respect to , where , , is the vector of the approximated line integrals. This can be achieved by using an analytical method such as filtered backprojection (FBP) [37], or by using an iterative technique [38], [39]. Unfortunately, the inverse problem (6) is ill-posed and direct inversion leads to noise amplification which is impractical for low-dose imaging. Moreover, the inversion relies on an idealized mathematical model that does not reflect the physics of the acquisition, especially by ignoring the polychromatic nature of the X-ray spectra.
B. Penalized Reconstruction
Alternatively, the reconstruction can be achieved for each energy bin by finding an estimate as the solution of an optimization problem of the form
| (7) |
where is a loss function (e.g., the Poisson negative log-likelihood for PCCT) that evaluates the goodness of fit between the data and , is a weight and is a penalty function or regularizer, generally convex and nonnegative, that promotes desired image properties while controlling the noise. The data fidelity term in (7) is convex when for all . Although many approaches were proposed to solve (7), most algorithms are somehow similar to the proximal gradient algorithm [40], [41], that is to say, given an image estimate at iteration , the next estimate is obtained via a reconstruction step followed by a smoothing step,
| (8) |
| (9) |
where is the gradient of the data fidelity loss evaluated at and is a suitable diagonal positive-definite matrix (typically, a diagonal majorizer of the Hessian of the data fidelity loss). The first step (8) is a gradient descent that guarantees a decrease of the data fidelity while the second step (9) is an image denoising operation. This type of approach encompasses optimization transfer techniques such as separable quadratic surrogate (SQS) [42], [43].
The choice of depends on the desired image properties. A popular choice consists in penalizing differences in the values of neighboring pixels with a smooth edge-preserving potential function and solving Eq. (9) is achieved with standard smooth optimization tools [42], [43]. Another popular choice is the compressed sensing (CS) approach, which has been widely used in medical imaging when using an undersampled measurement operator (e.g., sparse-view CT). CS consists of assuming that the signal to recover is sparse in some sense to recover it from far fewer samples than required by the Nyquist–Shannon sampling theorem. In the following paragraphs, we briefly discuss the synthesis and the analysis approaches.
In the synthesis approach, it is assumed that where is a dictionary matrix, i.e., an over-complete basis, consisting of atoms, and is a sparse vector of coefficients such that is represented by a fraction of columns of , or atoms. The reconstruction of the image is then given by
| (10) |
where can be either the semi-norm or its convex relaxation, the norm, and is a weight controlling the sparsity of . The optimization can be achieved by orthogonal matching pursuit [44] for and proximal gradient for . In some situations, imposing may be too restrictive and a relaxed constraint is often preferred. The reconstruction is then achieved by penalized reconstruction using a regulariser that prevents from deviating from , usually defined as
| (11) |
where is a weight. Solving Eq. (7) is achieved by alternating between minimization in (e.g., by performing several iterations of (8) and (9)) and minimization in (e.g., orthogonal matching pursuit [44] for and proximal gradient for ). This type of penalty forms the basis of learned penalties that we will address in Section III-D.
In the analysis (encoding) approach, it is assumed that is sparse, where is a sparsifying transform, and the penalty is
| (12) |
For example, in image processing, can be a wavelet transform or finite differences (discrete gradient). In the latter case and when , the corresponding penalty is referred to as total variation (TV)1. TV has been extensively used in image processing for its ability to represent piecewise constant objects [51]. Because is non-smooth, solving Eq. (9) requires variable splitting techniques such as proximal gradient, alternating direction method of multipliers (ADMM) [52] or the Chambolle-Pock algorithm [53].
C. Synergistic Penalties
Alternatively, the images can be simultaneously reconstructed. Introducing the spectral CT multichannel image, the binned projection data and the expected binned projections, the images can be simultaneously reconstructed as
| (13) |
where is a synergistic penalty function that promotes structural and/or functional dependencies between the multiple images and a proximal gradient algorithm to solve Eq. (13) at iteration to update is
| (14) |
| (15) |
where Eq. (15) corresponds to a synergistic smoothing step. The paradigm shift here is that allowing the channels to “talk to each other” can reduce the noise as each channel participates in the reconstruction of all the other ones. In the context of spectral CT, this suggests that the reconstruction of each image benefits from the entire measurement data . Here, we present a non-exhaustive list of existing approaches.
One class of approaches consists of enforcing structural similarities between the channels. Examples include joint total variation (JTV) which encourages gradient-sparse solutions (in the same way as the conventional TV) and also encourages joint sparsity of the gradients [54], [55]. Total nuclear variation encourages common edge locations and a shared gradient direction among image channels [56], [57]. All these works reported improved image quality with synergistic image processing as compared with single-image processing.
A second class of approaches consists of promoting similarities across channels by controlling the rank of the multichannel image. Given that the energy dependence of human tissues can be represented by the linear combination of two materials only (see Section IV), it is natural to expect a low rank in some sense in the spectral dimension. For dynamic CT imaging, Gao et al. [58] proposed a method, namely Robust Principle Component Analysis based 4-D CT (RPCA-4DCT), based on a low-rank (LR) + sparse decomposition of the multichannel image matrix ( time frames),
| (16) |
where is an LR matrix representing the information that is repeated across the channels and is a sparse matrix representing the variations in the form of outliers, and a synergistic penalty defined as
| (17) |
and the nuclear norm is a relaxation of the rank of a matrix, and showed that their approach outperforms TV-based (in both spatial and temporal dimensions) regularization. Gao et al. [59] then generalized this method for spectral CT with the prior rank intensity and sparsity model, which uses the rank of a tight-frame transform of the LR matrix to better characterize the multi-level and multi-filtered image coherence across the energy spectrum, in combination with energy-dependent intensity information, and showed their method outperformed conventional LR + sparse decomposition. This principle was further generalized by “folding” the multichannel image in a 3-way tensor (for 2-D imaging) and applying the generalized tensor nuclear norm regularizer to exploit structural redundancies across spatial dimensions (in addition to the spectral dimension) [60]–[65].
A third and different class of approaches consists of enforcing structural similarities of each with a reference low-noise high-resolution image , generally taken as the reconstruction from all combined energy bins. Instead of using a joint penalty , each channel is controlled by a penalty of the form
| (18) |
where is a “similarity measure” between and the reference image . The prior image-constrained compressed sensing (PICCS) [66], [67] approach uses , denoting the discrete gradient; the -norm can also be replaced with the semi-norm [68]. Variants of this approach include nonlocal similarity measures [69], [70] to preserve both high- and low-frequency components. More recently, Cueva et al. [71] proposed the directional TV approach for spectral CT, which enforces colinearity between the gradients of and , while preserving sparsity, and showed their approach outperforms TV.
To conclude, spectral CT reconstruction with synergistic penalties has been widely used to improve the quality of the reconstructed images. However, the success of this approach heavily depends on the selection of an appropriate synergistic penalty term, which is typically fixed and may not always accurately reflect the true underlying structure of the data.
D. Learned Penalties
Traditional regularization methods, such as those described in Sections III-B and III-C, impose a fixed handcrafted penalty on the reconstructed image based on certain assumptions about its structure, such as sparsity or smoothness. However, these assumptions may not always hold in practice, leading to suboptimal reconstructions. Learned penalty functions, on the other hand, can adaptively adjust the penalty term based on the specific characteristics of the data, allowing for more accurate and flexible reconstruction.
This subsection discusses learned synergistic penalties for multichannel image reconstruction. In particular, we will focus on penalties based on a generator , which is a trained mapping that takes as input a latent variable , which can be an image or a code, and returns a plausible multichannel image . The latent variable represents the patient which connects the different channels. The penalty function plays the role of a discriminator by promoting images originating from the generative model and by penalizing images that deviate from it, in a similar fashion to the relaxed synthesis model (11).
Most of this subsection will address DL, i.e., for some dictionary matrix , as it is the most prevalent learned penalty used in synergistic multichannel image reconstruction. Convolutional dictionary learning (CDL) will also be discussed in a short paragraph. Finally, we will discuss recent work that uses deep NN models.
In this subsection denotes a random spectral CT image whose joint distribution corresponds to the empirical distribution derived from a training dataset of spectral CT images , that is to say for all mapping ,
| (19) |
1). Dictionary Learning:
For simplicity this section will consider 2-D imaging (i.e., ), so that each image can be reshaped into a square matrix.
DL is a popular technique for regularizing the reconstruction process in medical imaging and especially in CT reconstruction [72]–[75]. The basic idea behind DL is to learn a dictionary matrix that can represent the image with a fraction of its columns. The dictionary operator requires a large number of atoms to accurately represent all possible images which increase the computational complexity of training. Therefore, to reduce the complexity, the image is generally split into smaller d-dimensional “patches” (possibly overlapping) with . For a given energy bin , the trained penalty to reconstruct a single attenuation image by penalized reconstruction (7) is given by
| (20) |
where is the trained dictionary matrix, is the patch extractor and each is the sparse vector of coefficients to represent the patch with . The training is generally performed by minimizing with respect to (with unit -norm constraints on its columns) over a training data set of high-quality images,
| (21) |
for example using the K-SVD algorithm introduced by Aharon et al. [76].
DL can also be used to represent images synergistically. Tensor dictionary learning (TDL) consists in folding the spectral images into a tensor and in training a spatio-spectral tensor dictionary to sparsely represent with a sparse core tensor , such that each atom conveys information across the spectral dimension. A common approach used to sparsely represent the sensor image is to use the Tucker decomposition [77], [78]. It was utilized in multispectral image denoising [79], [80] as well as in dynamic CT [81] (by replacing the spectral dimension by the temporal dimension). Denoting the spatio-spectral image patch extractor, each patch can be approximated by the Tucker decomposition as
| (22) |
where is the core tensor for the patch, and are the 2-D spatial dictionaries along each dimension and is the spectral dictionary (all of them consisting of orthogonal unit column vectors), and is the mode- tensor/matrix product (see for example Semerci et al. [61] for a definition of tensor-matrix product).
The Tucker decomposition requires a large number of atoms and therefore is cumbersome for DL in high dimensions. To remedy this, Zhang et al. [82] proposed to use the canonical polyadic decomposition (CPD), which consists of assuming that the core tensor is diagonal, i.e., and , which leads to the following approximation [78],
| (23) |
where for all , and are unit vectors, is a sparse vector corresponding to the diagonal of and ‘’ denotes the matrix outer product. Zhang et al. then used this decomposition to train spatio-spectral dictionaries combined with a K-CPD algorithm [83] from which the following penalty term is derived2:
| (24) |
with . The training is performed as
| (25) |
where is the spatio-spectral tensor obtained by folding the training multichannel image matrix , and the minimization is performed subject to the constraint . Wu et al. [84] proposed a similar approach with the addition of the semi-norm of the gradient images at each energy bin in order to enforce piecewise smoothness of the images, while Li et al. [85] added a PICCS-like penalty (18) to enforce joint sparsity of the gradients.
We can observe that the TDL regularizer with CPD can be rewritten as
| (26) |
where each column of is the matrix reshaped into a vector. This regularizer is a generalization of (20) to multichannel imaging with a collection of dictionaries and a unique sparse code for all energy bins . Similar representations were used in coupled DL in multimodal imaging synergistic reconstruction, such as in PET/magnetic resonance imaging (MRI) [86], [87], multi-contrast MRI [88] as well as super-resolution [89].
Patch-based DL may be inefficient as the atoms are shift-variant and may produce atoms that are shifted versions of each other. Moreover, using many neighboring/overlapping patches across the training images is not efficient in terms of sparse representation as sparsification is performed on each patch separately. Instead, CDL [90]–[92] consists in utilizing a trained dictionary of image filters to represent the image as a linear combination of sparse feature images convolved with the filters (synthesis model) that can be used in a penalty function similar to Eq. (20), without patch extraction. Bao et al. [93] used this approach for CT MBIR. Alternatively, convolutional analysis operator learning (CAOL) consists in training sparsifying convolutions, which can then be used as a penalty function for MBIR [94]. There are a few applications of CDL and CAOL in multichannel imaging and multi-energy CT (see [95] for a review). Degraux et al. [96] proposed a multichannel CDL model to represent two images simultaneously (intensity-depth imaging), using a collection of pairs of image filters. Gao et al. [97] proposed a more general model with common and unique filters. More recently, Perelli et al. [98] proposed a multichannel CAOL for DECT joint reconstruction, which uses pairs of image filters to jointly sparsify the low- and high-energy images, and demonstrated their method outperforms JTV-based synergistic reconstruction.
2). Deep-Learned Penalties:
The synthesis model used in DL can be generalized by replacing the multichannel dictionaries with a trained multi-branch which maps a single input to a collection of images designed to represent the spectral CT image . Unlike dictionary learning, which uses a finite number of atoms to represent the data, deep NNs can learn parameters that can capture more intricate patterns and structures in the image data. A synergistic regularizer used in Eq. (7) can then be defined as
| (27) |
where is a penalty function for (not necessarily sparsity-promoting), which is the generalization of multichannel DL (26) using multiple NNs. Wang et al. [99] used this approach with a collection of U-nets trained in a supervised way to map the attenuation image at the lowest energy bin to the attenuation image at energy bin , i.e.,
| (28) |
and combined a standard Huber penalty (the function in Eq. (27)) for . The trained penalty “connects” the channels by a spectral image such that each originates from a single image that is smooth in the sense of . Wang et al. reported substantial noise reduction as compared with individually reconstructed images and JTV synergistic reconstruction.
The training of the generative model can also be unsupervised, for example as a multichannel auto-encoder (AE), i.e.,
| (29) |
where is a multichannel encoder, i.e., that encodes a collection of images into a single latent vector, parametrized with . In this approach, is encouraged not to deviate from the “manifold” of plausible images . Pinton et al. [100] and Gautier et al. [101] used this approach respectively for PET/CT and PET/MRI using a multi-branch variational AE, and reported considerable noise reduction by reconstructing the images synergistically as opposed to reconstructing the images individually. A patched-based version of this penalty with a K-Sparse AE (i.e., with ) was proposed by Wu et al. [102] for single-channel CT. Duff et al. [103] proposed a similar approach with a Wasserstein generative adversarial network (W-GAN).
An alternative approach, namely the deep image prior introduced by Ulyanov et al. [104], consist of fixing the input and to optimize with respect to , in such a way that the reconstruction does not require pre-training of the NN. A multichannel version of this approach using a multi-branch NN with a single input was proposed for DECT [105].
Although deep-learned penalties have been successfully applied in image reconstruction, their application to spectral CT has been relatively limited and remains an active area of research. Future work should focus on developing more efficient and accurate deep-learned penalties that are specifically tailored to the unique challenges and opportunities of spectral CT.
E. Deep Learning-based Reconstruction
Another paradigm shift has been the development of end-to-end learning architectures that directly map the raw projection data to the reconstructed images. This approach, known as learned reconstruction, has two main categories: direct reconstruction and unrolling techniques. Direct reconstruction involves training a single NN to perform the reconstruction task, while unrolling techniques aim to mimic the iterative algorithm by “unrolling” its iterations into layers. These techniques have shown great potential in image reconstruction, where the acquisition of data at different energy levels provides additional information about the material composition of the imaged object. In this section, we review recent advances of unrolling-based architectures for image reconstruction and their extension to synergistic spectral CT reconstruction. Direct methods have not yet been deployed for spectral CT and will be discussed in Section VI.
In the following denotes a random spectral CT image/binned projections pair whose joint distribution corresponds to the empirical distribution derived from training pairs such that for all the spectral CT multichannel image is reconstructed from .
Unrolling techniques, or learned iterative schemes, have become increasingly popular for image reconstruction in recent years, due to their ability to leverage the flexibility and scalability of deep neural networks while retaining the interpretability and adaptability of classical iterative methods. Unrolling-based techniques aim at finding a deep architecture that approximates an iterative algorithm.
For all energy bins , the iteration of an algorithm to reconstruct the image can be written as
| (30) |
where is an image-to-image mapping that intrinsically depends on and that updates the image at layer to layer . The parameter typically comprises algorithm hyperparameters such as step lengths and penalty weights but also NN weights. For example, Eq. (8) and Eq. (9) are unrolled with where and . The Q-layer reconstruction architecture , , to reconstruct from is given as
| (31) |
where is a given initial image and the right-hand side depends on by means of of , and the trained parameter is obtained by supervised training as
| (32) |
Alternative to Eq. (30) and (31), for example incorporating memory from previous iterates at each layer, can be found in Arridge et al. [106]. By utilizing components of iterative algorithms such as the backprojector , unrolling-based architectures can map projection data to images without suffering from scaling issues. Many works from the literature derived unrolling architecture from existing model-based algorithms and we will only cite a non-exhaustive list; we refer the reader to Monga et al. [107] for a review of unrolling techniques until 2021. One of the first unrolling architectures, namely ADMM-net, was proposed by Yang et al. [108] for CS MRI and consists in a modified ADMM algorithm [52] where basics operation (finite-difference operator, soft-thresholding, etc.) are replaced by transformations such as convolution layers with parameters that are trained end-to-end. Other works rapidly followed for regularized inverse problems in general and image reconstruction in particular. Learned proximal operators, which consist of replacing the update (9) with a trainable convolutional neural network (CNN) [109], [110]. In a similar fashion, Chun et al., proposed BCD-Net [111] and its accelerated version Momentum-Net [112] which consists in unrolling a variable-splitting algorithm and replace the image regularization step with a CNN. Adler et al. [113] proposed a trainable unrolled version of the primal-dual (Chambolle-Pock) algorithm [53].
A synergistic reconstruction algorithm such as given by Eq. (14) and Eq. (15) may also be unrolled in a trainable deep multi-branch architecture by merging the mappings at each layer into a single multichannel mapping that depends on the entire binned projection dataset and on some parameter . The update from layer to layer is given by
| (33) |
where the mapping utilizes the entire data and updates the images simultaneously, thus allowing the information to pass between channels. For example, the layer corresponding to Eq. (14) and Eq. (15) is with and . The corresponding Q-layer reconstruction architecture , is given by
| (34) |
for some initialization , and the trained parameter is obtained by supervised training similar to Eq. (32) but using the data at all energy bins simultaneously:
| (35) |
A simplified representation of this architecture is shown in Fig. 2.
Fig. 2.
Representation of the synergistic unrolling architecture described in Eq. (33) with 3 energy bins k =1,2,3.
At the time we are writing this paper, very few research addressed synergistic reconstruction using unrolling-based architectures. We can cite the recent work SOUL-Net by Chen et al. [114] which proposes an ADMM-based architecture to solve the joint problem (13) with the nuclear norm (for LR penalty, cf. Section III-C) and TV. Chen et al. modified the singular value thresholding step for nuclear norm minimization by adding a ReLu function with trainable parameters, and replaced the TV minimization with a CNN combined with an attention-based network. They showed that their method outperforms “conventional” LR + sparse decomposition methods.
Unrolling techniques have shown great promise as a flexible and powerful tool for single-channel image reconstruction. Although these techniques have been applied successfully to a variety of imaging modalities, their application to multichannel synergistic reconstruction in spectral CT remains relatively limited and challenging, due to the high-dimensional nature of the data and the need for accurate modeling of the spectral correlations. However, unrolling techniques have been proposed for projection-based and one-step material decomposition, see Section IV.
IV. Material Decomposition
Spectral CT techniques such as DECT and PCCT are often used to characterize the materials of the scanned patient or object by decomposing the linear attenuation coefficient into material images. This process of material decomposition is based on the assumption that the energy dependence of the linear attenuation coefficient in each pixel can be expressed as a linear combination of a small number of basis functions [115]. The linear attenuation can then be modeled as
| (36) |
where represents the energy-dependent basis function and is the material image. These basis functions describe physical effects such as photoelectric absorption and Compton scattering [115] or the linear attenuation coefficients of representative materials of the scanned object such as water and bone for patients. With this model, two basis functions are sufficient to describe the variations of the linear attenuation coefficients of human tissues with energy [116]–[118]. One or more basis function(s) may also be used to represent a specific contrast agent, e.g., a material with a K-edge discontinuity in its attenuation coefficient in the range of diagnostic energies (30–140 keV) [119]. The material images can be represented in the discrete domain as a vector using the pixel basis functions (see Eq. (3)) with each pixel of the unknown image decomposed into the chosen material basis. The discrete object model for the basis decomposition is then
| (37) |
where is the weight of the basis function in the pixel. Injecting (37) into (2) links the material decomposition to the expected value (e.g. the number of detected X-ray photons for PCCT)
| (38) |
where . Material decomposition aims at estimating the decomposed CT images by matching the expected values , , with the measurements with different efficient spectra .
This problem is the combination of two sub-problems: tomographic reconstruction and spectral unmixing. The two problems can be solved sequentially or jointly and most techniques of the literature fall into one of the following categories: image-based, projection-based or one-step material decomposition.
A. Image-based Material Decomposition
Image-based algorithms decompose the multichannel CT image into material images . While each channel is often obtained by direct methods such as FBP, an alternative procedure is the reconstruction of each channel from by solving the MBIR problem in Eq. (7) or the joint reconstruction of from by solving the synergistic MBIR problem in Eq. (13). The discretized version of Eq. (37) is
| (39) |
with and the energy of the attenuation image . The images may be decomposed by solving in each pixel the linear inverse problem
| (40) |
where , , is the same matrix for all voxels characterizing the image-based decomposition problem. It is generally calibrated with spectral CT images of objects of known attenuation coefficients. Given that and are small, the pseudo-inverse of can be easily computed and applied quickly after the tomographic reconstruction of . Image-based material decomposition faces two challenges: (1) the spectral CT images are affected by higher noise than conventional CT (if the same total dose is split across energy bins) which will be enhanced by the poor conditioning of and (2) the spectral CT images will suffer from beam-hardening artifacts since the efficient spectra are not truly monochromatic in most cases, i.e., is actually voxel and object dependent.
Machine learning algorithms have been used for image-based decomposition to mitigate noise and beam-hardening artifacts. Some techniques learn an adequate regularization [120]–[125] while using the linear model in Eq. (40). These techniques are similar in essence to those described in Section III-D1 except that dictionary learning uses decomposed images for spatially regularizing the decomposed images.
NNs may be used instead to improve the linear model in Eq. (40) [126]. As in many other fields of research on image processing, deep CNNs have demonstrated their ability to solve image-based decomposition with a more satisfactory solution than the one produced by a pixel-by-pixel approach. Several deep learning architectures, previously designed to solve other image processing tasks, have been deployed for image-based decomposition. Most works are based on a supervised learning approach where a dataset of manually segmented basis material images are available: fully convolutional network [127], U-Net [128]–[133], Butterfly-Net [134], visual geometry group [132], [135], Incept-net [136], [137], generative adversarial network (GAN) [138], Dense-net [139]. These contributions differ on the type of architecture adopted and the complexity of the network which is measured by the number of trainable parameters. They also differ in which inputs are used by the network, e.g., reconstructed multichannel CT images [133] or pre-decomposed CT images [131]. The network output is generally the decomposed CT images but it may also be other images, e.g., the elemental composition [132], quantities used for radiotherapy planning such as the image of the electron density [140] or the virtual non-calcium image [137].
B. Projection-based Material Decomposition
The main limitation of image-based approaches is that the input multichannel CT image is generally flawed by beam hardening. If several energy measurements are available for the same ray ( for all ), with a dual-layer DECT or a PCCT, an alternative approach is projection-based decomposition [115], [119] which aims at estimating projections , of the decomposed CT images ,
| (41) |
from the measurements given the forward model
| (42) |
where and . In this context, the expected value becomes a function of (or = ) instead of . Given the decomposed projections , the images are obtained by solving the following inverse problem
| (43) |
where multichannel reconstruction algorithm, e.g. those described in Sections III-B and III-C can be deployed to reconstruct from .
Similar to image-based decomposition, projection-based decomposition can be solved pixel by pixel in the projection domain by solving
| (44) |
The number of inputs and unknowns is the same for each projection pixel, but it is more complex because the exponential in Eq. (42) induces a non-linear relationship between and . Moreover, this inverse problem (44) is non-convex [141] (unless, obviously, if the exponential is linearized) and fully-connected NNs have been used to solve it [142], [143]. Such networks can also be used to process input data for spectral distortions before material decomposition [144] or to modify the model described by Eq. (42) to account for pixel-to-pixel variations [145] or pulse pile-up [146].
However, these approaches cannot reduce noise compared to conventional estimation of most likely solutions [119] without accounting for spatial variations. The idea of spatially regularizing pixel-based material decomposition has first been investigated with variational approaches [147], [148] solving
| (45) |
As in image-based algorithms, DL [149], [150] has been investigated to improve the spatial regularization as well as CNNs to learn features of the projections with U-Net [129], [130], ResUnet [151], stacked auto-encoder [152], perceptron [153], GAN [154] and ensemble learning [155], [156].
A promising alternative to these supervised techniques, which are learning the physical model from the data, is to solve (45) by combining iterative reconstruction with learning algorithms in so-called learned gradient-descent using unrolling algorithms [157] detailed in Section III-E. Other approaches such as proposed by Zhang et al. [158] combine multiple NNs both for learning the material decomposition in the projection domain with an additional refinement network in the image domain to enhance the reconstructed image quality.
C. One-step Material Decomposition
One limitation of projection-based decomposition is that some statistical information is lost in decomposed projections which could be useful to reconstruct the most likely material maps . The noise correlations between the decomposed sinograms may be accounted for in the subsequent tomographic reconstruction [159], [160] but it cannot fully characterize the noise of the measurements , in particular with more than two energy bins (). Several groups have investigated an alternative solution combining material decomposition and tomographic reconstruction in a one-step algorithm which reconstructs the material maps from the measurements by solving the optimization problem
| (46) |
Compared to Eq. (7), solving (46) is a far more difficult problem, similar to projection-based algorithms but with a larger number of unknowns and inputs . Several iterative solutions have been proposed to address this problem by optimizing the most likely material maps given the measurements with spatial regularization. One of the main differences between these algorithms is the optimization algorithm, from non-linear conjugate gradient [161] to SQS algorithms [162]–[164] and primal-dual algorithms [165], [166].
The nature of this problem is such that all algorithms based on machine learning have used part of the physical model in their architecture. Generally, combining physics knowledge and deep learning for material decomposition is implemented through unrolling methods [167] (Section III-E). Eguizabal et al. [168] adapted the projection-based unrolling algorithm of [157] to one-step reconstruction. The same group has used machine learning to improve the physical model in Eq. (38) by modeling charge sharing [169]. Another approach is to insert a backprojection step into the network architecture, i.e. the adjoint of the line integral operator in Eq. (38), to account for this knowledge in the network architecture [170], [171]. Finally, machine learning may be used at each iteration for denoising the images, e.g. with a dictionary approach [172]. A self-supervised approach named Noise2Noise prior [173], which does not require manually segmented ground truth materials images, has been applied to one-step decomposition using a training dataset consisting of sinograms paired with their noisy counterpart obtained by sinogram splitting.
The different approaches for material decomposition differ on many levels, from computational cost to the accuracy of the decomposed images. For example, Abascal et al. [129] compared projection-based and image-based algorithms using variational approaches and machine learning. They observed the best image quality with an image-based material decomposition approach, as illustrated in Fig. 3. However, the recent Grand Challenge on Deep-Learning spectral Computed Tomography [174] demonstrated that many different approaches are still under investigation. Nine out of the ten best scorers used machine learning and most combined it with a model of the DECT acquisition. The development of such algorithms in clinical scanners will depend on both their practicality, e.g. the computational time, and the accuracy of the material decomposition of real data.
Fig. 3.
Material decomposition of simulated PCCT acquisitions of a patient phantom (left) with projection-based (middle) and image-based (right) U-Net CNNs. The two materials of the decomposition are soft tissue (top row) and bone (bottom row). Figure adapted from Abascal et al. [129] and distributed under a Creative Commons Attribution 4.0 License, see https://creativecommons.org/licenses/by/4.0/.
V. Data Pre-Processing and Image Post-Processing
CT technology has been the front-line imaging tool in emergency rooms due to its fast, non-invasive, and high-resolution features, with millions of scans performed annually worldwide. However, due to the increased cancer incidence from radiation exposure, “as low as reasonably achievable” is the central principle to follow in radiology practice. Recent advances in CT technology and deep learning techniques have led to great developments in reducing radiation doses in CT scans [175]. For example, aided by deep learning techniques, much progress has been made in low-dose or few-view CT reconstruction without sacrificing significant image quality. Furthermore, the use of DECT technology allows further cuts in radiation dose by replacing previous non-contrast CT scans with virtual unenhanced images in clinical practice [176].
While many prior-regularized iterative reconstruction techniques described in Section III inherently suppress noise and artifact, network-based post-processing techniques are also popular for removing noise and artifacts from already reconstructed low-dose spectral images and are covered here. Moreover, PCCT with PCDs is widely viewed as a comprehensive upgrade to DECT since it produces less noise, better spectral separation, and higher spatial resolution while requiring less radiation dose [29], [30]. However, the PCD often experiences increased nonuniformity and spectral distortion due to chargesharing and pulse pile-up effects compared to the traditional energy-integrating detectors (EID), and the correction of these imperfections in PCD images is included here. Finally, we also review deep learning techniques that enhance clinical diagnosis with spectral CT, which includes virtual monoenergetic image synthesis, virtual noncontrast image generation, iodine dose reduction, virtual calcium suppression, and other applications. The overview of this section is summarized in Fig. 4.
Fig. 4.
Overview of sub-topics in Section V. The data pre-processing section covers deep correction methods for spectral distortion (e.g., falsely increased counts in the low energy bin due to the charge sharing effect, and non-linear responses due to the pulse pileup effect) and non-uniformity in PCD projection images. The image post-processing sections discuss deep post-processing methods to enhance DECT and PCCT imaging and their impacts on clinical diagnosis.
A. PCCT Data Pre-processing
PCDs offer much smaller pixel size compared to EIDs and also possess energy discrimination ability that can greatly enhance CT imaging with significantly higher spatial and spectral resolution. However, PCD measurements are often distorted by undesired charge sharing and pulse pileup effects, which can limit the accuracy of attenuation values and material decomposition. Since accurately modeling these effects is highly complex, deep learning methods are being actively explored for distortion correction in a data driven manner. The initial trial is introduced in Touch et al. [144] where a simple fully-connected NN with two hidden layers of five neurons each was adopted mainly for charge sharing correction. Later the same network structure but with more neurons was used by Feng et al. [177] to compensate pulse pileup distortion, and similarly in [178], [179] for spectral distortion correction. A large CNN model was first introduced in Li et al. [180] to leverage inter-pixel information for both corrections of charge sharing and pulse pileup effects. The model included a dedicated generator with a pixel-wise fully-connected subnetwork for intra-pixel distortion caused by pulse pileup and a convolutional sub-network for inter-pixel cross-talk correction, and was trained using the W-GAN framework for spectral correction. More recently, Holbrook et al. [181] used multienergy CT scans with an EID to calibrate the PCD spectral distortion, and adopted a U-Net to map the distorted PCD projections into monochromatic projections generated by multienergy CT scans after material decomposition. Ma et al. [182] introduced CNN-LSTM to correct pulse pileup distortion in Xray source spectrum measurements, while Smith et al. [183] used a spatial-temporal CNN for charge sharing compensation.
There are also several interesting studies on artifact correction for PCCT using deep learning methods. Erath et al. [184] utilized a U-Net for scatter correction in the sinogram domain to compensate for the Moire artifacts caused by coarsé anti-scatter grids relative to the small detector pixel size, resulting in improved image quality and HU value accuracy. Due to the complexity of PCDs, their pixels tend to suffer more nonuniformity due to detector imperfections compared to EIDs, making the ring artifact issues more prominent in PCCT. To address this issue, Hein et al. [185] trained a U-Net with the perceptual loss for the correction of ring artifacts caused by pixel nonuniformity [186], while Fang et al. [187] used two U-Nets in both projection domain and image domain for ring artifacts removal.
B. Image Post-processing
1). Image Denoising:
In CT imaging, it is important to limit the radiation dose to patients, but reducing the dose often gives rise to image noise, which can strain radiologists’ interpretation. To address this issue, various image denoising methods have been developed that aim to recover a clean version from a noisy image by leveraging prior knowledge of the image to maintain sufficient image quality for clinical evaluation,
| (47) |
The development of CT noise reduction techniques has a long history with its root dating back shortly after the invention of CT. While our focus is on deep learning and spectral CT, it is important to briefly cover classic post-processing denoising techniques and deep learning techniques for single energy CT, as they can still be applied to spectral CT in a channel-by-channel manner. We will then dive into recent trends of self-supervised learning deep denoising methods, as well as deep methods that incorporate the correlations between energy channels.
Spatial filtering methods leverage the statistical nature of noise fluctuations and are achieved through local averaging or nonlocal averaging means [188]–[190]; optimization-based denoising methods, on the other hand, incorporate image model preassumptions such as domain sparsity, piecewise linearity, or gradient smoothness as regularization. Some well-known methods in this category include TV [191], DL [72], [192], wavelet based denoising [193], block-matching and 3-D filtering (BM3D) [194], and others. A good discussion of these classic denoising techniques is provided by Diwakar et al. in their review paper [195]. Different from the explicitly defined prior knowledge in traditional methods, the development of deep learning techniques, particularly CNNs, provides a data-driven approach to learn the implicit distribution knowledge from large amounts of images, offering a one-step solution to the denoising problem (Eq. (47)), i.e.,
| (48) |
where denotes the network function with optimized parameters after training. Since they are way more powerful than the traditional methods, deep methods will soon dominate the research field of CT image denoising. Initially, these methods were primarily trained in a supervised fashion using paired noisy and clean images, as generally depicted by Eq. (49), and the successful examples include REDCNN [196], wavelet network [197] and stacked competitive network [198].
| (49) |
where denotes a general loss function for network training, is the clean image corresponding to the noisy one and the expectation is taken over pairs from the training dataset. Following the idea, various network structures and loss functions have been explored. Representative network structures include U-Net [199]–[201], DenseNet [202], GAN [203]– [206], ResNet [207], [208], Residual dense network [209], Quadratic neural network [210], transformer [211], diffusion model [212], and more. Commonly used loss functions include mean squared error (MSE), mean absolute error, structural similarity index [213]–[215], adversarial loss [203], [216], TV loss [217], [218], perceptual loss [203], [219], edge incoherence [220], identity loss [206], [221], [222], projection loss [215], and more. For more detailed information, we refer readers to the latest two review papers on low-dose CT denoising [223], [224].
The issue of missing paired labels was soon realized when researchers attempted to apply supervised methods in practice. To address this, a number of unsupervised or self-supervised methods have been proposed. For instance, cycle-GAN based techniques are able to utilize unpaired data for training by promoting cycle consistency between domains [205], [206], [222], [225]. However, these GAN-based methods have been criticized for potentially generating erroneous structures. Poisson Unbiased Risk Estimator (PURE) and Weighted Stein’s Unbiased Risk Estimator (WSURE) are alternative methods that convert the supervised MSE loss calculation into a form that only relies on the noisy input, the network output, and its divergence [226]. This approach forms an unsupervised training framework where the divergence term is approximated using Monte-Carlo perturbation method [227]. Noise2Noise is another method that enables us to train the network with paired noise-noise images which are equivalent to being trained with original noise-clean pairs,
| (50) |
where and are different noisy realizations of the same image, e.g., two independent CT scans of the same object. Building on this idea, several recent variant methods have been developed for self-supervised low-dose CT denoising by generating noisy pairs via various approaches [228]–[236]. For instance, Noise2Inverse proposes to partition projection data into several sets and enforcing consistency between corresponding reconstructions [234], while Noise2Context promotes similarity between adjacent CT slices in 3-D thin-layer CT [232]; Half2Half adopts the thinning technique [237] to split a full dose real CT scan into two pseudo half dose scans [230].
Spectral CT powerfully extends the conventional single energy CT by introducing an extra energy dimension. However, the splitting of photons into different energy bins increases the noise level of the projection at each bin compared to conventional CT with the same overall radiation dose. Therefore, to achieve optimal denoising performance for spectral CT, it is necessary to leverage inter-bin information, similar to the approach taken in learned synergistic reconstruction (Section III-E), as described below,
| (51) |
Several recent papers have explored this direction. ULTRA [238] incorporates an -norm and anisotropic total variation loss to train a residual U-Net with multichannel inputs from PCCT scans. Noise2Sim [235] constructs noisy pairs using the Noise2Noise principle and replaces each pixel from the original noisy image with one of its -nearest pixels searched from the spatial dimension (including adjacent slices) and measured by non-local means. The multichannel image is fed to the network as a whole, and its value from different bins can be constructed independently to fully leverage the self-similarities within the spectral CT scans. By this means, comparable or even better performance has been demonstrated on experimental PCCT scans against the supervised learning methods. S2MS [231] proposes another interesting approach to leverage the inter-channel information by converting the linear attenuation map from each channel to a channel-independent density map, which forms different noisy realizations of the density images from multiple channels. Promising results from this self-supervised learning idea are demonstrated on a simulation study.
Besides developing various deep denoising methods, researchers have also investigated the effects of noise reduction on the downstream tasks [238], [239]. For example, Evans et al. [239] compared the material decomposition results of multi-bin PCCT images before and after denoising with BM3D and Noise2Sim through phantom studies. They found that image denoising improves the accuracy of material concentration quantification results, but not material classification results. In the clinical domain, there are several Food and Drug Administration (FDA)-approved deep denoising methods from multiple vendors (e.g., the TrueFidelity from GE Healthcare, the Advanced Intelligent Clear-IQ Engine (AiCE) from Canon, PixelShine from Algomedica, ClariCT.AI from ClariPI Inc., etc), and numerous studies have been performed to investigate their impacts on clinical significance. For ease of notation, we use deep learning image reconstruction (DLIR) to refer specially to these FDA-approved methods in clinical applications. Noda et al. [240] showed that with DLIR, the radiation dose of whole-body CT can be reduced by up to 75% while maintaining similar image quality and lesion detection rate compared to standard-dose CT reconstruction with iterative reconstruction through a study cohort of 59 patients. This conclusion is also supported in other studies where DLIR and iterative reconstruction of the same patient scans are compared, showing that DLIR provides significantly preferred image quality and reduced noise [241], [242].
For the diagnosis with DECT, the pancreatic cancer diagnostic acceptability and conspicuity can be significantly improved, and the use of DLIR reduces the variation in iodine concentration values while maintaining their accuracy [243]. Fukutomi et al. [244] suggests similar results in terms of iodine concentration quantification through both phantom and clinical studies. The stability of iodine quantification accuracy with DLIR has also been investigated in the context of radiation dose variation. For example, Kojima et al. [245] found that the accuracy is not affected by the radiation dose when the dose index is greater than 12.3 mGy. For a more detailed assessment of DLIR in clinical practice, a recent review paper by Szczykutowicz et al. [246] provides a good starting point. It is also worth noting that the aforementioned studies with PCCT [239] and DECT [244] lead to different conclusions about the impacts of denoising on iodine/material concentration quantification, which could be attributed to the different energy discrimination mechanisms between PCCT and DECT, as the number of energy bins and spectral separation can significantly influence the accuracy and stability of material decomposition performance [30].
2). Artifacts Correction:
Besides noise, image artifact is another factor that affects the quality of CT image for diagnostic evaluation. Few-view or limited-angle reconstruction is an effective method to reduce the radiation dose, but it can introduce globally distributed artifacts that are difficult to remove. To be concise and avoid overlap with Section III, here we only cover recent progress on post-processing-based artifact reduction approaches via deep learning for spectral CT. The networks are often trained in a supervised manner for this application and directly applied to FBP reconstructions to remove artifacts, which can be similarly described as Eq. (49) and Eq. (48) with and being few-view/limited-angle reconstruction and full-view/full-angle reconstruction respectively. For example, to reduce few-view reconstruction artifacts and accelerate reconstruction for scans at multiple energy points (i.e., 32 channels), Mustafa et al. [247] proposed a U-Net-based approach that maps few-view FBP reconstruction images to computationally intensive full-view iterative reconstruction images with TV regularization. The 32-channel FBP images were fed to the network simultaneously and transformed to high-quality 32-channel reconstructions in one step, majorly reducing the computational cost. More recently, Lee et al. [248] developed a multi-level wavelet convolutional neural network, using a U-Net architecture with the wavelet transform as the down-sampling/up-sampling operations, that effectively captures and removes globally distributed few-view artifacts. The network simultaneously processes multichannel images to leverage inter-channel information, and demonstrates promising results both numerically and experimentally with an edged silicon strip PCD. To address limited-angle artifacts for cone beam DECT, Zhang et al. [249] proposed the TIME-Net, which utilizes a transformer module with global attention. In addition, the two complementary limited-angle scans at two energies are fused together to form a prior reconstruction, then the features extracted from the prior reconstruction, high-energy reconstruction, and low-energy reconstruction are fused in latent space to leverage inter-channel information with the network.
In dual-source DECT scanners, the high-energy imaging chain (i.e., tube B with a tin filter, typically at 140 keV) often has a restricted field of view (FOV) (e.g., 33cm) due to physical constraints compared to the other chain (e.g., 50cm for tube A), which can be problematic for larger patients and affect diagnosis. To outpaint the missing regions and match the size of normal FOV, Liu et al. [250] proposed a self-supervised method that maps the low-energy image to the high-energy image with a loss function only focusing on image values within the restricted FOV. The outpainting is then automatically completed leveraging the shift-invariant nature of CNNs. Similarly, Schwartz et al. [251] proposed a method for FOV extension that involves feeding both the high-energy image and the low-energy image in the network, along with a high-energy estimation from the low-energy image via a piecewise-linear transfer function. The trained network was applied to patient data for renal lesion evaluation and showed reliable results in terms of HU value and lesion classification accuracy in the extended regions.
C. Image Generation for Clinical Applications
With the recent development of DECT and PCCT techniques, spectral imaging is reshaping the clinical utilization of CT. These techniques enable the generation of multiple types of images that enhance diagnosis and improve disease management, such as virtual monochromatic images (VMIs), virtual unenhanced images, bone suppression images, and material decomposition maps. A good number of research studies have been performed in these areas using deep learning approaches.
1). Single-Energy to Dual-energy Mapping:
Despite the great possibilities offered by DECT and PCCT, their accessibility remains limited in comparison to conventional single-energy CT, largely due to the high cost involved. To bridge the gap, Zhao et al. [252] successfully demonstrated the feasibility of using deep learning to predict high-energy CT images from given low-energy CT images in a retrospective study. Shortly, Lyu et al. [253] proposed a material decomposition CNN capable of generating high-quality DECT images from a low-energy scan combined with a single view high-energy projection, leveraging the anatomical consistency and energy-domain correlation between two energy images in DECT. The feasibility of this method has been validated with patient studies, showing great potential for simplifying DECT hardware and reducing radiation exposure during DECT scans.
2). Virtual Monochromatic Image:
VMIs are widely used as the basis for routine diagnosis due to their ability to reduce beam-hardening and metal artifacts, and enhance iodine conspicuity. They are obtained by linearly combining the basis material volume fraction maps [115], [254] obtained after material decomposition, as described by the material decomposition model in Section IV. To enhance readability and clarity, Eq. (36), which outlines this model, is replicated here in a spatially discrete form:
| (52) |
where denotes the volume fraction map of the material basis, stands for the linear attenuation coefficient of the corresponding material at energy , and is the total number of material basis. However, the synthesis of VMIs relies on material decomposition results and is therefore limited to DECT and PCCT, which may not be available in less developed areas. Similar to section V-C1, a number of approaches have been explored to directly synthesize the VMIs from single-energy CT scans. Cong et al. [255] first used a modified ResNet for VMI generation from single polychromatic CT scans, then developed a sinogram domain method [256] synthesizing VMIs with a fully-connected NN for virtual monochromatic energy sinogram prediction from single polychromatic measurements. Kawahara et al. [257] employed a GAN to generate VMIs from equivalent keVCT images, while Koike et al. [258] used a U-Net for a similar purpose in imaging of head and neck cancers. More interestingly, Fink et al. [259] found that using VMIs synthesized from single-energy CT images for pulmonary embolism classification provides better performance compared to working directly on the original single-energy images.
On the other hand, VMI synthesis is a downstream task after image reconstruction and material decomposition, during which deep denoising plays a role and potentially affects VMI quality in clinical practice. Extensive studies have investigated this effect through quantitative assessment and/or subjective reader studies. Kojima et al. [245] examined VMI CT number accuracy at various radiation doses, finding that accuracy remains unaffected except at extremely low radiation doses (6.3 mGy). Sato et al. [260] compared VMIs from DLIR with routine baselines from hybrid iterative reconstruction for contrast-enhanced abdomeninal DECT imaging, concluding that vessel and lesion conspicuity of VMIs and iodine density images are improved with DLIR. Xu et al. [261] reached a similar conclusion, and particularly they found that 40 keV VMIs from DLIR poses better CNR and similar or improved image quality compared to 50 keV VMI from hybrid iterative reconstruction, suggesting that 40 keV VMI with DLIR could be a new standard for routine low-keV VMI reconstruction. The study for carotid DECT angiography by Jiang et al. [262] also supports the conclusion that DLIR improves the image quality and diagnostic performance of VMIs compared to hybrid iterative reconstruction. This superiority is further confirmed in DECT angiography with reduced iodine dose (200 mgI/kg) in terms of image quality and arterial depiction by Noda et al. [243]. Additionally, the effect of direct denoising on VMIs has been investigated. In a study of Lee et al. [263] the post-processed VMI using ClariCT.AI (a FDA-approved vendor-agnostic imaging denoising software) is compared with original standard VMI in the assessment of hypoenhancing hepatic metastasis. The results suggest denoising leads to better image quality and lesion detectability. A similar conclusion was achieved by Seo et al. [264] with the same post-denoising method for the evaluation of hypervascular liver lesions.
3). Contrast Agent Dose Reduction:
Iodine-enhanced CT is essential for diagnosing various diseases. However, iodine-based contrast media can cause significant side effects, including allergic reactions in certain patients, and dose-dependent kidney injury and thyroid dysfunction. To investigate the possibility of reducing iodine administration dose while maintaining diagnostic accuracy, Haubold et al. [265] trained a GAN to selectively enhance iodine contrast. They ultimately achieved a 50% contrast dose saving ratio, confirmed by a visual Turing test involving three radiologists assessing pathological consistency. Noda et al. [266] explored the potential of leveraging vendor DLIR for simultaneous iodine and radiation dose reduction in thoraco-abdomino-pelvic DECT imaging. They compared the 40 keV VMIs from DLIR of double low-dose (50% iodine, 50% radiation) scans with VMIs from the hybrid iterative reconstruction of standard dose scans. The diagnostic image quality was achieved in 95% of participants in the double low-dose group, suggesting the feasibility of maintaining diagnostic quality at half doses of radiation and iodine using DLIR.
4). Others:
Several other intriguing deep post-processing techniques for spectral CT include virtual non-contrast image synthesis, virtual non-calcium image synthesis, and spectral CT-based thermometry. Virtual non-contrast images can replace non-contrast scans in a DECT scanning protocol, thus saving radiation dose. However, pure physics-based two-measurement material decomposition algorithms exhibit limited accuracy and stability in the presence of three materials. Poirot et al. [267] employed a CNN to leverage the anatomic information, bridging the gap between material decomposition-derived virtual non-contrast images and the real non-contrast images to generate higher fidelity images.
Virtual non-calcium images are useful for visualizing bone marrow, osteolytic lesions, and even the diagnosis of multiple myeloma [268], [269]. Like virtual non-contrast images, they also suffer from excessive noise and artifacts resulting from material decomposition. Gong et al. [137] proposed a custom dual-task CNN that directly maps the input of spectral CT images to material type maps and corresponding mass density maps. The experimental results demonstrate significantly reduced noise and artifacts in virtual non-calcium images and great visibility of bone marrow lesions.
CT-based thermometry provides a non-invasive method for estimating temperature inside the human body by monitoring the attenuation value changes associated with temperature-dependent radiodensity. Heinrich et al. [270] explored the potential of improving temperature sensitivity with VMIs from DLIR of DECT scans compared to conventional single-energy CT images. Their results show that VMIs significantly enhances temperature sensitivity for different materials, particularly for bone with a boost of 211%. The application of DLIR and hybrid iterative reconstruction has no effect on temperature measurement, suggesting the great potential for dose reduction with deep learning techniques. More recently, Wang et al. [271] incorporated an advanced PCD with 4 energy bin measurements for robust material decomposition and a fully-connected NN for temperature prediction. They observed a non-linear relationship between thermal sensitivity and the concentration of CaCl2 solution in the experiment, achieving final thermometry accuracies of 3.97°C and 1.8°C for 300 mmol/L CaCl2 solution and a milk-based protein shake, respectively.
VI. Perspectives
Advances in spectral CT is a major frontier of the medical CT field, which combines cutting-edge hardware for photon-counting detection and AI-empowered software for deep learning-based reconstruction. As we have reviewed above, photon-counting spectral CT promises to significantly improve the medical CT performance in terms of spatial resolution, spectral resolution, tissue contrast, and dose efficiency. The distinguished capability of photon-counting CT in material decomposition is clinically attractive to perform novel multicontrast-enhanced studies and boost CT, not only in anatomical imaging but also functional or even cellular imaging tasks. All of these can be implemented using machine learning methods or coupled with machine learning methods. Most of such machine learning methods are deep neural networks, involving each key step in the whole imaging workflow.
Looking ahead, the convergence of photon-counting and deep-learning techniques will surely establish spectral CT as the new standard of medical CT. To realize the huge potential of photon-counting spectral CT, there remain challenges to be addressed before task-specific methods and protocols can be successfully translated into clinical practice. These challenges include but are not limited to the following aspects.
Direct Reconstruction:
Deep NNs have been explored to reconstruct images from sinograms in a number of studies. In this approach, a neural network is trained on a large set of sinogram-image pairs until the network predicts realistic reconstructed images. Here, the NN learns to reconstruct the image and at the same time to reduce noise and to incorporate any other corrections desirable for reconstruction. Early methods developed for tomographic reconstruction using deep networks include AUTOMAP [272] for magnetic resonance reconstruction as well as LEARN [273] and iCT [274] for CT reconstruction. To tackle the computational complexity, more sophisticated and efficient networks were developed [275]–[278].
Direct reconstruction techniques may be extended to multichannel reconstruction including photon-counting spectral CT reconstruction. One possible way would be to have multichannel networks incorporating data in multiple energy bins or an ensemble of networks with weight sharing for each energy. Importantly, correlations among these data in these channels should be utilized; for example, as a term in the loss function.
Locally linear embedding Motion Correction:
The much-reduced pixel size of PCDs enables CT imaging at ultrahigh resolution, which is one major advantage of PCCTs over traditional EID-based CT and critical to resolve anatomical and pathologic details, such as cochlear features, lung nodules, and coronary plaques. As resolution drastically improves, the sensitivity to patient motion and geometric misalignment becomes high and can be the limiting factor of image resolution. This increased sensitivity also challenges the assumption of smooth patient movement across views [279]–[281].
To address the issue, Li et al. [282] developed a rigid patient motion compensation method for high-resolution helical PCCT based on locally linear embedding. Their method is in a coarse-to-fine searching framework to boost efficiency, along with several accuracy improving steps masking bad pixel, unreliable volume and patient bed respectively. The method was evaluated on patient wrist scans in a clinical trial, revealing fine bony structures previously hidden by motion blur, as shown in Fig. 5. Subsequently, Li et al. [283] proposed a unified reference-free all-in-one motion correction method for robotic CT with arbitrary scanning trajectories using a nine-degree-of-freedom model, which is capable of addressing rigid patient motion, system misalignment, and coordination errors simultaneously. The effectiveness of the method has been verified on experimental robotic-arm-based PCCT scans of a sacrificed mouse demonstrating a great resolution boost and artifacts reduction.
Fig. 5.
High-resolution PCCT scan of a patient wrist from a clinical trial (90 μm voxel) before and after motion correction (Adapted from Li et al. [282] with permission).
Diffusion Models:
As a score-matching-based generative approach, the diffusion models (DMs) have recently drawn a major attention of the community as they effectively compete or even outperform GANs for image generation and other tasks [284], and have been broadly adapted for medical imaging [285], including PCCT image generation [286]. They involve gradually degrading a sample of interest (i.e., an image) with subtle Gaussian noise until the sample becomes a random Gaussian field, learning the noising process in terms of a score function, and then, by inversion from a Gaussian noise realization, generate a meaningful sample [287]. Specifically, the inverse process uses the gradient of the log-density of the prior (the score) which is approximated with a NN trained for score matching, and generates an image according to the a-priori probability distribution of the training dataset.
DMs can be used to solve inverse problems by adding a data fidelity gradient descent step in the inverse diffusion, or by using the pseudo a-posteriori probability distribution conditioned to the observed data, which work in an unsupervised manner. These methods have been used in various inverse problems such as deblurring on RGB multichannel images [288]. Moreover, the DMs are independent of the measurement model, and the same approaches can be used in multi-energy spectral CT reconstruction or one-step material decomposition under different imaging geometries and sampling conditions. Recently,
Hardware Refinement:
Over the past years, photon-counting detectors have been greatly refined. There are more efforts on CZT detectors, but deep-silicon detectors are also of great interest. While CZT detectors and alike are more compact, the silicon technology is more mature, reliable and cost-effective with the potential to give more quantitative spectral imaging results. A detailed comparison is yet to be seen. Since the photon-counting detector pitches are substantially smaller than that of the energy-integrating detectors, the spatial resolution of CT images can be accordingly improved, coupled with a reduced X-ray source focal spot. However, a small focal spot usually means a low X-ray flux. Hence, the balance must be made between image resolution, noise and imaging speed. It is underlined that while the hardware refinement in either detectors or sources is important, this kind of research will be more often performed by leading companies than academic groups. Since this review is more focused on computational aspects of spectral CT, in the following we discuss more AI-related challenges.
Big Data Construction:
It is well known that big data is a prerequisite for data-driven research. Clearly, it is not easy to have big PCCT data for several reasons, including limited accessibility to PCCT scans, patient privacy, industrial confidentiality, and so on. We believe that this issue must be addressed using simulation tools, and ideally done in a healthcare metaverse. Such an idea was discussed as the first use case in a recent perspective article [289]. Along that direction, virtual twins of physical PCCT scanner models can scan patient avatars to produce simulated data. Along a complementary direction, a limited number of real PCCT scans can be used to train a generative model for realistic image augmentation. For example, it was recently shown that the diffusion model can be used to synthesize realistic data with suppressed privacy leakage [290]. This will facilitate federated learning at the level of datasets.
AI model Development:
When sufficiently informative PCCT data are available, more advanced AI models should be developed to address current weaknesses of deep reconstruction networks in the CT field. The well-known problems of deep networks include stability, generalizability, uncertainty, interpretability, fairness, and more. As briefly mentioned in our review, a unique opportunity in deep learning-based PCCT imaging is raw data correction for charge-sharing, pile-up and other effects. These effects are very complicated, nonlinear and stochastic, but deep learning-based solutions are few and there will be more in the future. Furthermore, large models are gaining great attention, with ChatGPT as a precursor of the next generation of AI methods, i.e., as the first step into the future of artificial general intelligence (AGI). It is believed that large models, multi-modal large models in particular, will further improve the PCCT performance.
High-performance and High-efficiency Computing:
Deep learning with large models takes computational resources. Parallel/cloud computing, model distillation and hybrid (combination of classic and deep learning) reconstruction methods can be synergistic to develop practical PCCT methods. Special hardware such as FPGAs [291] could be adapted in PCCT tasks for imaging speed and energy efficiency.
Clinical Translation:
The development of accurate and robust PCCT methods should lead to diverse clinical applications, from screening and diagnosis to treatment planning and prognosis. PCCT can be also used to guide minimally invasive procedures, such as biopsy and ablation, by providing real-time information over a region of interest [292]. The integration of PCCT (and DECT) with other imaging modalities, such as MRI and PET, would be beneficial as well, leading to a better understanding of anatomical forms and pathological functions.
Hybrid PET/CT Spectral Imaging:
The integration of spectral CT with PET has the potential to open novel clinical applications. However, such an integrated system either requires a costly hardware upgrade or is associated with increased radiation exposure. Most existing spectral CT imaging methods are based on a single modality that uses X-rays. Alternatively, it is possible to explore a combination of X-ray and γ-ray for spectral imaging [293]. The concept of this PET-enabled spectral CT method exploits a standard time-of-flight PET emission scan to derive high-energy γ-ray CT attenuation images and combines the images with low-energy X-ray CT images to form dual-energy or multi-energy imaging. This method has the potential to make spectral CT imaging more readily available on clinical PET/CT scanners. The enabling algorithm of this hybrid spectral imaging method is the reconstruction of γ-ray attenuation images from PET emission data using the maximum-likelihood attenuation and activity algorithm [293], [294]. While the counting statistics of PET emission data are relatively low, machine learning-based approaches have been developed to further improve image reconstruction, for example, using the kernel method alone [293], [295] or in combination with deep neural networks [296]–[298]. These reconstruction approaches are directly based on single subjects without requiring pretraining from a large number of datasets. Alternatively, many other big data-based deep learning techniques that are described in Section III, Section IV, and Section V may be applied to the development of hybrid PET/CT spectral imaging.
VII. Conclusion
In conclusion, this review has systematically reviewed spectral CT with an emphasis on photon-counting and deep learning techniques. This field has evolved from traditional DECT with an established status in medical imaging to contemporary PCCT with promising results and new utilities. Several remaining challenges have been discussed. The future of this technology looks exciting, with numerous opportunities for us to explore so that our imaging dreams can be turned into reality.
ACKNOWLEDGMENT
All authors declare that they have no known conflicts of interest in terms of competing financial interests or personal relationships that could have an influence or are relevant to the work reported in this paper.
This work did not involve human subjects or animals in its research.
This work was supported by the French National Research Agency (ANR) under grant No ANR-20-CE45-0020, and the National Institutes of Health (NIH) under grant Nos R01EB026646, R01CA233888, R01CA237267, R01HL151561, R42GM142394, R21CA264772, R01EB031102, R01EB032716, and R21EB027346.
ACRONYMS
- 2-D
2-dimensional
- 3-D
3-dimensional
- ADMM
alternating direction method of multipliers
- AE
auto-encoder
- AI
artificial intelligence
- BM3D
block-matching and 3-D filtering
- CAOL
convolutional analysis operator learning
- CDL
convolutional dictionary learning
- CNN
convolutional neural network
- CNR
contrast-to-noise ratio
- CPD
canonical polyadic decomposition
- CS
compressed sensing
- CT
computed tomography
- CZT
cadmium zinc telluride
- DECT
dual-energy computed tomography
- DL
dictionary learning
- DLIR
deep learning image reconstruction
- DM
diffusion model
- EID
energy-integrating detectors
- FBP
filtered backprojection
- FDA
Food and Drug Administration
- FOV
field of view
- GAN
generative adversarial network
- JTV
joint total variation
- kVp
peak kilovoltage
- LR
low-rank
- MBIR
model-based iterative reconstruction
- MRI
magnetic resonance imaging
- MSE
mean squared error
- NN
neural network
- PCCT
photon-counting computed tomography
- PCD
photon-counting detectors
- PET
positron emission tomography
- PICCS
prior image-constrained compressed sensing
- SQS
separable quadratic surrogate
- TDL
tensor dictionary learning
- TV
total variation
- VMI
virtual monochromatic image
- W-GAN
Wasserstein generative adversarial network
Footnotes
This definition of TV corresponds to anisotropic TV. The alternative form, isotropic TV consists of summing the -norm of the gradient at each pixel, and had been widely used in CT reconstruction [45]–[48]. Both TV penalties can be addressed by proximal gradient [49]. Alternatively, the semi-norm can also be used [50].
Contributor Information
Alexandre Bousse, LaTIM, Inserm UMR 1101, Université de Bretagne Occidentale, 29238 Brest, France..
Venkata Sai Sundar Kandarpa, LaTIM, Inserm UMR 1101, Université de Bretagne Occidentale, 29238 Brest, France..
Simon Rit, Univ Lyon, INSA-Lyon, Université Claude Bernard Lyon 1, UJM-Saint Étienne, CNRS, Inserm, CREATIS UMR 5220, U1294, F-69373, Lyon, France..
Alessandro Perelli, Department of Biomedical Engineering, School of Science and Engineering, University of Dundee, DD1 4HN, UK.
Mengzhou Li, Biomedical Imaging Center, Rensselaer Polytechnic Institute, Troy, New York, USA..
Guobao Wang, Department of Radiology, University of California Davis Health, Sacramento, USA..
Jian Zhou, CTIQ, Canon Medical Research USA, Inc., Vernon Hills, 60061, USA..
Ge Wang, Biomedical Imaging Center, Rensselaer Polytechnic Institute, Troy, New York, USA..
References
- [1].Taguchi K and Iwanczyk JS, “Vision 20/20: Single photon counting x-ray detectors in medical imaging,” Medical physics, vol. 40, no. 10, p. 100901, 2013. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [2].Taguchi K, Ballabriga R, Campbell M, and Darambara DG, “Photon counting detector computed tomography,” IEEE Transactions on Radiation and Plasma Medical Sciences, vol. 6, no. 1, pp. 1–4, 2021. [Google Scholar]
- [3].Taguchi K, Polster C, Segars WP, Aygun N, and Stierstorfer K, “Model-based pulse pileup and charge sharing compensation for photon counting detectors: A simulation study,” Medical physics, vol. 49, no. 8, pp. 5038–5051, 2022. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [4].Tai WY, Lee D, Zhan X, Taguchi K, and Zbijewski W, “Effects of bowtie scatter on material decomposition in photon-counting CT,” in Medical Imaging 2023: Physics of Medical Imaging, SPIE, vol. 12463, 2023, pp. 26–31. [Google Scholar]
- [5].Litjens G, Kooi T, Bejnordi BE, Setio AAA, Ciompi F, Ghafoorian M, Van Der Laak JA, Van Ginneken B, and Sanchez CI, “A survey on deep learning in medical image analysis,”´ Medical image analysis, vol. 42, pp. 60–88, 2017. [DOI] [PubMed] [Google Scholar]
- [6].Sahiner B, Pezeshk A, Hadjiiski LM, Wang X, Drukker K, Cha KH, Summers RM, and Giger ML, “Deep learning in medical imaging and radiation therapy,” Medical physics, vol. 46, no. 1, e1–e36, 2019. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [7].Domingues I, Pereira G, Martins P, Duarte H, Santos J, and Abreu PH, “Using deep learning techniques in medical imaging: A systematic review of applications on CT and PET,” Artificial Intelligence Review, vol. 53, pp. 4093–4160, 2020. [Google Scholar]
- [8].GENANT HK and BOYD D, “Quantitative bone mineral analysis using dual energy computed tomography,” Investigative Radiology, vol. 12, no. 6, pp. 545–551, Nov. 1977. [DOI] [PubMed] [Google Scholar]
- [9].Adam SZ, Rabinowich A, Kessner R, and Blachar A, “Spectral CT of the abdomen: Where are we now?” Insights into Imaging, vol. 12, no. 1, p. 138, Sep. 2021. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [10].Greffier J, Villani N, Defez D, Dabli D, and Si-Mohamed S, “Spectral CT imaging: Technical principles of dual-energy CT and multi-energy photon-counting CT,” en, Diagn. Interv. Imaging, Nov. 2022. [DOI] [PubMed] [Google Scholar]
- [11].Forghani R, De Man B, and Gupta R, “Dual-energy computed tomography: Physical principles, approaches to scanning, usage, and implementation: Part 1,” en, Neuroimaging Clin. N. Am, vol. 27, no. 3, pp. 371–384, Aug. 2017. [DOI] [PubMed] [Google Scholar]
- [12].Fornaro J, Leschka S, Hibbeln D, Butler A, Anderson N, Pache G, Scheffel H, Wildermuth S, Alkadhi H, and Stolzmann P, “Dual- and multi-energy CT: Approach to functional imaging,” en, Insights Imaging, vol. 2, no. 2, pp. 149–159, Jan. 2011. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [13].Johnson TRC, “Dual-energy CT: General principles,” en, AJR Am. J. Roentgenol, vol. 199, no. 5 Suppl, S3–8, Nov. 2012. [DOI] [PubMed] [Google Scholar]
- [14].Krauss B, “Dual-energy computed tomography: Technology and challenges,” en, Radiol. Clin. North Am, vol. 56, no. 4, pp. 497–506, Jul. 2018. [DOI] [PubMed] [Google Scholar]
- [15].McCollough CH, Leng S, Yu L, and Fletcher JG, “Dual- and multi-energy CT: Principles, technical approaches, and clinical applications,” en, Radiology, vol. 276, no. 3, pp. 637–653, Sep. 2015. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [16].Rajiah P, Abbara S, and Halliburton SS, “Spectral detector CT for cardiovascular applications,” en, Diagn. Interv. Radiol, vol. 23, no. 3, pp. 187–193, May 2017. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [17].Johnson T, Fink C, Schonberg SO, and Reiser MF, Eds., Dual energy CT in clinical practice (Diagnostic Imaging), en. Berlin, Germany: Springer, Nov. 2010. [Google Scholar]
- [18].Megibow AJ, Kambadakone A, and Ananthakrishnan L, “Dual-energy computed tomography: Image acquisition, processing, and workflow,” en, Radiol. Clin. North Am, vol. 56, no. 4, pp. 507–520, Jul. 2018. [DOI] [PubMed] [Google Scholar]
- [19].Flohr TG, McCollough CH, Bruder H, Petersilka M, Gruber K, Suss C, Grasruck M, Stierstorfer K, Krauss B, Raupach R, Primak AN, Kuttner A, Achenbach S, Becker C, Kopp A, and Ohnesorge BM, “First performance evaluation of a dual-source CT (DSCT) system,” en, Eur. Radiol., vol. 16, no. 2, pp. 256–268, Feb. 2006. [DOI] [PubMed] [Google Scholar]
- [20].Carmi R, Naveh G, and Altman A, “Material separation with dual-layer CT,” in IEEE Nuclear Science Symposium Conference Record, IEEE, 2005. [Google Scholar]
- [21].Vlassenbroek A, “Dual layer CT,” in Dual Energy CT in Clinical Practice, Springer Berlin; Heidelberg, Sep. 2010, pp. 21–34. [Google Scholar]
- [22].Philips, Advantages of a spectral detector-based ct system [White paper], https://www.philips.com/c-dam/b2bhc/master/landing-pages/spectralct/452299150831_WP_Advantages_of_spectral_detector_CT.pdf, 2013.
- [23].Gabbai M, Leichter I, Mahgerefteh S, and Sosna J, “Spectral material characterization with dual-energy CT: Comparison of commercial and investigative technologies in phantoms,” en, Acta Radiol, vol. 56, no. 8, pp. 960–969, Aug. 2015. [DOI] [PubMed] [Google Scholar]
- [24].Li L, Zhao T, and Chen Z, “First dual MeV energy x-ray CT for container inspection: Design, algorithm, and preliminary experimental results,” IEEE Access, vol. 6, pp. 45534–45542, 2018. [Google Scholar]
- [25].Boedeker K, Hayes M, Zhou J, Zhang R, and Yu Z, Deep learning spectral ct-faster, easier and more intelligent [White paper], Canon Medical Systems Corporation, https://global.medical.canon/products/computed-tomography/spectral, 2019. [Google Scholar]
- [26].Xi Y, Cong W, Harrison D, and Wang G, “Grating oriented line-wise filtration (GOLF) for dual-energy x-ray CT,” Sensing and Imaging, vol. 18, no. 1, Aug. 2017. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [27].Bornefalk H and Danielsson M, “Photon-counting spectral computed tomography using silicon strip detectors: A feasibility study,” Physics in Medicine & Biology, vol. 55, no. 7, p. 1999, Mar. 2010. [DOI] [PubMed] [Google Scholar]
- [28].Ballabriga R, Alozy J, Campbell M, Frojdh E, Heijne E, Koenig T, Llopart X, Marchal J, Pennicard D, Poikela T, Tlustos L, Valerio P, Wong W, and Zuber M, “Review of hybrid pixel detector readout asics for spectroscopic x-ray imaging,” Journal of Instrumentation, vol. 11, no. 01, P01007, 2016. [Google Scholar]
- [29].Danielsson M, Persson M, and Sjölin M, “Photon-counting x-ray detectors for CT,” Physics in Medicine & Biology, vol. 66, no. 3, 03TR01, 2021. [DOI] [PubMed] [Google Scholar]
- [30].Willemink MJ, Persson M, Pourmorteza A, Pelc NJ, and Fleischmann D, “Photon-counting CT: Technical principles and clinical prospects,” en, Radiology, vol. 289, no. 2, pp. 293–312, Nov. 2018. [DOI] [PubMed] [Google Scholar]
- [31].Samei E, Rajagopal J, and Jones E, “Hallway conversations in physics,” en, AJR Am. J. Roentgenol, vol. 215, no. 5, W50–W52, Nov. 2020. [DOI] [PubMed] [Google Scholar]
- [32].Shanblatt E, O’Doherty J, Petersilka M, Wolber P, Fung G, and Ramirez-Giraldo JC, Whitepaper: The technology behind photon-counting CT, Siemens Healthineers, https://www.siemens-healthineers.com/en-us/computed-tomography/technologies-and-innovations/photon-counting-ct, 2022. [Google Scholar]
- [33].Niu T, Dong X, Petrongolo M, and Zhu L, “Iterative image-domain decomposition for dual-energy CT,” Medical physics, vol. 41, no. 4, p. 041901, 2014. [DOI] [PubMed] [Google Scholar]
- [34].Feng J, Yu H, Wang S, and Liu F, “Image-domain based material decomposition by multi-constraint optimization for spectral CT,” Ieee Access, vol. 8, pp. 155450–155458, 2020. [Google Scholar]
- [35].Shikhaliev PM, “Projection x-ray imaging with photon energy weighting: Experimental evaluation with a prototype detector,” Physics in Medicine & Biology, vol. 54, no. 16, p. 4971, 2009. [DOI] [PubMed] [Google Scholar]
- [36].Arridge SR, Ehrhardt MJ, and Thielemans K, “(an overview of) synergistic reconstruction for multimodality/multichannel imaging methods,” Philosophical Transactions of the Royal Society A, vol. 379, no. 2200, p. 20200205, 2021. [DOI] [PubMed] [Google Scholar]
- [37].Natterer F, The Mathematics of Computerized Tomography (Classics in Applied Mathematics). SIAM, 2001. [Google Scholar]
- [38].Gordon R, Bender R, and Herman GT, “Algebraic Reconstruction Techniques (ART) for Three-Dimensional Electron Microscopy and X-Ray Photography,” Journal of Theoretical Biology, vol. 29, no. 3, pp. 471–481, 1970. [DOI] [PubMed] [Google Scholar]
- [39].Gilbert P, “Iterative methods for the three-dimensional reconstruction of an object from projections,” Journal of theoretical biology, vol. 36, no. 1, pp. 105–117, 1972. [DOI] [PubMed] [Google Scholar]
- [40].Beck A and Teboulle M, “A fast iterative shrinkage-thresholding algorithm for linear inverse problems,” SIAM journal on imaging sciences, vol. 2, no. 1, pp. 183–202, 2009. [Google Scholar]
- [41].Combettes PL and Pesquet J-C, “Proximal splitting methods in signal processing,” in Fixed-point algorithms for inverse problems in science and engineering, Springer, 2011, pp. 185–212. [Google Scholar]
- [42].Erdogan H and Fessler JA, “Monotonic algorithms for transmission tomography,” vol. 18, no. 9, pp. 801–814, 1999. [DOI] [PubMed] [Google Scholar]
- [43].Elbakri IA and Fessler JA, “Statistical image reconstruction for polyenergetic X-ray computed tomography,” IEEE transactions on medical imaging, vol. 21, no. 2, pp. 89–99, 2002. [DOI] [PubMed] [Google Scholar]
- [44].Pati YC, Rezaiifar R, and Krishnaprasad PS, “Orthogonal matching pursuit: Recursive function approximation with applications to wavelet decomposition,” in Proceedings of 27th Asilomar conference on signals, systems and computers, IEEE, 1993, pp. 40–44. [Google Scholar]
- [45].Ritschl L, Bergner F, Fleischmann C, and Kachelrieß M, “Improved total variation-based CT image reconstruction applied to clinical data,” Physics in Medicine & Biology, vol. 56, no. 6, p. 1545, 2011. [DOI] [PubMed] [Google Scholar]
- [46].Sidky EY, Jørgensen JH, and Pan X, “Convex optimization problem prototyping for image reconstruction in computed tomography with the chambolle–pock algorithm,” Physics in Medicine & Biology, vol. 57, no. 10, p. 3065, 2012. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [47].Liu Y, Liang Z, Ma J, Lu H, Wang K, Zhang H, and Moore W, “Total variation-stokes strategy for sparse-view x-ray CT image reconstruction,” IEEE transactions on medical imaging, vol. 33, no. 3, pp. 749–763, 2013. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [48].Niu S, Gao Y, Bian Z, Huang J, Chen W, Yu G, Liang Z, and Ma J, “Sparse-view x-ray CT reconstruction via total generalized variation regularization,” Physics in Medicine & Biology, vol. 59, no. 12, p. 2997, 2014. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [49].Beck A and Teboulle M, “Fast gradient-based algorithms for constrained total variation image denoising and deblurring problems,” IEEE transactions on image processing, vol. 18, no. 11, pp. 2419–2434, 2009. [DOI] [PubMed] [Google Scholar]
- [50].Xu L, Lu C, Xu Y, and Jia J, “Image smoothing via l0 gradient minimization,” in Proceedings of the 2011 SIGGRAPH Asia conference, 2011, pp. 1–12. [Google Scholar]
- [51].Strong D and Chan T, “Edge-preserving and scale-dependent properties of total variation regularization,” Inverse problems, vol. 19, no. 6, S165, 2003. [Google Scholar]
- [52].Boyd S, Parikh N, Chu E, Peleato B, and Eckstein J, “Distributed optimization and statistical learning via the alternating direction method of multipliers,” Foundations & Trends in Machine Learning, vol. 3, no. 1, pp. 1–122, 2010. [Google Scholar]
- [53].Chambolle A and Pock T, “A first-order primal-dual algorithm for convex problems with applications to imaging,” Journal of mathematical imaging and vision, vol. 40, no. 1, pp. 120–145, 2011. [Google Scholar]
- [54].Sapiro G and Ringach DL, “Anisotropic diffusion of multivalued images with applications to color filtering,” IEEE transactions on image processing, vol. 5, no. 11, pp. 1582–1586, 1996. [DOI] [PubMed] [Google Scholar]
- [55].Blomgren P and Chan TF, “Color TV: Total variation methods for restoration of vector-valued images,” IEEE transactions on image processing, vol. 7, no. 3, pp. 304–309, 1998. [DOI] [PubMed] [Google Scholar]
- [56].Lefkimmiatis S, Roussos A, Unser M, and Maragos P, “Convex generalizations of total variation based on the structure tensor with applications to inverse problems,” in International Conference on Scale Space and Variational Methods in Computer Vision, Springer, 2013, pp. 48–60. [Google Scholar]
- [57].Rigie DS and La Rivière PJ, “Joint reconstruction of multi-channel, spectral CT data via constrained total nuclear variation minimization,” Physics in Medicine & Biology, vol. 60, no. 5, p. 1741, 2015. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [58].Gao H, Cai J-F, Shen Z, and Zhao H, “Robust principal component analysis-based four-dimensional computed tomography,” Physics in Medicine & Biology, vol. 56, no. 11, p. 3181, 2011. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [59].Gao H, Yu H, Osher S, and Wang G, “Multi-energy CT based on a prior rank, intensity and sparsity model (PRISM),” Inverse problems, vol. 27, no. 11, p. 115012, 2011. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [60].Chu J, Li L, Chen Z, Wang G, and Gao H, “ Multi-energy CT reconstruction based on low rank and sparsity with the split-bregman method (mlrss),” in 2012 IEEE Nuclear Science Symposium and Medical Imaging Conference Record (NSS/MIC), IEEE, 2012, pp. 2411–2414. [Google Scholar]
- [61].Semerci O, Hao N, Kilmer ME, and Miller EL, “Tensor-based formulation and nuclear norm regularization for multienergy computed tomography,” IEEE Transactions on Image Processing, vol. 23, no. 4, pp. 1678–1693, 2014. [DOI] [PubMed] [Google Scholar]
- [62].Li L, Chen Z, Wang G, Chu J, and Gao H, “A tensor PRISM algorithm for multienergy CT reconstruction and comparative studies,” Journal of X-ray science and technology, vol. 22, no. 2, pp. 147–163, 2014. [DOI] [PubMed] [Google Scholar]
- [63].Li L, Chen Z, Cong W, and Wang G, “Spectral CT modeling and reconstruction with hybrid detectors in dynamic-threshold-based counting and integrating modes,” IEEE transactions on medical imaging, vol. 34, no. 3, pp. 716–728, 2014. [DOI] [PubMed] [Google Scholar]
- [64].Xia W, Wu W, Niu S, Liu F, Zhou J, Yu H, Wang G, and Zhang Y, “Spectral CT reconstruction–ASSIST: Aided by self-similarity in image-spectral tensors,” IEEE Transactions on Computational Imaging, vol. 5, no. 3, pp. 420–436, 2019. [Google Scholar]
- [65].He Y, Zeng L, Xu Q, Wang Z, Yu H, Shen Z, Yang Z, and Zhou R, “Spectral CT reconstruction via low-rank representation and structure preserving regularization,” Physics in Medicine & Biology, vol. 68, no. 2, p. 025011, 2023. [DOI] [PubMed] [Google Scholar]
- [66].Chen G-H, Tang J, and Leng S, “Prior image constrained compressed sensing (PICCS): A method to accurately reconstruct dynamic CT images from highly undersampled projection data sets,” Medical physics, vol. 35, no. 2, pp. 660–663, 2008. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [67].Yu Z, Leng S, Li Z, and McCollough CH, “Spectral prior image constrained compressed sensing (spectral PICCS) for photon-counting computed tomography,” Physics in Medicine & Biology, vol. 61, no. 18, p. 6707, 2016. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [68].Wang S, Wu W, Feng J, Liu F, and Yu H, “Low-dose spectral CT reconstruction based on image-gradient l0-norm and adaptive spectral PICCS,” Physics in Medicine & Biology, vol. 65, no. 24, p. 245005, 2020. [DOI] [PubMed] [Google Scholar]
- [69].Zhang Y, Xi Y, Yang Q, Cong W, Zhou J, and Wang G, “Spectral CT reconstruction with image sparsity and spectral mean,” IEEE transactions on computational imaging, vol. 2, no. 4, pp. 510–523, 2016. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [70].Yao L, Zeng D, Chen G, Liao Y, Li S, Zhang Y, Wang Y, Tao X, Niu S, Lv Q, et al. , “Multi-energy computed tomography reconstruction using a nonlocal spectral similarity model,” Physics in Medicine & Biology, vol. 64, no. 3, p. 035018, 2019. [DOI] [PubMed] [Google Scholar]
- [71].Cueva E, Meaney A, Siltanen S, and Ehrhardt MJ, “Synergistic multi-spectral CT reconstruction with directional total variation,” Philosophical Transactions of the Royal Society A, vol. 379, no. 2204, p. 20200198, 2021. [DOI] [PubMed] [Google Scholar]
- [72].Xu Q, Yu H, Mou X, Zhang L, Hsieh J, and Wang G, “Low-dose X-ray CT reconstruction via dictionary learning,” IEEE transactions on medical imaging, vol. 31, no. 9, pp. 1682–1697, 2012. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [73].Zhang C, Zhang T, Li M, Peng C, Liu Z, and Zheng J, “Low-dose CT reconstruction via l1 dictionary learning regularization using iteratively reweighted least-squares,” Biomedical engineering online, vol. 15, no. 1, pp. 1–21, 2016. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [74].Komolafe TE, Wang K, Du Q, Hu T, Yuan G, Zheng J, Zhang C, Li M, and Yang X, “Smoothed l0-constraint dictionary learning for low-dose X-ray CT reconstruction,” IEEE Access, vol. 8, pp. 116961–116973, 2020. [Google Scholar]
- [75].Xu M, Hu D, Luo F, Liu F, Wang S, and Wu W, “Limited-angle x-ray CT reconstruction using image gradient ℓ0-norm with dictionary learning,” IEEE Transactions on Radiation and Plasma Medical Sciences, vol. 5, no. 1, pp. 78–87, 2020. [Google Scholar]
- [76].Aharon M, Elad M, and Bruckstein A, “K-SVD: An algorithm for designing overcomplete dictionaries for sparse representation,” IEEE Transactions on signal processing, vol. 54, no. 11, pp. 4311–4322, 2006. [Google Scholar]
- [77].Tucker LR, “Some mathematical notes on three-mode factor analysis,” Psychometrika, vol. 31, no. 3, pp. 279–311, 1966. [DOI] [PubMed] [Google Scholar]
- [78].Zubair S and Wang W, “Tensor dictionary learning with sparse tucker decomposition,” in 2013 18th international conference on digital signal processing (DSP), IEEE, 2013, pp. 1–6. [Google Scholar]
- [79].Liu X, Bourennane S, and Fossati C, “Denoising of hyperspectral images using the parafac model and statistical performance analysis,” IEEE Transactions on Geoscience and Remote Sensing, vol. 50, no. 10, pp. 3717–3724, 2012. [Google Scholar]
- [80].Peng Y, Meng D, Xu Z, Gao C, Yang Y, and Zhang B, “Decomposable nonlocal tensor dictionary learning for multispectral image denoising,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2014, pp. 2949–2956. [Google Scholar]
- [81].Tan S, Zhang Y, Wang G, Mou X, Cao G, Wu Z, and Yu H, “Tensor-based dictionary learning for dynamic tomographic reconstruction,” Physics in Medicine & Biology, vol. 60, no. 7, p. 2803, 2015. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [82].Zhang Y, Mou X, Wang G, and Yu H, “Tensor-based dictionary learning for spectral CT reconstruction,” IEEE transactions on medical imaging, vol. 36, no. 1, pp. 142–154, 2016. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [83].Duan G, Wang H, Liu Z, Deng J, and Chen Y-W, “K-CPD: Learning of overcomplete dictionaries for tensor sparse coding,” in Proceedings of the 21st International Conference on Pattern Recognition (ICPR2012), IEEE, 2012, pp. 493–496. [Google Scholar]
- [84].Wu W, Zhang Y, Wang Q, Liu F, Chen P, and Yu H, “Low-dose spectral CT reconstruction using image gradient ℓ0–norm and tensor dictionary,” Applied Mathematical Modelling, vol. 63, pp. 538–557, 2018. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [85].Li X, Sun X, Zhang Y, Pan J, and Chen P, “Tensor dictionary learning with an enhanced sparsity constraint for sparse-view spectral CT reconstruction,” in Photonics, MDPI, vol. 9, 2022, p. 35. [Google Scholar]
- [86].Sudarshan VP, Chen Z, and Awate SP, “Joint PET+MRI patch-based dictionary for bayesian random field PET reconstruction,” in International Conference on Medical Image Computing and Computer-Assisted Intervention, Springer, 2018, pp. 338–346. [Google Scholar]
- [87].Sudarshan VP, Egan GF, Chen Z, and Awate SP, “Joint PET-MRI image reconstruction using a patch-based joint-dictionary prior,” Medical image analysis, vol. 62, p. 101669, 2020. [DOI] [PubMed] [Google Scholar]
- [88].Song P, Weizman L, Mota JF, Eldar YC, and Rodrigues MR, “Coupled dictionary learning for multi-contrast MRI reconstruction,” IEEE transactions on medical imaging, vol. 39, no. 3, pp. 621–633, 2019. [DOI] [PubMed] [Google Scholar]
- [89].Yang J, Wang Z, Lin Z, Cohen S, and Huang T, “Coupled dictionary training for image super-resolution,” IEEE transactions on image processing, vol. 21, no. 8, pp. 3467–3478, 2012. [DOI] [PubMed] [Google Scholar]
- [90].Wohlberg B, “Efficient algorithms for convolutional sparse representations,” IEEE Transactions on Image Processing, vol. 25, no. 1, pp. 301–315, 2015. [DOI] [PubMed] [Google Scholar]
- [91].Chun IY and Fessler JA, “Convolutional dictionary learning: Acceleration and convergence,” IEEE Transactions on Image Processing, vol. 27, no. 4, pp. 1697–1712, 2017. [DOI] [PubMed] [Google Scholar]
- [92].Garcia-Cardona C and Wohlberg B, “Convolutional dictionary learning: A comparative review and new algorithms,” IEEE Transactions on Computational Imaging, vol. 4, no. 3, pp. 366–381, 2018. [Google Scholar]
- [93].Bao P, Xia W, Yang K, Chen W, Chen M, Xi Y, Niu S, Zhou J, Zhang H, Sun H, et al. , “Convolutional sparse coding for compressed sensing CT reconstruction,” IEEE transactions on medical imaging, vol. 38, no. 11, pp. 2607–2619, 2019. [DOI] [PubMed] [Google Scholar]
- [94].Chun IY and Fessler JA, “Convolutional analysis operator learning: Acceleration and convergence,” IEEE Transactions on Image Processing, vol. 29, pp. 2108–2122, 2019. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [95].Garcia-Cardona C and Wohlberg B, “Convolutional dictionary learning for multi-channel signals,” in 2018 52nd Asilomar Conference on Signals, Systems, and Computers, IEEE, 2018, pp. 335–342. [Google Scholar]
- [96].Degraux K, Kamilov US, Boufounos PT, and Liu D, “Online convolutional dictionary learning for multimodal imaging,” in 2017 IEEE International Conference on Image Processing (ICIP; ), 2017, pp. 1617–1621. [Google Scholar]
- [97].Gao F, Deng X, Xu M, Xu J, and Dragotti PL, “Multi-modal convolutional dictionary learning,” IEEE Transactions on Image Processing, vol. 31, pp. 1325–1339, 2022. [DOI] [PubMed] [Google Scholar]
- [98].Perelli A, Garcia SA, Bousse A, Tasu J-P, Efthimiadis N, and Visvikis D, “Multi-channel convolutional analysis operator learning for dual-energy CT reconstruction,” Physics in Medicine & Biology, vol. 67, no. 6, p. 065001, 2022. [DOI] [PubMed] [Google Scholar]
- [99].Wang Z, Bousse A, Vermet F, Pinton NJ, Froment J, Vedel B, Tasu J-P, and Visvikis D, “Synergistic multi-energy reconstruction with a deep penalty “connecting the energies”,” in IEEE Nuclear Science Symposium Medical Imaging Conference and Room Temperature Semiconductor Conference, 2022. [Online]. Available: https://hal.ehesp.fr/IBNM/hal-03955092v1. [Google Scholar]
- [100].Pinton NJ, Bousse A, Wang Z, Cheze-Le-Rest C, Maxim V, Comtat C, Sureau F, and Visvikis D, “Synergistic PET/CT reconstruction using a joint generative model,” in International Conference on Fully Three-Dimensional Image Reconstruction in Radiology and Nuclear Medicine, Stony Brook, NY, USA, 2023, pp. D2–1–3. [Google Scholar]
- [101].Gautier V, Comtat C, Sureau F, Bousse A, Friot-Giroux L, Maxim V, and Sixou B, “VAE constrained MR guided PET reconstruction,” in International Conference on Fully Three-Dimensional Image Reconstruction in Radiology and Nuclear Medicine, Stony Brook, NY, USA, 2023, pp. 74–77. [Google Scholar]
- [102].Wu D, Kim K, El Fakhri G, and Li Q, “Iterative low-dose CT reconstruction with priors trained by artificial neural network,” IEEE transactions on medical imaging, vol. 36, no. 12, pp. 2479–2486, 2017. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [103].Duff M, Campbell ND, and Ehrhardt MJ, “Regularising inverse problems with generative machine learning models,” arXiv preprint arXiv:2107.11191, 2021. [Google Scholar]
- [104].Ulyanov D, Vedaldi A, and Lempitsky V, “Deep image prior,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 9446–9454. [Google Scholar]
- [105].Gong K, Kim K, Wu D, Kalra MK, and Li Q, “Low-dose dual energy CT image reconstruction using non-local deep image prior,” in 2019 IEEE Nuclear Science Symposium and Medical Imaging Conference (NSS/MIC), IEEE, 2019, pp. 1–2. [Google Scholar]
- [106].Arridge SR, Maass P, Oktem O, and Sch onlieb C-B, “Solving inverse problems using data-driven models,” Acta Numerica, vol. 28, pp. 1–174, 2019. [Google Scholar]
- [107].Monga V, Li Y, and Eldar YC, “Algorithm unrolling: Interpretable, efficient deep learning for signal and image processing,” IEEE Signal Processing Magazine, vol. 38, no. 2, pp. 18–44, 2021. [Google Scholar]
- [108].Yang Y, Sun J, Li H, and Xu Z, “Deep ADMM-Net for compressive sensing MRI,” in Proceedings of the 30th International Conference on Neural Information Processing Systems, 2016, pp. 10–18. [Google Scholar]
- [109].Meinhardt T, Moller M, Hazirbas C, and Cremers D, “Learning proximal operators: Using denoising networks for regularizing inverse imaging problems,” in Proceedings of the IEEE International Conference on Computer Vision, 2017, pp. 1781–1790. [Google Scholar]
- [110].Gupta H, Jin KH, Nguyen HQ, McCann MT, and Unser M, “CNN-based projected gradient descent for consistent CT image reconstruction,” IEEE transactions on medical imaging, vol. 37, no. 6, pp. 1440–1453, 2018. [DOI] [PubMed] [Google Scholar]
- [111].Chun Y and Fessler JA, “Deep bcd-net using identical encoding-decoding CNN structures for iterative image recovery,” in 2018 IEEE 13th Image, Video, and Multidimensional Signal Processing Workshop (IVMSP), IEEE, 2018, pp. 1–5. [Google Scholar]
- [112].Chun IY, Huang Z, Lim H, and Fessler J, “Momentum-net: Fast and convergent iterative neural network for inverse problems,” IEEE transactions on pattern analysis and machine intelligence, 2020. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [113].Adler J and Oktem O, “Learned primal-dual reconstruction,”IEEE transactions on medical imaging, vol. 37, no. 6, pp. 1322–1332, 2018. [DOI] [PubMed] [Google Scholar]
- [114].Chen X, Xia W, Yang Z, Chen H, Liu Y, Zhou J, and Zhang Y, “SOUL-net: A sparse and low-rank unrolling network for spectral CT image reconstruction,” arXiv preprint arXiv:2207.12639, 2022. [DOI] [PubMed] [Google Scholar]
- [115].Alvarez RE and Macovski A, “Energy-selective reconstructions in x-ray computerised tomography,” Physics in Medicine & Biology, vol. 21, no. 5, p. 733, 1976. [DOI] [PubMed] [Google Scholar]
- [116].Bornefalk H, “XCOM intrinsic dimensionality for low-Z elements at diagnostic energies,” Med Phys, vol. 39, no. 2, pp. 654–657, Feb. 2012. DOI: 10.1118/1.3675399. [DOI] [PubMed] [Google Scholar]
- [117].Alvarez R, “Dimensionality and noise in energy selective x-ray imaging,” Med Phys, vol. 40, no. 11, p. 111909, Nov. 2013. DOI: 10.1118/1.4824057. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [118].Tang X and Ren Y, “On the conditioning of basis materials and its impact on multimaterial decomposition-based spectral imaging in photon-counting CT,” Med Phys, vol. 48, no. 3, pp. 1100–1116, 2021. DOI: 10.1002/mp.14708. [DOI] [PubMed] [Google Scholar]
- [119].Roessl E and Proksa R, “K-edge imaging in x-ray computed tomography using multi-bin photon counting detectors,” Phys Med Biol, vol. 52, no. 15, pp. 4679–4696, Aug. 2007. DOI: 10.1088/0031-9155/52/15/020. [DOI] [PubMed] [Google Scholar]
- [120].Wu W, Yu H, Chen P, Luo F, Liu F, Wang Q, Zhu Y, Zhang Y, Feng J, and Yu H, “Dictionary learning based image-domain material decomposition for spectral CT,” Phys Med Biol, vol. 65, no. 24, p. 245006, Dec. 2020. DOI: 10.1088/1361-6560/aba7ce. [DOI] [PubMed] [Google Scholar]
- [121].Wu W, Chen P, Wang S, Vardhanabhuti V, Liu F, and Yu H, “Image-domain material decomposition for spectral CT using a generalized dictionary learning,” IEEE Trans Radiat Plasma Med Sci, vol. 5, no. 4, pp. 537–547, Jul. 2021, ISSN: 2469–7303. DOI: 10.1109/TRPMS.2020.2997880. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [122].Zhang T, Yu H, Xi Y, Wang S, and Liu F, “Spectral CT image-domain material decomposition via sparsity residual prior and dictionary learning,” IEEE Transactions on Instrumentation and Measurement, vol. 72, pp. 1–13, 2023, ISSN: 1557–9662. DOI: 10.1109/TIM.2022.3221120.37323850 [DOI] [Google Scholar]
- [123].Li Z, Ravishankar S, Long Y, and Fessler J, “Image-domain material decomposition using data-driven sparsity models for dual-energy CT,” in 2018 IEEE 15th International Symposium on Biomedical Imaging (ISBI 2018), Apr. 2018, pp. 52–56. DOI: 10.1109/ISBI.2018.8363521. [DOI] [Google Scholar]
- [124].Li Z, Chun I, and Long Y, “Image-domain material decomposition using an iterative neural network for dual-energy CT,” in 2020 IEEE 17th International Symposium on Biomedical Imaging (ISBI), Apr. 2020, pp. 651–655. DOI: 10.1109/ISBI45749.2020.9098590. [DOI] [Google Scholar]
- [125].Li Z, Long Y, and Chun I, “An improved iterative neural network for high-quality image-domain material decomposition in dual-energy CT,” Med Phys, 2022. DOI: 10.1002/mp.15817. [DOI] [PubMed] [Google Scholar]
- [126].Feng C, Kang K, and Xing Y, “Fully connected neural network for virtual monochromatic imaging in spectral computed tomography,” Journal of Medical Imaging, vol. 6, no. 1, p. 011006, 2019. DOI: 10.1117/1.JMI.6.1.011006. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [127].Xu Y, Yan B, Zhang J, Chen J, Zeng L, and Wang L, “Image decomposition algorithm for dual-energy computed tomography via fully convolutional network,” Computational and Mathematical Methods in Medicine, vol. 2018, p. 2527516, 2018, ISSN: 1748–670X. DOI: 10.1155/2018/2527516. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [128].Clark D, Holbrook M, and Badea C, “Multi-energy CT decomposition using convolutional neural networks,” in Medical Imaging 2018: Physics of Medical Imaging, Lo JY, Schmidt TG, and Chen G-H, Eds., International Society for Optics and Photonics, vol. 10573, 2018, 105731O. DOI: 10.1117/12.2293728. [DOI] [Google Scholar]
- [129].Abascal J, Ducros N, Pronina V, Rit S, Rodesch P-A, Broussaud T, Bussod S, Douek P, Hauptmann A, Arridge S, and Peyrin F, “Material decomposition in spectral CT using deep learning: A Sim2Real transfer approach,” IEEE Access, vol. 9, pp. 25632–25647, 2021, ISSN: 2169–3536. DOI: 10.1109/ACCESS.2021.3056150. [DOI] [Google Scholar]
- [130].Holbrook M, “System development and preclinical applications of photon counting micro-CT,” Ph.D. dissertation, Duke University, 2021. [Online]. Available: https://hdl.handle.net/10161/23753. [Google Scholar]
- [131].Fang W and Li L, “Attenuation image referenced (AIR) effective atom number image calculation for MeV dual-energy container CT using image-domain deep learning framework,” Results in Physics, vol. 35, p. 105406, 2022, ISSN: 2211–3797. DOI: 10.1016/j.rinp.2022.105406. [Online]. Available: https://www.sciencedirect.com/science/article/pii/S2211379722001693. [DOI] [Google Scholar]
- [132].Fujiwara D, Shimomura T, Zhao W, Li K-W, Haga A, and Geng L-S, “Virtual computed-tomography system for deep-learning-based material decomposition,” Phys Med Biol, vol. 67, no. 15, p. 155008, Jul. 2022. DOI: 10.1088/1361-6560/ac7bcd. [DOI] [PubMed] [Google Scholar]
- [133].Nadkarni R, Allphin A, Clark D, and Badea C, “Material decomposition from photon-counting CT using a convolutional neural network and energy-integrating CT training labels,” Phys Med Biol, vol. 67, no. 15, p. 155003, Jul. 2022. DOI: 10.1088/1361-6560/ac7d34. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [134].Zhang W, Zhang H, Wang L, Wang X, Hu X, Cai A, Li L, Niu T, and Yan B, “Image domain dual material decomposition for dual-energy CT using butterfly network,” Med Phys, vol. 46, no. 5, pp. 2037–2051, 2019. DOI: 10.1002/mp.13489. [DOI] [PubMed] [Google Scholar]
- [135].Chen Z and Li L, “Robust multimaterial decomposition of spectral CT using convolutional neural networks,” Optical Engineering, vol. 58, no. 1, p. 013104, 2019. DOI: 10.1117/1.OE.58.1.013104. [DOI] [Google Scholar]
- [136].Gong H, Tao S, Rajendran K, Zhou W, McCollough C, and Leng S, “Deep-learning-based direct inversion for material decomposition,” Med Phys, vol. 47, no. 12, pp. 6294–6309, 2020. DOI: 10.1002/mp.14523. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [137].Gong H, Baffour F, Glazebrook K, Rhodes N, Tiegs-Heiden C, Thorne J, Cook J, Kumar S, Fletcher J, McCollough C, and Leng S, “Deep learning-based virtual noncalcium imaging in multiple myeloma using dual-energy CT,” Med Phys, vol. 49, no. 10, pp. 6346–6358, 2022. DOI: 10.1002/mp.15934. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [138].Shi Z, Li H, Cao Q, Wang Z, and Cheng M, “A material decomposition method for dual-energy CT via dual interactive Wasserstein generative adversarial networks,” Med Phys, vol. 48, no. 6, pp. 2891–2905, 2021. DOI: 10.1002/mp.14828. [DOI] [PubMed] [Google Scholar]
- [139].Wu X, He P, Long Z, Guo X, Chen M, Ren X, Chen P, Deng L, An K, Li P, Wei B, and Feng P, “Multi-material decomposition of spectral CT images via Fully Convolutional DenseNets,” J X-ray Sci Technol, vol. 27, no. 3, pp. 461–471, 2019, ISSN: 1095–9114. DOI: 10.3233/XST-190500. [DOI] [PubMed] [Google Scholar]
- [140].Su K-H, Kuo J-W, Jordan DW, Hedent SV, Klahr P, Wei Z, Helo RA, Liang F, Qian P, Pereira GC, Rassouli N, Gilkeson RC, Traughber BJ, Cheng C-W, and Muzic RF, “Machine learning-based dual-energy ct parametric mapping,” Phys Med Biol, vol. 63, no. 12, p. 125001, Jun. 2018. DOI: 10.1088/1361-6560/aac711. [DOI] [PubMed] [Google Scholar]
- [141].Abascal J, Ducros N, and Peyrin F, “Nonlinear material decomposition using a regularized iterative scheme based on the Bregman distance,” Inverse Probl, vol. 34, p. 124003, 2018, ISSN: 0266–5611. DOI: 10.1088/1361-6420/aae1e7. [DOI] [Google Scholar]
- [142].Lee W-J, Kim D-S, Kang S-W, and Yi W-J, “Material depth reconstruction method of multi-energy X-ray images using neural network,” in 2012 Annual International Conference of the IEEE Engineering in Medicine and Biology Society, Aug. 2012, pp. 1514–1517. DOI: 10.1109/EMBC.2012.6346229. [DOI] [PubMed] [Google Scholar]
- [143].Zimmerman K and Schmidt T, “Experimental comparison of empirical material decomposition methods for spectral CT,” Phys Med Biol, vol. 60, no. 8, pp. 3175–3191, Apr. 2015. DOI: 10.1088/0031-9155/60/8/3175. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [144].Touch M, Clark D, Barber W, and Badea C, “A neural network-based method for spectral distortion correction in photon counting x-ray CT,” Phys Med Biol, vol. 61, no. 16, p. 6132, Jul. 2016. DOI: 10.1088/0031-9155/61/16/6132. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [145].Zimmerman K, Sharma G, Parchur A, Joshi A, and Schmidt T, “Experimental investigation of neural network estimator and transfer learning techniques for K-edge spectral CT imaging,” Med Phys, vol. 47, no. 2, pp. 541–551, 2020. DOI: 10.1002/mp.13946. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [146].Jenkins P and Gilat Schmidt T, “Experimental study of photon-counting CT neural network material decomposition under conditions of pulse pileup,” Journal of Medical Imaging, vol. 8, no. 1, pp. 1–14, 2021. DOI: 10.1117/1.JMI.8.1.013502. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [147].Brendel B, Bergner F, Brown K, and Koehler T, “Penalized likelihood decomposition for dual layer spectral CT,” in Fourth international conference on image formation in X-ray computed tomography, Bamberg, Germany, 2016. [Google Scholar]
- [148].Ducros N, Abascal J-J, Sixou B, Rit S, and Peyrin F, “Regularization of nonlinear decomposition of spectral x-ray projection images,” Med Phys, vol. 44, e174–e187, 9 Sep. 2017, ISSN: 2473–4209. DOI: 10.1002/mp.12283. [DOI] [PubMed] [Google Scholar]
- [149].Mechlem K, Allner S, Ehn S, Mei K, Braig E, Münzel D, Pfeiffer F, and Noel P, “A post-processing algorithm for spectral CT material selective images using learned dictionaries,” Biomed Phys Eng Express, vol. 3, no. 2, p. 025009, 2017. DOI: 10.1088/2057-1976/aa6045. [DOI] [Google Scholar]
- [150].Mechlem K, Sellerer T, Ehn S, Munzel D, Braig E, Herzen J, Noel PB, and Pfeiffer F, “Spectral angiography material decomposition using an empirical forward model and a dictionary-based regularization,” IEEE Trans. Med. Imag, p. 1, 2018, ISSN: 0278–0062. DOI: 10.1109/TMI.2018.2840841. [DOI] [PubMed] [Google Scholar]
- [151].Jiang X, Zhang X, Stayman JW, and Gang GJ, “Multi-material decomposition with triple layer flat-panel detector cbct using model-based and deep learning approaches,” in International Conference on Fully Three-Dimensional Image Reconstruction in Radiology and Nuclear Medicine, Stony Brook, NY, USA, Jul. 2023. [Google Scholar]
- [152].Xu Y, Yan B, Chen J, Zeng L, and Li L, “Projection decomposition algorithm for dual-energy computed tomography via deep neural network,” J Xray Sci Technol, vol. 26, no. 3, pp. 361–377, 2018, ISSN: 1095–9114. DOI: 10.3233/XST-17349. [DOI] [PubMed] [Google Scholar]
- [153].Lu Y, Kowarschik M, Huang X, Xia Y, Choi J-H, Chen S, Hu S, Ren Q, Fahrig R, Hornegger J, and Maier A, “A learning-based material decomposition pipeline for multi-energy x-ray imaging,” Med Phys, vol. 46, no. 2, pp. 689–703, 2019. DOI: 10.1002/mp.13317. [DOI] [PubMed] [Google Scholar]
- [154].Geng M, Tian Z, Jiang Z, You Y, Feng X, Xia Y, Yang K, Ren Q, Meng X, Maier A, and Lu Y, “PMS-GAN: Parallel multistream generative adversarial network for multi-material decomposition in spectral computed tomography,” IEEE Trans. Med. Imag, vol. 40, no. 2, pp. 571–584, Feb. 2021, ISSN: 1558–254X. DOI: 10.1109/TMI.2020.3031617. [DOI] [PubMed] [Google Scholar]
- [155].Lu Y, Geret J, Unberath M, Manhart M, Ren Q, Fahrig R, Hornegger J, and Maier A, “Projection-based material decomposition by machine learning using image-based features for computed tomography,” in Fully 3D Image Reconstruction in Radiology and Nuclear Medicine, Newport, USA, 2015, pp. 448–451. [Google Scholar]
- [156].Lu Y, Kowarschik M, Huang X, Chen S, Ren Q, Fahrig R, Hornegger J, and Maier A, “Material decomposition using ensemble learning for spectral X-ray imaging,” IEEE Trans Radiat Plasma Med Sci, vol. 2, no. 3, pp. 194–204, May 2018, ISSN: 2469–7303. DOI: 10.1109/TRPMS.2018.2805328. [DOI] [Google Scholar]
- [157].Eguizabal A, Oktem O, and Persson M, “Deep learning for material decomposition in photon-counting CT,” Aug. 2022. [Google Scholar]
- [158].Zhang R, Zhou J, and Yu Z, “A cascaded deep-learning reconstruction method for sparse-view kv-switching dual-energy CT,” in Medical Imaging 2020: Physics of Medical Imaging, SPIE, vol. 11312, 2020, pp. 499–505. [Google Scholar]
- [159].Xu Q, Sawatzky A, Anastasio M, and Schirra C, “Sparsity-regularized image reconstruction of decomposed K-edge data in spectral CT,” Phys Med Biol, vol. 59, no. 10, N65–N79, May 2014. DOI: 10.1088/0031-9155/59/10/N65. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [160].Liu Y, Yu Z, and Zou Y, “Impact of covariance modeling in dual-energy spectral CT image reconstruction,” in Medical Imaging 2015: Physics of Medical Imaging, International Society for Optics and Photonics, vol. 9412, 2015, 94123Q. [Google Scholar]
- [161].Cai C, Rodet T, Legoupil S, and Mohammad-Djafari A, “A full-spectral bayesian reconstruction approach based on the material decomposition model applied in dual-energy computed tomography,” Med Phys, vol. 40, p. 111916, 11 Nov. 2013, ISSN: 2473–4209. DOI: 10.1118/1.4820478. [DOI] [PubMed] [Google Scholar]
- [162].Long Y and Fessler J, “Multi-material decomposition using statistical image reconstruction for spectral CT,” IEEE Trans. Med. Imag, vol. 33, no. 8, pp. 1614–1626, 2014. DOI: 10.1109/TMI.2014.2320284. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [163].Weidinger T, Buzug T, Flohr T, Kappler S, and Stierstorfer K, “Polychromatic iterative statistical material image reconstruction for photon-counting computed tomography,” Int J Biomed Imaging, vol. 2016, p. 5871604, 2016. DOI: 10.1155/2016/5871604. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [164].Mechlem K, Ehn S, Sellerer T, Braig E, Munzel D, Pfeiffer F, and Noel P, “Joint statistical iterative material image reconstruction for spectral computed tomography using a semi-empirical forward model,” IEEE Trans. Med. Imag, vol. 37, pp. 68–80, 1 Jan. 2018, ISSN: 1558–254X. DOI: 10.1109/TMI.2017.2726687. [DOI] [PubMed] [Google Scholar]
- [165].Foygel Barber R, Sidky E, Gilat Schmidt T, and Pan X, “An algorithm for constrained one-step inversion of spectral CT data,” Phys Med Biol, vol. 61, no. 10, pp. 3784–3818, May 2016. DOI: 10.1088/0031-9155/61/10/3784. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [166].Tairi S, Anthoine S, Boursier Y, Dupont M, and Morel C, “ProMeSCT: A proximal metric algorithm for spectral CT,” IEEE Trans Radiat Plasma Med Sci, vol. 5, no. 4, pp. 548–558, Jul. 2021, ISSN: 2469–7303. DOI: 10.1109/TRPMS.2020.3036028. [DOI] [Google Scholar]
- [167].Xia W, Shan H, Wang G, and Zhang Y, “Physics-/model-based and data-driven methods for low-dose computed tomography: A survey,” IEEE Signal Processing Magazine, vol. 40, no. 2, pp. 89–100, 2023. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [168].Eguizabal A, Oktem Ö, and Persson M, “A deep learning one-step solution to material image reconstruction in photon counting spectral CT,” in Medical Imaging 2022: Physics of Medical Imaging, Zhao W and Yu L, Eds., International Society for Optics and Photonics, vol. 12031, 2022, 120310Y. DOI: 10.1117/12.2612426. [DOI] [Google Scholar]
- [169].Ström E, Persson M, Eguizabal A, and Öktem OO, “Photon-counting CT reconstruction with a learned forward operator,” IEEE Transactions on Computational Imaging, pp. 1–16, 2022, ISSN: 2333–9403. DOI: 10.1109/TCI.2022.3183405. [DOI] [Google Scholar]
- [170].Su T, Sun X, Yang J, Mi D, Zhang Y, Wu H, Fang S, Chen Y, Zheng H, Liang D, and Ge Y, “DIRECT-Net: A unified mutualdomain material decomposition network for quantitative dual-energy CT imaging,” Med Phys, vol. 49, no. 2, pp. 917–934, 2022. DOI: 10.1002/mp.15413. [DOI] [PubMed] [Google Scholar]
- [171].Zhu J, Su T, Zhang X, Yang J, Mi D, Zhang Y, Gao X, Zheng H, Liang D, and Ge Y, “Feasibility study of three-material decomposition in dual-energy cone-beam CT imaging with deep learning,” Phys Med Biol, vol. 67, no. 14, p. 145012, Jul. 2022. DOI: 10.1088/1361-6560/ac7b09. [DOI] [PubMed] [Google Scholar]
- [172].Zeegers M, Kadu A, van Leeuwen T, and Batenburg K, “ADJUST: A dictionary-based joint reconstruction and unmixing method for spectral tomography,” Inverse Probl, vol. 38, no. 12, p. 125002, Oct. 2022. DOI: 10.1088/1361-6420/ac932e. [DOI] [Google Scholar]
- [173].Fang W, Wu D, Kim K, Kalra M, Singh R, Li L, and Li Q, “Iterative material decomposition for spectral CT using self-supervised Noise2Noise prior,” Phys Med Biol, vol. 66, no. 15, p. 155013, Jul. 2021. DOI: 10.1088/1361-6560/ac0afd. [DOI] [PubMed] [Google Scholar]
- [174].Sidky E and Pan X, “Report on the AAPM deep-learning spectral CT Grand Challenge,” Medical Physics, 2023. DOI: 10.1002/mp.16363. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [175].Vliegenthart R, Fouras A, Jacobs C, and Papanikolaou N, “Innovations in thoracic imaging: CT, radiomics, AI and x-ray velocimetry,” Respirology, vol. 27, no. 10, pp. 818–833, 2022. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [176].Virarkar MK, Vulasala SSR, Gupta AV, Gopireddy D, Kumar S, Hernandez M, Lall C, and Bhosale P, “Virtual non-contrast imaging in the abdomen and the pelvis: An overview,” in Seminars in Ultrasound, CT and MRI, Elsevier, vol. 43, 2022, pp. 293–310. [DOI] [PubMed] [Google Scholar]
- [177].Feng R, Rundle D, and Wang G, “Neural-networks-based photon-counting data correction: Pulse pileup effect,” arXiv preprint arXiv:1804.10980, 2018. [Google Scholar]
- [178].Fang Z, Li S, Hu W, Zhang J, and Chen S, “Spectral distortion correction of photon counting detector based on neural network,” in MIPPR 2019: Remote Sensing Image Processing, Geographic Information Systems, and Other Applications, SPIE, vol. 11432, 2020, pp. 333–340. [Google Scholar]
- [179].Shi Z, Wang Z, Kong F, Cao Q, Wang N, and Qi J, “An x-ray crosstalk correction method using FCNN for a novel energy resolving scheme in spectral CT,” Physics in Medicine & Biology, vol. 66, no. 11, p. 115022, 2021. [DOI] [PubMed] [Google Scholar]
- [180].Li M, Rundle DS, and Wang G, “X-ray photon-counting data correction through deep learning,” arXiv preprint arXiv:2007.03119, 2020. [Google Scholar]
- [181].Holbrook MD, Clark DP, and Badea CT, “Deep learning based spectral distortion correction and decomposition for photon counting CT using calibration provided by an energy integrated detector,” in Medical Imaging 2021: Physics of Medical Imaging, SPIE, vol. 11595, 2021, pp. 505–513. [Google Scholar]
- [182].Ma X-K, Huang H-Q, Huang B-R, Shen Z-W, Wang Q-T, Xiao Y-Y, Zhong C-L, Xin H, Sun P, Jiang K-M, et al. , “Xray spectra correction based on deep learning CNN-LSTM model,” Measurement, vol. 199, p. 111510, 2022. [Google Scholar]
- [183].Smith A, Atlas J, and Atharifard A, “Correcting charge sharing distortions in photon counting detectors utilising a spatial-temporal CNN,” in Image and Vision Computing: 37th International Conference, IVCNZ; 2022, Auckland, New Zealand, November 24–25, 2022, Revised Selected Papers, Springer, 2023, pp. 75–90. [Google Scholar]
- [184].Erath J, Magonov J, Maier J, Fournié E, Petersilka M, Stierstorfer K, and Kachelrieß M, “Deep scatter estimation for coarse anti-scatter grids as used in photon-counting CT,” in 7th International Conference on Image Formation in X-Ray Computed Tomography, SPIE, vol. 12304, 2022, pp. 112–117. [Google Scholar]
- [185].Hein D, Liappis K, Eguizabal A, and Persson M, “Deep learning ring artifact correction in photon-counting spectral CT with perceptual loss,” in 7th International Conference on Image Formation in X-Ray Computed Tomography, SPIE, vol. 12304, 2022, pp. 615–621. [Google Scholar]
- [186].Getzin M, Li M, Rundle DS, Butler AP, and Wang G, “Non-uniformity correction for MARS photon-counting detectors,” in 15th International Meeting on Fully Three-Dimensional Image Reconstruction in Radiology and Nuclear Medicine, SPIE, vol. 11072, 2019, pp. 503–507. [Google Scholar]
- [187].Fang W, Li L, and Chen Z, “Removing ring artefacts for photon-counting detectors using neural networks in different domains,” IEEE Access, vol. 8, pp. 42447–42457, 2020. [Google Scholar]
- [188].Kachelriess M, Watzke O, and Kalender WA, “Generalized multidimensional adaptive filtering for conventional and spiral single-slice, multi-slice, and cone-beam CT,” Medical physics, vol. 28, no. 4, pp. 475–490, 2001. [DOI] [PubMed] [Google Scholar]
- [189].Manduca A, Yu L, Trzasko JD, Khaylova N, Kofler JM, McCollough CM, and Fletcher JG, “Projection space denoising with bilateral filtering and CT noise modeling for dose reduction in CT,” Medical physics, vol. 36, no. 11, pp. 4911–4919, 2009. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [190].Li Z, Yu L, Trzasko JD, Lake DS, Blezek DJ, Fletcher JG, McCollough CH, and Manduca A, “Adaptive nonlocal means filtering based on local noise level for CT denoising,” Medical physics, vol. 41, no. 1, p. 011908, 2014. [DOI] [PubMed] [Google Scholar]
- [191].Tian Z, Jia X, Yuan K, Pan T, and Jiang SB, “Low-dose CT reconstruction via edge-preserving total variation regularization,” Physics in Medicine & Biology, vol. 56, no. 18, p. 5949, 2011. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [192].Mairal J, Bach F, Ponce J, and Sapiro G, “Online dictionary learning for sparse coding,” in Proceedings of the 26th annual international conference on machine learning, 2009, pp. 689–696. [Google Scholar]
- [193].Jiao C, Wang D, Lu H, Zhang Z, and Liang JZ, “Multiscale noise reduction on low-dose CT sinogram by stationary wavelet transform,” in 2008 IEEE Nuclear Science Symposium Conference Record, IEEE, 2008, pp. 5339–5344. [Google Scholar]
- [194].Makinen Y, Azzari L, and Foi A, “Collaborative filtering of correlated noise: Exact transform-domain variance for improved shrinkage and patch matching,” IEEE Transactions on Image Processing, vol. 29, pp. 8339–8354, 2020. [DOI] [PubMed] [Google Scholar]
- [195].Diwakar M and Kumar M, “A review on CT image noise and its denoising,” Biomedical Signal Processing and Control, vol. 42, pp. 73–88, 2018. [Google Scholar]
- [196].Chen H, Zhang Y, Kalra MK, Lin F, Chen Y, Liao P, Zhou J, and Wang G, “Low-dose CT with a residual encoder-decoder convolutional neural network,” IEEE transactions on medical imaging, vol. 36, no. 12, pp. 2524–2535, 2017. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [197].Kang E, Min J, and Ye JC, “A deep convolutional neural network using directional wavelets for low-dose x-ray CT reconstruction,” Medical physics, vol. 44, no. 10, e360–e375, 2017. [DOI] [PubMed] [Google Scholar]
- [198].Du W, Chen H, Wu Z, Sun H, Liao P, and Zhang Y, “Stacked competitive networks for noise reduction in low-dose CT,” PloS one, vol. 12, no. 12, e0190069, 2017. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [199].Shan H, Zhang Y, Yang Q, Kruger U, Kalra MK, Sun L, Cong W, and Wang G, “3-D convolutional encoder-decoder network for low-dose CT via transfer learning from a 2-D trained network,” IEEE transactions on medical imaging, vol. 37, no. 6, pp. 1522–1534, 2018. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [200].Gunduzalp D, Cengiz B, Unal MO, and Yildirim I, “3D U-NetR: Low dose computed tomography reconstruction via deep learning and 3 dimensional convolutions,” arXiv preprint arXiv:2105.14130, 2021. [Google Scholar]
- [201].Liu Y and Zhang Y, “Low-dose CT restoration via stacked sparse denoising autoencoders,” Neurocomputing, vol. 284, pp. 80–89, 2018. [Google Scholar]
- [202].Ming J, Yi B, Zhang Y, and Li H, “Low-dose CT image denoising using classification densely connected residual network,” KSII Transactions on Internet and Information Systems (TIIS), vol. 14, no. 6, pp. 2480–2496, 2020. [Google Scholar]
- [203].Yang Q, Yan P, Zhang Y, Yu H, Shi Y, Mou X, Kalra MK, Zhang Y, Sun L, and Wang G, “Low-dose CT image denoising using a generative adversarial network with wasserstein distance and perceptual loss,” IEEE transactions on medical imaging, vol. 37, no. 6, pp. 1348–1357, 2018. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [204].Wolterink JM, Leiner T, Viergever MA, and Išgum I, “Generative adversarial networks for noise reduction in low-dose CT,” IEEE transactions on medical imaging, vol. 36, no. 12, pp. 2536–2545, 2017. [DOI] [PubMed] [Google Scholar]
- [205].Tang C, Li J, Wang L, Li Z, Jiang L, Cai A, Zhang W, Liang N, Li L, Yan B, et al. , “Unpaired low-dose CT denoising network based on cycle-consistent generative adversarial network with prior image information,” Computational and mathematical methods in medicine, vol. 2019, 2019. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [206].You C, Li G, Zhang Y, Zhang X, Shan H, Li M, Ju S, Zhao Z, Zhang Z, Cong W, et al. , “CT super-resolution GAN constrained by the identical, residual, and cycle learning ensemble (GAN-CIRCLE),” IEEE transactions on medical imaging, vol. 39, no. 1, pp. 188–203, 2019. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [207].Yang W, Zhang H, Yang J, Wu J, Yin X, Chen Y, Shu H, Luo L, Coatrieux G, Gui Z, et al. , “Improving low-dose CT image using residual convolutional network,” Ieee Access, vol. 5, pp. 24698–24705, 2017. [Google Scholar]
- [208].Yin X, Zhao Q, Liu J, Yang W, Yang J, Quan G, Chen Y, Shu H, Luo L, and Coatrieux J-L, “Domain progressive 3D residual convolution network to improve low-dose CT imaging,” IEEE transactions on medical imaging, vol. 38, no. 12, pp. 2903–2913, 2019. [DOI] [PubMed] [Google Scholar]
- [209].Nishii T, Kobayashi T, Tanaka H, Kotoku A, Ohta Y, Morita Y, Umehara K, Ota J, Horinouchi H, Ishida T, et al. , “Deep learning–based post hoc CT denoising for myocardial delayed enhancement,” Radiology, vol. 305, no. 1, pp. 82–91, 2022. [DOI] [PubMed] [Google Scholar]
- [210].Fan F, Shan H, Kalra MK, Singh R, Qian G, Getzin M, Teng Y, Hahn J, and Wang G, “Quadratic autoencoder (q-ae) for low-dose CT denoising,” IEEE transactions on medical imaging, vol. 39, no. 6, pp. 2035–2050, 2019. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [211].Wang D, Wu Z, and Yu H, “Ted-net: Convolution-free t2t vision transformer-based encoder-decoder dilation network for low-dose CT denoising,” in Machine Learning in Medical Imaging: 12th International Workshop, MLMI 2021, Held in Conjunction with MICCAI 2021, Strasbourg, France, September 27, 2021, Proceedings 12, Springer, 2021, pp. 416–425. [Google Scholar]
- [212].Xia W, Lyu Q, and Wang G, “Low-dose CT using denoising diffusion probabilistic model for 20x speedup,” arXiv preprint arXiv:2209.15136, 2022. [Google Scholar]
- [213].Wang Z, Bovik AC, Sheikh HR, and Simoncelli EP, “Image quality assessment: From error visibility to structural similarity,” IEEE transactions on image processing, vol. 13, no. 4, pp. 600–612, 2004. [DOI] [PubMed] [Google Scholar]
- [214].Zhang H, Liu B, Yu H, and Dong B, “Metainv-net: Meta inversion network for sparse view CT image reconstruction,” IEEE Transactions on Medical Imaging, vol. 40, no. 2, pp. 621–634, 2020. [DOI] [PubMed] [Google Scholar]
- [215].Unal MO, Ertas M, and Yildirim I, “An unsupervised reconstruction method for low-dose ct using deep generative regularization prior,” Biomedical Signal Processing and Control, vol. 75, p. 103598, 2022. [Google Scholar]
- [216].Goodfellow I, Pouget-Abadie J, Mirza M, Xu B, Warde-Farley D, Ozair S, and Bengio Y, “Generative adversarial networks, 1–9,” arXiv preprint arXiv:1406.2661, 2014. [Google Scholar]
- [217].Feng Z, Li Z, Cai A, Li L, Yan B, and Tong L, “A preliminary study on projection denoising for low-dose CT imaging using modified dual-domain u-net,” in 2020 3rd International Conference on Artificial Intelligence and Big Data (ICAIBD), IEEE, 2020, pp. 223–226. [Google Scholar]
- [218].Chen M, Pu Y-F, and Bai Y-C, “Low-dose CT image denoising using residual convolutional network with fractional tv loss,” Neurocomputing, vol. 452, pp. 510–520, 2021. [Google Scholar]
- [219].Ledig C, Theis L, Huszar F, Caballero J, Cunningham A, Acosta A, Aitken A, Tejani A, Totz J, Wang Z, et al. , “Photorealistic single image super-resolution using a generative adversarial network,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2017, pp. 4681–4690. [Google Scholar]
- [220].Shan H, Padole A, Homayounieh F, Kruger U, Khera RD, Nitiwarangkul C, Kalra MK, and Wang G, “Competitive performance of a modularized deep neural network compared to commercial algorithms for low-dose CT image reconstruction,” Nature Machine Intelligence, vol. 1, no. 6, pp. 269–276, 2019. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [221].Zhu J-Y, Park T, Isola P, and Efros AA, “Unpaired image-to-image translation using cycle-consistent adversarial networks,” in Proceedings of the IEEE international conference on computer vision, 2017, pp. 2223–2232. [Google Scholar]
- [222].Kang E, Koo HJ, Yang DH, Seo JB, and Ye JC, “Cycle-consistent adversarial denoising network for multiphase coronary CT angiography,” Medical physics, vol. 46, no. 2, pp. 550–562, 2019. [DOI] [PubMed] [Google Scholar]
- [223].Kulathilake KSH, Abdullah NA, Sabri AQM, and Lai KW, “A review on deep learning approaches for low-dose computed tomography restoration,” Complex & Intelligent Systems, pp. 1–33, 2021. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [224].Immonen E, Wong J, Nieminen M, Kekkonen L, Roine S, Törnroos S, Lanca L, Guan F, and Metsälä E, “The use of deep learning towards dose optimization in low-dose computed tomography: A scoping review,” Radiography, vol. 28, no. 1, pp. 208–214, 2022. [DOI] [PubMed] [Google Scholar]
- [225].Park HS, Baek J, You SK, Choi JK, and Seo JK, “Unpaired image denoising using a generative adversarial network in x-ray CT,” IEEE Access, vol. 7, pp. 110414–110425, 2019. [Google Scholar]
- [226].Kim K, Soltanayev S, and Chun SY, “Unsupervised training of denoisers for low-dose CT reconstruction without full-dose ground truth,” IEEE Journal of Selected Topics in Signal Processing, vol. 14, no. 6, pp. 1112–1125, 2020. [Google Scholar]
- [227].Lehtinen J, Munkberg J, Hasselgren J, Laine S, Karras T, Aittala M, and Aila T, “Noise2noise: Learning image restoration without clean data,” arXiv preprint arXiv:1803.04189, 2018. [Google Scholar]
- [228].Hasan AM, Mohebbian MR, Wahid KA, and Babyn P, “Hybrid-collaborative noise2noise denoiser for low-dose CT images,” IEEE Transactions on Radiation and Plasma Medical Sciences, vol. 5, no. 2, pp. 235–244, 2020. [Google Scholar]
- [229].Won D, Jung E, An S, Chikontwe P, and Park SH, “Self-supervised learning based CT denoising using pseudo-CT image pairs,” arXiv preprint arXiv:2104.02326, 2021. [Google Scholar]
- [230].Yuan N, Zhou J, and Qi J, “Half2half: Deep neural network based CT image denoising without independent reference data,” Physics in Medicine & Biology, vol. 65, no. 21, p. 215020, 2020. [DOI] [PubMed] [Google Scholar]
- [231].Zhang C, Chang S, Bai T, and Chen X, “S2MS: Self-supervised learning driven multi-spectral CT image enhancement,” in 7th International Conference on Image Formation in X-Ray Computed Tomography, SPIE, vol. 12304, 2022, pp. 473–479. [Google Scholar]
- [232].Zhang Z, Liang X, Zhao W, and Xing L, “Noise2context: Context-assisted learning 3d thin-layer for low-dose CT,” Medical Physics, vol. 48, no. 10, pp. 5794–5803, 2021. [DOI] [PubMed] [Google Scholar]
- [233].Wu D, Kim K, and Li Q, “Low-dose CT reconstruction with noise2noise network and testing-time fine-tuning,” Medical Physics, vol. 48, no. 12, pp. 7657–7672, 2021. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [234].Hendriksen AA, Pelt DM, and Batenburg KJ, “Noise2inverse: Self-supervised deep convolutional denoising for tomography,” IEEE Transactions on Computational Imaging, vol. 6, pp. 1320–1335, 2020. [Google Scholar]
- [235].Niu C, Li M, Fan F, Wu W, Guo X, Lyu Q, and Wang G, “Noise suppression with similarity-based self-supervised deep learning,” IEEE Transactions on Medical Imaging, 2022. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [236].Unal MO, Ertas M, and Yildirim I, “Self-supervised training for low-dose CT reconstruction,” in 2021 IEEE 18th International Symposium on Biomedical Imaging (ISBI), IEEE, 2021, pp. 69–72. [Google Scholar]
- [237].Elhamiasl M and Nuyts J, “Simulating lower-dose scans from an available CT scan,” in 15th International Meeting on Fully Three-Dimensional Image Reconstruction in Radiology and Nuclear Medicine, SPIE, vol. 11072, 2019, pp. 153–157. [Google Scholar]
- [238].Wu W, Hu D, Niu C, Broeke LV, Butler AP, Cao P, Atlas J, Chernoglazov A, Vardhanabhuti V, and Wang G, “Deep learning based spectral CT imaging,” Neural Networks, vol. 144, pp. 342–358, 2021. [DOI] [PubMed] [Google Scholar]
- [239].Evans CJ, Li M, Niu C, Wang G, and Roeder RK, “Effects of image denoising on quantitative material decomposition in photon-counting spectral computed tomography,” in Medical Imaging 2022: Physics of Medical Imaging, SPIE, vol. 12031, 2022, pp. 748–755. [Google Scholar]
- [240].Noda Y, Kaga T, Kawai N, Miyoshi T, Kawada H, Hyodo F, Kambadakone A, and Matsuo M, “Low-dose whole-body CT using deep learning image reconstruction: Image quality and lesion detection,” The British Journal of Radiology, vol. 94, no. 1121, p. 20201329, 2021. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [241].Parakh A, Cao J, Pierce TT, Blake MA, Savage CA, and Kambadakone AR, “Sinogram-based deep learning image reconstruction technique in abdominal CT: Image quality considerations,” European Radiology, vol. 31, no. 11, pp. 8342–8353, 2021. [DOI] [PubMed] [Google Scholar]
- [242].Fair E, Profio M, Kulkarni N, Laviolette PS, Barnes B, Bobholz S, Levenhagen M, Ausman R, Griffin MO, Duvnjak P, et al. , “Image quality evaluation in dual-energy CT of the chest, abdomen, and pelvis in obese patients with deep learning image reconstruction,” Journal of Computer Assisted Tomography, vol. 46, no. 4, pp. 604–611, 2022. [DOI] [PubMed] [Google Scholar]
- [243].Noda Y, Kawai N, Nagata S, Nakamura F, Mori T, Miyoshi T, Suzuki R, Kitahara F, Kato H, Hyodo F, et al. , “Deep learning image reconstruction algorithm for pancreatic protocol dual-energy computed tomography: Image quality and quantification of iodine concentration,” European Radiology, vol. 32, no. 1, pp. 384–394, 2022. [DOI] [PubMed] [Google Scholar]
- [244].Fukutomi A, Sofue K, Ueshima E, Negi N, Ueno Y, Tsujita Y, Yabe S, Yamaguchi T, Shimada R, Kusaka A, et al. , “Deep learning image reconstruction to improve accuracy of iodine quantification and image quality in dual-energy CT of the abdomen: A phantom and clinical study,” European Radiology, vol. 33, no. 2, pp. 1388–1399, 2023. [DOI] [PubMed] [Google Scholar]
- [245].Kojima T, Shirasaka T, Kondo M, Kato T, Nishie A, Ishigami K, and Yabuuchi H, “A novel fast kilovoltage switching dual-energy CT with deep learning: Accuracy of CT number on virtual monochromatic imaging and iodine quantification,” Physica Medica, vol. 81, pp. 253–261, 2021. [DOI] [PubMed] [Google Scholar]
- [246].Szczykutowicz TP, Toia GV, Dhanantwari A, and Nett B, “A review of deep learning CT reconstruction: Concepts, limitations, and promise in clinical practice,” Current Radiology Reports, vol. 10, no. 9, pp. 101–115, 2022. [Google Scholar]
- [247].Mustafa W, Kehl C, Olsen UL, Gregersen SKS, Malmgren-Hansen D, Kehres J, and Dahl AB, “Sparse-view spectral CT reconstruction using deep learning,” arXiv preprint arXiv:2011.14842, 2020. [Google Scholar]
- [248].Lee M, Kim H, Cho H-M, and Kim H-J, “Ultra-low-dose spectral CT based on a multi-level wavelet convolutional neural network,” Journal of Digital Imaging, vol. 34, pp. 1359–1375, 2021. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [249].Zhang Y, Hu D, Yan Z, Zhao Q, Quan G, Luo S, Zhang Y, and Chen Y, “TIME-Net: Transformer-integrated multi-encoder network for limited-angle artifact removal in dual-energy CBCT,” Medical Image Analysis, vol. 83, p. 102650, 2023. [DOI] [PubMed] [Google Scholar]
- [250].Liu C-K and Huang H-M, “Unsupervised deep learning based image outpainting for dual-source, dual-energy computed tomography,” Radiation Physics and Chemistry, vol. 188, p. 109635, 2021. [Google Scholar]
- [251].Schwartz FR, Clark DP, Ding Y, Ramirez-Giraldo JC, Badea CT, and Marin D, “Evaluating renal lesions using deep-learning based extension of dual-energy fov in dual-source CT—a retrospective pilot study,” European journal of radiology, vol. 139, p. 109734, 2021. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [252].Zhao W, Lv T, Gao P, Shen L, Dai X, Cheng K, Jia M, Chen Y, and Xing L, “A deep learning approach for dual-energy CT imaging using a single-energy CT data,” in 15th International Meeting on Fully Three-Dimensional Image Reconstruction in Radiology and Nuclear Medicine, SPIE, vol. 11072, 2019, pp. 355–359. [Google Scholar]
- [253].Lyu T, Zhao W, Zhu Y, Wu Z, Zhang Y, Chen Y, Luo L, Li S, and Xing L, “Estimating dual-energy CT imaging from single-energy CT data with material decomposition convolutional neural network,” Medical image analysis, vol. 70, p. 102001, 2021. [DOI] [PubMed] [Google Scholar]
- [254].Yu L, Christner JA, Leng S, Wang J, Fletcher JG, and McCollough CH, “Virtual monochromatic imaging in dual-source dual-energy CT: Radiation dose and image quality,” Medical physics, vol. 38, no. 12, pp. 6371–6379, 2011. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [255].Cong W, Xi Y, Fitzgerald P, De Man B, and Wang G, “Virtual monoenergetic CT imaging via deep learning,” Patterns, vol. 1, no. 8, p. 100128, 2020. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [256].Cong W, Xi Y, De Man B, and Wang G, “Monochromatic image reconstruction via machine learning,” Machine learning: science and technology, vol. 2, no. 2, p. 025032, 2021. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [257].Kawahara D, Ozawa S, Kimura T, and Nagata Y, “Image synthesis of monoenergetic CT image in dual-energy CT using kilovoltage CT with deep convolutional generative adversarial networks,” Journal of Applied Clinical Medical Physics, vol. 22, no. 4, pp. 184–192, 2021. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [258].Koike Y, Ohira S, Teraoka Y, Matsumi A, Imai Y, Akino Y, Miyazaki M, Nakamura S, Konishi K, Tanigawa N, et al. , “Pseudo low-energy monochromatic imaging of head and neck cancers: Deep learning image reconstruction with dual-energy CT,” International Journal of Computer Assisted Radiology and Surgery, vol. 17, no. 7, pp. 1271–1279, 2022. [DOI] [PubMed] [Google Scholar]
- [259].Fink MA, Seibold C, Kauczor H-U, Stiefelhagen R, and Kleesiek J, “Jointly optimized deep neural networks to synthesize monoenergetic images from single-energy CT angiography for improving classification of pulmonary embolism,” Diagnostics, vol. 12, no. 5, p. 1224, 2022. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [260].Sato M, Ichikawa Y, Domae K, Yoshikawa K, Kanii Y, Yamazaki A, Nagasawa N, Nagata M, Ishida M, and Sakuma H, “Deep learning image reconstruction for improving image quality of contrast-enhanced dual-energy CT in abdomen,” European Radiology, vol. 32, no. 8, pp. 5499–5507, 2022. [DOI] [PubMed] [Google Scholar]
- [261].Xu JJ, Lönn L, Budtz-Jørgensen E, Hansen KL, and Ulriksen PS, “Quantitative and qualitative assessments of deep learning image reconstruction in low-kev virtual monoenergetic dual-energy CT,” European Radiology, vol. 32, no. 10, pp. 7098–7107, 2022. [DOI] [PubMed] [Google Scholar]
- [262].Jiang C, Jin D, Liu Z, Zhang Y, Ni M, and Yuan H, “Deep learning image reconstruction algorithm for carotid dual-energy computed tomography angiography: Evaluation of image quality and diagnostic performance,” Insights into Imaging, vol. 13, no. 1, p. 182, 2022. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [263].Lee T, Lee JM, Yoon JH, Joo I, Bae JS, Yoo J, Kim JH, Ahn C, and Kim JH, “Deep learning-based image reconstruction of 40-kev virtual monoenergetic images of dual-energy CT for the assessment of hypoenhancing hepatic metastasis,” European Radiology, vol. 32, no. 9, pp. 6407–6417, 2022. [DOI] [PubMed] [Google Scholar]
- [264].Seo JY, Joo I, Yoon JH, Kang HJ, Kim S, Kim JH, Ahn C, and Lee JM, “Deep learning-based reconstruction of virtual monoenergetic images of kvp-switching dual energy CT for evaluation of hypervascular liver lesions: Comparison with standard reconstruction technique,” European Journal of Radiology, vol. 154, p. 110390, 2022. [DOI] [PubMed] [Google Scholar]
- [265].Haubold J, Hosch R, Umutlu L, Wetter A, Haubold P, Radbruch A, Forsting M, Nensa F, and Koitka S, “Contrast agent dose reduction in computed tomography with deep learning using a conditional generative adversarial network,” European Radiology, vol. 31, pp. 6087–6095, 2021. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [266].Noda Y, Kawai N, Kawamura T, Kobori A, Miyase R, Iwashima K, Kaga T, Miyoshi T, Hyodo F, Kato H, et al. , “Radiation and iodine dose reduced thoraco-abdomino-pelvic dual-energy CT at 40 kev reconstructed with deep learning image reconstruction,” The British Journal of Radiology, vol. 95, no. 1134, p. 20211163, 2022. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [267].Poirot MG, Bergmans RH, Thomson BR, Jolink FC, Moum SJ, Gonzalez RG, Lev MH, Tan CO, and Gupta R, “Physics-informed deep learning for dual-energy computed tomography image processing,” Scientific reports, vol. 9, no. 1, p. 17709, 2019. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [268].Kosmala A, Weng AM, Heidemeier A, Krauss B, Knop S, Bley TA, and Petritsch B, “Multiple myeloma and dual-energy CT: Diagnostic accuracy of virtual noncalcium technique for detection of bone marrow infiltration of the spine and pelvis,” Radiology, vol. 286, no. 1, pp. 205–213, 2018. [DOI] [PubMed] [Google Scholar]
- [269].Kosmala A, Weng AM, Krauss B, Knop S, Bley TA, and Petritsch B, “Dual-energy CT of the bone marrow in multiple myeloma: Diagnostic accuracy for quantitative differentiation of infiltration patterns,” European radiology, vol. 28, pp. 5083–5090, 2018. [DOI] [PubMed] [Google Scholar]
- [270].Heinrich A, Schenkl S, Buckreus D, Güttler FV, and Teichgräber UK, “CT-based thermometry with virtual monoenergetic images by dual-energy of fat, muscle and bone using FBP, iterative and deep learning–based reconstruction,” European Radiology, vol. 32, pp. 424–431, 2022. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [271].Wang N, Li M, and Haverinen P, “Photon-counting computed tomography thermometry via material decomposition and machine learning,” Visual Computing for Industry, Biomedicine, and Art, vol. 6, no. 1, pp. 1–6, 2023. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [272].Zhu B, Liu JZ, Cauley SF, Rosen BR, and Rosen MS, “Image reconstruction by domain-transform manifold learning,” Nature, vol. 555, no. 7697, pp. 487–492, 2018. [DOI] [PubMed] [Google Scholar]
- [273].Chen H, Zhang Y, Chen Y, Zhang J, Zhang W, Sun H, Lv Y, Liao P, Zhou J, and Wang G, “Learn: Learned experts’ assessment-based reconstruction network for sparse-data ct,” IEEE transactions on medical imaging, vol. 37, no. 6, pp. 1333–1347, 2018. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [274].Li Y, Li K, Zhang C, Montoya J, and Chen G-H, “Learning to reconstruct computed tomography images directly from sinogram data under a variety of data acquisition conditions,” IEEE transactions on medical imaging, vol. 38, no. 10, pp. 2469–2481, 2019. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [275].Thaler F, Hammernik K, Payer C, Urschler M, and Štern D, “Sparse-view ct reconstruction using wasserstein gans,” in Machine Learning for Medical Image Reconstruction: First International Workshop, MLMIR 2018, Held in Conjunction with MICCAI 2018, Granada, Spain, September 16, 2018, Proceedings, Springer, 2018, pp. 75–82. [Google Scholar]
- [276].Xie H, Shan H, and Wang G, “Deep encoder-decoder adversarial reconstruction (DEAR) network for 3D CT from few-view data,” Bioengineering, vol. 6, no. 4, p. 111, 2019. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [277].Jiao F, Gui Z, Li K, Shangguang H, Wang Y, Liu Y, and Zhang P, “A dual-domain cnn-based network for ct reconstruction,” IEEE Access, vol. 9, pp. 71091–71103, 2021. [Google Scholar]
- [278].Kandarpa V, Perelli A, Bousse A, and Visvikis D, “LRR-CED: Low-resolution reconstruction-aware convolutional encoder–decoder network for direct sparse-view CT image reconstruction,” Physics in Medicine & Biology, vol. 67, no. 15, p. 155007, 2022. [DOI] [PubMed] [Google Scholar]
- [279].Jang S, Kim S, Kim M, Son K, Lee K-Y, and Ra JB, “Head motion correction based on filtered backprojection in helical CT scanning,” IEEE transactions on medical imaging, vol. 39, no. 5, pp. 1636–1645, 2019. [DOI] [PubMed] [Google Scholar]
- [280].Sun T, Kim J-H, Fulton R, and Nuyts J, “An iterative projection-based motion estimation and compensation scheme for head x-ray CT,” Medical physics, vol. 43, no. 10, pp. 5705–5716, 2016. [DOI] [PubMed] [Google Scholar]
- [281].Sisniega A, Stayman JW, Yorkston J, Siewerdsen J, and Zbijewski W, “Motion compensation in extremity cone-beam ct using a penalized image sharpness criterion,” Physics in Medicine & Biology, vol. 62, no. 9, p. 3712, 2017. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [282].Li M, Lowe C, Butler A, Butler P, and Wang G, “Motion correction via locally linear embedding for helical photon-counting CT,” in 7th International Conference on Image Formation in X-Ray Computed Tomography, SPIE, vol. 12304, 2022, pp. 559–567. DOI: 10.1117/12.2646714. [DOI] [Google Scholar]
- [283].Li M, Bohacova J, Uher J, Cong W, Rubinstein J, and Wang G, “Motion correction for robot-based x-ray photon-counting CT at ultrahigh resolution,” in Developments in X-Ray Tomography XIV, SPIE, vol. 12242, 2022, pp. 172–181. DOI: 10.1117/12.2635263. [DOI] [Google Scholar]
- [284].Dhariwal P and Nichol A, “Diffusion models beat gans on image synthesis,” Advances in Neural Information Processing Systems, vol. 34, pp. 8780–8794, 2021. [Google Scholar]
- [285].Kazerouni A, Aghdam EK, Heidari M, Azad R, Fayyaz M, Hacihaliloglu I, and Merhof D, “Diffusion models in medical imaging: A comprehensive survey,” Medical Image Analysis, p. 102846, 2023. [DOI] [PubMed] [Google Scholar]
- [286].Hein D and Persson M, “Generation of photon-counting spectral ct images using a score-based diffusion model,” in International Conference on Fully Three-Dimensional Image Reconstruction in Radiology and Nuclear Medicine, Stony Brook, NY, USA, Jul. 2023. [Google Scholar]
- [287].Ho J, Jain A, and Abbeel P, “Denoising diffusion probabilistic models,” Advances in Neural Information Processing Systems, vol. 33, pp. 6840–6851, 2020. [Google Scholar]
- [288].Chung H, Kim J, Mccann MT, Klasky ML, and Ye JC, “Diffusion posterior sampling for general noisy inverse problems,” arXiv preprint arXiv:2209.14687, 2022. [Google Scholar]
- [289].Wang G, Badal A, Jia X, Maltz JS, Mueller K, Myers KJ, Niu C, Vannier M, Yan P, Yu Z, et al. , “Development of metaverse for intelligent healthcare,” Nature Machine Intelligence, pp. 1–8, 2022. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [290].Shi Y and Wang G, “Conversion of the Mayo LDCT data to synthetic equivalent through the diffusion model for training denoising networks with a theoretically perfect privacy,” arXiv preprint arXiv:2301.06604, 2023. [Google Scholar]
- [291].Cong J, Liu B, Neuendorffer S, Noguera J, Vissers K, and Zhang Z, “High-level synthesis for FPGAs: From prototyping to deployment,” IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems, vol. 30, no. 4, pp. 473–491, 2011. [Google Scholar]
- [292].Li M, Fang Z, Cong W, Niu C, Wu W, Uher J, Bennett J, Rubinstein JT, and Wang G, “Clinical micro-CT empowered by interior tomography, robotic scanning, and deep learning,” IEEE Access, vol. 8, pp. 229018–229032, 2020. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [293].Wang G, “PET-enabled dual-energy CT: Image reconstruction and a proof-of-concept computer simulation study,” Physics in Medicine and Biology, vol. 65, no. 24, p. 245028, 2020. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [294].Rezaei A, Defrise M, Bal G, Conti M, Watson C, and Nuyts J, “Simultaneous reconstruction of activity and attenuation in time-of-flight PET,” IEEE Transactions on Medical Imaging, vol. 31, no. 12, pp. 2224–2233, 2012. [DOI] [PubMed] [Google Scholar]
- [295].Wang G and Qi J, “PET image reconstruction using kernel method,” IEEE Transactions on Medical Imaging, vol. 34, no. 1, pp. 61–71, 2015. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [296].Li S and Wang G, “Modified kernel MLAA using autoencoder for PET-enabled dual-energy CT,” Philosophical Transactions of the Royal Society A, vol. 379, no. 2204, p. 20200204, 2021. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [297].Li S and Wang G, “Neural MLAA for PET-enabled dual-energy CT imaging.,” Proc. SPIE Medical Imaging 2021: Physics of Medical Imaging, vol. 115951G, p. 20200204, 2021. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [298].Li S and Wang G, “Deep kernel representation for image reconstruction in PET.,” IEEE Transactions on Medical Imaging, vol. 41, no. 11, pp. 3029–3038, 2022. [DOI] [PMC free article] [PubMed] [Google Scholar]





