Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2018 Apr 21.
Published in final edited form as: Phys Med Biol. 2017 Apr 21;62(8):3284–3298. doi: 10.1088/1361-6560/aa6392

Assessment of Vectorial Total Variation Penalties on Realistic Dual-Energy CT Data

David S Rigie 1, Adrian A Sanchez 1, Patrick J La Rivière 1
PMCID: PMC5575889  NIHMSID: NIHMS898593  PMID: 28350547

Abstract

Vectorial extensions of total variation have recently been developed for regularizing the reconstruction and denoising of multi-channel images, such as those arising in spectral computed tomography. Early studies have focused mainly on simulated, piecewise-constant images whose structure may favor total-variation penalties. In the current manuscript, we apply vectorial total variation to real dual-energy CT data of a whole turkey in order to determine if the same benefits can be observed in more complex images with anatomically realistic textures. We consider the total nuclear variation (TVN) as well as another vectorial total variation based on the Frobenius norm (TVF) and standard channel-by-channel total variation (TVS). We performed a series of 3D TV denoising experiments comparing the three TV variants across a wide range of smoothness parameter settings, optimizing each regularizer according to a very-high-dose “ground truth” image. Consistent with the simulation studies, we find that both vectorial TV variants achieve a lower error than the channel-by-channel TV and are better able to suppress noise while preserving actual image features. In this real data study, the advantages are subtler than in the previous simulation study, although the TVN penalty is found to have clear advantages over either TVS or TVF when comparing material images formed from linear combinations of the denoised energy images.

1. Background

Due to the recent progress in dual- and multi-energy x-ray CT, there has been growing interest in developing effective regularizers for multi-channel image reconstruction and denoising. In multi-energy scanning modes, the total x-ray dose is subdivided among two or more energy channels, thus reducing the signal-to-noise ratio (SNR) of each one. Furthermore, a pre- or post-reconstruction material decomposition step is usually performed, which greatly amplifies noise [1]. Some of these deleterious effects may be partially mitigated by including strong image priors into the reconstruction or post-processing steps. Several different approaches appear in the literature, including methods based on robust principal component analysis [2], patch-based penalties [3], and vectorial extensions of the total variation (TV) [4, 5, 6]. These varied techniques share a common goal of leveraging redundant structure across image channels (e.g. shared edges) in order to improve noise suppression while mitigating feature loss.

Previously, we applied a vectorial extension of the total variation penalty, which has been referred to as the “total nuclear variation” (TVN), to spectral CT reconstruction [6]. For clarity, we will recap several salient points from that manuscript. TVN extends the total variation to multi-channel images by penalizing the nuclear norm of the Jacobian. The resulting functional encourages multi-channel images to have shared edges and gradient vectors that point in the same direction, while being insensitive to contrast or scale differences. Furthermore, TVN is convex and reduces to the usual total variation when there is only a single image channel. In [6], we described a flexible image reconstruction framework based on a data-constrained TV minimization approach that allowed us to directly assess the performance of TVN compared to a channel-by-channel application of the conventional TV. By jointly processing several image channels with edge-coupling regularization, improved reconstructions were demonstrated from the same noisy data in a series of simulation studies with the numerical XCAT phantom. Furthermore, the TVN was applied to both energy and material images with similar gains relative to the channel-by-channel TV. Overall, TVN allowed for a greater reduction in image noise with less edge blurring due to its ability to pool edge information across all image channels.

In this paper, we will build on those findings by investigating whether the same benefits can be demonstrated on more realistic CT data. In doing so, we aim to address several limitations of the previous study. Firstly, we will evaluate the TVN on images that have a realistic texture. Although the XCAT phantom utilized in the previous study is based closely on real CT images of a human cadaver, its contrast is piecewise. This could be problematic for evaluating any sort of TV regularization, since TV is known to favor this kind of structure [7]. Furthermore, one of the potential benefits of the TVN is that it promotes aligned gradient vectors across energy channels, but due to its piecewise nature, the XCAT phantom also trivially satisfies this condition. Other authors have applied a very similar regularizer to the joint reconstruction of PET-MR patient data with favorable results [8], but to our knowledge, no studies have reported on TVN for real multi-energy CT images.

To address the question of whether the previous findings are translatable to realistic multi-energy CT data, we will investigate the proposed regularizer on dual-energy images of a turkey acquired on a diagnostic, clinical scanner. While the previous study focused on image reconstruction, the TVN can also be applied to other image restoration problems [4, 5]. Since we did not have access to the projection data or geometry parameters of our scanner, we focus solely on denoising in this work. Because denoising is much less computationally expensive than reconstruction, this enabled us to conduct a thorough exploration of the parameter space, which was not feasible in the previous study. In addition to assessing the TVN on realistic data, we also address several other issues of practical interest. Firstly, in order to isolate the importance of edge-coupling and gradient alignment, we also study another multi-channel variant of the TV [9] that encourages joint sparsity of the gradient magnitude images, without explicit encouragement of aligned gradient vectors. Additionally, we demonstrate that the TVN can be efficiently extended to 3D image volumes, which is of practical relevance for reconstructing or denoising multi-detector and cone beam CT data.

2. Theory

2.1. Notation and definitions

In this section, we outline the notation that we will use for the remainder of the document. We adopt a somewhat unconventional set of indexing rules to maintain as much clarity as possible in describing algebraic operations on vectors with multiple dimensions of spatial and spectral information.

Single-Channel images

Consider a discretized 3D image uI where I = ℝM·N·P is a finite dimensional vector space equipped with an inner product

u,vI=i,j,ku(i,j,k)v(i,j,k),u,vI. (1)

We use the convention u(i, j, k) to refer to a particular voxel, where i, j, and k specify the row, column, and slice indices and u(i, j, k) ∈ ℝ is a discretized version of some continuous image function u(x, y, z). The integers M, N, and P specify the total number of rows, columns, and slices, respectively. The quantity ∇u is a vector in the space G = I × I, where the operator ∇u : IG represents a discrete approximation to the gradient. At each voxel location we define the quantity (∇u)(i, j, k) ∈ ℝ3 as

(u)(i,j,k)=((u)x(i,j,k)(u)y(i,j,k)(u)z(i,j,k)), (2)

where

(u)x(i,j,k)={u(i,j,k+1)-u(i,j,k)k<P0k=P (3)
(u)y(i,j,k)={u(i,j+1,k)-u(i,j,k)j<N0j=N (4)
(u)z(i,j,k)={u(i+1,j,k)-u(i,j,k)i<M0i=M. (5)

We also define an inner product in G,

ν,zG=α{x,y,z}i,j,kνα(i,j,k)zα(i,j,k),ν,zG. (6)

Note that we use bold font to indicate that each spatial location (i, j, k) maps to a vector of gradient values. Further, we will need a discrete divergence operator div z : GI, which is chosen to be the negative transpose of the gradient operator, defined by

u,zG=-u,divzI. (7)

Lastly, we define the mixed ℓ1/ℓ2-norm in G as

z1,2=i,j,kz(i,j,k)2zG, (8)

indicating that we take an ℓ1-norm over the spatial indices (i, j, k) and an ℓ2-norm of each 3-vector z(i, j, k). This mixed-norm notation is often used in the literature on sparse regression and occasionally to compactly define the isotropic total variation [10],

TV(u)=u1,2. (9)

Multi-channel images

Now, we consider a discrete image u ∈ 𝒥 with L spectral channels,

u(i,j,k)=(u1(i,j,k)u2(i,j,k)uL(i,j,k)), (10)

where 𝒥 = ℝL·M·N·P. The quantity u is a discretized version of some continuous vector field u(x, y, z), such as a multi-energy image volume. We also define an inner product in 𝒥,

u,vJ=i,j,ku(i,j,k)v(i,j,k)u,vJ, (11)

where the subscript ℓ denotes a particular image channel. Since u is the discrete analog of a vector field, we can also define the discrete Jacobian, Ju : 𝒥 → 𝒢, which generalizes the gradient operator to vector fields. In particular we have

(Ju)(i,j,k)=((u1)x(i,j,k)(u1)y(i,j,k)(u1)z(i,j,k)(u2)x(i,j,k)(u2)y(i,j,k)(u2)z(i,j,k)(uL)x(i,j,k)(uL)y(i,j,k)(uL)z(i,j,k)). (12)

At every voxel (i, j, k), the L × 3 sub-matrix (Ju) (i, j, k) fully characterizes the first-order derivatives of u, with each row consisting of the gradient vector of one of the L image channels. The quantity Ju is in vector space 𝒢 = 𝒥 × 𝒥 = ℝ(L×3)·(M·N·P) with an inner product

V,ZG=α{x,y,z}i,j,kVα(i,j,k)Zα(i,j,k)V,ZG. (13)

An element V ∈ 𝒢 is a discretized version of a tensor field V (x, y, z), so we use uppercase font to indicate that every spatial location (i, j, k) maps to a matrix. Again, we need an analog of the divergence operator Div Z : 𝒢 → 𝒥 that is the negative transpose of J. It is constructed to satisfy

Ju,ZG=-u,DivZJ. (14)

We define the mixed ℓ1/nuclear-norm as

Z1,=i,j,kZ(i,j,k)ZG, (15)

where ||Z(i, j, k)|| is referred to as the “nuclear norm” of matrix Z(i, j, k) and is equal to the sum of its singular values. Additionally, we define the mixed ℓ1/Frobenius norm as

Z1,F=i,j,kZ(i,j,k)FZG, (16)

where the Frobenius norm of a generic M × N matrix is given by

XF=i=1Mj=1NXij2. (17)

Denoising Model

In this section, we will describe a generic denoising model as well as the three variants of the TV that we have investigated. Our denoising model is specified by the following optimization problem:

argminu12u-g22+λTV(u), (18)

where g is the noisy image, λ controls the trade-off between data fidelity and smoothness, and TV refers generally to some variant of the total variation. The simplest way to generalize the TV to multi-channel images is to apply the standard isotropic TV separately to each energy channel; we refer to this as the channel-by-channel TV, denoted by TVS. Additionally, we can define a “vectorial” TV that simultaneously penalizes edges across all image channels, which is defined by the voxel-wise application of some matrix norm to the discrete Jacobian of the multi-channel image function. Two interesting choices are the Frobenius norm, which encourages joint sparsity in the image gradients, and the Nuclear norm, which encourages joint sparsity and directional alignment of gradient vectors [5, 6]. We denote these two vectorial TV variants by TVF and TVN, respectively. All three of these regularizers are summarized in Table 1.

Table 2.

Relative error and UQI (in parenthesis) of each VTV variant at its respective optimal λ value. These correspond to the minima of the error curves in Figure 1, ε(λopt).

Regularizer 80 kVp 140 kVp
TVS 5.501e-2 (0.9981) 3.101e-2 (0.9994)
TVF 4.820e-2 (0.9985) 3.113e-2 (0.9994)
TVN 3.876e-2 (0.9990) 3.027e-2 (0.9994)

Primal-Dual Saddlepoint Problem

As in [6], the denoising model can be expressed as a saddlepoint problem, which is given by

minumaxZJu,ZG-δK(Z)+12λu-g22, (19)

where δ𝒦 is a generic set indicator function defined by

δK(X){0XKXK. (20)

The set 𝒦 is equal to 𝒮, ℱ, or 𝒩 for TVS, TVF, and TVN, respectively. The sets 𝒮, ℱ, and 𝒩 are defined below:

S={ZG:Z(i,j,k)21i,j,k,} (21)
F={ZG:Z(i,j,k)F1i,j,k} (22)
N={ZG:σmax(Z(i,j,k))1i,j,k}. (23)

Algorithm Updates

As in [6], we solve this optimization problem using Chambolle and Pock’s first-order, primal-dual algorithm (CPPD) [11]. However, unlike the reconstruction problem described in [6], this saddle point problem is strongly convex in the primal variable u, with convexity constant λ−1, so we can use their accelerated algorithm. This version of CPPD achieves a convergence rate of O (1/N2) rather than O(1/N). The updates are given by Algorithm 1.

The primal and dual stepsizes τ and σ must be initialized so that τσ||J||2 ≤ 1, where ||J|| is the spectral radius (also the maximum singular value) of the Jacobian operator J. For 3D image arrays, It can be shown that J12 [12]. We find that convergence is always fastest when στ = 1/12; however there still exists freedom to choose the ratio of primal to dual stepsizes. The observed rate of convergence can be dramatically impacted by this choice, but we are unaware of any good method to automatically select these parameters. We have used τ = 0.05 for all of the results in this paper. We iterate until the primal-dual gap falls below 10−6, which we have found, empirically, to be a reasonable stopping point where no appreciable voxel value changes can be observed in the windows of interest.

Algorithm 1.

TV Denoising Update Equations

1: Initialize: σkτk = 1/12
2: repeat
3: Z(k+1) = Π𝒦 (Z(k) + σkJ ū(k))
4: u(k+1) = (λu(k) + τkg + λτkDiv Z(k+1))/(λ + τk)
5: θk=λ/(λ+2τk), τk = θkτk, σk = σkk
6: ū(k+1) = u(k+1) + θk (u(k+1) u(k))
7: until ▷ stop when convergence criteria met

Implementation of Projection Operators

The algorithm updates involve projecting the dual variable Z onto one of the convex sets, 𝒮, ℱ, or 𝒩, corresponding to TVS, TVF, or TVN. Pseudocode for each of these projection operators is given by Algorithm 2, Algorithm 3, and Algorithm 4, respectively. For TVS and TVF, the projection operators are very straightforward and only involve simple conditionals and vector normalization. The projection operator Π𝒩 for TVN is slightly more complicated and requires the computation of eigenvalues and eigenvectors of a real, symmetric 3×3 matrix for every image voxel. Luckily, analytic methods exist for diagonalizing 3×3 Hermitian matrices, so the computational burden is not too significant; we rely on the implementation described by Kopp [13]. We also note that for the special case of only two energy channels, there is an alternative implementation of Π𝒩 that only involves a 2 × 2 eigendecomposition, but we present this version because it generalizes to an arbitrary number of image channels, which is important for photon-counting spectral CT systems.

Algorithm 2.

Implementation of Projection Operator Π𝒮(Z)

1: for all i, j, k,do
2: xZ(i, j, k)
3: if ||x||2 ≥ 1 then
4:   xx/||x||2
5: end if
6: Zℓ(i, j, k) ← x
7: end for
Algorithm 3.

Implementation of Projection Operator Π(Z)

1: for all i, j, k do
2: XZ(i, j, k)
3: if ||X||F ≥ 1 then
4:   XX/||X||F
5: end if
6:   Z(i, j, k) ← X
7: end for

3. Denoising Experiments

Data Acquisition

A series of scans were acquired of a frozen supermarket turkey at 80 and 140 kVp on a 256-slice multidetector CT (Brilliance iCT, Philips Healthcare) in axial mode. At each kVp, we acquired both a “low-dose” scan (100mAs) and ten high dose scans (410 mAs for 80 kVp and 425 mAs for 120/140 kVp). The high-dose scans were repeated so that the resulting images could be averaged to create a very low-noise “ground truth” image. All 33 scans were acquired with the exact same table position and no other parameters were altered between acquisitions besides the kVp and mAs.

Algorithm 4.

Implementation of Projection Operator Π𝒩(Z)

1: for all i, j, k do
2: XZ(i, j, k)
3:  (λ, V ) ← Eig(XTX) ▷ compute eigenvalues and eigenvectors [13]
4: diag(λ)
5: if Tr (Σ) > 1 then
6:   Σp ← min(1, Σ) ▷ element-wise minimum
7:   XXV Σ+ΣpV T
8: end if
9: Z(i, j, k) ← X
10: end for

All of the scans were reconstructed with FBP (high resolution mode, Filter E) via the manufacturer’s software and using the exact same parameters: 512 × 512 × 128 image volumes with 0.562mm voxels and a 0.625mm slice thickness. Linear interpolation was used along the z-axis (axial direction) to synthesize a volume with isotropic resolution (0.562mm) to avoid complications with the discrete gradient and divergence operators. Since the scan was acquired in axial mode, the slices toward the axial extremes of the acquired data volume exhibited severe cone beam artifacts. A smaller volume consisting of only the central 32 slices was extracted in order to minimize these effects.

Comparison of TVS, TVF, and TVN

In order to compare these TV variants, we synthesized a dual kVp image pair by extracting the same 3D volume from the 80 and 140 kVp low-dose (100mAs) datasets. Then we denoised this dual-energy data over a wide range of λ values with each of TVS, TVF, and TVN; we first identified λmin and λmax, corresponding to images that appeared under- and over-smoothed and sampled 100 equally spaced values within this interval. Each result was then compared to the corresponding 10X averaged high-dose image, and the normalized Euclidean distance was computed as an error metric. This is defined as

ε(λ)=u(λ)-uref2uref2, (24)

where u*(λ) is the denoised image with parameter λ, and uref is the corresponding 10X high-dose reference image. In the case of TVF and TVN, the dual-energy data are processed jointly due to the interchannel coupling of the regularizer, but we report the error metrics individually for each channel. This is shown in Figure 1.

Figure 1.

Figure 1

Relative error curves based on (24) are shown for the three different VTV variants studied at both 80 kVp (left) and 140 kVp (right). Optimal values of λ are identified by the minima of these curves.

In general, the error curves should not be compared at any one specific common value of λ because the functions TVS, TVF, and TVN have inherently different scales. Instead the curves should be evaluated in terms of the possible family of solutions they each represent and, most importantly, their minimum-error solution. As can be seen in Figure 1, the channel-coupling of the VTV has little effect on the 140 kVp image; however, it does improve the accuracy of the noisier 80 kVp image, where it is clear that TVN outperforms TVF, which in turn outperforms TVS. The minima of the various error curves is reported for both the 80 and 140 kVp channels for all regularizers in Table 2. In addition to relative error, which can be thought of as a normalized root-mean-square error (RMSE), we also report the universal quality index (UQI) proposed in [14]. For two images x and y, the UQI is defined as

Table 1.

Three multi-channel variants of the TV studied in this work. Note, TVS induces no inter-channel coupling and is equivalent to separately denoising each image channel.

Defintion Edge Coupling Gradient Alignment
TVS(u) = TVS(u)==1LTV(u)
TVF(u) = ||Ju||1,F
TVN(u) = ||Ju||1,
UQI=4σxyx¯y¯(σx2+σy2)[(x¯)2+(y¯)2], (25)

where

x¯=1Ni=1Nxi,y¯=1Ni=1Nyiσx2=1N-1i=1N(xi-x¯)2,σy2=1Ni=1N(yi-y¯)2σxy=1N-1i=1N(xi-x¯)(yi-y¯).

In particular, the UQI is computed between the high dose reference images and the denoised images and reported in Table 2 in parenthesis. The dynamic range of UQI is [1, 1], where higher values are better.

Figure 2 shows a comparison between the denoised 80 kVp images between different VTV variants, each shown at its optimal lambda value (the minimizer of the error curve); Figure 3 shows the same comparison for an ROI. It can be seen that the channel-coupling does improve the preservation of details and suppression of noise. The differences between TVF and TVN are quite subtle with TVN arguably having slightly cleaner edges, but the improvement over channel-by-channel TV (TVS) is somewhat more pronounced. Figures 4 and 5 shows the corresponding results for the 140 kVp channel. In that case, differences between regularizers are practically indiscernible. This observation is consistent with our previous findings that the channel-coupling effect of these VTV penalties is primarily helpful to the noisier image channels because it allows edge information to be transferred from the less noisy channel.

Figure 2.

Figure 2

Reconstructed 80 kVp images corresponding to low-dose data (a), 10x averaged high-dose data (b) and denoised low-dose images with TVS, TVF, and TVN regularization, respectively (c–e). The parameter λ is selected to optimize each regularizer according to figure 1, and the display window is [−200 HU, 200 HU].

Figure 3.

Figure 3

ROI from reconstructed 80 kVp images corresponding to low-dose data (a), 10x averaged high-dose data (b) and denoised low-dose images with TVS, TVF, and TVN regularization, respectively (c–e). The parameter λ is selected to optimize each regularizer according to figure 1, and the display window is [−200 HU, 200 HU].

Figure 4.

Figure 4

Reconstructed 140 kVp images corresponding to low-dose data (a), 10x averaged high-dose data (b) and denoised low-dose images with TVS, TVF, and TVN regularization, respectively (c–e). The parameter λ is selected to optimize each regularizer according to figure 1, and the display window is [−200 HU, 200 HU].

Figure 5.

Figure 5

ROI from reconstructed 80 kVp images corresponding to low-dose data (a), 10× averaged high-dose data (b) and denoised low-dose images with TVS, TVF, and TVN regularization, respectively (c–e). The parameter λ is selected to optimize each regularizer according to figure 1, and the display window is [−200 HU, 200 HU].

4. Water-Subtracted Images

In principle, one can use dual-energy CT measurements to solve for quantitative material density maps; however, this sort of material decomposition requires either accurate knowledge of the input spectrum or a significant amount of calibration data. Alternatively, one can simply take linear combinations of the reconstructed 80 and 140 kVp images, where the coefficients are chosen to cancel out a given material [15]. For example, merely subtracting the 140 kVp image from the 80 kVp image should approximately cancel out any voxels that have an effective atomic number very similar to water, provided that the original images have already been converted to Hounsfield units (HU). Though the resultant image lacks quantitative accuracy, it is qualitatively very similar to a bone density image.

After denoising the low-dose 80/140 kVp images with the various regularizers (TVS, TVF, and TVN), we formed water-subtracted images from each image pair. We find that the differences between the various regularizers is far more pronounced in this water-subtracted image than in either the 80 or 140 kVp images, which can be seen in figures 7 and 8. Compared with TVS, TVN does a better job at preserving fine details in the bony structures while also achieving a less noisy texture. Visually, TVN also appears to slightly outperform TVF in this regard, indicating that the effect of encouraging aligned gradient vectors across image channels may be important, despite the fact that TVF and TVN barely show any visible differences at all in the raw 80/140 kVp images. These differences are quantified and confirmed in table 3, wherein the contrast-to-noise ratio (CNR) is computed between an inclusion in the vertebral structure and the nearby bone for each water-subtracted image. We define CNR as

Figure 7.

Figure 7

Water-subtracted images corresponding to low-dose data (a), 10× averaged high-dose data (b) and denoised low-dose images with TVS, TVF, and TVN regularization, respectively (c–e). The parameter λ is selected to optimize each regularizer according to figure 1.

Figure 8.

Figure 8

ROI from water-subtracted kVp images corresponding to low-dose data (a), 10× averaged high-dose data (b) and denoised low-dose images with TVS, TVF, and TVN regularization, respectively (c–e). The parameter λ is selected to optimize each regularizer according to figure 1.

Table 3.

A comparison of the contrast-to-noise ratio (CNR) between an inclusion in the vertebral structure and the adjacent solid bone across TVS, TVF, and TVN in the water-subtracted image.

Image CNR
Low Dose (Original) 2.4
10× Dose (Reference) 6.0
TVS 3.0
TVF 3.2
TVN 5.7
CNR=2(HU¯bone-HU¯inclusion)2σbone2+σinclusion2, (26)

where HU¯inclusion is the mean HU in the ROI shown in figure 6 and HU¯bone is an identically sized ROI placed over bone. σinclusion2 and σbone2 are the variances in these ROIs.

Figure 6.

Figure 6

The ROI placed over the inclusion is depicted by the yellow rectangle.

As in Table 2, we also report the relative error and UQI metrics for the water- subtracted images in Table 4.

Table 4.

Relative error and UQI of each VTV variant at its respective optimal λ value for the water-subtracted images.

Regularizer Relative Error Universal Quality Index (UQI)
TVS 0.6260 0.8044
TVF 0.3768 0.8885
TVN 0.3261 0.9258

5. Discussion

In this paper, we assessed the utility of the vectorial TV for realistic CT images. Consistent with [6], we found that the edge-coupling properties of TVN can help to preserve details and also improve quantitative accuracy, compared with channel-by-channel TV. The advantages were most striking when examining material images formed as linear combinations of the denoised energy images. In this study, the benefits were more subtle than in the reconstruction study utilizing the numerical XCAT phantom. This is not particularly surprising since that phantom is piecewise constant, which makes it possible to operate in a regime of very strong regularization.

In addition, we looked at TVF, which also encourages common edges but is ambivalent to gradient-vector alignment between image channels. Although TVN led to more accurate images (relative to the reference images) and arguably cleaner edges, especially for material images, TVF has some practical advantages, as it is simpler to implement and slightly less computationally expensive. However, we have demonstrated that TVN can be implemented efficiently in 3D for an arbitrary number of energy channels using the analytic matrix diagonalization method described in [13]. For this denoising study, TVN involves slightly longer computation times than TVF or TVS ( 20% longer run time in our implementation), but this difference is negligible for image reconstruction because the the projection and backprojection steps are the dominant factors determining the per-iteration runtime.

Interestingly, the benefits of the inter-channel coupling were most apparent when there was a very large disparity in the noise levels between the low- and high-energy image channels. Intuitively, this makes sense because both TVF and TVN allow for edge information to be transferred from the low-noise image into noisy image. This could be useful for making multi-energy CT acquisitions more robust to scenarios where certain channels receive low count rates, such as dual kVp imaging of obese patients.

6. Conclusion

The edge-coupling properties of TVF and TVN lead to improved image quality and accuracy compared to the conventional total variation when denoising multi-channel images. On real data, with realistic anatomical texture and appearance, the improvement is not as pronounced as it was in previous numerical experiments with piecewise constant phantoms[6]. However, TVN consistently outperforms TVF and TVS in terms of root-mean-square error, universal quality index, and subjective visual assessment.

Acknowledgments

This work was supported in part by Toshiba Medical Research Institute USA, NIH grant R01EB017293, the National Institute of Biomedical Imaging and Bioengineering of the National Institutes of Health under grant number T32 EB002103, and the National Institute of General Medical Sciences of the National Institutes of Health under Award Number R25GM109439 (Project Title: University of Chicago Initiative for Maximizing Student Development (IMSD)). Partial funding for the computation in this work was provided by NIH Grant Nos. S10 RRO21039 and P30 CA14599. Its contents are solely the responsibility of the authors and do not necessarily represent the official views of the National Institutes of Health or Toshiba Medical Research Institute USA.

References

RESOURCES