Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2017 Jun 8.
Published in final edited form as: Comput Vis Image Underst. 2016 Oct 6;155:173–194. doi: 10.1016/j.cviu.2016.10.002

Simultaneous Tumor Segmentation, Image Restoration, and Blur Kernel Estimation in PET Using Multiple Regularizations

Laquan Li 1, Jian Wang 1, Wei Lu 2,3, Shan Tan 1
PMCID: PMC5463621  NIHMSID: NIHMS831838  PMID: 28603407

Abstract

Accurate tumor segmentation from PET images is crucial in many radiation oncology applications. Among others, partial volume effect (PVE) is recognized as one of the most important factors degrading imaging quality and segmentation accuracy in PET. Taking into account that image restoration and tumor segmentation are tightly coupled and can promote each other, we proposed a variational method to solve both problems simultaneously in this study. The proposed method integrated total variation (TV) semi-blind de-convolution and Mumford-Shah segmentation with multiple regularizations. Unlike many existing energy minimization methods using either TV or L2 regularization, the proposed method employed TV regularization over tumor edges to preserve edge information, and L2 regularization inside tumor regions to preserve the smooth change of the metabolic uptake in a PET image. The blur kernel was modeled as anisotropic Gaussian to address the resolution difference in transverse and axial directions commonly seen in a clinic PET scanner. The energy functional was rephrased using the Γ-convergence approximation and was iteratively optimized using the alternating minimization (AM) algorithm. The performance of the proposed method was validated on a physical phantom and two clinic datasets with non-Hodgkin’s lymphoma and esophageal cancer, respectively. Experimental results demonstrated that the proposed method had high performance for simultaneous image restoration, tumor segmentation and scanner blur kernel estimation. Particularly, the recovery coefficients (RC) of the restored images of the proposed method in the phantom study were close to 1, indicating an efficient recovery of the original blurred images; for segmentation the proposed method achieved average dice similarity indexes (DSIs) of 0.79 and 0.80 for two clinic datasets, respectively; and the relative errors of the estimated blur kernel widths were less than 19% in the transversal direction and 7% in the axial direction.

Keywords: image restoration, tumor segmentation, blur kernel estimation, variational method, TV regularization, L2 regularization

1. Introduction

Positron emission tomography (PET) with 18F-fluorodeoxyglucose (FDG) plays an important role in cancer diagnosis, staging, treatment planning and outcome assessment (Juweid and Cheson, 2006; Rohren et al., 2004). PET imaging can provide functional and metabolic information of the human body, and allows earlier diagnosis and better management of oncology patients, compared to anatomical imaging techniques such as magnetic resonance imaging (MRI) and computed tomography (CT).

Many oncology applications using PET rely on accurate tumor segmentation (Zaidi and El Naqa, 2010). For example, in radiation therapy treatment planning an accurate delineation of tumor can ensure the maximum amount of dose delivered to the cancer cells in the tumor and potentially reduce side effects to surrounding healthy tissues (Pan and Mawlawi, 2008); it is known that spatial-temporal PET features extracted from pre- and post-treatment PET images are useful predictors of pathologic tumor response to neoadjuvant chemoradiation therapy, and most of these features rely on accurate delineation of the tumor regions (Tan et al., 2013).

In the past decades, a number of automatic and semiautomatic approaches have been proposed for PET tumor segmentation (Foster et al., 2014; Zaidi and El Naqa, 2010). Thresholding with a fixed percentage such as 42% or 50% of the maximum standardized uptake value (SUVmax) in the tumor region are the most frequently used one in clinic oncology using PET. However, the optimum threshold varies greatly with many factors such as tumor size and shape, image noise and contrast, and others (Biehl et al., 2006; Erdi et al., 1997; Hatt et al., 2011a; Hatt et al., 2011b). A fixed threshold may not always work well for different PET images. Various advanced methods were also used for tumor delineation in PET imaging, such as the stochastic modelling-based methods (Aristophanous et al., 2007; Hatt et al., 2009), learning-based methods (Belhassen and Zaidi, 2010; Kerhet et al., 2009), deformable models (Hsu et al., 2008; Li et al., 2008), Graph-based methods (Bađc et al., 2011), etc. Some of these methods were adopted directly from the general field of computer vision and image processing, without considering the characteristics of the PET images.

Many degrading factors can affect image quality and quantitative metabolic assessment in PET imaging. Among others, the partial volume effect (PVE) is recognized as one of the most important degrading factors, due to the limited spatial resolution of PET scanners, especially for tumors with small sizes (Pan and Mawlawi, 2008; Zaidi and El Naqa, 2010). PVE renders an object boundary blurry in PET imaging (Pan and Mawlawi, 2008), which seriously affect the tumor segmentation accuracy. To remedy this problem, some specifically designed segmentation methods have been proposed for PET images.

In order to calculate a reasonable threshold for tumor segmentation in PET, adaptive thresholding methods have been proposed (Biehl et al., 2006; Daisne et al., 2003; Drever et al., 2006; Erdi et al., 1997; Jentzen et al., 2007; Nestle et al., 2005; Schaefer et al., 2008). Rather than a direct consideration of PVE, these methods computed the optimal threshold using different image features, such as the source-to-background ratio (SBR), full width at half maximum (FWHM), mean background intensity, mean lesion intensity, and volume size (Foster et al., 2014). Generally, these methods required prior calibrations, and their parameters needed to be re-optimized through phantom experiments for a new scanner and different image reconstruction and correction algorithms. Tan et al. (2015) showed that the problem to calculate the optimal threshold for PET tumor segmentation is itself ill-posed, or undetermined. Another major limitation of thresholding methods is that these methods might not work well for tumors with heterogeneous metabolic uptake.

Some other segmentation methods considering PVE have also been studied for PET in the past several years (De Bernardi et al., 2009; El Naqa et al., 2004; Geets et al., 2007; Li et al., 2008; Riddell et al., 1999). Most of these methods recovered the ‘clean’ image first by partial volume correction (PVC) techniques, and the tumors were then delineated on the recovered image (Geets et al., 2007). Various PVC methods have been developed for PET imaging. Srinivas et al. (2009) provided a practical look-up table based on recovery coefficient (RC) for the PVC of spherical lesions. Brix et al. (1997) put forward a reconstruction-based PVC method. This method incorporated scanner characteristics into the process of iterative image reconstruction to improve the spatial resolution of a whole-body PET system. Rousset et al. (1998) designed a geometric transfer matrix (GTM) method to correct PVE for PET brain imaging. The GTM method modeled the geometric interactions between the PET system and the brain activity distribution. Moreover, some de-convolution based image restoration methods have been used for PVE correction (Barbee et al., 2010; Boussion et al., 2009; Kirov et al., 2008; Teo et al., 2007; Tohka and Reilhac, 2008). Teo et al. (2007) proposed an iterative de-convolution technique based on Van Cittert’s method (Van Cittert, 1931) to correct PVE. Tohka and Reilhac (2008) studied three iterative de-convolution algorithms for PVC. The three de-convolution methods included Richardson–Lucy (Lucy, 1974; Richardson, 1972), reblurred Van Cittert iteration according to Carasso (1999), and reblurred Van Cittert iteration with the total variation (TV) regularization (Rudin et al., 1992). Kirov et al. (2008) presented a post-reconstruction PVE correction based on iterative de-convolution, and further modified the regularization procedure (Alenius and Ruotsalainen, 1997) depending on the local topology of the PET image to avoid an increase in variance. In these de-convolution methods the point spread functions (PSFs) were assumed to be known or measured in advance.

Unlike these existing methods considering image restoration (PVC) and tumor segmentation as two separate processes, in this study we proposed a new variational framework for simultaneous image restoration, tumor segmentation and blur kernel estimation, specifically designed for PET imaging.

The idea of simultaneous image restoration and segmentation is itself not new. Bar et al. (2006) integrated semi-blind image de-convolution (parametric blur-kernel) with Mumford-Shah segmentation in a variational framework for two-dimensional image restoration and segmentation. Zheng and Hellwich (2006) built a blind image de-convolution and segmentation model by extending the Mumford-Shah regularization in the context of Bayesian estimation. A cost term for the estimation of the blur kernels via a newly introduced prior solution space was added to the Mumford-Shah model. Cai (2015) established a multiphase segmentation model by combining the piecewise constant Mumford-Shah segmentation model with a new data fidelity term coming from an image restoration model. This method can be used for noisy, blurred vector-valued images; and the noise type and blur kernel have to be known previously. Ayasso and Mohammad-Djafari (2010) presented a method to simultaneously restore and segment the nondestructive testing images. They studied a family of heterogeneous Gauss–Markov fields with Potts region labels model in a Bayesian estimation framework. Paul et al. (2013) derived a new class of data-fitting energies that coupled image restoration and segmentation using a statistical framework of generalized linear models for two-region segmentation of scalar-valued images. Most of these methods were proposed for general image processing applications in two-dimensionality, and they generally used either TV regularization (Paul et al., 2013) or L2 regularization (Bar et al., 2006; Zheng and Hellwich, 2006) on the restored image, and used an isotropic Gaussian function to model the PSF (Bar et al., 2006) or assumed that the PSF was already known (Ayasso and Mohammad-Djafari, 2010; Cai, 2015; Paul et al., 2013).

However, these existing methods may not be used directly in PET because a PET image has its own specific properties and the clinic radiation oncology has its own specific requirements for PET image processing as well. First of all, in methods based on the Mumford-Shah model (Bar et al., 2006; Zheng and Hellwich, 2006), the regularization for image intensity was applied to only the non-edge areas, but not to target edges. As a result, the intensity over these un-regularized edges in the restored image was sometimes unstable. Our experimental results have confirmed that, without regularizations over the tumor edges, some abnormal large uptake values can appear in the restored PET image. This might not be a problem for general image processing applications. However, in radiation oncology, it is well-known that the maximum uptake in a tumor region is a widely used feature for cancer diagnosis and treatment outcome prediction (Tan et al., 2013). These abnormal large uptake values on tumor edges might lead to an incorrect interpretation of PET images in patient management. Secondly, existing methods generally used only one single regularization (e.g., TV in (Paul et al., 2013) and L2 in (Bar et al., 2006; Zheng and Hellwich, 2006)). However, a PET image consists of components with different characteristics, requiring multiple regularizations in the variational segmentation and restoration framework. For example, tumor edges in PET are significantly blurred due to PVE, which asks for regularizations with the advantage of preserving edges such as TV. On the other hand, the metabolic uptake inside the tumor region of a PET image may vary smoothly and is somehow heterogeneous. The TV regularization may not work well for the non-edge component of a typical PET image, because minimizing an objective function with the TV regularization enforces neighboring pixels to have similar uptake values thus tends to produce piecewise constant areas in the restored image (Chan et al., 2000), which is the well-known staircase effect. For the non-edge component, the L2 regularization should be more appropriate than TV. Thirdly, the clinic PET scanner can have different resolutions in different directions, which asks for an anisotropic PSF model rather than an isotropic one, as used in most existing methods. Last but not the least, most existing methods were implemented and tested on two-dimensional images. The strong spatial correlation among neighboring slices in PET requires three-dimensional (3D) processing.

To the best of our knowledge, there is no published method of simultaneous restoration and segmentation specifically designed for PET images. The proposed method in this study addressed both problems simultaneously by integrating semi-blind de-convolution and Mumford-Shah segmentation model, with regularization terms and blur kernel estimation adapted to PET imaging. The novelty of the proposed method can be summarized as follows:

  1. In order to suppress the abnormal large uptake values appearing over edges in the restoration result, regularizations were applied not only to the non-edge component but also to the tumor edge component.

  2. Multiple regularizations were employed in the simultaneous restoration and segmentation variational framework, with each regularization adapted to a different image component in PET. Particularly, the TV regularization was used over tumor edges for its good edge-preserving ability, while the L2 regularization was used inside the tumor region to preserve the smooth change of the metabolic uptake in a PET image.

  3. The blur kernel was modeled as anisotropic Gaussian to address the resolution difference in transverse and axial directions of the PET scanner.

  4. The algorithm was implemented in three-dimensionality, which made full use of the spatial correlation among neighboring slices and potentially improved the accuracy of restoration and segmentation.

The proposed method was iteratively optimized using an alternating minimization (AM) algorithm. The performance of the proposed method was validated on a physical phantom and two datasets of clinic patients with non-Hodgkin’s lymphoma and esophageal cancer, respectively. Experimental results showed that the proposed method improved the segmentation performance when compared with all other tested methods. The blur kernel estimation and desired restored image can be obtained simultaneously.

2. Methods and materials

2.1. Fundamentals

2.1.1. De-convolution

Let Ω ⊂ ℝ3 be a closed and bounded region and f : Ω → ℝ be a sharp or ideal 3D image defined on Ω. The PET imaging can be expressed as the following model (Boussion et al., 2009; El Naqa et al., 2006)

g=fh+n, (1)

where h is the PSF of the scanner, n is the additive noise, g is the observed PET image defined on Ω, and * is the 3D convolution operator. In PET imaging, the noise is mainly caused by inaccurate counting of the photon pairs (Humm et al., 2003), and the noise property is related to the used reconstruction methods (Tong et al., 2010). A common assumption is that overall the noise may be characterized as Gaussian (Ollinger and Fessler, 1997).

Normally, only the blurred image g is given, PSF and the sharp image f are unknown. We may want to recover both the sharp image and PSF, which is known as blind de-convolution. A blind de-convolution problem is ill-posed with respect to both f and h. You and Kaveh (1996) proposed to simultaneously regularize f and h and solved the blind de-convolution problem by minimizing the following energy functional

F(f,h)=12fh-gL2(Ω)2+α1Ω|f|2dA+α2Ω|h|2dA, (2)

where α1 and α2 are both positive parameters, and dA denotes dxdydz (dx, dy and dz represent the infinitesimal elements of x, y and z, respectively). Unfortunately, one potential drawback of the L2 regularization in Eq.(2) is that it over-smoothes object edges (Perrone and Favaro, 2014). In view of the superiority of the TV regularization on preserving edges, Chan and Wong (1998) regularized both the recovered image and PSF by TV norm. Their blind de-convolution model can be expressed as

F(f,h)=12fh-gL2(Ω)2+α1Ω|f|dA+α2Ω|h|dA. (3)

The parameters α1 and α2 control the smoothness of the recovered image and PSF, respectively.

2.1.2. Mumford-Shah segmentation

In Ref (Mumford and Shah, 1989), Mumford and Shah proposed an energy minimization method to find an optimal piecewise smooth function f to approximate the observed image g. The energy functional was defined as

F(f,K)=12Ω(f-g)2dA+βΩ/K|f|2dA+αKdσ, (4)

where α and β are both positive parameters, K ⊂ Ω represents the edge set of the objects, Ω/K denotes the non-edge areas of the image, and f : Ω → ℝ is continuous or even differentiable in Ω/K but discontinuous across K. The first term is the fidelity term, which requires f to be an “optimal” piecewise smooth approximation of g; the second term is the piecewise-smoothness term, which forces the solution f to be smooth in Ω/K; the third term stands for the total surface measure of the edge set K, which contributes to the smoothness of the edge set K.

The minimization of Eq.(4) is difficult due to the non-convex property of this functional and the unknown discontinuity edge set K in the integral domains (Scherzer, 2011). Ambrosio and Tortorelli (1990) introduced the Γ-convergence to solve this problem. They approximated an irregular functional F(f, K) by a sequence of regular functionals Fε (f, v) with a small positive constant ε which meets the equation of limε0Fε(f,v)=F(f,K), and the minimizing of F can thus be replaced by the minimizing of Fε. A continuous function v was used to represent the so-called edge map, i.e., v(x, y, z) ≈ 0 if (x, y, z) is close to object edges, and v(x, y, z) ≈ 1 if (x, y, z) is far away from object edges. Thus, Eq.(4) can be replaced with the Γ-convergence form:

Fε(f,v)=12Ω(f-g)2dA+βΩv2|f|2dA+αΩ(ε|v|2+14ε(v-1)2)dA. (5)

2.2. Coupling restoration with segmentation

2.2.1. A novel model of simultaneous restoration, segmentation and blur kernel estimation

The blurred edges of a tumor in a PET image can become sharp after using a de-convolution technique, which would be helpful to segmentation; conversely, if the precise contours can be obtained by segmentation, it would be easier to estimate the blur kernel, and thus be helpful to image restoration. Considering that both segmentation and restoration problems can benefit each other, we integrated tumor restoration with segmentation using a variational framework in PET.

The proposed method combined the de-convolution model (Eq.(3)) with the Mumford-Shah segmentation model (Eq.(4)), and utilized multiple regularization terms for the restored image. The blur kernel was constrained to be an anisotropic Gaussian parameterized by different widths in transversal and axial directions to be adapted to the different spatial resolution in different directions of a PET scanner. The TV regularization has the advantage of preserving edge information, but suffers from the well-known staircase effect; the L2 regularization can avoid staircase effect, but may result in over-smoothed object edges. In view of this, we used the TV regularization over tumor edges and the L2 regularization in non-edge areas in the restored images. Our functional was defined as

F(f,K,σ1,σ2)=12Ω(hσ1σ2f-g)2dxdydz+βΩ/K|f|2dxdydz+ηK|f|dxdydz+αKdσ+γΩ|hσ1,σ2|2dxdydz. (6)

It is difficult to minimize the above functional because of the non-convex property and the unknown discontinuity edge set K. We rephrased the proposed model using Γ-convergence approximation:

Fε(f,v,σ1,σ2)=12Ω(hσ1,σ2f-g)2dxdydz+βΩv2|f|2dxdydz+ηΩ(1-v2)|f|dxdydz+αΩ(ε|v|2+(v-1)24ε)dxdydz+γΩ|hσ1,σ2|2dxdydz, (7)

where α, β, η, γ and ε are positive parameters; v represents the edge map; hσ1,σ2 is an anisotropic Gaussian kernel parameterized by its widths σ1 (transversal direction) and σ2 (axial direction):

hσ1,σ2=1(2π)3/2σ12σ2e-(x2+y22σ12+z22σ22). (8)

In order to enhance the robustness and convergence of the algorithm, we imposed the following constraints on hσ1,σ2 and f:

Ωhσ1,σ2(x,y,z)dxdydz=1,hσ1,σ2(x,y,z)0,f(x,y,z)0. (9)

The first term in Eq.(7) is the data fidelity term which contributes to the approximation of the restored image f and the given blurred image g. The second term in Eq.(7) is a piecewise-smoothness term (L2 regularization) constraining the smoothness of the non-edge areas; and the third term (TV regularization) in Eq.(7) is the TV regularization for object edges, which contributes to the preservation of edge information. The fourth term in Eq.(7) is used to constrain the length of the edge set. The last term in Eq.(7) represents the regularization of the blur kernel in which the convenient L2 norm was used because the Gaussian kernel hσ1,σ2 is a smooth function. This term is necessary to reduce the fundamental ambiguity in the division of the apparent blur between the restored image and the blur kernel (Bar et al., 2006).

2.2.2. Alternating minimization

It can be observed that the objective function (Eq.(7)) is not jointly convex with respect to f, v and hσ1,σ2, but it is convex with respect to any one of the three functions if the other two are estimated and fixed. In view of the above, the AM algorithm (Chan and Shen, 2005; Chan and Wong, 1998; You and Kaveh, 1996) was used to iteratively optimize the energy functional in our study. The AM algorithm is a classical method for solving composite minimization problems. The basic idea is to minimize the objective function only with respect to one variable and keep the others fixed in each iteration. The convergence of the algorithm was proved in Ref (Chan and Shen, 2005).

In our study, the restored image f was initialized as the blurred image g; the edge function v was initialized with a function of constant 1; σ1 and σ2 were initialized to a small number to make the blur kernel hσ1,σ2 approximate the delta function kernel.

The first and second steps in our AM algorithm were to minimize the objective function (Eq.(7)) with respect to the edge map v and the restored image f. These two steps were achieved by solving the Euler-Lagrange equations subject to Neumann boundary conditions (∂v/∂N=0, ∂f/∂N=0):

δFεδv=(2β|f|2-2η|f|)v+α·v-12ε-2εα2v=0, (10)
δFεδf=(hσ1,σ2(f-g)hσ1,σ2(-x,-y,-z)-2βDiv(v2f)-ηDiv((1-v2)f|f|)=0. (11)

The detailed derivation of the two Euler-Lagrange equations is presented in Appendix A.

During the (i+1)th iteration of the AM algorithm, according to the relationship between the Euler-Lagrange equation (Eq.(10)) and the gradient descent flow, we updated the edge map v by the following iterations

vt+1i+1=vti+1-εv[(2β|fi|2-2η|fi|)vti+1+α·vti+1-12ε-2εα2vti+1], (12)

where the step length εv > 0 and t denotes the iteration number of the gradient descent flow. This iteration was repeated until the mismatch of the current and the previous iteration results was less than a predefined small positive constant. Similarly, according to the Euler-Lagrange equation Eq.(11), we updated the restored image f by

ft+1i+1=fti+1-εf[(hσ1,σ2ifti+1-g)hσ1,σ2i(-x,-y,-z)-2βDiv((vi+1)2fti+1)-ηDiv((1-(vi+1)2)fti+1|fti+1|)]. (13)

The step length εf >0 and the termination condition was the same as the iteration of v. The restored image f was truncated at zero:

fi+1(x,y,z)={fi+1(x,y,z),iffi+1(x,y,z)>00,otherwise. (14)

The next step of the AM algorithm was to minimize the objective function (Eq.(7)) with respect to the blur kernel widths σ1 and σ2 by solving the following two equations:

Fεσ1=Ω[(hσ1,σ2f-g)(hσ1,σ2σ1f)+γσ1|hσ1,σ2|2]dxdydz=0, (15)
Fεσ2=Ω[(hσ1,σ2f-g)(hσ1,σ2σ2f)+γσ2|hσ1,σ2|2]dxdydz=0, (16)

where

hσ1,σ2σ1=1(2π)3/2σ12σ2e-(x2+y22σ12+z22σ22)(x2+y2σ13-2σ1), (17)
hσ1,σ2σ2=1(2π)3/2σ12σ2e-(x2+y22σ12+z22σ22)(z2σ23-1σ2), (18)

and

σ1|hσ1,σ2|2=x2+y24π3σ14σ22e-(x2+y2σ12+z2σ22)(x2+y4σ17-4σ15)+z24π3σ14σ22e-(x2+y2σ12+z2σ22)(x2+y2σ13σ24-2σ1σ24), (19)
σ2|hσ1,σ2|2=x2+y24π3σ14σ22e-(x2+y2σ12+z2σ22)(z2σ14σ23-1σ14σ2)+z24π3σ14σ22e-(x2+y2σ12+z2σ22)(z2σ27-3σ25). (20)

The detailed derivation of Eqs.(17) and (19) is presented in Appendix B. Eqs.(15) and (16) were solved using the bisection method.

During the (i+1)th iteration of the AM algorithm, the blur kernel hσ1,σ2i+1 was obtained as:

hσ1,σ2i+1(x,y,z)={hσ1,σ2i+1(x,y,z),ifhσ1,σ2i+1(x,y,z)>00,otherwise, (21)
hσ1,σ2i+1(x,y,z)=hσ1,σ2i+1(x,y,z)Ωhσ1,σ2i+1(x,y,z)dxdydz. (22)

The AM algorithm can be summarized as follows:

Initialization: f = g, v = 1, σ1 = σ2 = ε1, σ1prev = σ2prev ≫ 1
Iteration: while (|σ1prevσ1| > ε2 or |σ2prevσ2| > ε2) repeat
  1. Update the edge map v using the gradient descent algorithm, as in Eq.(12).

  2. Update the restored image f using the gradient descent algorithm, as in Eq.(13), and constrain f with Eq.(14).

  3. Set σ1prev = σ1, and update σ1 by solving Eq.(15) using the bisection method.

  4. Set σ2prev = σ2, and update σ2 by solving Eq.(16) using the bisection method.

  5. Update the blur kernel hσ1,σ2 with the new σ1 and σ2 using Eqs.(21) and (22).

end

Here ε2 was set as a small positive constant to control the degree of convergence of the AM algorithm.

2.3. Validation study

2.3.1. Datasets for validation

The performance of the proposed algorithm was validated on a physical phantom and two clinic datasets with different tumor types and imaged using different scanners. This retrospective study was approved by the University of Maryland Baltimore Institutional Review Board.

In the physical phantom, six spheres used as target volumes were placed inside a cylindrical container to mimic clinical head and neck tumor (Li et al., 2008). The volumes of these six spheres were 20, 16, 12, 6, 1 and 0.5 ml, respectively. These spheres were filled with 11C (half life=20.38 min) solution at an activity concentration of 3.627 μCi/ml as sources and the cylindrical container was filled with 18F-FDG (half life=109.77 min) solution at an activity concentration of 0.224 μCi/ml as the background. The phantom was scanned continuously on a Biograph-40 True Point/True View (Siemens Medical Solutions Inc., Knoxville, TN) PET/CT scanner for 120 min. The spatial resolution for a point source in the transversal direction and axial direction were 4.5 mm and 5.7 mm (FWHM, at 10 mm from the center of the field of view), respectively. Images were reconstructed using the iterative Ordered Subset Expectation Maximization algorithm (OSEM), and the voxel size was 4.0728×4.0728×2.027 mm3. Because of the different decay rate of 11C and 18F-FDG, PET images with SBRs of 8:1, 6:1, 4:1 and 2:1, respectively, were obtained for our experiments.

The first clinic dataset was acquired from eleven patients with non-Hodgkin’s lymphoma. Each patient was imaged on a GE Advance PET tomograph (General Electric Medical Systems, Milwaukee, Wisconsin, USA) (Betrouni et al., 2012). The transverse and axial resolution were 3.8 mm and 4.0 mm, respectively, at the center of the field of view (DeGrado et al., 1994). Each patient received the intravenous (IV) injection of a 370MBq activity of 18F-FDG one hour before the beginning of a whole-body imaging scan. The scanning process consisted of a 5-min emission scan and a 2-min transmission scan (external source of germanium), and the final PET images were reconstructed in transverse slices using the OSEM algorithm with a voxel size of 4.3×4.3×4.3 mm3 (Betrouni et al., 2012). For each PET image, only one tumor was selected for test and each tumor was manually delineated three times by five qualified nuclear medicine physicians (Betrouni et al., 2012). Therefore, each tumor was performed fifteen manual delineations. The ground truth for each tumor was computed using a voting strategy based on the fifteen manual delineations: a voxel was considered as the tumor voxel if and only if it was enclosed in the target region at least eight times.

The second clinic dataset was acquired from twenty patients with esophageal cancer. These patients were subjected to PET/CT scans at University of Maryland Medical Center from 2006 to 2009, and all imaging scans were completed on an integrated 16-slice Gemini PET/CT scanner (Philips Medical Systems, Cleveland, OH). The spatial resolution near the center was about 4.9 mm (FWHM) in both transverse and axial directions (Lodge et al., 2004). Following an institution standard protocol, the patients fasted for at least 4 hours and then received the intravenous injection of 12–14 mCi 18F-FDG. One hour after the injection, whole-body PET and CT imaging scan was started. PET image were attenuation corrected and reconstructed using a maximum likelihood algorithm. The voxel size of PET and CT images were respectively 4.0×4.0×4.0 mm3 and 0.98×0.98×4.0 mm3. The ground truth were manually delineated by an experienced radiation oncologist using the complementary visual features of PET and CT. The general location and nature of a lesion were determined by the visual interpretation of the PET image and the boundary was determined by both PET and CT images.

For each sphere or tumor, a rectangular region of interest (ROI) was manually defined that encloses the whole selected sphere or tumor. The proposed method and all comparison algorithms were implemented in ROI.

2.3.2. Algorithms for comparison

For comparison, the deformable models — Mumford-Shah (MS) segmentation algorithm (Mumford and Shah, 1989), Active Contours (AC) (Chan and Vese, 2001) and Geodesic Active Contours (GAC) (Caselles et al., 1997), the widely used thresholding methods — 42% and 50% of the SUVmax (42% threshold and 50% threshold) and Otsu automatic thresholding (Otsu, 1975), the clustering algorithm — Fuzzy C-means Clustering (FCM), and graph theory method — Graph Cuts (GC) (Boykov and Funka-Lea, 2006) were also tested in our experiments. In addition, in order to test the effects of the multiple regularizations in our proposed algorithm, the simultaneous restoration and segmentation method (Bar’s method) proposed by Bar et al. (2006) was tested. Note that this method does not use any regularizations on the edge set.

The parameters of all methods were manually optimized. To adjust the parameters of the proposed method, we first chose one image from each dataset. The parameter α controls the length of the object edges. The segmented foreground areas decrease when α increases. For PET images with a large tumor volume, a smaller value should be assigned to α. Parameters β and η control the smoothness of the restored images. When β increases, the restored image becomes smoother. When η decreases, the restored edges become sharper. When the SBR of the PET image is low, β and η should be set to a relative smaller value. The parameter γ controls the width of the Gaussian kernel. A larger γ will lead to a wider kernel. If the observed image is severely blurred, a relative larger value of γ should be better for estimating the blur kernel. We adjusted the values of β, η and γ depending on the restored image, and adjusted α depending on the segmentation result. We evaluated the parameter adjustment according to the accuracy of the segmentation result. When the segmentation accuracy no longer increased, we assumed that the parameters were optimized and the proposed method achieved the best performance. Because the real blur kernels of different images kept unchanged in the same dataset, the resulting β, η and γ were used for the remaining images in the same dataset. Only the parameter α might still need to be adjusted according to the tumor volume. For all three experimental data sets, the parameters of the proposed method were tuned in the following range: α changed between 5×10−8 and 9×10−7, β changed between 2×10−2 and 5×10−2, η was between 3×10−4 and 5×10−3, γ altered between 10 and 60, ε1=0.8, and ε2=10−4.

2.3.3. Criteria for performances evaluation

The segmentation accuracy of all tested algorithms was evaluated using three criteria.

Dice similarity index (DSI) measures the similarity (spatial overlap) between the segmented and ground truth foreground objects. Denote the segmented foreground objects set as VA and the ground truth foreground set as VR. DSI is computed as follows (Dewalle-Vignion et al., 2011; Prieto et al., 2012):

DSI(VA,VR)=2|VAVR||VA|+|VR|, (23)

where |X| represents the size of the set X. A bigger value of DSI indicates a better segmentation. DSI reaches the maximum value 1 if two sets completely overlap, and reaches the minimum value 0 if the two sets have no spatial overlap.

Classification error (CE) measures the spatial location bias of the segmented foreground objects compared to the ground truth foreground (Hatt et al., 2009):

CE(VA,VR)=|VFP|+|VFN||VR|, (24)

where |VFP| represents the number of false positive errors (background voxels that were wrongly assigned to tumor), |VFN| represents the number of false negative errors (tumor voxels that were wrongly assigned to background).

Volume error (VE) measures the difference between the segmented foreground volume and the ground truth foreground volume (Dewalle-Vignion et al., 2011):

VE(VA,VR)=abs(|VA|-|VR|)|VR|. (25)

A bigger value of CE and VE means a worse segmentation. The values of CE and VE can be bigger than 1. In order to enhance the observability, all the values bigger than 1 were truncated as 1 in our experiments.

We used the recovery coefficient (RC) to quantify the restoration accuracy. RC is generally used to quantify the degree of blurring of a blurred image and hence can be used to evaluate the restoration accuracy (Geworski et al., 2000; Kessler et al., 1984; Srinivas et al., 2009). RC is defined as:

RC=Measuredsphereactivity-MeasuredbkgactivityTruesphereactivity-Truebkgactivity, (26)

where bkg is the abbreviation for background. A smaller RC value means that the image is more blurred. For restored image, the closer the RC gets to 1, the better the restoration result is.

For blur kernel estimation, we calculated the relative errors between the corresponding FWHMs of the estimated blur kernel widths and the real FWHMs of the PET scanner.

3. Results

3.1. A toy simulation

The proposed algorithm was first tested on a toy computer simulation. A non-blurred volume was produced with a uniform background of intensity of 50 and a 3D sphere of uniform activity concentration embedded in its center. Without any loss of generality, the voxel size was set to be 1.0×1.0×1.0 mm3. The sphere radius was set to be 2 mm, 4 mm and 6 mm, respectively. For each radius, three different activity concentrations were set, which made SBR vary as 2:1, 3:1, and 4:1. Each non-blurred volume was convolved with a 3D isotropic Gaussian PSF of FWHM of 4 mm and 6 mm, respectively. A total of 18 blurred volumes were constructed. The task was to restore the non-blurred volume, delineate the sphere, and estimate the PSF, given a blurred volume.

Fig. 1 shows DSIs of tested methods for sphere segmentation, averaged over all 18 blurred volumes. The proposed method (DSI=0.80) was much better than other tested methods according to this evaluation criterion. Bar’s method (DSI=0.74) was the second best among all methods. In those without taking into account PVE, GC (DSI=0.71) and AC (DSI=0.70) were relatively better. The two fixed threshold methods — 42% threshold (DSI=0.25) and 50% threshold (DSI=0.33) performed the worst.

Figure 1.

Figure 1

DSI for sphere segmentation, averaged over all 18 blurred volumes.

Fig. 2 (a) presents DSI as a function of FWHM, averaged over all sphere radii and SBRs. Comparing with the FWHM of 4mm, the FWHM of 6mm made images more blurred. As a result, all tested methods performed worse for the FWHM of 6mm. The proposed method performed the best for both FWHM levels and it was more stable over the change of FWHM. Fig. 2 (b) shows DSI as a function of the sphere radius, and Fig. 2 (c) shows DSI as a function of SBR. The proposed method was robust to the change of both sphere radii and SBRs. Although Bar’s method, Otsu, FCM, AC, GAC and GC algorithms performed almost as well as the proposed method for the largest spheres (6mm), their performance dropped sharply with the decrease of the sphere radius. The two fixed threshold methods were especially sensitive to both sphere radius and SBR.

Figure 2.

Figure 2

(a) DSI at different FWHM levels (averaged over all sphere radii and SBRs); (b) DSI at different sphere radius (averaged over all SBRs and FWHM levels); (c) DSI at different SBRs (averaged over all sphere radii and FWHM levels).

One slice of a blurred volume with FWHM 6mm, SBR 2:1, radius 4mm is shown in Fig. 3 (a). The horizontal profile (Fig. 3 (b)) along the red line in Fig. 3 (a) looks like a Gaussian function, which indicates that the edge of the constructed volume was blurred. The restored image (Fig. 3 (d)), which was almost exactly the same as the ground truth (Fig. 3 (g)), and the corresponding horizontal profile (Fig. 3 (e)) both suggested that the proposed method obtained a good restoration result. The estimated widths of the blur kernel (PSF) were σ1=2.78 mm (Fig. 3 (h)) and σ2=2.77 mm (Fig. 3 (i)), which were very close to the true value 2.55 mm (corresponding to the FWHM of 6 mm); and the estimated blur kernel (Fig. 3 (f)) was visually very similar to the true one (Fig. 3 (c)).

Figure 3.

Figure 3

Figure 3

(a) Blurred volume with FWHM 6mm, SBR 2:1, radius 4mm; (b) horizontal profile along the red line in the blurred volume; (c) original Gaussian kernel; (d) restored volume; (e) horizontal profile along the red line in the restored volume; (f) estimated blur kernel; (g) ground truth; (h) convergence curve of σ1; (i) convergence curve of σ2.

To quantitatively evaluate the restoration accuracy of the proposed algorithm, RCs of the eighteen blurred volumes and the recovered volumes are listed in Table 1. According to this Table, the original blurred volumes had low RCs (all RCs were less than or equal to 0.65). For fixed FWHM and radius, RC did not vary with SBRs. The blur degree increased when the sphere radius decreased. RCs of the recovered volumes were more close to 1 than those of the original blurred volumes, which suggested the good restoration performance of the proposed method. As expected, the bigger the sphere, the better the restoration effect.

Table 1.

RCs of the original blurred volumes and the recovered volumes of the proposed method, respectively.

FWHM (mm) SBR Radius (mm) Blurred volumes Recovered volumes (Ours)
RC RC
4 2:1 2 0.21 1.38
4 0.51 1.37
6 0.65 1.03
3:1 2 0.21 1.17
4 0.51 1.17
6 0.65 0.99
4:1 2 0.21 1.07
4 0.51 1.13
6 0.65 1
6 2:1 2 0.09 1.38
4 0.32 1.08
6 0.49 1.06
3:1 2 0.09 0.92
4 0.32 1.05
6 0.49 1.03
4:1 2 0.09 0.73
4 0.32 1.03
6 0.49 0.99

3.2. Experiments on physical phantom

Fig. 4 displays DSI, CE and VE for tested algorithms, averaged over all sphere sizes and SBRs in the physical phantom study. The proposed method performed the best according to the three accuracy evaluation criteria (DSI=0.77, CE=0.41 and VE=0.20). Bar’s method (DSI=0.74, CE=0.44 and VE=0.27) was the second best and followed by FCM (DSI=0.72, CE=0.47 and VE=0.36) and GC method (DSI=0.71, CE=0.54 and VE=0.38). Three thresholding methods (50% and 42% threshold, and Otsu) had worse performance than those advanced ones.

Figure 4.

Figure 4

(a) DSI, (b) CE and (c) VE, for physical phantoms averaged over all spheres and SBRs.

DSIs for six spheres, averaged over four SBRs, is shown in Fig. 5. Quite obviously, the proposed method was the most robust to the change of the sphere size. For example, for the smallest sphere (0.5ml), the proposed method had a DSI of 0.55, while the DSI of Bar’s method was only 0.51 and all other methods were less than 0.4. FCM had a slightly better segmentation result for the three largest spheres than the proposed method. For the three smallest spheres, however, the performance of FCM went down very quickly, and was significantly worse than that of the proposed method.

Figure 5.

Figure 5

DSI for different spheres, averaged over all SBRs.

Fig. 6 displays DSIs at different SBR levels, averaged over all spheres, demonstrating the good robustness of the proposed method over SBR variation. Like in the toy simulation, three thresholding methods — 50% and 42% threshold, and Otsu — were very sensitive to SBR variation. Especially for SBR 2:1, the performances of these three methods deteriorated drastically.

Figure 6.

Figure 6

DSI at different SBR levels, averaged over all sphere sizes.

For a visual inspection, results of the proposed method for the 6ml sphere at SBR 4:1 are shown in Fig. 7. The horizontal profile (Fig. 7 (b)) and edge map v (Fig. 7 (c)) of the observed sphere (Fig. 7 (a)) suggested that the original sphere edge had been blurred. The second row of Fig. 7 displays the restored image (Fig. 7(d)) and its horizontal profile (Fig. 7(e)) and the edge map v (Fig. 7(f)), all showing the good restoration effect of the proposed method. The estimated blur kernel widths were 1.54 mm in the transversal direction (σ1), and 2.60 mm in the axial direction (σ2), corresponding to FWHMs of 3.63 mm and 6.12 mm, respectively. Both estimated FWHMs were very close to those of the real PET scanner (FWHM=4.5 mm in the transversal direction, and 5.7 mm in the axial direction) and the relative errors were 19% and 7%, respectively.

Figure 7.

Figure 7

(a) Observed 6ml sphere at SBR 4:1; (b) horizontal profile along the red line in the observed sphere; (c) edge map v of the observed sphere; (d) restored image; (e) horizontal profile along the red line in the restored image; (f) edge map v of the restored image; (g) estimated blur kernel (σ1=1.54 mm, σ2=2.60 mm); (h) convergence curve of σ1; (i) convergence curve of σ2.

In order to observe the restoration effect more clearly, Fig. 8 displays several restoration results. From the third column of this figure it can be seen that, without regularization on the edges, some abnormal high intensities appeared on object edges in the restoration results of Bar’s method, such as the voxels ‘A’, ‘B’ and ‘C’. The intensities of ‘A’, ‘B’ and ‘C’ were 73118, 91223 and 72963, respectively, while the intensities of the remaining voxels in these three slices were less than 35906, 40138 and 42862, respectively. There were not voxels with abnormal intensity in the results of the proposed method, thanks to the TV regularization on object edges.

Figure 8.

Figure 8

(a), (d) and (g): Observed 20ml, 16ml and 12ml spheres at SBR 4:1, respectively; (b), (e) and (h): restored 20ml, 16ml and 12ml spheres obtained by the proposed method, respectively; (c), (f) and (i): the restored 20ml, 16ml and 12ml spheres obtained by Bar’s method, respectively.

Table 2 lists RCs of the original blurred PET images and the recovered images of the proposed method. RCs of the original blurred PET spheres indicated that the blur degree was highly dependent on the sphere volume and only weakly influenced by SBRs, which was consistent with the toy simulation. Since small spheres were more susceptible to PVE, their blur degrees were greater than those of the big spheres. The proposed method could recover the ‘clean’ spheres effectively for both large and small spheres.

Table 2.

RCs of the original blurred PET images and the recovered images of the proposed method, respectively.

SBR Volume 20 ml 16 ml 12 ml 6 ml 1 ml 0.5 ml
Method
8:1 Original PET spheres 0.82 0.80 0.75 0.65 0.29 0.16
Recovered spheres (ours) 1.06 1.07 1.06 1.03 0.86 1.03
6:1 Original PET spheres 0.83 0.81 0.76 0.63 0.27 0.14
Recovered spheres (ours) 1.07 1.08 1.05 1.09 1.07 1.01
4:1 Original PET spheres 0.83 0.82 0.76 0.64 0.26 0.16
Recovered spheres (ours) 1.13 1.13 1.15 1.14 1.03 1.08
2:1 Original PET spheres 0.70 0.72 0.65 0.37 0.07 0.01
Recovered spheres (ours) 1.16 0.93 0.92 1.18 1.69 0.44

3.3. Experiments on non-Hodgkin’s lymphoma

The average DSI, CE and VE of the eleven non-Hodgkin’s lymphoma patients are displayed in Fig. 9. As this figure shows, the proposed method (DSI=0.79, CE=0.41 and VE=0.08) was superior to all other tested methods. FCM (DSI=0.75, CE=0.55 and VE=0.51) performed well according to DSI, but poorly according to CE and VE. The three thresholding methods still performed the worst.

Figure 9.

Figure 9

Figure 9

(a) DSI, (b) CE and (c) VE, averaged over all patients with non-Hodgkin’s lymphoma.

In order to visually inspect the segmentation effect of different methods, Fig. 10 presents the manual contour (closed blue curve in Fig. 10 (a)) and several automatic segmentation results (closed blue curves in Fig. 10 (b–f)) for one patient. For this patient, the proposed method had the most accurate segmentation (Fig. 10 (b), DSI=0.93, CE=0.14 and VE=0.01) and its contour coincided well with the manual contour. The lesions delineated by other methods (Fig. 10 (c–f)) were obviously smaller than the manual contour.

Figure 10.

Figure 10

Visual comparison for one patient with non-Hodgkin’s lymphoma (a) manual Contour, (b) the proposed method (DSI=0.93), (c) 42% thresholding (DSI=0.59), (d) FCM (DSI=0.70), (e) AC (DSI=0.82), and (f) GC (DSI=0.76).

Fig. 11 (a) shows one slice of the original PET image of a non-Hodgkin’s lymphoma patient. The horizontal profile (Fig. 11 (b), along the red line in the original PET image) and the edge map (Fig. 11 (c)) demonstrated that it was difficult to localize the edge for the original PET image. The three images of the second row — Fig. 11 (d–f) — suggested that the proposed method could get a restored result with more clear edges, compared to the original PET image. The estimated blur kernel widths σ1 (the transversal direction) and σ2 (the axial direction) converged to 1.43 mm and 1.75 mm, leading to FWHMs of 3.37 mm and 4.12 mm, respectively, which were very close to the real FWHMs (3.8 mm in the transversal direction, and 4.0 mm in the axial direction) with relative errors of 11% and 3%.

Figure 11.

Figure 11

Figure 11

(a) Original PET image (Zoomed in) of a non-Hodgkin’s lymphoma patient; (b) horizontal profile along the red line in the original PET image; (c) edge map of the original PET image; (d) restored image; (e) horizontal profile along the red line in the restored image; (f) edge map of the restored image; (g) the estimated blur kernel (σ1=1.43 mm, σ2=1.75 mm); (h) convergence curve of σ1; (i) convergence curve of σ2.

Fig. 12 (a) shows one PET image slice of a patient with non-Hodgkin’s lymphoma. Bar’s method still resulted in some abnormal large intensities on object edges, such as voxel ‘A’ in Fig. 12 (c). The intensity (standardized uptake value (SUV)) of ‘A’ was 22.19, while the intensities of other voxels in this slice were less than 11.44.

Figure 12.

Figure 12

(a) One slice of a PET image of a patient with non-Hodgkin’s lymphoma; (b) the restored image obtained by the proposed method; (c) the restored image obtained by Bar’s method.

3.4. Experiments on esophageal cancer

Fig. 13 shows the averaged DSI, CE and VE over all patients in the esophageal cancer dataset. Again, the proposed method performed the best with DSI=0.80, CE=0.42, and VE=0.06, followed by GC (DSI=0.75, CE=0.50, VE=0.39), Bar’s method (DSI=0.73, CE=0.53, VE=0.23) and FCM (DSI=0.72, CE=0.43, VE=0.39). For this dataset, 50% threshold (DSI=0.60, CE=0.60, VE=0.58) performed the worst.

Figure 13.

Figure 13

Figure 13

(a) DSI, (b) CE, and (c) VE, averaged over all patients with esophageal cancer.

The manual contour (Fig. 14 (a)) and several automatic segmentation results (Fig. 14 (b–f)) of one patient with esophageal cancer are displayed in Fig. 14. The contour obtained by the proposed method (Fig. 14 (b)) was in good agreement with the manual contour. Other methods performed significantly worse. The 50% threshold (Fig. 14 (c)) and Otsu (Fig. 14 (d)) had only DSI of 0.39 and 0.38, and the lesions delineated by the two methods were obviously smaller than the actual lesion (Fig. 14 (a)).

Figure 14.

Figure 14

Visual comparison (axial view) for one patient with esophageal cancer (a) Manual Contour, (b) the proposed method (DSI=0.87), (c) 50% thresholding (DSI=0.39), (d) Otsu (DSI=0.38), (e) FCM (DSI=0.67), and (f) GC (DSI=0.71).

Fig. 15 displays the results of the proposed method for an esophageal cancer patient. The original PET volume was blurred due to PVE, and it was hard to find tumor edges. The proposed method improved the quality of the blurred PET image, which can be seen from the restored image (Fig. 15 (d)), the corresponding horizontal profile (Fig. 15 (e)) and the edge map v (Fig. 15 (f)). The blur kernel widths σ1 converged to 1.78 mm which corresponds to FWHM of 4.19 mm in the transversal direction, and σ2 converged to 1.99 mm which corresponds to FWHM of 4.69 mm in the axial direction. Both estimated FWHMs were very close to the real FWHMs (the real FWHMs in both transversal and axial direction were 4.9 mm) and the relative errors were 14% and 4%, respectively.

Figure 15.

Figure 15

(a) Original PET image (zoomed in) of an esophageal cancer patient; (b) horizontal profile along the red line in the original PET image; (c) edge map of the original PET image; (d) restored image; (e) horizontal profile along the red line in the restored image; (f) edge map of the restored image; (g) the estimated blur kernel (σ1=1.78 mm, σ2=1.99 mm); (h) convergence curve of σ1; (i) convergence curve of σ2.

Fig. 16 displays the restoration results of the PET images of two esophageal cancer patients. The bright voxels ‘A’ and ‘B’ in Fig. 16 (c) and Fig. 16 (f) were the abnormal large intensities produced by Bar’s method. The intensity (SUV) of ‘A’ and ‘B’ were 66.65 and 22.94, respectively, which were considerably larger than all the other voxels in these two slices. Fortunately, no voxels with abnormal intensity were found in the restored results of the proposed methods.

Figure 16.

Figure 16

(a) and (d): one slice of the PET image of two esophageal cancer patients; (b) and (e): the restored images obtained by the proposed method; (c) and (f): the restored images obtained by Bar’s method.

3.5. Algorithm Robustness

In this section, we examined the robustness of the proposed method to the four parameters (α, β, η and γ) in Eq. (6) and Eq.(7). The experiments were tested with one parameter deviated from the manually tuned value while the other three kept unchanged.

Table 3 lists the segmentation robustness of the proposed method over α, β, η and γ, respectively, in given intervals, for one esophageal patient. As we can see, the segmentation performance of the proposed method is stable in the given intervals to the change of η and γ. For example, the values of DSI and CE kept unchanged when γ changed from 10 to 50. The segmentation accuracy was slightly affected by the parameters α and β. But these effects were still reasonable and acceptable. For example, DSI only changed from 0.83 to 0.90 when α changed by 10 times from 3×10−8 to 3×10−7, and the DSI changed from 0.81 to 0.90 when β changed from 2×10−2 to 6×10−2.

Table 3.

Segmentation performance of the proposed method with different α, β, η and γ in a given interval.

α 3×10−8 5×10−8 8×10−8 1×10−7 3×10−7
DSI 0.83 0.87 0.89 0.90 0.86
VE 0.30 0.12 0.05 0.05 0.22
CE 0.39 0.26 0.22 0.19 0.25
β 2×10−2 3×10−2 4×10−2 5×10−2 6×10−2
DSI 0.81 0.87 0.89 0.90 0.88
VE 0.30 0.17 0.11 0.05 0.07
CE 0.32 0.23 0.21 0.19 0.24
η 2×10−3 3×10−3 4×10−3 5×10−3 6×10−3
DSI 0.86 0.90 0.89 0.89 0.87
VE 0.06 0.05 0.11 0.12 0.18
CE 0.28 0.19 0.21 0.20 0.24
γ 10 20 30 40 50
DSI 0.90 0.90 0.90 0.90 0.90
VE 0.05 0.05 0.05 0.06 0.05
CE 0.19 0.19 0.19 0.19 0.19

Fig. 17 displays the restored images of the proposed method with different parameters. Compared with the original PET image (Fig. 17 (a)), each set of parameters obtained a clearer restored image. By observing the three images of each column, we can see that the variation of the parameters in the given interval affected the restoration results slightly. The second column (Fig. 17 (b)–(d)) shows that a larger α obtained a restored image with sharper edges but smaller foreground area (tumor area). With the increase of β (Fig. 17 (e)–(g)), the restored image became more smooth. Fig. 17 (h)–(j) shows that a larger η resulted in smoother tumor edges. The parameter γ can adjust the width of the blur kernel. The larger the γ, the bigger the estimated blur kernel widths (σ1 and σ2). Only when the corresponding FWHM values of the σ1 and σ2 were close to the spatial resolutions of the PET scanner can we achieved the desired results.

Figure 17.

Figure 17

The restored images of the proposed method with different parameters. (a) Original PET image of an esophageal cancer patient; the second column (b)–(d): the restored image with different α; the third column (e)–(g): the restored image with different β; the fourth column (h)–(j): the restored image with different η; the fifth column (k)–(m): the restored image with different γ.

3.6. Execution time

Each iteration of the alternating minimization of the proposed method consists of two gradient descent algorithms and two bisection methods. In the whole iterative process, the most time-consuming part was the gradient descent algorithm due to its disadvantage of the slow convergence. For an image with size of N×N×N, the computational complexity of our algorithm is O(M×N3), where M denotes the iteration number of the AM algorithm. Fortunately, tumor segmentation in radiation oncology was generally conducted in a ROI with small size (small N) rather than the whole PET image. Our current algorithm was implemented in Matlab, and the typical execution time was about several minutes for one patient in our experiments (3.40 GHz CPU, 8GB RAM). The current code still has room to speed up.

4. Discussion

Image restoration and segmentation are both well-known difficult problems in image processing. Due to the limited spatial resolution of clinical PET systems, PET imaging can be largely affected by PVE, especially for tumors with small sizes (Pan and Mawlawi, 2008; Zaidi and El Naqa, 2010). Tumor boundaries are commonly blurred due to PVE (Fig. 8 (a), (d), (g); Fig. 12 (a); Fig. 16 (a), (d)), which affects tumor segmentation accuracy (Soret et al., 2007). One strategy to handle PVE is to correct PVE firstly, and then to segment tumor on the restored image. De-convolution is a widely used method for PVC. However, de-convolution is itself a well-known ill-posed inverse problem. In some circumstances, the system PSF may be unknown, which makes the de-convolution even more difficult. On the contrary, if a precise segmentation result can be available, the blur kernel can be estimated based on the segmentation result, which will be helpful to image de-convolution. As a result, image restoration and segmentation can promote each other, and it is meaningful to integrate the two problems together.

In Ref (Bar et al., 2006), the authors integrated image de-convolution with segmentation in a variational framework. In their model, a piecewise-smoothness term (L2 regularization) was used to constrain the regions away from object edges, and no constraint was used for the edge set, which sometimes resulted in abnormal large intensity voxels on object edges (Fig. 8 (c), (f), (i); Fig. 12 (c); and Fig. 16 (c), (f)) and further influenced the estimation of the blur kernel and degraded the quality of the restored image.

In this study, we conducted PET image restoration, tumor segmentation and blur kernel estimation simultaneously on a novel variational method. In the proposed method, the restored image was constrained by two regularizations — TV regularization was used over tumor edges and L2 regularization was used in non-edge regions. The multiple regularizations preserved not only edges but also smooth regions. The experimental results showed that the proposed method could handle PVE and improve segmentation performance. As shown in Figs.1, 4, 9 and 13, the proposed method performed the best according to all the three accuracy evaluation criteria — DSI=0.80, CE=0.43 and VE=0.28 in the toy simulation, DSI=0.77, CE=0.41 and VE=0.20 for the physical phantom, DSI=0.79, CE=0.41 and VE=0.08 for the non-Hodgkin’s lymphoma dataset, and DSI=0.80, CE=0.42 and VE=0.06 for the esophageal cancer dataset — among all tested methods. Figs. 2, 5 and 6 showed that the proposed method was robust to the change of object sizes and SBRs.

Besides the good segmentation performance, the proposed method can effectively estimate the blur kernel (Fig. 3 (f), Fig. 7 (g), Fig. 11 (g), Fig. 15 (g)) and obtain the desired restored image (Fig. 3 (d), Fig. 7 (d), Fig. 11 (d), Fig. 15 (d), Table 1 and Table 2). In the objective functional (Eq.(7)), the blur kernel was constrained to the anisotropic Gaussians parameterized by different widths in transversal (σ1) and axial (σ2) directions, as shown in Eq.(8), which can handle PET images with different spatial resolution in different directions and improve the algorithm’s robustness. In the toy simulation, the estimated σ1 and σ2 were nearly equal the real values. In physical phantom and clinic datasets experiments, σ1 (transversal direction) and σ2 (axial direction) converged to different values and the corresponding FWHM values were very close to those of the real PET scanners. For the restored images, the RCs were close to 1 both in the toy simulation and in the phantom study.

There are four regularization terms in the objective energy functional (Eq.(6)), which was iteratively optimized using the AM algorithm through the Γ-convergence approximation (Eq.(7)). All these four regularization terms are important in the proposed variational framework. The second term in Eqs.(6) and (7) is the L2 regularization needed for both restoration and segmentation. For image restoration, it constrains the smoothness of the non-edge areas and removes the noise. For segmentation, it plays a crucial role in the evolving of the edge map v which is used to delineate the tumor at last. On tumor edges, the gradient values are high, which made the corresponding values of the edge map v gradually tend to 0 during the minimizing process of the energy functional. Without this term, the proposed model would not be able to achieve tumor segmentation. The third term is the TV regularization applied over target edges. It mainly contributes to preserving and sharpening the tumor edges. If this term was omitted, some abnormal large uptake values would appear on tumor edges. The fourth term is the length regularization term that constrains the length and smoothness of the edge set. If this term was omitted, the target edges in the segmentation result would be jagged and the segmentation accuracy would be reduced. The last term in Eq.(6) and (7) represents the regularization of the blur kernel. If this term was omitted, it would be difficult to control the widths of the blur kernel (Gaussian kernel). The estimation accuracy of the blur kernel may decrease, which would affect the image restoration and further reduce the segmentation accuracy.

Most of existing segmentation algorithms taking into account PVE need to take some prior calibrations through phantom experiments to make the parameters of these methods fit a specific imaging setting, such as some adaptive threshold approaches (Daisne et al., 2003; Drever et al., 2006; Jentzen et al., 2007; Schaefer et al., 2008), or need to know the PSF of the scanner in advance. These methods can be highly system-dependent. Different from these methods, the proposed method needs no prior calibrations or other prior information, and does not depend on the imaging scanners. In this sense, the proposed method is more efficient and has wider applications.

Although the proposed method had a good performance for PET image processing, there are still some issues that need to be resolved in the future. Firstly, the energy functional (Eq.(7)) is non-convex. Therefore, the optimization algorithms can be trapped into local minima and it is a common problem for most variational image processing algorithms. Obtaining the global minimum will be an important and challenging work in our future work. We will study the convex relaxation of the proposed model based on previous researches (Bresson et al., 2007; Chan et al., 2006; Paul et al., 2013), and solve the relaxed convex problem using the alternating split-Bregman (ASB) method (Goldstein et al., 2010; Paul et al., 2013). Secondly, the blur kernel was constrained to Gaussian in this study. Sometimes it may be not accurate. In future work, more types of blur kernel will be studied.

5. Conclusion

In this study, we proposed a novel variational method with multiple regularizations for simultaneous image restoration, tumor segmentation and blur kernel estimation in PET imaging. The alternate usage of the TV and L2 regularizations can preserve not only edges but also smooth regions. The proposed method can handle PVE and is independent to imaging systems and settings. Experimental results demonstrated that the proposed method had a good segmentation and restoration performance, and could accurately estimate the blur kernel as well.

Acknowledgments

Shan Tan and Laquan Li were supported in part by the National Natural Science Foundation of China, under Grant Nos. 61375018 and 61672253, and Fundamental Research Funds for the Central Universities, under Grant No. 2012QN086. Wei Lu was supported in part by the National Institutes of Health Grant No. R01 CA172638.

Appendix A

The detailed mathematical derivations of the Eqs.(10) and (11) corresponding to the objective functional (Eq.(7)) are given in this appendix. In the following derivation, let f(x,y,z) and v(x,y,z) be extended to the whole ℝ3 with the Neumann boundary condition, and hσ denotes hσ1,σ2, and the integration element dA denotes dxdydz.

(1) Variation with respect to v

The variation of Fε with respect to v can be expressed as

δFεδv=λFε(v+λψ)|λ=0=λ{βΩ(v+λψ)2|f|2dA+ηΩ(1-(v+λψ)2)|f|dA+αΩ(ε|(v+λψ)|2+(v+λψ-1)24ε)dA}|λ=0=0. (A.1)

Let

δFεIδv=λ{βΩ(v+λψ)2|f|2dA}|λ=0,δFεIIδv=λ{ηΩ(1-(v+λψ)2)|f|dA}|λ=0,δFεIIIδv=λ{αΩ(ε|(v+λψ)|2)dA}|λ=0,δFεIVδv=λ{αΩ((v+λψ-1)24ε)dA}|λ=0,

then

δFεIδv={βΩ2(v+λψ)ψ|f|2dA}|λ=0=2βΩv|f|2ψdA, (A.2)
δFεIIδv={ηΩ-2(v+λψ)ψ|f|2dA}|λ=0=-2ηΩv|f|ψdA, (A.3)
δFεIIIδv={αΩ(2ε(vx+λψx)ψx+2ε(vy+λψy)ψy+2ε(vz+λψz)ψz)dA}|λ=0=2αεΩ(vxψx+vyψy+vzψz)dA=2αε{Ωx(vxψ)dA+Ωy(vyψ)dA+Ωz(vzψ)dA-Ω(vxx+vyy+vzz)ψdA}. (A.4)

By using the divergence theorem, we get

Ωx(vxψ)dA+Ωy(vyψ)dA+Ωz(vzψ)dA=Ωψvdn, (A.5)

where n is the normal vector to the contour. With the Neumann boundary condition:

vn=0,(x,y,x)Ω,

we obtain

Ωψvdn=0. (A.6)

By substituting Eqs.(A.5) and (A.6) into Eq.(A.4), we have

δFεIIIδv=-2αεΩ(vxx+vyy+vzz)ψdA=-2αεΩ2v·ψdA. (A.7)
δFεIVδv={αΩ24ε(v+λψ-1)ψdA}|λ=0=αΩ(v-1)2εψdA. (A.8)

Substituting Eqs.(A.2), (A.3), (A.7) and (A.8) into Eq.(A.1), we have

δFεδv=δFεIδv+δFεIIδv+δFεIIIδv+δFεIVδv=Ω(2βv|f|2-2ηv|f|-2αε2v+α(v-1)2ε)ψdA=0. (A.9)

For Eq.(A.9), using the fundamental lemma of calculus of variations, we obtain Eq.(10):

(2β|f|2-2η|f|)v+α·v-12ε-2εα2v=0.

(2) Variation with respect to f

The variation of Fε with respect to f can be expressed as

δFεδf=λFε(f+λϕ)|λ=0=λ{12Ω(hσ(f+λϕ)-g)2dA+βΩv2((fx+λϕx)2+(fy+λϕy)2+(fz+λϕz)2)dA+ηΩ(1-v2)(fx+λϕx)2+(fy+λϕy)2+(fz+λϕz)2dA}|λ=0=0. (A.10)

Let

δFεIδf=λ{12Ω(hσ(f+λϕ)-g)2dA}|λ=0,δFεIIδf=λ{βΩv2((fx+λϕx)2+(fy+λϕy)2+(fz+λϕz)2)dA}|λ=0,δFεIIIδf=λ{ηΩ(1-v2)(fx+λϕx)2+(fy+λϕy)2+(fz+λϕz)2dA}|λ=0,

then

δFεIδf={12Ω2(hσ(f+λϕ)-g)(hσϕ)dA}|λ=0=Ω(hσf-g)(hσϕ)dA. (A.11)

Let l(x, y, z) = hσ * fg, then

δFεIδf=Ωl(x,y,z)(hσϕ)dA=Ωl(x,y,z)(Ωhσ(x-x,y-y,z-z)ϕ(x,y,z)dxdydz)dxdydz=Ω(Ωl(x,y,z)hσ(x-x,y-y,z-z)dxdydz)ϕ(x,y,z)dxdydz=Ω(l(x,y,z)hσ(-x,-y,-z))ϕ(x,y,z)dxdydz. (A.12)
δFεIIδf={βΩv2(2(fx+λϕx)ϕx+2(fy+λϕy)ϕy+2(fz+λϕz)ϕz)dA}|λ=0=2βΩv2(fxϕx+fyϕy+fzϕz)dA=2β{Ωx(v2ϕfx)dA+Ωy(v2ϕfy)dA+Ωz(v2ϕfz)dA-Ω(x(v2fx)+y(v2fy)+z(v2fz))ϕdA}. (A.13)

By using the divergence theorem, we get

Ωx(v2ϕfx)dA+Ωy(v2ϕfy)dA+Ωz(v2ϕfz)dA=Ωv2ϕfdn, (A.14)

where n is the normal vector to the contour. With the Neumann boundary condition:

fn=0,(x,y,x)Ω,

we obtain

Ωv2ϕfdn=0. (A.15)

By substituting Eqs.(A.14) and (A.15) into Eq.(A.13), we get

δFεIIδf=-2βΩ(x(v2fx)+y(v2fy)+z(v2fz))ϕdA=-2βΩ(Div(v2f))ϕdA. (A.16)
δFεIIIδf={ηΩ(1-v2)(fx+λϕx)ϕx+(fy+λϕy)ϕy+(fz+λϕz)ϕz(fx+λϕx)2+(fy+λϕy)2+(fz+λϕz)2dA}|λ=0=ηΩ(1-v)2fxϕx+fyϕy+fzϕz(fx)2+(fy)2+(fz)2dA=η{Ωx((1-v2)ϕfx|f|)dA+Ωy((1-v2)ϕfy|f|)dA+Ωz((1-v2)ϕfz|f|)dA-Ω(x((1-v2)fx|f|)+y((1-v2)fy|f|)+z((1-v2)fz|f|))ϕdA}. (A.17)

By using the divergence theorem, we get

Ωx((1-v2)ϕfx|f|)dA+Ωy((1-v2)ϕfy|f|)dA+Ωz((1-v2)ϕfz|f|)dA=Ω(1-v2)ϕf|f|dn. (A.18)

As with Eq.(A.14), using the Neumann boundary condition, we obtain

Ω(1-v2)ϕf|f|dn=0. (A.19)

Substituting Eqs.(A.18) and (A.19) into Eq.(A.17), we obtain

δFεIIIδf=-ηΩ(x((1-v2)fx|f|)+y((1-v2)fy|f|)+z((1-v2)fz|f|))ϕdA=-ηΩ(Div((1-v2)f|f|))ϕdA. (A.20)

Substituting Eqs.(A.12), (A.16) and (A.20) into Eq.(A.10), we obtain

δFεδf=δFεIδf+δFεIIδf+δFεIIIδf=Ω((hσf-g)hσ(-x,-y,-z)-2βDiv(v2f)-ηDiv((1-v2)f|f|))ϕdA=0. (A.21)

For Eq.(A.21), using the fundamental lemma of calculus of variations, we obtain Eq.(11):

(hσf-g)hσ(-x,-y,-z)-2βDiv(v2f)-ηDiv((1-v2)f|f|)=0.

Appendix B

In this appendix, we present the detailed derivation of Eqs.(17) and (19).

The anisotropic Gaussian kernel is defined as:

hσ1,σ2=1(2π)3/2σ12σ2e-(x2+y22σ12+z22σ22).

Following this definition, Eq.(17) can be obtained using the following steps:

hσ1,σ2σ1=1(2π)3/2σ2e-z22σ22σ1(1σ12e-x2+y22σ12)=1(2π)3/2σ2e-z22σ22(-2σ13e-x2+y22σ12+1σ12e-x2+y22σ12x2+y2σ13)=1(2π)3/2σ12σ2e-(x2+y22σ12+z22σ22)(x2+y2σ13-2σ1).

Similarly, we can derive Eq.(18).

Eq.(19) can be obtained as:

σ1|hσ1,σ2|2=σ1((xhσ1,σ2)2+(yhσ1,σ2)2+(zhσ1,σ2)2)=σ1(1(2π)3σ14σ22e-(x2+y2σ12+z2σ22)(x2+y2σ14+z2σ24))=σ1(x2+y2(2π)3σ18σ22e-(x2+y2σ12+z2σ22))+σ1(z2(2π)3σ14σ26e-(x2+y2σ12+z2σ22))=x2+y24π3σ14σ22e-(x2+y2σ12+z2σ22)(x2+y2σ17-4σ15)+z24π3σ14σ22e-(x2+y2σ12+z2σ22)(x2+y2σ13σ24-2σ1σ24).

In the same way, we can derive Eq.(20).

References

  1. Alenius S, Ruotsalainen U. Bayesian image reconstruction for emission tomography based on median root prior. European journal of nuclear medicine. 1997;24:258–265. doi: 10.1007/BF01728761. [DOI] [PubMed] [Google Scholar]
  2. Ambrosio L, Tortorelli VM. Approximation of functional depending on jumps by elliptic functional via t-convergence. Communications on Pure and Applied Mathematics. 1990;43:999–1036. [Google Scholar]
  3. Aristophanous M, Penney BC, Martel MK, Pelizzari CA. A Gaussian mixture model for definition of lung tumor volumes in positron emission tomography. Medical physics. 2007;34:4223–4235. doi: 10.1118/1.2791035. [DOI] [PubMed] [Google Scholar]
  4. Ayasso H, Mohammad-Djafari A. Joint NDT image restoration and segmentation using Gauss–Markov–Potts prior models and variational bayesian computation. Image Processing, IEEE Transactions on. 2010;19:2265–2277. doi: 10.1109/TIP.2010.2047902. [DOI] [PubMed] [Google Scholar]
  5. Bađc U, Yao J, Caban J, Turkbey E, Aras O, Mollura DJ. A graph-theoretic approach for segmentation of PET images, Engineering in Medicine and Biology Society, EMBC. 2011 Annual International Conference of the IEEE; IEEE; 2011. pp. 8479–8482. [DOI] [PMC free article] [PubMed] [Google Scholar]
  6. Bar L, Sochen N, Kiryati N. Semi-blind image restoration via Mumford-Shah regularization. Image Processing, IEEE Transactions on. 2006;15:483–493. doi: 10.1109/tip.2005.863120. [DOI] [PubMed] [Google Scholar]
  7. Barbee DL, Flynn RT, Holden JE, Nickles RJ, Jeraj R. A method for partial volume correction of PET-imaged tumor heterogeneity using expectation maximization with a spatially varying point spread function. Physics in medicine and biology. 2010;55:221. doi: 10.1088/0031-9155/55/1/013. [DOI] [PMC free article] [PubMed] [Google Scholar]
  8. Belhassen S, Zaidi H. A novel fuzzy C-means algorithm for unsupervised heterogeneous tumor quantification in PET. Medical physics. 2010;37:1309–1324. doi: 10.1118/1.3301610. [DOI] [PubMed] [Google Scholar]
  9. Betrouni N, Makni N, Dewalle-Vignion AS, Vermandel M. MedataWeb: A shared platform for multimodality medical images and Atlases. IRBM. 2012;33:223–226. [Google Scholar]
  10. Biehl KJ, Kong FM, Dehdashti F, Jin JY, Mutic S, El Naqa I, Siegel BA, Bradley JD. 18F-FDG PET definition of gross tumor volume for radiotherapy of non–small cell lung cancer: is a single standardized uptake value threshold approach appropriate? Journal of Nuclear Medicine. 2006;47:1808–1812. [PubMed] [Google Scholar]
  11. Boussion N, Le Rest CC, Hatt M, Visvikis D. Incorporation of wavelet-based denoising in iterative deconvolution for partial volume correction in whole-body PET imaging. European journal of nuclear medicine and molecular imaging. 2009;36:1064–1075. doi: 10.1007/s00259-009-1065-5. [DOI] [PubMed] [Google Scholar]
  12. Boykov Y, Funka-Lea G. Graph cuts and efficient ND image segmentation. International journal of computer vision. 2006;70:109–131. [Google Scholar]
  13. Bresson X, Esedoḡlu S, Vandergheynst P, Thiran JP, Osher S. Fast global minimization of the active contour/snake model. Journal of Mathematical Imaging and vision. 2007;28:151–167. [Google Scholar]
  14. Brix G, Doll J, Bellemann ME, Trojan H, Haberkorn U, Schmidlin P, Ostertag H. Use of scanner characteristics in iterative image reconstruction for high-resolution positron emission tomography studies of small animals. European Journal of Nuclear Medicine and Molecular Imaging. 1997;24:779–786. doi: 10.1007/BF00879667. [DOI] [PubMed] [Google Scholar]
  15. Cai X. Variational image segmentation model coupled with image restoration achievements. Pattern Recognition. 2015;48:2029–2042. [Google Scholar]
  16. Carasso AS. Linear and nonlinear image deblurring: A documented study. SIAM journal on numerical analysis. 1999;36:1659–1689. [Google Scholar]
  17. Caselles V, Kimmel R, Sapiro G. Geodesic active contours. International journal of computer vision. 1997;22:61–79. [Google Scholar]
  18. Chan T, Marquina A, Mulet P. High-Order Total Variation-Based Image Restoration. Siam Journal on Scientific Computing. 2000;22:503–516. [Google Scholar]
  19. Chan TF, Esedoglu S, Nikolova M. Algorithms for finding global minimizers of image segmentation and denoising models. SIAM journal on applied mathematics. 2006;66:1632–1648. [Google Scholar]
  20. Chan TF, Shen JJ. Image processing and analysis: variational, PDE, wavelet, and stochastic methods. Siam 2005 [Google Scholar]
  21. Chan TF, Vese LA. Active contours without edges. Image processing, IEEE transactions on. 2001;10:266–277. doi: 10.1109/83.902291. [DOI] [PubMed] [Google Scholar]
  22. Chan TF, Wong CK. Total variation blind deconvolution. Image Processing, IEEE Transactions on. 1998;7:370–375. doi: 10.1109/83.661187. [DOI] [PubMed] [Google Scholar]
  23. Daisne JF, Sibomana M, Bol A, Doumont T, Lonneux M, Grégoire V. Tri-dimensional automatic segmentation of PET volumes based on measured source-to-background ratios: influence of reconstruction algorithms. Radiotherapy and Oncology. 2003;69:247–250. doi: 10.1016/s0167-8140(03)00270-6. [DOI] [PubMed] [Google Scholar]
  24. De Bernardi E, Faggiano E, Zito F, Gerundini P, Baselli G. Lesion quantification in oncological positron emission tomography: A maximum likelihood partial volume correction strategy. Medical physics. 2009;36:3040–3049. doi: 10.1118/1.3130019. [DOI] [PubMed] [Google Scholar]
  25. DeGrado TR, Turkington TG, Williams JJ, Stearns CW, Hoffman JM, Coleman RE. Performance characteristics of a whole-body PET scanner. Journal of Nuclear Medicine. 1994;35:1398. [PubMed] [Google Scholar]
  26. Dewalle-Vignion A, Betrouni N, Lopes R, Huglo D, Stute S, Vermandel M. A new method for volume segmentation of PET images, based on possibility theory. Medical Imaging, IEEE Transactions on. 2011;30:409–423. doi: 10.1109/TMI.2010.2083681. [DOI] [PubMed] [Google Scholar]
  27. Drever L, Robinson DM, McEwan A, Roa W. A local contrast based approach to threshold segmentation for PET target volume delineation. Medical physics. 2006;33:1583–1594. doi: 10.1118/1.2198308. [DOI] [PubMed] [Google Scholar]
  28. El Naqa I, Bradley J, Deasy J, Biehl K, Laforest R, Low D. Improved analysis of PET images for radiation therapy. 14th International Conference on the Use of Computers in Radiation Therapy; Seoul, Korea. 2004. pp. 361–363. [Google Scholar]
  29. El Naqa I, Low DA, Bradley JD, Vicic M, Deasy JO. Deblurring of breathing motion artifacts in thoracic PET images by deconvolution methods. Medical physics. 2006;33:3587–3600. doi: 10.1118/1.2336500. [DOI] [PubMed] [Google Scholar]
  30. Erdi YE, Mawlawi O, Larson SM, Imbriaco M, Yeung H, Finn R, Humm JL. Segmentation of lung lesion volume by adaptive positron emission tomography image thresholding. Cancer. 1997;80:2505–2509. doi: 10.1002/(sici)1097-0142(19971215)80:12+<2505::aid-cncr24>3.3.co;2-b. [DOI] [PubMed] [Google Scholar]
  31. Foster B, Bagci U, Mansoor A, Xu Z, Mollura DJ. A review on segmentation of positron emission tomography images. Computers in biology and medicine. 2014;50:76–96. doi: 10.1016/j.compbiomed.2014.04.014. [DOI] [PMC free article] [PubMed] [Google Scholar]
  32. Geets X, Lee JA, Bol A, Lonneux M, Grégoire V. A gradient-based method for segmenting FDG-PET images: methodology and validation. European journal of nuclear medicine and molecular imaging. 2007;34:1427–1438. doi: 10.1007/s00259-006-0363-4. [DOI] [PubMed] [Google Scholar]
  33. Geworski L, Knoop BO, de Cabrejas ML, Knapp WH, Munz DL. Recovery correction for quantitation in emission tomography: a feasibility study. European journal of nuclear medicine. 2000;27:161–169. doi: 10.1007/s002590050022. [DOI] [PubMed] [Google Scholar]
  34. Goldstein T, Bresson X, Osher S. Geometric applications of the split Bregman method: segmentation and surface reconstruction. Journal of Scientific Computing. 2010;45:272–293. [Google Scholar]
  35. Hatt M, Cheze-le Rest C, Van Baardwijk A, Lambin P, Pradier O, Visvikis D. Impact of tumor size and tracer uptake heterogeneity in 18F-FDG PET and CT Non–Small Cell Lung Cancer tumor delineation. Journal of Nuclear Medicine. 2011a;52:1690–1697. doi: 10.2967/jnumed.111.092767. [DOI] [PMC free article] [PubMed] [Google Scholar]
  36. Hatt M, Cheze le Rest C, Turzo A, Roux C, Visvikis D. A fuzzy locally adaptive Bayesian segmentation approach for volume determination in PET. Medical Imaging, IEEE Transactions on. 2009;28:881–893. doi: 10.1109/TMI.2008.2012036. [DOI] [PMC free article] [PubMed] [Google Scholar]
  37. Hatt M, Le Rest CC, Albarghach N, Pradier O, Visvikis D. PET functional volume delineation: a robustness and repeatability study. European journal of nuclear medicine and molecular imaging. 2011b;38:663–672. doi: 10.1007/s00259-010-1688-6. [DOI] [PubMed] [Google Scholar]
  38. Hsu CY, Liu CY, Chen CM. Automatic segmentation of liver PET images. Computerized Medical Imaging and Graphics. 2008;32:601–610. doi: 10.1016/j.compmedimag.2008.07.001. [DOI] [PubMed] [Google Scholar]
  39. Humm JL, Rosenfeld A, Del Guerra A. From PET detectors to PET scanners. European journal of nuclear medicine and molecular imaging. 2003;30:1574–1597. doi: 10.1007/s00259-003-1266-2. [DOI] [PubMed] [Google Scholar]
  40. Jentzen W, Freudenberg L, Eising EG, Heinze M, Brandau W, Bockisch A. Segmentation of PET volumes by iterative image thresholding. Journal of Nuclear Medicine. 2007;48:108–114. [PubMed] [Google Scholar]
  41. Juweid ME, Cheson BD. Positron-emission tomography and assessment of cancer therapy. New England Journal of Medicine. 2006;354:496–507. doi: 10.1056/NEJMra050276. [DOI] [PubMed] [Google Scholar]
  42. Kerhet A, Small C, Quon H, Riauka T, Greiner R, McEwan A, Roa W. Artificial Intelligence in Medicine. Springer; 2009. Segmentation of lung tumours in positron emission tomography scans: a machine learning approach; pp. 146–155. [Google Scholar]
  43. Kessler RM, Ellis JR, Jr, Eden M. Analysis of emission tomographic scan data: limitations imposed by resolution and background. Journal of computer assisted tomography. 1984;8:514–522. doi: 10.1097/00004728-198406000-00028. [DOI] [PubMed] [Google Scholar]
  44. Kirov A, Piao J, Schmidtlein C. Partial volume effect correction in PET using regularized iterative deconvolution with variance control based on local topology. Physics in medicine and biology. 2008;53:2577. doi: 10.1088/0031-9155/53/10/009. [DOI] [PubMed] [Google Scholar]
  45. Li H, Thorstad WL, Biehl KJ, Laforest R, Su Y, Shoghi KI, Donnelly ED, Low DA, Lu W. A novel PET tumor delineation method based on adaptive region-growing and dual-front active contours. Medical physics. 2008;35:3711–3721. doi: 10.1118/1.2956713. [DOI] [PMC free article] [PubMed] [Google Scholar]
  46. Lodge M, Dilsizian V, Line B. Performance assessment of the Philips GEMINI PET/CT scanner. Journal of Nuclear Medicine. 2004;45:425. [Google Scholar]
  47. Lucy LB. An iterative technique for the rectification of observed distributions. The astronomical journal. 1974;79:745. [Google Scholar]
  48. Mumford D, Shah J. Optimal approximations by piecewise smooth functions and associated variational problems. Communications on pure and applied mathematics. 1989;42:577–685. [Google Scholar]
  49. Nestle U, Kremp S, Schaefer-Schuler A, Sebastian-Welsch C, Hellwig D, Rübe C, Kirsch CM. Comparison of different methods for delineation of 18F-FDG PET–positive tissue for target volume definition in radiotherapy of patients with non–small cell lung cancer. Journal of Nuclear Medicine. 2005;46:1342–1348. [PubMed] [Google Scholar]
  50. Ollinger JM, Fessler JA. Positron-emission tomography. IEEE Signal Processing Magazine. 1997;14:43–55. [Google Scholar]
  51. Otsu N. A threshold selection method from gray-level histograms. Automatica. 1975;11:23–27. [Google Scholar]
  52. Pan T, Mawlawi O. PET/CT in radiation oncology. Medical physics. 2008;35:4955–4966. doi: 10.1118/1.2986145. [DOI] [PubMed] [Google Scholar]
  53. Paul G, Cardinale J, Sbalzarini IF. Coupling image restoration and segmentation: a generalized linear model/Bregman perspective. International Journal of Computer Vision. 2013;104:69–93. [Google Scholar]
  54. Perrone D, Favaro P. Total variation blind deconvolution: The devil is in the details, Computer Vision and Pattern Recognition (CVPR). 2014 IEEE Conference on; IEEE; 2014. pp. 2909–2916. [Google Scholar]
  55. Prieto E, Lecumberri P, Pagola M, Gomez M, Bilbao I, Ecay M, Penuelas I, Marti-Climent JM. Twelve automated thresholding methods for segmentation of PET images: a phantom study. Physics in medicine and biology. 2012;57:3963. doi: 10.1088/0031-9155/57/12/3963. [DOI] [PubMed] [Google Scholar]
  56. Richardson WH. Bayesian-based iterative method of image restoration. JOSA. 1972;62:55–59. [Google Scholar]
  57. Riddell C, Brigger P, Carson R, Bacharach S. The watershed algorithm: a method to segment noisy PET transmission images. Nuclear Science, IEEE Transactions on. 1999;46:713–719. [Google Scholar]
  58. Rohren EM, Turkington TG, Coleman RE. Clinical applications of PET in oncology1. Radiology. 2004;231:305–332. doi: 10.1148/radiol.2312021185. [DOI] [PubMed] [Google Scholar]
  59. Rousset OG, Ma Y, Evans AC. Correction for partial volume effects in PET: principle and validation. Journal of Nuclear Medicine. 1998;39:904–911. [PubMed] [Google Scholar]
  60. Rudin LI, Osher S, Fatemi E. Nonlinear total variation based noise removal algorithms. Physica D: Nonlinear Phenomena. 1992;60:259–268. [Google Scholar]
  61. Schaefer A, Kremp S, Hellwig D, Rübe C, Kirsch CM, Nestle U. A contrast-oriented algorithm for FDG-PET-based delineation of tumour volumes for the radiotherapy of lung cancer: derivation from phantom measurements and validation in patient data. European journal of nuclear medicine and molecular imaging. 2008;35:1989–1999. doi: 10.1007/s00259-008-0875-1. [DOI] [PubMed] [Google Scholar]
  62. Scherzer O. Handbook of Mathematical Methods in Imaging. Vol. 1. Springer Science & Business Media; 2011. [Google Scholar]
  63. Soret M, Bacharach SL, Buvat I. Partial-volume effect in PET tumor imaging. Journal of Nuclear Medicine. 2007;48:932–945. doi: 10.2967/jnumed.106.035774. [DOI] [PubMed] [Google Scholar]
  64. Srinivas SM, Dhurairaj T, Basu S, Bural G, Surti S, Alavi A. A recovery coefficient method for partial volume correction of PET images. Annals of nuclear medicine. 2009;23:341–348. doi: 10.1007/s12149-009-0241-9. [DOI] [PubMed] [Google Scholar]
  65. Tan S, Kligerman S, Chen WG, Lu M, Kim G, Feigenberg S, D’Souza WD, Suntharalingam M, Lu W. Spatial-Temporal F-18 FDG-PET Features for Predicting Pathologic Response of Esophageal Cancer to Neoadjuvant Chemoradiation Therapy. International Journal of Radiation Oncology Biology Physics. 2013;85:1375–1382. doi: 10.1016/j.ijrobp.2012.10.017. [DOI] [PMC free article] [PubMed] [Google Scholar]
  66. Tan S, Li L, DSouza W, Lu W. Optimal tumor segmentation in PET using multiple thresholds. Annual Meeting of the Society of Nuclear Medicine and Molecular Imaging; Baltimore, MD, United States. 2015. pp. 1782–1782. [Google Scholar]
  67. Teo BK, Seo Y, Bacharach SL, Carrasquillo JA, Libutti SK, Shukla H, Hasegawa BH, Hawkins RA, Franc BL. Partial-volume correction in PET: validation of an iterative postreconstruction method with phantom and patient data. Journal of Nuclear Medicine. 2007;48:802–810. doi: 10.2967/jnumed.106.035576. [DOI] [PubMed] [Google Scholar]
  68. Tohka J, Reilhac A. Deconvolution-based partial volume correction in Raclopride-PET and Monte Carlo comparison to MR-based method. Neuroimage. 2008;39:1570–1584. doi: 10.1016/j.neuroimage.2007.10.038. [DOI] [PubMed] [Google Scholar]
  69. Tong S, Alessio AM, Kinahan PE. Noise and signal properties in PSF-based fully 3D PET image reconstruction: an experimental evaluation. Physics in Medicine & Biology. 2010;55:1453–1473. doi: 10.1088/0031-9155/55/5/013. [DOI] [PMC free article] [PubMed] [Google Scholar]
  70. Van Cittert P. Effect of slit width upon distribution of intensity in spectral lines. II. Z Phys. 1931;69:298–308. [Google Scholar]
  71. You YL, Kaveh M. A regularization approach to joint blur identification and image restoration. Image Processing, IEEE Transactions on. 1996;5:416–428. doi: 10.1109/83.491316. [DOI] [PubMed] [Google Scholar]
  72. Zaidi H, El Naqa I. PET-guided delineation of radiation therapy treatment volumes: a survey of image segmentation techniques. European journal of nuclear medicine and molecular imaging. 2010;37:2165–2187. doi: 10.1007/s00259-010-1423-3. [DOI] [PubMed] [Google Scholar]
  73. Zheng H, Hellwich O. Extended mumford-shah regularization in bayesian estimation for blind image deconvolution and segmentation. Proceedings of the 11th international conference on Combinatorial Image Analysis; Springer-Verlag; 2006. pp. 144–158. [Google Scholar]

RESOURCES