Skip to main content
International Journal of Biomedical Imaging logoLink to International Journal of Biomedical Imaging
. 2010 Sep 26;2010:105610. doi: 10.1155/2010/105610

Artificial Neural Network-Based System for PET Volume Segmentation

Mhd Saeed Sharif 1,*, Maysam Abbod 1, Abbes Amira 2, Habib Zaidi 3, 4,4
PMCID: PMC2948894  PMID: 20936152

Abstract

Tumour detection, classification, and quantification in positron emission tomography (PET) imaging at early stage of disease are important issues for clinical diagnosis, assessment of response to treatment, and radiotherapy planning. Many techniques have been proposed for segmenting medical imaging data; however, some of the approaches have poor performance, large inaccuracy, and require substantial computation time for analysing large medical volumes. Artificial intelligence (AI) approaches can provide improved accuracy and save decent amount of time. Artificial neural networks (ANNs), as one of the best AI techniques, have the capability to classify and quantify precisely lesions and model the clinical evaluation for a specific problem. This paper presents a novel application of ANNs in the wavelet domain for PET volume segmentation. ANN performance evaluation using different training algorithms in both spatial and wavelet domains with a different number of neurons in the hidden layer is also presented. The best number of neurons in the hidden layer is determined according to the experimental results, which is also stated Levenberg-Marquardt backpropagation training algorithm as the best training approach for the proposed application. The proposed intelligent system results are compared with those obtained using conventional techniques including thresholding and clustering based approaches. Experimental and Monte Carlo simulated PET phantom data sets and clinical PET volumes of nonsmall cell lung cancer patients were utilised to validate the proposed algorithm which has demonstrated promising results.

1. Introduction

Medical images can be obtained using various modalities such as positron emission tomography (PET), single-photon emission computed tomography (SPECT), computed tomography (CT), magnetic resonance imaging (MRI), and ultrasound (US). PET is a molecular imaging technique used to probe physiological functions at the molecular level rather than to look at anatomy through the use of trace elements such as carbon, oxygen, and nitrogen which have a high abundance within the human body. PET plays a central role in the management of oncological patients beside the other main components such as diagnosis, staging, treatment, prognosis, and followup. Owing to its high sensitivity and specificity, PET is effective in targeting specific functional or metabolic signatures that may be associated with various diseases. Among all diagnostic and therapeutic procedures, PET is unique in the sense that it is based on molecular and pathophysiological mechanisms and employs radioactively labeled biological molecules as tracers to study the pathophysiology of the tumour in vivo to direct treatment and assess response to therapy. The leading current area of clinical use of PET is in oncology, where 18F-fluorodeoxyglucose (FDG) remains the most widely used tracer. FDG-PET has already had a large valuable effect on cancer staging and treatment, and its use in clinical oncology practice continues to evolve [15].

The main challenge of PET is its low spatial resolution which results in the so-called partial volume effect. This effect should be reduced to the minimum level, so that the required information can be accurately quantified and extracted from the analysed volume. On the other hand, the increasing number of patient scans beside the widespread application of PET have raised the urgent need for effective volume analysis techniques to aid clinicians in clinical diagnosis and set the proper plan for treatment. Analysing and extracting the proper information from PET volumes can be performed by deploying image segmentation and classification approaches which provide richer information than that obtained directly from qualitative assessment alone performed on the original PET volumes [6]. The need for accurate and fast analysis for medical volume segmentation leads to exploit artificial intelligence (AI) techniques. These include artificial neural networks (ANN), expert systems, robotics, genetic algorithms, intelligent agents, logic programming, fuzzy logic, neurofuzzy, natural language processing, and automatic speech recognition [7, 8].

ANN is one of the powerful AI techniques that has the capability to learn from a set of data and construct weight matrices to represent the learning patterns. ANN has great success in many applications including pattern classification, decision making, forecasting, and adaptive control. Many research studies have been carried out in the medical field utilising ANN for medical image segmentation and classification with different medical imaging modalities. Multilayer perceptron (MLP) neural network (NN) have been used by [9] to identify breast nodule malignancy using sonographic images. A multiple classifier system using five NNs and five sets of texture features extraction for the characterization of hepatic tissue from CT images is presented in [10]. Kohonen self-organizing NN for segmentation and a multilayer backpropagation NN for classification for multispectral MRI images have been used in [11]. Kohonen NN was also used for image segmentation in [12]. Computer-aided diagnostic (CAD) scheme to detect lung nodules using a multiresolution massive training artificial neural network (MTANN) is presented in [13].

The aim of this paper is to develop a robust, efficient PET volume segmentation system using ANN. The proposed system is evaluated using different training algorithms and its performance assessed using different metrics. ANN outputs are also compared with the outputs of conventional approaches including thresholding and clustering using experimental PET phantom studies and clinical volumes of nonsmall cell lung cancer patients.

This paper is organised as follows. Section 2 presents mathematical background for the selected approaches. The materials and methods used are described in Section 3. Experimental results and their discussion are given in section 4, and finally some conclusions are presented in Section 5.

2. Mathematical Background

2.1. Mathematical Model of a Neuron

ANN is a mathematical model which emulates the activity of biological neural networks in the human brain. It consists of two or several layers each one has many interconnected group of neurons. Each neuron in the ANN has a number of inputs (the input vector P) and one output (Y). The input vector elements are multiplied by weights w 1,1, w 1,2,…, w 1,R, and the weighted values are fed to the summing junction. Their sum is simply the dot product (W.P) of the single-row matrix W and the vector P. The neuron has a bias b, which is summed with the weighted inputs to form the net input n. This sum, n, is the argument of the transfer function f [14]

n=i=1Rw1,i·pi+b, (1)
Y=f(W·P+b), (2)
Y(j)=f[i=1Rw1,i(j).pi(j)+b]. (3)

The learning process can be summarized in the following steps: (1) the initial weights are randomly assigned, (2) the neuron is activated by applying inputs vector and desired output (Y d), and (3) calculation of the actual output (Y) at iteration j=1 as illustrated in (3), where iteration j refers to the jth training example presented to the neuron. The following step is to update the weights to obtain the output consistent with the training examples, as illustrated in

w1,i(j+1)=w1,i(j)+Δw1,i(j), (4)

where Δw 1,i(j) is the weight correction at iteration j. The weight correction is computed by using the delta rule in

Δw1,i(j)=αpi(j)e(j), (5)

where α is the learning rate and e(j) is the error which can be given by

e(j)=Yd(j)Y(j). (6)

Finally, the iteration j is increased by one, and the previous two steps are repeated until the convergence is reached.

2.2. Thresholding

2.2.1. Hard Thresholding

Thresholding is the simplest precursory technique for image segmentation. This methodology attempts to determine an intensity value that can separate the slice g(x, y) into two parts [15]. All voxels with intensities f(x, y) larger than the threshold value T are allocated into one class, and all the others into another class.

g(x,y)={f(x,y)if  f(x,y)>T,0if  f(x,y)<T. (7)

Thresholding approach does not consider the spatial characteristics of a volume; it is sensitive to noise and intensities variation. Thresholding approach has been used extensively in the literature as ground truth to compare some of the proposed schemes for medical image segmentation [16, 17].

2.2.2. Soft Thresholding

Soft thresholding is more complex process compared to hard thresholding. This approach replaces each voxel which has a greater value than the threshold value by the difference between the threshold and the current voxel values. Soft thresholding could put into evidence some important regions as the region of interest (ROI) in this study.

2.2.3. Adaptive Thresholding

Otsu's method has been used as a third approach, which chooses the threshold that minimizes the intraclass variance of the black and white voxels in the volume [18]. Likewise, other variants of adaptive thresholding based on source-to-background ratio were also reported [4].

2.3. Multiresolution Analysis

Multiresolution analysis (MRA) is designed to give good time resolution and poor frequency resolution at high frequencies, and poor time resolution and good frequency resolution at low frequencies. It enables the exploitation of slice characteristics associated with a particular resolution level, which may not be detected using other analysis techniques [1921]. The wavelet transform for a function f(t) can be defined as follows:

Xψ(a,b)=f(t)ψ(a,b)(t)dt, (8)

where

ψ(a,b)(t)=1aψ(tba). (9)

The parameters a, b are called the scaling and shifting parameters, respectively [20, 22]. Haar wavelet filter will be used in the experimental study at different levels of decomposition. The Haar wavelet transform (HWT) of a two-dimensional slice can be performed using two approaches: the first one is called standard decomposition of a slice, where the one-dimensional HWT is applied to each row of voxel values followed by another one-dimensional HWT on the column of the processed slice. The other approach is called nonstandard decomposition, which alternates between the one-dimensional HWT operations on rows and columns. HWT serves as a prototype for all other wavelet transforms. Like all wavelet transforms, HWT decomposes a slice into four subimages of half the original size. HWT is conceptually simple, fast, memory efficient, and can be reversed without the edge effects that are associated with other wavelet transforms. HWT is a matrix-vector-based operation and can be formulated as follows:

H=12(1111), (10)
O=I×H, (11)
O=((IH)TH)T=HTIH, (12)
I=(H1)TOH1, (13)

where I is 2 × 2 input matrix, H contains the Haar coefficients, and O is the output matrix. Equations (12) and (13) show the transposed and reconstructed matrices, respectively. MRA has been used in the literature for different applications [2224].

2.4. Clustering

Clustering techniques aim to classify each voxel in a volume into the proper cluster, then these clusters are mapped to display the segmented volume. The most commonly used clustering technique is the K-means method, which clusters n voxels into K clusters (K less than n) [25]. This algorithm chooses the number of clusters, K, then randomly generates K clusters and determines the cluster centers. The next step is assigning each point in the volume to the nearest cluster center, and finally recompute the new cluster centers. The two previous steps are repeated until the minimum variance criterion is achieved. This approach is similar to the expectation-maximization algorithm for Gaussian mixture in which they both attempt to find the centers of clusters in the volume. Its main objective is to achieve a minimum intracluster variance V

V=i=1KxjSi(xjμi)2, (14)

where K is the number of clusters, S = 1,2,…, K, and μ i is the mean of all voxels in the cluster i. K-means approach has been used with other techniques for clustering medical images [26].

3. Materials and Methods

3.1. The Proposed System

The proposed medical volume segmentation system is illustrated in Figure 1. The 3D PET volume acquired from the scanner goes through the preprocessing block, which enhances the quality of slice features and removes most of the noise from each slice. The enhanced volume can be processed using three approaches, the first processing block is thresholding which removes the background and unnecessary information producing a volume consists of two classes the background and the ROI. The second approach is K-means clustering technique which classifies each slice in PET volume into an appropriate number of clusters. The third approach is ANN which is used in both spatial and wavelet domains. The preprocessed PET volume is fed first to the ANN which is trained to detect the tumour. In another block, the PET volume is transformed into the wavelet domain using HWT at different levels of decomposition. This transform decomposes the volume and produces the approximation, horizontal, vertical, and diagonal features for each slice. The approximation features are fed to another ANN for classifying and quantifying the tumour. The outputs of ANNs are compared in the next step with the outputs of the other two approaches, while the best outputs are selected and displayed. The system has been tested using experimental and simulated phantom studies and clinical oncological PET volumes of nonsmall cell lung cancer patients.

Figure 1.

Figure 1

Proposed system for PET volume segmentation.

3.2. Phantom Studies

In this study, PET volumes containing simulated tumour have been utilised. Two phantom data sets have been used. The first data set is obtained using NEMA IEC image quality body phantom which consists of an elliptical water-filled cavity with six spherical inserts suspended by plastic rods of volumes 0.5, 1.2, 2.6, 5.6, 11.5, and 26.5 ml (inner diameters of 10, 13, 17, 22, 28, and 37 mm). The voxel size is 4.07 mm × 4.07 mm × 5 mm, while the size of the obtained phantom volume is 168 × 168 × 66. This phantom was extensively used in the literature for assessment of image quality and validation of quantitative procedures [2730]. Other variants of multisphere phantoms have also been suggested [31]. The PET scanner used for acquiring the data is the Biograph 16 PET/CT scanner (Siemens Medical Solution, Erlangen, Germany) operating in 3D mode [32]. Following Fourier rebinning and model-based scatter correction, PET images were reconstructed using two-dimensional iterative normalized attenuation-weighted ordered subsets expectation maximization (NAW-OSEM). CT-based attenuation correction was used to reconstruct the PET emission data. The default parameters used were ordered OSEM iterative reconstruction with four iterations and eight subsets followed by a postprocessing Gaussian filter (kernel full-width half-maximal height, 5 mm).

The second data set consists of Monte Carlo simulations of the Zubal antropommorphic model where two volumes were generated [33]. The first volume contains a matrix with isotropic voxels, the size of this volume is 128 × 128 × 180. The second volume contains the same matrix of the first one but with nonisotropic voxels having a matrix size of 128 × 128 × 375. The voxel size in both volumes is 5.0625 mm × 5.0625 mm × 2.4250 mm. The second data volume has 3 tumours in the lungs whose characteristics are given in Table 1.

Table 1.

Tumours characteristics for the second data set.

Tumour Isotropic Nonisotropic
Number Position Size Position Size
1 slice 68 2 voxels slice 142 2 voxels
2 slice 57 3 voxels slice 119 3 voxels
3 slice 74 2 voxels slice 155 2 voxels

3.3. Clinical PET Studies

Clinical PET volumes of patients with histologically proven NSCLC (clinical Stage Ib-IIIb) who have undertaken a diagnostic whole-body PET/CT scan were used for assessment of the proposed segmentation technique [34]. Patients fasted no less than 6 hours before PET/CT scanning. The standard protocol involved intravenous injection of 18F-FDG followed by a physiologic saline (10 ml). The injected FDG activity was adjusted according to patient's weight using the following formula: A (Mbq) = weight (Kg) 4 + 20. After 45 min uptake time, free-breathing PET and CT images were acquired. The data were reconstructed using the same procedure described for the phantom studies. The maximal tumour diameters measured from the macroscopic examination of the surgical specimen served as ground truth for comparison with the maximum diameter estimated by the proposed segmentation technique. The voxel size is 5.31 mm × 5.31 mm × 5 mm, while the size of the obtained clinical volume is 128 × 128 × 178.

4. Results and Discussion

4.1. NEMA Image Quality Phantom

An experimental study has been run at the beginning to determine the best ANN design and algorithms. Multilayer feedforward NNs [8] consists of input layer (144 neurons), hidden layer (variant number of hidden neurons), and outputs layer (1) has been chosen first to determine the best number of hidden neurons. To evaluate the effect of the number of neurons in the hidden layer and achieve the best ANN performance for our application, different numbers of neurons in the hidden layer have been used. The maximum number of iterations used in the ANN is 1000. The experiment has been repeated 10 times for each chosen number of the hidden neurons, and the average was considered for that number. Hyperbolic tangent sigmoid transfer function has been used for all layers except the output layer where the linear activation function is used. The two activation functions are illustrated in Figure 2. Levenberg-Marquardt backpropagation training algorithm has been used during the evaluation of neurons numbers in the hidden layer [35] to validate the best design for the ANN, which is suitable for the proposed application. Figure 3 presents the number of neurons in the first hidden layer with the performance measured using mean-squared error (MSE) at 1000 iterations. The results obtained after this evaluation shows that the best number of the hidden neurons which corresponds to the smallest MSE, and good ANN outputs is 70 hidden neurons.

Figure 2.

Figure 2

Activation functions: (a) Tangent-sigmoid function, (b) Linear function.

Figure 3.

Figure 3

Evaluation of the number of neurons in the first hidden layer with neural network performance using MSE.

Using the achieved ANN structure, different training algorithms have been evaluated in the next step to achieve the best ANN performance. In this evaluation the same ANN structure, sufficient training cases and 1000 epochs have been considered. The following training algorithms have been used in this part of the study. BFGS quasi-Newton backpropagation [36], bayesian regulation backpropagation (BR) [37], conjugate gradient backpropagation with Powell-Beale restarts (CGB) [38], conjugate gradient backpropagation with Fletcher-Reeves updates (CGF) [39], conjugate gradient backpropagation with Polak-Ribire updates (CGP) [39], gradient descent backpropagation (GD) [40], gradient descent with momentum backpropagation (GDM) [41], gradient descent with adaptive learning rate backpropagation (GDA) [42], gradient descent with momentum and adaptive learning rate backpropagation (GDX) [43], Levenberg-Marquardt backpropagation (LM) [35], one-step secant backpropagation (OSS) [44], random order incremental training with learning functions (R), resilient backpropagation (RP) [45], and scaled conjugate gradient backpropagation (SCG) [46]. The average of the performance and the required time for each of these training algorithms are presented in Table 2, this experiment has been repeated for 10 times and the standard deviation for the performance achieved is 4.15E-07. The best outputs associated with the best performance was achieved using Levenberg-Marquardt backpropagation training algorithm. This algorithm is using a combination of techniques which allows the NN to be trained efficiently. This combination includes backpropagation, gradient descent approach, and Gauss-Newton technique [47, 48].

Table 2.

Evaluation of the effect of different training algorithms on the performance of multilayer feedforward NN with 144-70-1 topology.

Training function Time (sec) Performance (MSE)
BFG 58 0.000426
BR 35 3.69e-6
CGB 11 0.0128
CGF 11 0.0156
CGP 13 0.0160
GD 6 0.117
GDM 6 0.100
GDA 6 0.0950
GDX 6 0.0383
LM 27 1.32E-7
OSS 13 0.0198
R 226 0.0503
RP 6 0.0183
SCG 10 0.0154

After determining the main design parameters of ANN, a feedforward ANN with one hidden layer (70 hidden neurons), one outputs layer (1) has been used in the study of PET data sets. The training algorithm used with this network is Levenberg-Marquardt backpropagation algorithm. In this application, 70% of the first data set have been used for training (46 slices), 15% for validating (10 slices), and 15% for testing (10 slices). A window of 12 × 12 voxels has been used to scan each input slice. The size of this window is chosen to include all the spheres even the biggest one. The utilisation of this window has reduced the input features size fed into the ANN each time without losing the slice details in addition to reduce the required computational time. The input features of the ANN have been extracted in spatial and wavelet domains. For both domains an ANN with 144 inputs, 70 hidden neurons, and (1) outputs layer has been used. The input features in the spatial domain are the voxels of each processed slice. While the utilised wavelet filter decomposes each slice from the input volume and produces four types of coefficients. The approximation coefficients produced by the HWT represent the most detailed information about the analysed slice. The size of these coefficients (84 × 84) is half of the original size. The ANN achieved good performance with very small MSE, 2.39E-16.

An objective evaluation of the artificial intelligence system (AIS) outputs has been performed by comparing the sphere computed volume (CV) with its true original volume (TV). The experimental results have been repeated 10 times, and the average of the sphere volume measured using ANN is calculated. The standard deviation for the volume of sphere 1 is 0.0971, for sphere 2 is 0.1170, for sphere 3 is 0.1185, for sphere 4 is 0.1232, for sphere 5 is 0.1258, and for sphere 6 is 0.1293. The CV obtained from the ANN, and the percentage of the absolute relative error (ARE %) for each sphere are presented in Table 3. ANN has clearly detected all spheres, where spheres 1, 2, and 3 are accurately segmented whereas spheres 4 is overestimated, and sphere 5 and 6 are underestimated. It is worth mentioning that the proposed system has shown better performance compared to the thresholding and clustering based approaches which are used as ground truth. Adaptive, soft, and hard thresholding approaches have been also used to perform the segmentation. The best results obtained from these approaches is by using adaptive threshold method which is used for the comparison with the other assessed techniques. Figure 4 presents the obtained spheres volumes using three thresholding approaches. Table 3 illustrates a comparison between the assessed approaches in term of ARE percentage. Thresholding approach has overestimated the volume of all spheres, while the exploitation of K-means clustering approach underestimates the volume of all spheres, particularly spheres 5 and 6.

Table 3.

Comparison of sphere volumes and ARE between TV and CV for different segmentation algorithms assessed including thresholding, clustering and ANN.

Spheres Thresholding Clustering ANN
No. TV (ml) CV (ml) ARE % CV (ml) ARE % CV (ml) ARE %
1 26.52 26.74 0.81 25.73 2.97 26.54 0.08
2 11.49 11.62 1.07 11.01 4.22 11.51 0.10
3 5.58 5.96 6.97 5.28 5.35 5.63 0.95
4 2.57 2.81 9.08 2.40 6.89 2.63 2.13
5 1.15 1.29 11.98 1.03 10.03 1.10 4.54
6 0.52 0.61 15.76 0.46 11.52 0.49 5.88

Figure 4.

Figure 4

The segmented volume for all spheres in data set 1 using adaptive, soft, and hard thresholding approaches, respectively.

The segmented slices from thresholding, clustering and ANN in the wavelet domain are illustrated in Figure 5, where Figure 5(d) is zoomed for illustration purpose. The three-dimensional shaded surface for each segmented sphere obtained from ANN are plotted in Figure 6, where the voxel values are scaled in [0..1] on Z axis and voxels number is within [0..12] on the remaining two axes.

Figure 5.

Figure 5

Phantom data set 1: (a) Original PET image (168 × 168), (b) thresholded image (168 × 168), (C) clustered image (168 × 168), (d) segmented image (84 × 84) using ANN and MRA, zoomed by a factor of 2.

Figure 6.

Figure 6

Segmented sphere surface plot for phantom data set 1. Voxel values scaled in [0..1] on Z axis, and voxels number is within [0..12] on X and Y axes: (a) sphere 1, (b) sphere 2, (c) sphere 3, (d) sphere 4, (e) sphere 5, and (f) sphere 6.

4.2. Simulated Zubal Phantom

The proposed segmentation system was able to detect tumours in the second phantom data set with isotropic voxels. The first tumour with size of 2 voxels was clearly detected in slice 68. Figure 7 shows the segmented slices from thresholding, clustering, and ANN in the wavelet domain for this tumour. The second and third tumours with size 3 and 2 voxels, respectively, were also clearly detected in slice 57 and 74, respectively. Similar results have been achieved for detecting tumours in the second data set with nonisotropic voxels. On the other hand similar segmented slices have been obtained using ANN in the spatial domain, however, more computational time is required for processing all data sets in this domain.

Figure 7.

Figure 7

Phantom data set 2. (Tumour 1): (a) Original PET image (128 × 128), (b) thresholded image (128 × 128), (C) clustered image (128 × 128), (d) segmented image (64 × 64) using ANN and MRA, zoomed by a factor of 2, where tumour 1 (2 voxel) is detected.

4.3. Performance Evaluation

In the field of AI a number of performance metrics can be employed to evaluate the performance of ANN. A confusion matrix is a visualisation tool typically used in supervised and unsupervised learning approaches. Each row of the matrix represents the instances in a predicted class, while each column represents the instances in an actual class. One benefit of a confusion matrix is that it is easy to see if the system is confusing two classes (the tumour and the remaining tissues). The confusion matrix for the first data set shows that 1 voxel out of 65 ones in the first segmented sphere was misclassified, Figure 8(a). Where the percent in the green box refers to each class prediction accuracy. While the percent in the pink box refers to the misclassified voxels accuracy in each class. The gray boxes represent the percents of classified voxels numbers in each class in green, and the percent of the error in each class in red. The blue box represents the total percent of all classes in green and the total error in these classes in red. All the numbers in the confusion matrix are represented as a percentage.

Figure 8.

Figure 8

Confusion matrix: (a) for phantom data set 1, sphere 1, (b) phantom data set 2, tumour 1, (c) phantom data set 2, tumour 2, and (d) for phantom data set 2, tumour 3.

The confusion matrix for the second data set, tumour 1 is illustrated in Figure 8(b). The two voxels of this tumour were precisely classified in one class and the remaining voxels (4094) classified in the other class. The confusion matrix for the second data set, tumour 2, shows that the 3 voxels were precisely classified as a first class and the remaining voxels (4093) classified in the other class. The obtained result is presented in Figure 8(c), while Figure 8(d) illustrates the confusion matrix for the second data set, tumour 3. The two voxels of tumour 3 were precisely classified as a first class and the remaining voxels (4094) classified in the other class.

The other performance checking approach is receiver operating characteristic (ROC). This approach can be represented by plotting the fraction of true positives rate (TPR) versus the fraction of false positives rate (FPR), where the perfect point in the ROC curve is the point (0,1). The ROC curve for the first data set is located near the perfect point and the FPR for the sphere voxels number is near the 0 point. Perfect ROC has been obtained for the second data set and the FPR for tumour voxels number is 0.

4.4. Clinical PET Studies

The proposed approaches have been also tested on clinical PET volumes of nonsmall cell lung cancer patients. A subjective evaluation based on the clinical knowledge has been carried out for the output of the proposed approaches. The tumour in these slices has a maximum diameter on the y-axis of 90 mm (estimated by histology). The segmented tumour using ANN in spatial domain and wavelet domain (after scaling) has a diameter of 90.1 mm. The segmented volumes using the proposed approaches outlines a well defined contour as illustrated in Figure 9.

Figure 9.

Figure 9

Illustration of algorithm performance on clinical PET data showing: (a) original PET image (128 × 128), (b) thresholded image (128 × 128), (C) clustered image (128 × 128), and (d) segmented image (64 × 64) using ANN and MRA, zoomed by a factor of 2.

5. Conclusions

An artificial intelligence system based on multilayer artificial neural networks was proposed for PET volume segmentation. Different training algorithms have been utilised in this study to validate the best algorithm for the targeted application. Two PET phantom data sets and a clinical PET volume of nonsmall cell lung cancer patient have been used to evaluate the performance of the proposed system. Objective and subjective evaluation for the system outputs have been carried out. Confusion matrix and receiver operating characteristic were also used to judge the performance of the trained neural network. Experimental and simulated phantom results have shown a good performance for the ANN in detecting the tumours in spatial and wavelet domains for both phantom and clinical PET volumes. Accurate tumour quantification was also achieved through this system. Ongoing research is focusing on further validation of the proposed algorithm in a clinical setting and the exploitation of other artificial intelligence tools and feature extraction techniques.

Acknowledgment

This paper was supported by the Swiss National Science Foundation under Grant no. 3152A0-102143.

References

  • 1.Mankoff D, Muzi M, Zaidi H. Quantitative analysis in nuclear oncologic imaging. In: Zaidi H, editor. Quantitative Analysis in Nuclear Medicine Imaging. New York, NY, USA: Springer; 2006. pp. 494–536. [Google Scholar]
  • 2.Montgomery DWG, Amira A, Zaidi H. Fully automated segmentation of oncological PET volumes using a combined multiscale and statistical model. Medical Physics. 2007;34(2):722–736. doi: 10.1118/1.2432404. [DOI] [PubMed] [Google Scholar]
  • 3.Aristophanous M, Penney BC, Pelizzari CA. The development and testing of a digital PET phantom for the evaluation of tumor volume segmentation techniques. Medical Physics. 2008;35(7):3331–3342. doi: 10.1118/1.2938518. [DOI] [PubMed] [Google Scholar]
  • 4.Vees H, Senthamizhchelvan S, Miralbell R, Weber DC, Ratib O, Zaidi H. Assessment of various strategies for 18F-FET PET-guided delineation of target volumes in high-grade glioma patients. European Journal of Nuclear Medicine and Molecular Imaging. 2009;36(2):182–193. doi: 10.1007/s00259-008-0943-6. [DOI] [PubMed] [Google Scholar]
  • 5.Basu S. Selecting the optimal image segmentation strategy in the era of multitracer multimodality imaging: a critical step for image-guided radiation therapy. European Journal of Nuclear Medicine and Molecular Imaging. 2009;36(2):180–181. doi: 10.1007/s00259-008-1033-5. [DOI] [PubMed] [Google Scholar]
  • 6.Zaidi H, El Naqa I. PET-guided delineation of radiation therapy treatment volumes: a survey of image segmentation techniques. European Journal of Nuclear Medicine and Molecular Imaging. 2010;37:1–37. doi: 10.1007/s00259-010-1423-3. [DOI] [PubMed] [Google Scholar]
  • 7.Dreyfus G. Neural Networks Methodology and Applications. Berlin, Germany: Springer; 2005. [Google Scholar]
  • 8.Arbib MA. The Handbook of Brain Theory and Neural Networks. Cambridge, Mass, USA: Massachusetts Institute of Technology; 2003. [Google Scholar]
  • 9.Joo S, Moon WK, Kim HC. Computer-aidied diagnosis of solid breast nodules on ultrasound with digital image processing and artificial neural network. In: Proceedings of the 26th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBS ’04); 2004; pp. 1397–1400. [DOI] [PubMed] [Google Scholar]
  • 10.Mougiakakou SG, Valavanis I, Nikita KS, Nikita A, Kelekis D. Characterization of CT liver lesions based on texture features and a multiple neural network classification scheme. In: Proceedings of the 25th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBS ’03), vol. 2; 2003; pp. 1287–1290. [Google Scholar]
  • 11.Reddick WE, Glass JO, Cook EN, Elkin TD, Deaton RJ. Automated segmentation and classification of multispectral magnetic resonance images of brain using artificial neural networks. IEEE Transactions on Medical Imaging. 1997;16(6):911–918. doi: 10.1109/42.650887. [DOI] [PubMed] [Google Scholar]
  • 12.Reyes-Aldasoro CC, Aldeco A. Advances in Artificial Perception and Robotics CIMAT. 2000. Image segmentation and compression using neural networks. [Google Scholar]
  • 13.Suzuki K, Abe H, MacMahon H, Doi K. Image-processing technique for suppressing ribs in chest radiographs by means of massive training artificial neural network (MTANN) IEEE Transactions on Medical Imaging. 2006;25(4):406–416. doi: 10.1109/TMI.2006.871549. [DOI] [PubMed] [Google Scholar]
  • 14.Luger GF. Artificial Intelligence: Structures and Strategies for Complex Problem Solving. Pearson Education Inc.; 2009. [Google Scholar]
  • 15.Sahoo PK, Soltani S, Wong AKC. A survey of thresholding techniques. Computer Vision, Graphics and Image Processing. 1988;41(2):233–260. [Google Scholar]
  • 16.Kanakatte A, Gubbi J, Mani N, Kron T, Binns D. A pilot study of automatic lung tumor segmentation from positron emission tomography images using standard uptake values. In: Proceedings of the IEEE Symposium on Computational Intelligence in Image and Signal Processing (CIISP ’07); 2007; pp. 363–368. [Google Scholar]
  • 17.Sezgin M, Sankur B. Survey over image thresholding techniques and quantitative performance evaluation. Journal of Electronic Imaging. 2004;13(1):146–168. [Google Scholar]
  • 18.Otsu N. A threshold selection method from gray-level histograms. IEEE Transactions on Systems, Man, and Cybernetics. 1979;9(1):62–66. [Google Scholar]
  • 19.Mallat SG. A theory for multiresolution signal decomposition: the wavelet representation. IEEE Transactions on Pattern Analysis and Machine Intelligence. 1989;11(7):674–693. [Google Scholar]
  • 20.Gonzalez R, Woods R. Digital Image Processing. Upper Saddle River, NJ, USA: Prentice-Hall; 2001. [Google Scholar]
  • 21.Amira A, Chandrasekaran S, Montgomery DWG, Servan Uzun I. A segmentation concept for positron emission tomography imaging using multiresolution analysis. Neurocomputing. 2008;71(10–12):1954–1965. [Google Scholar]
  • 22.Rajpoot KM, Rajpoot NM. Hyperspectral colon tissue cell classification. In: Medical Imaging; 2004; Proceedings of SPIE. [Google Scholar]
  • 23.Fan G, Xia X-G. Wavelet-based texture analysis and synthesis using hidden Markov models. IEEE Transactions on Circuits and Systems I. 2003;50(1):106–120. [Google Scholar]
  • 24.Liu H, Chen Z, Chen X, Chen Y. Multiresolution medical image segmentation based on wavelet transform. In: Proceedings of the 27th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBS ’05); September 2005; pp. 3418–3421. [DOI] [PubMed] [Google Scholar]
  • 25.Jain AK, Dubes RC. Algorithms for Clustering Data. Upper Saddle River, NJ, USA: Prentice-Hall; 1988. [Google Scholar]
  • 26.Ng HP, Ong SH, Foong KWC, Goh PS, Nowinski WL. Medical image segmentation using k-means clustering and improved watershed algorithm. In: Proceedings of the 7th IEEE Southwest Symposium on Image Analysis and Interpretation; March 2006; pp. 61–65. [Google Scholar]
  • 27.Jonsson C, Odh R, Schnell P-O, Larsson SA. A comparison of the imaging properties of a 3- and 4-ring biograph PET scanner using a novel extended NEMA phantom. In: Proceedings of the IEEE Nuclear Science Symposium and Medical Imaging Conference (NSS-MIC ’07), vol. 4; November 2007; pp. 2865–2867. [Google Scholar]
  • 28.Thomas MDR, Bailey DL, Livieratos L. A dual modality approach to quantitative quality control in emission tomography. Physics in Medicine and Biology. 2005;50(15):N187–N194. doi: 10.1088/0031-9155/50/15/N03. [DOI] [PubMed] [Google Scholar]
  • 29.Bergmann H, Dobrozemsky G, Minear G, Nicoletti R, Samal M. An inter-laboratory comparison study of image quality of PET scanners using the NEMA NU 2-2001 procedure for assessment of image quality. Physics in Medicine and Biology. 2005;50(10):2193–2207. doi: 10.1088/0031-9155/50/10/001. [DOI] [PubMed] [Google Scholar]
  • 30.Herzog H, Tellmann L, Hocke C, Pietrzyk U, Casey ME, Kawert T. NEMA NU2-2001 guided performance evaluation of four siemens ECAT PET scanners. IEEE Transactions on Nuclear Science. 2004;51(5):2662–2669. [Google Scholar]
  • 31.Wilson JM, Turkington TG. Multisphere phantom and analysis algorithm for PET image quality assessment. Physics in Medicine and Biology. 2008;53(12):3267–3278. doi: 10.1088/0031-9155/53/12/013. [DOI] [PubMed] [Google Scholar]
  • 32.Zaidi H, Schoenahl F, Ratib O. Geneva PET/CT facility: design considerations and performance characteristics of two commercial (Biograph 16/64) scanners. European Journal of Nuclear Medicine and Molecular Imaging. 2007;34(supplement 2):p. S166. [Google Scholar]
  • 33.Tomeï S, Reilhac A, Visvikis D, et al. OncoPET-DB: a freely distributed database of realistic simulated whole body 18F-FDG PET images for oncology. IEEE Transactions on Nuclear Science. 2010;57(1):246–255. [Google Scholar]
  • 34.van Baardwijk A, Bosmans G, Boersma L, et al. PET-CT-based auto-contouring in non-small-cell lung cancer correlates with pathology and reduces interobserver variability in the delineation of the primary tumor and involved nodal volumes. International Journal of Radiation Oncology, Biology, Physics. 2007;68(3):771–778. doi: 10.1016/j.ijrobp.2006.12.067. [DOI] [PubMed] [Google Scholar]
  • 35.Kermani BG, Schiffman SS, Nagle HT. Performance of the Levenberg-Marquardt neural network training method in electronic nose applications. Sensors and Actuators B. 2005;110(1):13–22. [Google Scholar]
  • 36.Gill PE, Murray W, Wright MH. Practical Optimization. New York, NY, USA: Academic Press; 1981. [Google Scholar]
  • 37.MacKay D. Bayesian interpolation. Neural Computation. 1992;4(3):415–447. [Google Scholar]
  • 38.Powell MJD. Restart procedures for the conjugate gradient method. Mathematical Programming. 1977;12(1):241–254. [Google Scholar]
  • 39.Scales LE. Introduction to Non-Linear Optimization. Berlin, Germany: Springer; 1985. [Google Scholar]
  • 40.Bengio Y, Simard P, Frasconi P. Learning long-term dependencies with gradient descent is difficult. IEEE Transactions on Neural Networks. 1994;5(2):157–166. doi: 10.1109/72.279181. [DOI] [PubMed] [Google Scholar]
  • 41.Bhaya A, Kaszkurewicz E. Steepest descent with momentum for quadratic functions is a version of the conjugate gradient method. Neural Networks. 2004;17(1):65–71. doi: 10.1016/S0893-6080(03)00170-9. [DOI] [PubMed] [Google Scholar]
  • 42.Iranmanesh S. A differential adaptive learning rate method for back-propagation neural networks. In: Proceedings of the 10th WSEAS International Conference on Neural Networks; 2009. [Google Scholar]
  • 43.Magoulas GD, Vrahatis MN, Androulakis GS. Improving the convergence of the backpropagation algorithm using learning rate adaptation methods. Neural Computation. 1999;11(7):1769–1796. doi: 10.1162/089976699300016223. [DOI] [PubMed] [Google Scholar]
  • 44.Battiti R. First and second order methods for learning: between steepest descent and Newton's method. Neural Computation. 1992;4(2):141–166. [Google Scholar]
  • 45.Riedmiller M, Braun H. A direct adaptive method for faster backpropagation learning: the RPROP algorithm. In: Proceedings of the IEEE International Conference on Neural Networks (ICNN ’93); April 1993; pp. 586–591. [Google Scholar]
  • 46.Moller MF. A scaled conjugate gradient algorithm for fast supervised learning. Neural Networks. 1993;6(4):525–533. [Google Scholar]
  • 47.Yu X, Efe MO, Kaynak O. A backpropagation learning framework for feedforward neural networks. In: Proceedings of the IEEE International Symposium on Circuits and Systems (ISCAS ’01), vol. 3; May 2001; pp. 700–702. [Google Scholar]
  • 48.Fine TL. Feedforward Network Methodology. Berlin, Germany: Springer; 1999. [Google Scholar]

Articles from International Journal of Biomedical Imaging are provided here courtesy of Wiley

RESOURCES