Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2020 Apr 1.
Published in final edited form as: IEEE Trans Biomed Eng. 2018 Aug 30;66(4):1069–1081. doi: 10.1109/TBME.2018.2866764

A Novel Extension to Fuzzy Connectivity for Body Composition Analysis: Applications in Thigh, Brain, and Whole Body Tissue Segmentation

Ismail Irmakci 1, Sarfaraz Hussein 2, Aydogan Savran 3, Rita R Kalyani 4, David Reiter 5, Chee W Chia 6, Kenneth W Fishbein 7, Richard G Spencer 8, Luigi Ferrucci 9, Ulas Bagci 10,*
PMCID: PMC6511985  NIHMSID: NIHMS1525359  PMID: 30176577

Abstract

Magnetic resonance imaging (MRI) is the non-invasive modality of choice for body tissue composition analysis due to its excellent soft tissue contrast and lack of ionizing radiation. However, quantification of body composition requires an accurate segmentation of fat, muscle and other tissues from MR images, which remains a challenging goal due to the intensity overlap between them. In this study, we propose a fully automated, data-driven image segmentation platform that addresses multiple difficulties in segmenting MR images such as varying inhomogeneity, non-standardness, and noise, while producing high-quality definition of different tissues. In contrast to most approaches in the literature, we perform segmentation operation by combining three different MRI contrasts and a novel segmentation tool which takes into account variability in the data. The proposed system, based on a novel affinity definition within the fuzzy connectivity (FC) image segmentation family, prevents the need for user intervention and reparametrization of the segmentation algorithms. In order to make the whole system fully automated, we adapt an affinity propagation clustering algorithm to roughly identify tissue regions and image background. We perform a thorough evaluation of the proposed algorithm’s individual steps as well as comparison with several approaches from the literature for the main application of muscle/fat separation. Furthermore, whole-body tissue composition and brain tissue delineation were conducted to show the generalization ability of the proposed system. This new automated platform outperforms other state-of-the-art segmentation approaches both in accuracy and efficiency.

Index Terms: MRI, Fat Segmentation, Muscle Segmentation, Affinity Propagation, Muscle Quantification, Fat Quantification, Whole-body tissue classification, Brain tissue segmentation

I. INTRODUCTION

BODY composition changes are observed in aging and may be related to the development of disorders such as metabolic syndrome and diabetes. Sarcopenia, the progressive loss of muscle mass with age tends to be associated with higher amount of fat (adipose) in the thigh muscle (Figure 1). It is noteworthy that the rate of decline in muscle strength with age exceeds what would be predicted based on muscle mass alone. This suggests that perhaps age affects muscle quality by facilitating fat infiltration. These facts propel current investigations, and research into the non-invasive quantification of body tissue composition is the subject of a great deal of current research [1]. Magnetic resonance imaging (MRI) provides excellent soft tissue contrast (Figure 1) without exposing patients to ionizing radiation, making it the preferred modality for body composition studies1. Previously, we developed segmentation methods for abdominal, thoracic, and brown fat quantification [1]. Here, we focus on fat/muscle quantification in thigh regions using multiple MRI image contrasts.

Fig. 1:

Fig. 1:

Fat and muscle are important outcome measures in the evaluation of several health conditions. (a) depicts the muscular representation of the whole-body where thigh regions are highlighted. (b) is the coronal view of the whole-body representation which shows bone, fat, and muscle distribution in thigh region (zoomed version is in (c)). Pre-defined thigh region (c), (d) and (e) show MRI correspondences of the starting and ending positions of the thigh. MRI is a non-invasive imaging modality, allowing visualization of tissues with its excellent soft tissue contrast and optimal discrimination between different tissue types.

A. Unmet clinical need and summary of challenges

Conventional methods of manual or semi-automated segmentation methods either fail to provide high sensitivity and specificity or require substantial user interaction with the segmentation software. Because these methods are time consuming and have limited inter-raters reliability, they cannot be used for large-scale studies including those for routine clinical evaluations. There are other difficulties such as noise, inhomogeneity, and variable intensity that are not completely avoidable in MRI methodology that may affect any image-based quantification task (i.e., this is specifically called “intensity nonstandardness”). It is highly desirable to have an automated method that can create standardized thigh MR images where non-anatomical uncertainties are minimized while providing accurate and efficient quantification of fat and muscle tissue compartments in the thigh.

To the best of our knowledge, there is no algorithm reported to conduct fat and muscle segmentation within a very short time interval (i.e., seconds) and with high precision. Efficiency is of particular importance because it may be one of the major reasons that many algorithms are not adopted into clinical settings or large research studies. Last, but not least, benefits of multi-contrast MRI have not been fully explored for fat and muscle quantification as yet. Here, we prove the complementary nature of multi-contrast imaging quantitatively.

B. Background and related works

Prescott et al. [2] used level set segmentation methods by utilizing various initialization templates to perform fat/muscle segmentation of thigh MR images. This idea was built upon the success of atlas-based segmentation methods in brain images and anatomical organs where tissue distribution played a significant role [3], [4]. In another study, Ahmad et al. [5] performed a semi-automated segmentation of quadriceps muscles in thigh MRI. In addition to requiring user input, this method relied heavily on proper segmentation from a single slice and carefully defined prior information.

Shape models were also extensively reported in the literature, particularly when tissue boundary separation was difficult. Essafi et al.’s [6] study can be categorized under this setting. The proposed algorithm used diffusion wavelets and geometric constraints to optimize the position of anatomical landmarks which helped separation of different tissues. The method achieved superior performance in modeling the heterogeneity of the muscle topology. However, it is unknown how this method will perform under different MRI inhomogeneities and noise, which were not tested in their study. In a similar fashion, the same objective of modeling tissue heterogeneity was addressed in [7], where a point distribution model along with higher order Markov random field (MRF) model were used to represent tissue boundaries. Yet, the performances of such methods often suffered when boundary points were not localized reliably. Moreover, segmentation outcome was extremely sensitive to initialization of the model assembly, as reported also in our previous works [8], [9].

Recent investigations in this field have sought to address the challenges described above using different methods. For instance, Andrews et al. [10], included a generalized log-ratio notion in their quantification algorithm to represent prior information of anatomical volume and tissue adjacency in a probabilistic space. Random decision forest algorithm was used to learn intensity and texture characteristics of muscle and fat tissues. Later, this information was combined with a shape model in order to perform the final segmentation. Makrogiannis et al. [11] used a Gaussian mixture model (GMM) for modeling the muscle, fat, and inter-muscle adipose tissue (IMAT) jointly from water and fat-suppressed MRI images. Due to the nature of this problem, clustering-based approaches were better suited for quantifying IMAT. However, clustering-based approaches were not suitable for subcutaneous fat (i.e., fat under the skin) and muscle separation, and a separate step for their delineations was required. To address this challenge, the authors used active contour-based approach with user-interventions to quantify fat and muscle distributions from multiple MR image contrasts. A subsequent study by the same authors on the same data set included a modified active contour based delineation method, which wasn’t completely successful in segmenting fat/muscle in thigh regions for individual MRI modalities [12].

C. Overview of the proposed approach and our contributions

We propose a generic image segmentation method based on a novel fuzzy connectivity segmentation algorithm. Specifically, we design a novel computational platform for joint segmentation of fat and muscle from multi-contrast MR images of the thigh. Later, we show that the proposed algorithm is capable of classifying tissues in whole body MRI and brain MRIs as well. The proposed algorithm has two modules: automated seed selection (i.e., detection/localization) and delineation (i.e., precise boundary identification). Figure 2 illustrates the overview of the proposed method with its two modules for the main application in this paper: muscle/fat separation from multi-contrast MRIs. Prior to segmentation, each MR image is preprocessed, and the three image contrasts are combined to form an input tensor (i.e., a vector including inputs from multiple sources is called as MRI or input tensor from now on).

Fig. 2:

Fig. 2:

An overview of the proposed framework is illustrated. First, the MRI images are pre-processed and cleaned using the approaches explained in Section II-B. The cleaned images are then combined to create an input tensor, which is then fed into a seed selection scheme. We have discussed two seeding strategies i.e. affinity propagation (AP) and morphology based operators. The final stage devises two alternative fusion strategies for multi-object segmentation from multi-contrast MRI: fusion at the affinity space and fusion at the decision level (last column). While affinity combination uses evidence based inference from probability density function and segmentation results of each contrast image, decision fusion uses Karnaugh map based multi-label fusion strategy to refine the final tissue segmentation.

Briefly, pre-processing includes consecutive filtering operations that remove bias-field (inhomogeneity) [13], minimize noise [14], and stabilize intensity non-standardness (i.e., acquisition-to-acquisition signal intensity variations) [13], [15], [16]. Resulting MRI images, referred to as “clean images”, are used for tissue segmentation and quantification. In the first module, we propose to use affinity propagation (AP) (and alternatively morphology based) seed generation approaches to identify sample locations for fat and muscle tissues as well as the background.

The second module describes our efforts for segmenting fat and muscle tissues based on sampled seeds in the first module to make the segmentation algorithm fully automated. We utilize the clean volumetric MRI in three different image contrasts jointly, and apply a novel affinity function to drive multi-object fuzzy connectivity (FC) image segmentation. This new affinity function adaptively combines individual affinities pertaining to each MRI contrast.

Our contributions

  • Our proposed method is completely data-driven and fully automated.

  • Unlike the most previous works, we consider multiple MRI contrasts in a unified segmentation platform to optimize complementary information from those individual contrasts. Compared to other studies that use multiple MRI contrast, our proposed study is different and unique since we perform the segmentation by combining strengths of other contrast images through a multi-object affinity function within the FC image segmentation algorithm.

  • Seed-based segmentation algorithms use different strategies spanning from manual to automated approaches. In our implementation, we optimize the seed selection procedure without the need for extensive search. Instead, we adapt an unsupervised clustering method (i.e., AP) to roughly and efficiently identify objects of interest prior to the delineation procedure.

The remainder of this manuscript is organized as follows. Section II describes the details of the proposed method along with potential alternatives. Section III introduces experiments and performance evaluations. The manuscript ends with a discussion and concluding remarks in Section IV.

II. MATERIALS AND METHODS

A. Data and reference standards

Muscle/Fat separation application:

MRI were acquired using a 3T Philips Achieva MRI scanner (Philips Healthcare, Best, The Netherlands) equipped with a Q-body radiofrequency coil for transmission and reception. Three separate image volumes were obtained using a spoiled gradient echo for readout, with coverage from the proximal to distal ends of the femur using 80 slices in the foot to head direction, a field of view (FOV) of 440 × 296 × 400 mm3 and a voxel size of 1 × 1 mm2 in-plane, and slice thickness varies from 1 mm to 3 mm in different scans (one particular scan was with 5 mm slice thickness). Acquisition parameters included: repetition time (TR) = 7.7 ms, echo time (TE) = 2.4ms, number of signal averages (NSA) = 2, flip angle (FA) = 25 degrees, and the bandwidth of 452 Hz. The three image acquisitions were tailored to yield images containing signal from water-and-fat, fat-only, and water-only, respectively. The first image containing water-and-fat was obtained with the FFE sequence described above. For fat-only and water-only images, water and fat suppression was obtained using spectral pre-saturation with inversion recovery (SPIR). Fifty subjects’ 150 MR images (three contrast images for each subject) were retrospectively collected from BLSA (Baltimore Longitudinal Study of Aging) database [17]. Experiments were conducted on a regular PC with Intel Core i5 at 2.9 GHz configuration with 16 GB RAM at 1867 MHz speed. The details about the experimental cohort including age, weight, height, and BMI of the subjects are included in Table I.

TABLE I:

Details about the experimental cohort used in the experiments with age (years), weight (kg), height (cm) and Body Mass Index (BMI).

Experimental Cohort
Characteristics Range mean ± std.
Age [44, 89] years 71 ± 11 years
Weight [47.7, 121] kg 77.5 ± 17 kg
Height [148, 187.1] cm 169.32 ± 9.26 cm
BMI [18.67, 45.68] 26.87 ± 5.02

Whole-Body tissue quantification:

Whole-body MR imaging was conducted using 3-D dual-echo, spoiled gradient echo sequences with Dixon-type fat-water separation (Wollenweber et al. [18]). The following protocol was used for determining the images. FOV was 48 cm × 48 cm, TE1/TE2 was 1.15 ms / 2.3 ms, FA/TR=12°/4.33 ms, acquisition matrix = 256 × 128 with 3.8 mm slices, and volumetric images of in-phase, out-phase, fat, and water were reconstructed using Dixon/IDEAL processing on the GE MR systems with a breath hold acquisition with a target of under 20 s. We have used 3 whole body MRI scans to do our segmentation feasibility study.

Brain tissue delineation:

For brain tissue segmentation, the publicly available BRAINWEB dataset was used in our experiments. The dataset comprises simulated phantom images which were constructed from a high-resolution (1-mm isotropic voxels) data which was generated after registering T1-weighted scans having gradient-echo acquisitions with TR/TE/FA = 18 ms/10 ms/30°. The specificity and sensitivity results of the whole body and brain tissue segmentations are given in Table IV in results section.

TABLE IV:

Segmentation performance of the proposed approach for whole body and brain tissue segmentation.

Tissues Number of Images Dice Score Sensitivity Specificity
Brain Tissue 20 - - -
Gray Matter - 0.93 0.92 0.94
White Matter - 0.93 0.95 0.91
Whole-Body 3 - - -
Fat - 0.90 0.89 0.99
Muscle - 0.88 0.87 0.99

B. Pre-processing MR images

MR images exhibit (1) acquisition-to-acquisition signal intensity variations (called intensity non-standardness), (2) inherent noise, and (3) intensity non-uniformity (i.e., bias-field or inhomogeneity). In our earlier work, we have shown that for precise quantification, all three of these signal variations should be corrected [9], [15], [19]–[21]. As the bias correction promotes noise, denoising filters should be used after the inhomogeneity correction step. Hence, the last step is the intensity standardization for correcting non-standardness among images [9].

1). Inhomogeneity correction:

Inhomogeneity occurs due to low spatial frequency intensity variations. Many MRI vendors include simple inhomogeneity correction methods during image acquisition process [15], [20]–[22]. However, those methods are often based on physical phantoms that are not truly representative of patient anatomy and position. In our experiments, we use post-processing filtering method based on nonparametric nonuniform intensity normalization, called N4ITK [23], which maximizes the high-frequency content of the tissue intensity distribution. We use the following parameters in N4ITK for reproducible results: order of spline (for fitting)=3, number of histogram bins=200, full width at half maximum for Gaussian convolution parameter=0.15.

2). Denoising:

Noise is ubiquitous in MRI scans. In addition to inherent MRI noise, image enhancement algorithms such as inhomogeneity correction can further intensify the noise. To this end, we have used the ball-scale based diffusive filtering method that preserves boundary sharpness and fine structures as reported in [9], [14], [19], which is found to have superior performance than anisotropic diffusion based filtering methods [14], [24].

3). Intensity standardization:

Due to acquisition-to-acquisition signal intensity variations, MR image intensities do not possess a tissue-specific numeric meaning between subsequent scans even for images acquired from the same subject, scanner, body region, or pulse sequence [16]. This phenomenon is called non-standardness. Non-standardness has been largely ignored in the literature until recently since many vendors have started to offer various intensity standardization algorithms for their image reconstruction processes [25]–[27]. Basically, these algorithms (called intensity standardization) are preprocessing techniques, which map image intensities into a standard intensity scale. The mapping is done through a piece-wise linear function which is obtained from a training step. In the current study, we have followed the intensity standardization approach as reported in [27]. For reproducible results, we set the min and max intensity values into 1 and 4095, respectively, as suggested by [16]. In addition, we set only a single foreground histogram mode to be determined automatically from the imaging data.

Figure 3 shows water-only, water-fat, and fat-only MR images obtained directly from the scanner in the first row. Images include inhomogeneity and noise distortion as expected. In the second row, MR images have been cleaned through the pre-processing steps in this order: inhomogeneity correction, denoising, and intensity standardization. Substantial improvements in the images can be readily observed in clean images. Yellow arrows show inhomogeneity and noisy regions prior to pre-processing step. As a quantitative example, we have also provided the coefficient of variance (CoV) values for each image before and after pre-processing steps in Figure 7 of the experiments and results section. As noted, the pre-processing approach improves not only visual interpretation but also reduces intensity variation for each tissue.

Fig. 3:

Fig. 3:

First row includes water-only, water-fat, and fat-only MR images with inherent noise and bias-field (marked with yellow arrows). The second row shows cleaned MRI correspondences after the preprocessing steps: inhomogeneity correction, denoising, and intensity standardization. We trained standardization parameters (i.e., histogram landmarks) separately for different image modality.

Fig. 7:

Fig. 7:

Intensity variations prior to pre-processing steps and afterwards are represented as coefficient of variance (CoV). CoV of water-only (MRI1), water-fat (MRI2), fat-only (MRI3) images are summarized both for fat and muscle tissues, respectively. All images in the dataset (N=150) were used for this analysis.

C. Background on FC image segmentation

FC defines the “hanging togetherness” between any two voxels p and q within an image [28]. A binary adjacency relationship (μα) determines adjacent voxels in α-adjacency. For any two adjacent voxels pi and pi+1, their local hanging togetherness is generally defined using an affinity function μk (pi, pi+1). This allows us to create probability maps when one of the voxel’s label is known. When two voxels are more closely related, their affinity becomes greater and this can be used for labeling purposes (i.e., segmentation). To generalize this concept from adjacent two voxels to a set of voxels, FC considers a “path”. A path π between p and q is identified as a sequence of adjacent voxels π = 〈p0 = p, p1,…, pl = q〉, where for an arbitrary path π, the strength of the path is defined as the “minimum affinity” along that path:

μ(π)=min0ilμk(pi,pi+1). (1)

Assuming P(p,q) is the set of all possible paths between p and q, then FC between them is the strength of the strongest path (indicating closely connected voxels with strong affinities as belonging to the same object):

μκ(p,q)=maxπP(p,q)μ(π). (2)

Defining a proper affinity function for two adjacent voxels is the key for a successful FC operation. Generally, the affinity function consists of three components: adjacency-based affinity μd, homogeneity-based affinity (μψ), and object-based affinity (μϕ) which are defined as:

μk(pi,pi+1)=μd(pi,pi+1)μψ(pi,pi+1)μϕ(pi,pi+1). (3)

A wide range of mathematical functions can be used for affinities [29]. Herein, we use Euclidean distance for μd, and adopt the following form of μψ and μϕ for representing tissue class intensity properties:

μψ(pi,pi+1)=exp(|f(pi)f(pi+1)|22σψ2), (4)

And

μϕ(pi,pi+1)=min(exp(|f(pi)m|22σψ2),exp(|f(pi+1)m|22σϕ2)), (5)

where σψ and σϕ control the variance, and m controls the expected mean intensity of the target object. In our implementation, these values are all determined in the training step.

FC segmentation is obtained by generating a fuzzy object O corresponding to a set of seed points sS. The fuzzy object membership value at a voxel p is then determined as the maximum FC value of all seed points:

μO(p)=maxsSμκ(p,s). (6)

The final object is derived from a probability map (i.e., the fuzzy object O) by thresholding.

D. Novel Affinity Function for FC

The FC algorithm is based on affinity functions. The most prominent affinities explained so far in the literature are adjacency-based (μd), homogeneity-based (μψ), and object-feature based (μϕ). In this work, we propose to combine different affinities of different MRI contrasts and utilize the combined affinity function set to conduct segmentation. Figure 4 shows fat and muscle intensity distribution (after images are corrected for noise and inhomogeneity) for water-only (MRI1), water-fat (MRI2), and fat-only (MRI3) images. As it is depicted, the standard deviations of intensities pertaining to muscle are quite different across different contrasts. Similarly, fat intensity patterns have different mean and variance across different image contrasts. It is also worth mentioning that fat-only and water-fat contrasts show large overlaps in fat histograms as would be expected for non-fat suppressed image types. However, infiltrated fat inside the muscle tissue can still be observed in the Figure 4 (first row, the difference between blue and green density distributions indicates infiltrated fat). For precise muscle analysis, it would be more appropriate to choose MRI2 as it follows a balanced Gaussian distribution which is easier to analyze and incorporate into segmentation and/or tissue characterization algorithms (Figure 4, second row).

Fig. 4:

Fig. 4:

Probability density estimations of fat and muscle intensities in water-only (MRI1), water-fat (MRI2), and fat-only (MRI3) images are shown. All images in the dataset (N=150) were used for this analysis. In the first row, although water-fat and fat-only have similar variances, the intensity distribution in water-only is quite different from the other two. In the second row, water-only, water-fat and fat-only have different distributions with different variances. Complementary information is noted therein, with water-fat contrast is better suited for Gaussian modeling, which is highly desirable in many statistical analysis settings.

Based on all these qualitative observations and density histograms where intensity patterns follow distinct distributions, we propose to combine the affinity functions derived from each MRI contrast to quantify muscle and fat tissues jointly. Moreover, we also test the alternative hypothesis where individual segmentation from each image contrast will be combined as a decision fusion process. Here in this subsection, we confine ourselves to unifying affinity functions. In multi-contrast MRI of the thigh, let us have a fuzzy object membership affinity for each contrast: μOMRI1,μOMRI2,μOMRI3. Assume that there is a fuzzy object membership function μOnew  that combines the complementary strength of muscle and fat tissue intensity distributions from different MRI contrasts. One straightforward way to combine multiple affinities is the weighted summation approach as:

μOnew=i=13wiμOMRIi, s.t.i=13wi=1,

where each affinity is weighted based on the effects of that particular MRI contrast on the final segmentation. Assuming that the segmentation accuracy is measured by dice similarity coefficient (DSC), then w can be set as:

wi=DSC(MRIi)i=13DSC(MRIi)

This step is conducted off-line, as once the weights are determined from each contrast based on the goodness of segmentation results; the system uses the same set of weights for the segmenting different images.

E. Fusion of Multiple Segmentations (FC-Fusion)

As mentioned in the previous section, we also explore an alternative segmentation method based on fusion of multiple segmentations (namely, label or decision fusion). In decision fusion process, for designing FC fusion segmentation, we use a truth table for different input combinations. Figure 5 illustrates this process by indicating 0 or 1 for each entry of the truth label where 1 in output column indicates that both tissues (muscle and fat) can be recovered from the quantification process and 1 in left columns simply indicates if that particular image is used in fusion process or not. As can be seen from output column, to get fusion segmentation results, we need to use either MRI1 and MRI3 or sole MRI2 so as to maximize benefits of multi-contrast MRI. This indicates that, in the absence of MRI2, we need to use both MRI1 and MRI3 to generate the segmentation output. The truth table is simplified using a Karnaugh map (K-Map) to formulate the decision fusion and the resulting output map is determined as (MRI2 + MRI1.MRI3). The main reason for fusion of FC segmentation results is to compensate for any segmentation errors while performing segmentation over each modality separately.

Fig. 5:

Fig. 5:

Truth table for different combinations of multicontrast MRI. The final segmentation output is 1 either if MRI2 (water-fat) is available or MRI1 (water-only) and MRI3 (fat-only) are available. The simplified output is obtained using Karnaugh Map which gives FC-Fusion segmentation result.

F. Automatic seed selection

The proposed algorithm needs foreground and background seeds to initiate segmentation. Seed selection is critical in many medical image segmentation problems where literature is vast in proposing automatic seed selection procedures [30]. In this study, we devise three strategies for background and foreground seed sampling: manual, morphological filtering, and affinity propagation based seeding. We also evaluate these strategies’ comparative performances in the proposed delineation framework.

1). Manual seed selection:

Although our proposed method is fully automated, sometimes user interaction may be necessary for very challenging cases. In addition, it is often desirable to compare the fully automated method with interactive methods for efficiency purposes. Manual interaction is the most popular and perhaps the most reliable foreground and background seed sampling method. The strength of the manual seed selection comes from the superior recognition performance of human observers compared to automated methods. Sensitivity and specificity of manual seed selection in our experiments are 100% as observers are not required to differentiate vague boundaries between fat and muscle. However, it is also important to note that manual seed selection is the localization/detection part of the segmentation process and it is completely different from manual annotation of the objects, which is often called “manual segmentation”. The drawback of this strategy is the computational overhead when multiple seeds are required to complete segmentation. As summarized in Table V, manual seeding takes around 27 seconds, which is more than three times of the proposed segmentation’s processing time. This duration may increase substantially for other methods, such as graph-cut segmentation where the user may need to put several seeds at locations where the segmentation algorithm is underachieving.

TABLE V:

Computational efficiencies of the proposed framework’s subsequent steps (seeding and segmentation) are summarized for fat and muscle tissues, respectively.

Method Fat Muscle Total
FC Segmentation 3 sec 5 sec 8 sec
FC-Fusion Segmentation 15 sec 15 sec 30 sec
Proposed Segmentation 3 sec 5 sec 8 sec
Manual-Seeding 12 sec 15 sec 27 sec
Morphology-Seeding 1 sec 2 sec 3 sec
Proposed AP-Seeding 1.17 sec 1.16 sec 2.33 sec

2). Morphological image filtering for seed generation:

Clustering and morphology filtering based methods have been shown successful in segmenting tissues. Due to the unique challenges of the problem at hand, it may be desirable to have more advanced methods than manual seeding. In our particular problem, clustering and morphology filtering based methods can identify tissue classes up to a certain level, even if they fail to segment whole tissue labeling. Figure 2 second row shows how this information can be obtained through morphology based filtering methods such as thresholding, erosion, and connected component analysis. Parameters of thresholding and connected component analysis are optimized for fat and muscle tissues only in water-fat images due to a better contrast between fat and muscle tissues. Once fat and muscle tissues are roughly identified by morphological filtering, we detect sample voxels for muscle and fat tissue locations from the largest connected components of the output. We also ensure that the sampled seeds have neighboring voxels which have the same label. This constraint avoids seed selection from uncertain locations (i.e., boundaries) which are vulnerable to mis-segmentations.

3). Affinity propagation for seed generation:

Affinity propagation (AP) is an unsupervised learning task which is used to partition data into meaningful and similar groups while considering the similarity between pairs of data points [31]. Although the AP clustering algorithm is used successfully in many areas like image segmentation [32], auto-detection of genes; automatic seed generation via AP clustering is a novel contribution. Moreover, the proposed seed generation methodology through AP is based on multi-objects (i.e., fat and muscle) on multiple MR image contrasts. Herein, we first estimate the probability density functions of multiple MR image contrasts and roughly localize fat and muscle regions by sampling the most discriminative intensity values (Figure 2). To test the reliability of the sampled intensity locations, we use a connected component analysis to check the size of the object that the sampled voxel may belong to. Among the three MR image contrasts, experimental results and qualitative judgments show that water-fat and fat-only images provide the best contrast for seeding procedure which is described in the experiments and results section.

III. EXPERIMENTS AND RESULTS

A. Evaluation metrics

We have evaluated the experimental results both qualitatively and quantitatively. For qualitative evaluations, two participating expert interpreters use their visual assessment to evaluate the segmentation results. Furthermore, we have used dice similarity coefficient (DSC) to evaluate how well the segmented regions overlap with the ground truth provided by the participating interpreters. Intra- and inter-operator agreement rates between expert interpreters were found to be 85.7% and 83.5%, respectively. We have also used the coefficient of variance (CoV) to determine how effectively intensity variations can be eliminated. Lower CoVs indicate improved quality in images due to pre-processing (i.e., inhomogeneity correction).

B. Model evidence for MRI information content, and effects of preprocessing

To quantitatively study the information content of different MRI image contrasts, we calculate DSC pertaining to each image type. As shown in Figure 6, the highest DSC is obtained when water-fat (MRI2) is used. The lowest segmentation accuracy is obtained with water-only images (MRI1) as they contain the least discriminative information about the fat tissue.

Fig. 6:

Fig. 6:

Evidence obtained from the individual delineations of fat and muscle tissue show that highest accuracy was obtained when water-fat (MRI2) is used. Especially in muscle delineation, the standard deviation of the water-fat image is extremely small compared to the other modalities. Fat-only images (MRI3) gave similar accuracy as that of water-fat images in fat delineation but with larger variation. Water-only images (MRI1) seem to contain least discriminative information about the fat tissue. All images in the dataset (N=150) were used for this analysis.

It can be seen in Figure 7 that pre-processing makes CoV of fat and muscle tissues smaller. Since intensity variations are minimized prior to segmentation procedure with the pre-processing framework, segmentation quality is improved. In summary, the pre-processing will be substantially affecting the output not only for FC based segmentation methods but also for any segmentation approach that utilizes image intensities.

C. Evaluation of the Segmentation Accuracy

When segmentation accuracies are compared, the most important property is to have high mean accuracy with low standard deviation. We have compared our two proposed methods with the baseline FC segmentation as well (Table II). Furthermore, we have quantitatively tested the effectiveness of the seeding procedure in terms of accuracy and robustness with respect to different seeding strategies. For automated seeding, we have proposed two procedures: AP and morphology based seeding strategies. The proposed method achieves the best results with high mean and low standard deviation as summarized in Table II. All FC based segmentation results with the proposed AP seed generation are substantially better than other segmentation results, which are either based on manual or morphology based seed generation. In the proposed method, the segmentation with the automated AP seed generation gives the best DSC because the combination of weighted affinity functions further improved the segmentation quality. This is primarily due to the improved nonparametric modeling of the tissue class with AP clustering method.

TABLE II:

Dice similarity coefficients (DSC) are listed for the proposed method(s). Two proposed methods (FC-Fusion and novel affinity based FC) are compared with the baseline method (FC segmentation) with respect to seeding methodologies. Higher DSC values indicate better segmentation accuracies. Novel affinity function based FC algorithm outperformed baseline and alternatively created algorithm (FC-fusion). The best seeding procedure was AP-seeding in all cases. All images in the dataset (N=150) were used for this analysis.

Method Fat (%) Muscle (%)
FC Segmentation DSC (mean/std) DSC (mean/std)
Manual-Seeding 75.60 / 11.18 52.59 / 18.14
Morphology-Seeding 54.90 / 15.98 71.52 / 20.09
With AP-Seeding 77.26 / 11.57 72.88 / 18.59
FC-Fusion Segmentation DSC (mean/std) DSC (mean/std)
Manual-Seeding 77.51 / 17.46 82.07 / 9.27
Morphology-Seeding 77.39 / 18.84 84.99 / 17.64
With AP-Seeding 81.42 / 14.27 87.33 / 13.08
Proposed Segmentation DSC (mean/std) DSC (mean/std)
Manual-Seeding 81.07 / 14.75 81.70 / 11.44
Morphology-Seeding 80.66 / 10.36 85.27 / 20.36
With AP-Seeding 84.06 / 12.23 87.86 / 11.78
Proposed Segmentation % (mean/std) % (mean/std)
Sensitivity of AP-Seeding 79.27 / 12.46 84.18 / 16.75
Specificity of AP-Seeding 98.16 / 8.50 96.78 / 10.10

Once segmentations are finalized, we have conducted a linear regression analysis for volumetric quantification of computed and true fat/muscle tissues. The correlation between the computed muscle-fat volumes and the ground truth volumes is shown in Figure 8. As it can be observed, the computed volumes are highly correlated with the ground truth volumes (R2 > 0.8 and R2 > 0.95 for muscle and fat volumes, respectively).

Fig. 8:

Fig. 8:

Correlation between muscle volumes and fat volumes with respect to ground truth measured volumes.

D. Comparison to other methods

As briefly summarized in Section I-B, available approaches for segmentation of fat and muscle for thigh MRI can be categorized under boundary based methods [11], region based methods such as Markov random field (MRF) [33], graph-cut [34] and clustering [35], and machine learning based methods such as random (decision) forest (RF) [36], multilayer perceptron (MLP) [37], SVM [38], BayesNET [39], Hoeffding trees [40], Random committee [41], and AdaBoost [42]. We compare our proposed method with several existing methods both qualitatively and quantitatively. Table III lists the quantitative comparisons of the proposed method’s DSCs for fat and muscle segmentation with Boykov’s graph-cut [34], active contour [11], [12] and machine learning based methods of random forest [36] and random committee [41].

TABLE III:

Dice similarity coefficients (DSC) are listed for the proposed and alternative methods to show their efficacy in delineating fat and muscle tissues. Higher DSC values indicate better segmentation accuracies. All algorithms were reproduced, optimized, and applied to our data for a fair comparison. All images in the dataset (N=150) were used for this analysis.

Method Fat (%) Muscle (%)
Graph-Cut DSC (mean/std) DSC (mean/std)
Boykov’s MaxFlow [34] 73.50 / 16.75 72.20 / 15.01
Active Contour [11], [12] DSC (mean/std) DSC (mean/std)
Active contour/snake 71.41 / 16.13 74.73 / 14.86
Clustering DSC (mean/std) DSC (mean/std)
Random Forest [36] 73.41 / 15.38 71.19 / 18.41
Random Committee [41] 74.83 / 13.39 71.90 / 16.42
Proposed Segmentation DSC (mean/std) DSC (mean/std)
With AP-Seeding 84.06 / 12.23 87.86 / 11.78

We report statistical validation of FC based (proposed) and other methods’ segmentation using DSC in Table III. Proposed FC based algorithm showed higher DSCs compared to other methods. Among the widely used methods, Boykov’s Max Flow Min Cut works better than active contour for fat segmentation but worse for muscle segmentation. While analyzing the classifier based segmentation results, although random forest and random committee show quite similar DSCs, random committee is a little better as it has higher mean and lower deviation. Among the proposed FC based-methods, the best results were obtained when fusion was conducted at the affinity function level: arranging the weight of the affinity functions for each imaging modality (i.e., image contrast of the MRI) significantly improved the DSC for fat and muscle to around 84% and 87%, respectively (Table II). The results from our proposed and alternative segmentation methods prove the efficacy and feasibility of our proposed solution.

For qualitative comparison with other methods, Figure 9 illustrates an example of the segmentation performance for several different methods. We have selected a representative image slice from a single image contrast of the thigh MRI in Figure 9(a), and segmentation results of region based, boundary based, and machine learning based methods are shown in Figure 9(b-m). The proposed algorithm shows improved segmentation results in Figure 9(i) compared to others. Several of the other methods were not able to capture large fat regions in the subcutaneous compartment while leakages (i.e., over-segmentation to other tissues) were inevitable in most cases.

Fig. 9:

Fig. 9:

Comparison of fat and muscle tissue segmentation results for a given MRI slice (a). Machine learning algorithms: SVM in (b) [38], BayesNET in (c) [39], Hoeffding tree in (d) [40], Random committee in (e) [41], Random forest in (f) [36], AdaBoost in (g) [42] and Multilayer perceptron in (h) [37] are trained separately and optimized results are used for comparison with respect to the proposed algorithm (i). The selected MRI slice was used to qualitatively illustrate the impact of image inhomogeneity and fat tissue distribution on segmentation accuracy of each algorithm; the proposed approach (i) shows a dramatic improvement (no over- or under-segmentation issue observed) compared to all other methods. Other segmentation methods such as level set (j) [45] and MRF (k) [33] are also illustrated. The uncertainty map of the given slice (calculated based on homogeneity of the tissue distribution during FC formulation) indicates potential locations where an algorithm may fail (m).

E. Feasibility Studies: Brain and Whole-Body Tissue Segmentation

To evaluate the generalizability of our proposed approach both for different MRI type and body regions, we performed two additional experiments. In the first experiment, we applied our algorithm into Dixon MRI sequences for whole body tissue segmentation. Briefly, for a given sample whole-body MRI (Dixon, Figure 10(a)-left), its 3-class tissue segmentation (Figure 10(a)-right) was obtained within seconds. In our implementation, we considered fat, muscle, and lung fields in the segmentation. Resulting tissue delineations were visually evaluated (and found to be clinically feasible) by the participating experts of this study.

Fig. 10:

Fig. 10:

The generalizability of the proposed approach is shown via (a) whole-body tissue segmentation and (b) brain tissue segmentation. The whole-body tissue segmentation is performed on Dixon MRI sequences, whereas in (b), brain tissue segmentation with varying levels of noise and smoothing is depicted.

In the second experiment, our proposed method was validated using 20 simulated brain MRI scans (T1-weighted) [43], with the ground truth known. Various noise and smoothing levels were also used to evaluate the robustness of the proposed method. Three smoothing levels (small, medium, and large) were applied to the images using a Gaussian smoothing filter. Three noise levels (small, medium, and large) were applied to the images as well. With the proposed approach, considering all 20 brain MRI scans, Gray Matter (GM) and White Matter (WM) segmentation accuracies were excellent with the sensitivity and specificity rates above 90%; Cerebrospinal fluid (CSF) segmentation was good with the sensitivity and specificity always above 85%. The literature demonstrates that CSF is challenging to segment, with sensitivities generally in the 80% - 88% range and specificities between 51% to 88% [44]. Figure 10(b) shows segmented brain tissues. These feasibility results (both whole body and brain tissue segmentations) demonstrate the promising generic aspects of our proposed method in segmentation and quantification of different body regions, and type of MRI scans. The segmentation results are shown in Table IV.

F. Computational efficiency

Proposed segmentation and seeding (AP) methods have achieved the lowest computational time while guaranteeing higher accuracy as listed in Table V. The slowest segmentation procedure took only 57 seconds, of which 27 seconds were spent on seeding and 30 seconds were used for fusion strategy. The shortest segmentation duration was 10.33 seconds provided by the proposed AP-seeding based FC segmentation algorithm integrating adjusted affinities. Manual seeding was restricted to a real clinical setting where a few seeds were often desired to initiate the segmentation; otherwise, seeding process itself can be time-consuming. Increasing the number of seeds led to similar segmentation results as that obtained through AP based seeding method but at the expense of an increase in the analysis time.

G. Quantification from a single image slice

In clinical practice, radiologists frequently use a single slice to estimate the whole volume for a region in an object. For instance, in abdominal fat quantification, instead of segmenting the whole abdomen and separating subcutaneous tissue from visceral, which is a challenging problem, a single slice at the umbilical level is often selected and used to quantify abdominal fat volume [1], [46]. The main reasons for this practice are two-fold. First is a lack of software algorithms that can work efficiently and accurately. This is because many algorithms fail to provide exact quantification, and thus user interaction is necessary for a better quantification. It is time-consuming for radiologists to correct each segmentation failure for volumetric images. Second, the anatomy does not change abruptly from slice to slice. Thus, it is a reasonable assumption that the content of the tissue distribution in a single slice will be correlated with the whole volume. Then, based on a known correlation between a single slice and volume, creating a linear regression equation will estimate the volumetric tissue distribution from a single slice. Inspired by this clinical practice and to explore the inter-relationship of thigh MRI slices, we perform a Pearson correlation test between fat and muscle volume extracted from a single slice with its volumetric surrogate truths. Figure 11 shows correlation value for each slice when estimating the volumetric distribution of fat and muscle, respectively. As indicated, each slice contributed more than R = 0.97 correlation value, and the highest correlations R > 0.99 were obtained when mid slice of the thigh region was selected. This indicates that the proposed system can be used for much faster muscle and fat volume estimation from a single slice for practical purposes.

Fig. 11:

Fig. 11:

Fat and muscle volumes are correlated with a single slice from its stack. Pearson correlation is maximized when the single slice is in the mid-thigh region.

IV. DISCUSSIONS AND CONCLUDING REMARKS

A. Limitations and Future Works

Our study has a few limitations to be noted. With its success in almost all tasks of visual domain and its extensive use for medical image analysis, deep learning can serve as an alternative to our proposed approach. The joint segmentation of multi-contrast images in deep neural network framework can lead to accurate and robust performance. In addition, the parameters in our proposed approach were learned offline, which can be estimated in an online manner for reliable performance. Moreover, in the context of skeletal muscle composition studies, more advanced MRI acquisition methods like Dixon imaging can be used which provide an improved correspondence between the image signal and the fat and water fractions within a given voxel. Lastly, it is worth mentioning that partial volume can be an issue to be solved for separating water and fat signals when their contributions are roughly equal at a given voxel. Especially when the resolution is limited, this effect should be corrected. Although we did not observe large errors due to partial volumes, conventional partial volume correction methods for MR images such as [47] can be applied to minimize such errors in volumetric quantification. Furthermore, Dixon can be used for fat/muscle quantification in sub-voxel level unlike our method where we used voxel level quantification.

Our future study will involve exploring another important imaging marker, called biomechanical muscle quality (BMQ), for a potential association of aging and age-related metabolic diseases. In order to calculate BMQ, we will identify the intra-muscular adipose tissue (IMAT), which is associated with a reduction in force production and suggested to affect BMQ. For this purpose, a unified framework to perform segmentation and quantification of the three tissues (muscle, fat, and IMAT) can be utilized in the clinical evaluation of BMQ. The current work also suggests a potential for application of deep learning based methods for improved multi-contrast image segmentation approaches. Inspired by the recent success of purely machine learning methods and the current methods’ limited achievement on IMAT quantification (i.e., around 70–80% success bands [11], [12], [35]), we aim to explore the joint segmentation of these three tissues using a deep learning framework in our future study. Moreover, given a limited number of images, we performed evaluations on a small dataset for whole-body tissue classification as well as brain tissue segmentation. Although the results are quite promising, with a larger dataset, deep learning based approaches can be explored for improved segmentation of whole-body and brain tissues.

B. Summary

Conventional methods for the tissue delineation and quantification mostly rely on manual or semi-automated algorithms, having low accuracy and efficiency. In the case of multi-contrast MRI scans, conventional approaches use isolated frameworks for each imaging modality; thus, leading to suboptimal results. Given the clinical importance of fat and muscle segmentation and quantification in thigh MRI, as well their utility in other research settings, we proposed a novel approach to address unique challenges associated with these images and showed the generalization ability of the proposed system for the whole-body tissue composition and the brain tissue delineation. We provided a complete quantification framework that automatically minimizes extreme challenges (inhomogeneity, noise, and non-standardness) of the MRI automatically. To maintain the unsupervisory and automated nature of our proposed approach, we performed automatic seed generation using an affinity propagation clustering algorithm. In the delineation step, to the best of our knowledge, the proposed approach is the first to explore the utility of multiple affinity functions to perform a unified segmentation of multiple objects from multi-contrast MRI. Our automated seed selection approach also evades the need for filtering approaches which can be prone to subjectivities.

ACKNOWLEDGMENT

This work was supported in part by the Intramural Research Program of the National Institute on Aging of the National Institutes of Health (NIH).

Footnotes

1

Figure was created with the BioDigital Human Visualization Platform ©.

Contributor Information

Ismail Irmakci, Ege University, Izmir, Turkey. Irmakci’s research is supported by scientific council of Turkey (TUBITAK-BIDEB 2214/A)..

Sarfaraz Hussein, Center for Research in Computer Vision at University of Central Florida, Orlando, FL..

Aydogan Savran, Ege University, Izmir, Turkey. Irmakci’s research is supported by scientific council of Turkey (TUBITAK-BIDEB 2214/A)..

Rita R. Kalyani, Department of Medicine, Johns Hopkins University School of Medicine, Baltimore, MD.

David Reiter, National Institute on Aging, National Institutes of Health (NIH), Baltimore, MD..

Chee W. Chia, National Institute on Aging, National Institutes of Health (NIH), Baltimore, MD.

Kenneth W. Fishbein, National Institute on Aging, National Institutes of Health (NIH), Baltimore, MD.

Richard G. Spencer, National Institute on Aging, National Institutes of Health (NIH), Baltimore, MD.

Luigi Ferrucci, National Institute on Aging, National Institutes of Health (NIH), Baltimore, MD..

Ulas Bagci, Center for Research in Computer Vision at University of Central Florida, Orlando, FL..

REFERENCES

  • [1].Hussein S, Green A, Watane A, Reiter D, Chen X, Papadakis GZ, Wood B, Cypess A, Osman M, and Bagci U, “Automatic segmentation and quantification of white and brown adipose tissues from pet/ct scans,” IEEE Transactions on Medical Imaging, vol. 36, no. 3, pp. 734–744, 2017. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [2].Prescott JW, Best TM, Swanson MS, Haq F, Jackson RD, and Gurcan MN, “Anatomically anchored template-based level set segmentation: application to quadriceps muscles in mr images from the osteoarthritis initiative,” Journal of digital imaging, vol. 24, no. 1, pp. 28–43, 2011. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [3].Okada T, Shimada R, Hori M, Nakamoto M, Chen Y-W, Nakamura H, and Sato Y, “Automated segmentation of the liver from 3d ct images using probabilistic atlas and multilevel statistical shape model,” Academic radiology, vol. 15, no. 11, pp. 1390–1403, 2008. [DOI] [PubMed] [Google Scholar]
  • [4].Aljabar P, Heckemann RA, Hammers A, Hajnal JV, and Rueckert D, “Multi-atlas based segmentation of brain images: atlas selection and its effect on accuracy,” Neuroimage, vol. 46, no. 3, pp. 726–738, 2009. [DOI] [PubMed] [Google Scholar]
  • [5].Ahmad E, Yap MH, Degens H, and McPhee JS, “Atlas-registration based image segmentation of mri human thigh muscles in 3d space,” in SPIE Medical Imaging. International Society for Optics and Photonics, 2014, pp. 90 371L–90 371L. [Google Scholar]
  • [6].Essafi S, Langs G, Deux J-F, Rahmouni A, Bassez G, and Paragios N, “Wavelet-driven knowledge-based mri calf muscle segmentation,” in 2009 IEEE International Symposium on Biomedical Imaging: From Nano to Macro. IEEE, 2009, pp. 225–228. [Google Scholar]
  • [7].Wang C, Teboul O, Michel F, Essafi S, and Paragios N, “3d knowledge-based segmentation using pose-invariant higher-order graphs,” in International Conference on Medical Image Computing and Computer-Assisted Intervention. Springer, 2010, pp. 189–196. [DOI] [PubMed] [Google Scholar]
  • [8].Chen X, Summers RM, Cho M, Bagci U, and Yao J, “An automatic method for renal cortex segmentation on ct images: evaluation on kidney donors,” Academic radiology, vol. 19, no. 5, pp. 562–570, 2012. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [9].Bagci U, Chen X, and Udupa JK, “Hierarchical scale-based multi-object recognition of 3-d anatomical structures,” IEEE Transactions on Medical Imaging, vol. 31, no. 3, pp. 777–789, 2012. [DOI] [PubMed] [Google Scholar]
  • [10].Andrews S and Hamarneh G, “The generalized log-ratio transformation: learning shape and adjacency priors for simultaneous thigh muscle segmentation,” IEEE transactions on medical imaging, vol. 34, no. 9, pp. 1773–1787, 2015. [DOI] [PubMed] [Google Scholar]
  • [11].Makrogiannis S, Fishbein K, Moore A, Spencer R, and Ferrucci L, “Image-based tissue distribution modeling for skeletal muscle quality characterization,” IEEE Transactions on Biomedical Engineering, vol. 63, no. 4, pp. 805–813, 2016. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [12].Makrogiannis S, Serai S, Fishbein KW, Schreiber C, Ferrucci L, and Spencer RG, “Automated quantification of muscle and fat in the thigh from water-, fat-, and nonsuppressed mr images,” Journal of Magnetic Resonance Imaging, vol. 35, no. 5, pp. 1152–1161, 2012. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [13].Madabhushi A and Udupa JK, “Interplay between intensity standardization and inhomogeneity correction in mr image processing,” IEEE Transactions on Medical Imaging, vol. 24, no. 5, pp. 561–576, 2005. [DOI] [PubMed] [Google Scholar]
  • [14].Saha PK and Udupa JK, “Scale-based diffusive image filtering preserving boundary sharpness and fine structures,” IEEE Transactions on Medical Imaging, vol. 20, no. 11, pp. 1140–1155, 2001. [DOI] [PubMed] [Google Scholar]
  • [15].Bagci U, Udupa JK, and Chen X, “Intensity non-standardness affects computer recognition of anatomical structures,” in SPIE Medical Imaging. International Society for Optics and Photonics, 2011, pp. 79 642M–79 642M. [Google Scholar]
  • [16].Nyúl LG, Udupa JK et al. , “On standardizing the mr image intensity scale,” image, vol. 1081, 1999. [DOI] [PubMed] [Google Scholar]
  • [17].Ferrucci L, “The Baltimore Longitudinal Study of Aging (BLSA): a 50-year-long journey and plans for the future,” 2008. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [18].Wollenweber S, Ambwani S, Delso G, Lonn A, Mullick R, Wiesinger F, Piti Z, Tari A, Novak G, and Fidrich M, “Evaluation of an atlas-based PET head attenuation correction using PET/CT & MR patient data,” IEEE Transactions on Nuclear Science, vol. 60, no. 5, pp. 3383–3390, 2013. [Google Scholar]
  • [19].Bagci U, Udupa JK, and Chen X, “Orientation estimation of anatomical structures in medical images for object recognition,” in SPIE Medical Imaging. International Society for Optics and Photonics, 2011, pp. 79 622L–79 622L. [Google Scholar]
  • [20].Bagci U, Udupa JK, and Bai L, “The influence of intensity standardization on medical image registration,” in SPIE Medical Imaging. International Society for Optics and Photonics, 2010, pp. 76 251X–76 251X. [Google Scholar]
  • [21].Bağcı U, Udupa JK, and Bai L, “The role of intensity standardization in medical image registration,” Pattern Recognition Letters, vol. 31, no. 4, pp. 315–323, 2010. [Google Scholar]
  • [22].Axel L, Costantini J, and Listerud J, “Intensity correction in surface-coil mr imaging,” American Journal of Roentgenology, vol. 148, no. 2, pp. 418–420, 1987. [DOI] [PubMed] [Google Scholar]
  • [23].Tustison NJ, Avants BB, Cook PA, Zheng Y, Egan A, Yushkevich PA, and Gee JC, “N4itk: improved n3 bias correction,” IEEE transactions on medical imaging, vol. 29, no. 6, pp. 1310–1320, 2010. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [24].Bagci U, Udupa J, and Chen X, “Ball-scale based multi-object recognition in a hierarchical framework,” in Proc. of SPIE Medical Imaging, vol. 7623, 2010, pp. 7 623 451–76 234 512. [Google Scholar]
  • [25].Robitaille N, Mouiha A, Crépeault B, Valdivia F, and Duchesne S, “Tissue-based mri intensity standardization: application to multicentric datasets,” Journal of Biomedical Imaging, vol. 2012, p. 4, 2012. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [26].Jäger F, Nyúl L, Frericks B, Wacker F, and Hornegger J, “Whole body mri intensity standardization,” in Bildverarbeitung für die Medizin 2007. Springer, 2007, pp. 459–463. [Google Scholar]
  • [27].Nyúl LG, Udupa JK, and Zhang X, “New variants of a method of mri scale standardization,” IEEE transactions on medical imaging, vol. 19, no. 2, pp. 143–150, 2000. [DOI] [PubMed] [Google Scholar]
  • [28].Udupa JK and Samarasekera S, “Fuzzy connectedness and object definition: theory, algorithms, and applications in image segmentation,” Graphical models and image processing, vol. 58, no. 3, pp. 246–261, 1996. [Google Scholar]
  • [29].Ciesielski KC, Udupa JK, Falcão AX, and Miranda PA, “Fuzzy connectedness image segmentation in graph cut formulation: A linear-time algorithm and a comparative analysis,” Journal of Mathematical Imaging and Vision, vol. 44, no. 3, pp. 375–398, 2012. [Google Scholar]
  • [30].Pham DL, Xu C, and Prince JL, “Current methods in medical image segmentation 1,” Annual review of biomedical engineering, vol. 2, no. 1, pp. 315–337, 2000. [DOI] [PubMed] [Google Scholar]
  • [31].Frey BJ and Dueck D, “Clustering by passing messages between data points,” science, vol. 315, no. 5814, pp. 972–976, 2007. [DOI] [PubMed] [Google Scholar]
  • [32].Foster B, Bagci U, Xu Z, Dey B, Luna B, Bishai W, Jain S, and Mollura DJ, “Segmentation of pet images for computer-aided functional quantification of tuberculosis in small animal models,” IEEE Transactions on Biomedical Engineering, vol. 61, no. 3, pp. 711–724, 2014. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [33].Rue H and Held L, Gaussian Markov random fields: theory and applications. CRC Press, 2005. [Google Scholar]
  • [34].Chen X, Udupa JK, Bagci U, Zhuge Y, and Yao J, “Medical image segmentation by combining graph cuts and oriented active appearance models,” IEEE Transactions on Image Processing, vol. 21, no. 4, pp. 2035–2046, 2012. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [35].Tan C, Li K, Yan Z, Yang D, Zhang S, Yu HJ, Engelke K, Miller C, and Metaxas D, “A detection-driven and sparsity-constrained deformable model for fascia lata labeling and thigh inter-muscular adipose quantification,” Computer Vision and Image Understanding, 2016. [Google Scholar]
  • [36].Breiman L, “Random forests,” Machine learning, vol. 45, no. 1, pp. 5–32, 2001. [Google Scholar]
  • [37].Rumelhart DE, Hinton GE, and Williams RJ, “Learning internal representations by error propagation,” DTIC Document, Tech. Rep, 1985. [Google Scholar]
  • [38].Cortes C and Vapnik V, “Support-vector networks,” Machine learning, vol. 20, no. 3, pp. 273–297, 1995. [Google Scholar]
  • [39].Nielsen TD and Jensen FV, Bayesian networks and decision graphs. Springer Science & Business Media, 2009. [Google Scholar]
  • [40].Hulten G, Spencer L, and Domingos P, “Mining time-changing data streams,” in Proceedings of the seventh ACM SIGKDD international conference on Knowledge discovery and data mining. ACM, 2001, pp. 97–106. [Google Scholar]
  • [41].Lira MM, de Aquino RR, Ferreira AA, Carvalho MA, Neto ON, and Santos GS, “Combining multiple artificial neural networks using random committee to decide upon electrical disturbance classification,” in Neural Networks, 2007. IJCNN 2007. International Joint Conference on. IEEE, 2007, pp. 2863–2868. [Google Scholar]
  • [42].Freund Y and Schapire RE, “A desicion-theoretic generalization of online learning and an application to boosting,” in European conference on computational learning theory. Springer, 1995, pp. 23–37. [Google Scholar]
  • [43].Collins DL, Zijdenbos AP, Kollokian V, Sled JG, Kabani NJ, Holmes CJ, and Evans AC, “Design and construction of a realistic digital brain phantom,” IEEE transactions on medical imaging, vol. 17, no. 3, pp. 463–468, 1998. [DOI] [PubMed] [Google Scholar]
  • [44].de Boer R, Vrooman HA, Ikram MA, Vernooij MW, Breteler MM, van der Lugt A, and Niessen WJ, “Accuracy and reproducibility study of automatic mri brain tissue segmentation methods,” Neuroimage, vol. 51, no. 3, pp. 1047–1056, 2010. [DOI] [PubMed] [Google Scholar]
  • [45].Lim PH, Bagci U, and Bai L, “Introducing willmore flow into level set segmentation of spinal vertebrae,” IEEE Transactions on Biomedical Engineering, vol. 60, no. 1, pp. 115–122, 2013. [DOI] [PubMed] [Google Scholar]
  • [46].Green AL, Bagci U, Hussein S, Kelly PV, Muzaffar R, Neuschwander-Tetri BA, and Osman MM, “Brown adipose tissue detected by pet/ct imaging is associated with less central obesity,” Nuclear Medicine Communications, vol. 38, no. 7, pp. 629–635, 2017. [DOI] [PubMed] [Google Scholar]
  • [47].Shattuck DW, Sandor-Leahy SR, Schaper KA, Rottenberg DA, and Leahy RM, “Magnetic resonance image tissue classification using a partial volume model,” NeuroImage, vol. 13, no. 5, pp. 856–876, 2001. [DOI] [PubMed] [Google Scholar]

RESOURCES