Abstract
Magnetic resonance imaging (MRI) is the non-invasive modality of choice for body tissue composition analysis due to its excellent soft tissue contrast and lack of ionizing radiation. However, quantification of body composition requires an accurate segmentation of fat, muscle and other tissues from MR images, which remains a challenging goal due to the intensity overlap between them. In this study, we propose a fully automated, data-driven image segmentation platform that addresses multiple difficulties in segmenting MR images such as varying inhomogeneity, non-standardness, and noise, while producing high-quality definition of different tissues. In contrast to most approaches in the literature, we perform segmentation operation by combining three different MRI contrasts and a novel segmentation tool which takes into account variability in the data. The proposed system, based on a novel affinity definition within the fuzzy connectivity (FC) image segmentation family, prevents the need for user intervention and reparametrization of the segmentation algorithms. In order to make the whole system fully automated, we adapt an affinity propagation clustering algorithm to roughly identify tissue regions and image background. We perform a thorough evaluation of the proposed algorithm’s individual steps as well as comparison with several approaches from the literature for the main application of muscle/fat separation. Furthermore, whole-body tissue composition and brain tissue delineation were conducted to show the generalization ability of the proposed system. This new automated platform outperforms other state-of-the-art segmentation approaches both in accuracy and efficiency.
Index Terms: MRI, Fat Segmentation, Muscle Segmentation, Affinity Propagation, Muscle Quantification, Fat Quantification, Whole-body tissue classification, Brain tissue segmentation
I. INTRODUCTION
BODY composition changes are observed in aging and may be related to the development of disorders such as metabolic syndrome and diabetes. Sarcopenia, the progressive loss of muscle mass with age tends to be associated with higher amount of fat (adipose) in the thigh muscle (Figure 1). It is noteworthy that the rate of decline in muscle strength with age exceeds what would be predicted based on muscle mass alone. This suggests that perhaps age affects muscle quality by facilitating fat infiltration. These facts propel current investigations, and research into the non-invasive quantification of body tissue composition is the subject of a great deal of current research [1]. Magnetic resonance imaging (MRI) provides excellent soft tissue contrast (Figure 1) without exposing patients to ionizing radiation, making it the preferred modality for body composition studies1. Previously, we developed segmentation methods for abdominal, thoracic, and brown fat quantification [1]. Here, we focus on fat/muscle quantification in thigh regions using multiple MRI image contrasts.
A. Unmet clinical need and summary of challenges
Conventional methods of manual or semi-automated segmentation methods either fail to provide high sensitivity and specificity or require substantial user interaction with the segmentation software. Because these methods are time consuming and have limited inter-raters reliability, they cannot be used for large-scale studies including those for routine clinical evaluations. There are other difficulties such as noise, inhomogeneity, and variable intensity that are not completely avoidable in MRI methodology that may affect any image-based quantification task (i.e., this is specifically called “intensity nonstandardness”). It is highly desirable to have an automated method that can create standardized thigh MR images where non-anatomical uncertainties are minimized while providing accurate and efficient quantification of fat and muscle tissue compartments in the thigh.
To the best of our knowledge, there is no algorithm reported to conduct fat and muscle segmentation within a very short time interval (i.e., seconds) and with high precision. Efficiency is of particular importance because it may be one of the major reasons that many algorithms are not adopted into clinical settings or large research studies. Last, but not least, benefits of multi-contrast MRI have not been fully explored for fat and muscle quantification as yet. Here, we prove the complementary nature of multi-contrast imaging quantitatively.
B. Background and related works
Prescott et al. [2] used level set segmentation methods by utilizing various initialization templates to perform fat/muscle segmentation of thigh MR images. This idea was built upon the success of atlas-based segmentation methods in brain images and anatomical organs where tissue distribution played a significant role [3], [4]. In another study, Ahmad et al. [5] performed a semi-automated segmentation of quadriceps muscles in thigh MRI. In addition to requiring user input, this method relied heavily on proper segmentation from a single slice and carefully defined prior information.
Shape models were also extensively reported in the literature, particularly when tissue boundary separation was difficult. Essafi et al.’s [6] study can be categorized under this setting. The proposed algorithm used diffusion wavelets and geometric constraints to optimize the position of anatomical landmarks which helped separation of different tissues. The method achieved superior performance in modeling the heterogeneity of the muscle topology. However, it is unknown how this method will perform under different MRI inhomogeneities and noise, which were not tested in their study. In a similar fashion, the same objective of modeling tissue heterogeneity was addressed in [7], where a point distribution model along with higher order Markov random field (MRF) model were used to represent tissue boundaries. Yet, the performances of such methods often suffered when boundary points were not localized reliably. Moreover, segmentation outcome was extremely sensitive to initialization of the model assembly, as reported also in our previous works [8], [9].
Recent investigations in this field have sought to address the challenges described above using different methods. For instance, Andrews et al. [10], included a generalized log-ratio notion in their quantification algorithm to represent prior information of anatomical volume and tissue adjacency in a probabilistic space. Random decision forest algorithm was used to learn intensity and texture characteristics of muscle and fat tissues. Later, this information was combined with a shape model in order to perform the final segmentation. Makrogiannis et al. [11] used a Gaussian mixture model (GMM) for modeling the muscle, fat, and inter-muscle adipose tissue (IMAT) jointly from water and fat-suppressed MRI images. Due to the nature of this problem, clustering-based approaches were better suited for quantifying IMAT. However, clustering-based approaches were not suitable for subcutaneous fat (i.e., fat under the skin) and muscle separation, and a separate step for their delineations was required. To address this challenge, the authors used active contour-based approach with user-interventions to quantify fat and muscle distributions from multiple MR image contrasts. A subsequent study by the same authors on the same data set included a modified active contour based delineation method, which wasn’t completely successful in segmenting fat/muscle in thigh regions for individual MRI modalities [12].
C. Overview of the proposed approach and our contributions
We propose a generic image segmentation method based on a novel fuzzy connectivity segmentation algorithm. Specifically, we design a novel computational platform for joint segmentation of fat and muscle from multi-contrast MR images of the thigh. Later, we show that the proposed algorithm is capable of classifying tissues in whole body MRI and brain MRIs as well. The proposed algorithm has two modules: automated seed selection (i.e., detection/localization) and delineation (i.e., precise boundary identification). Figure 2 illustrates the overview of the proposed method with its two modules for the main application in this paper: muscle/fat separation from multi-contrast MRIs. Prior to segmentation, each MR image is preprocessed, and the three image contrasts are combined to form an input tensor (i.e., a vector including inputs from multiple sources is called as MRI or input tensor from now on).
Briefly, pre-processing includes consecutive filtering operations that remove bias-field (inhomogeneity) [13], minimize noise [14], and stabilize intensity non-standardness (i.e., acquisition-to-acquisition signal intensity variations) [13], [15], [16]. Resulting MRI images, referred to as “clean images”, are used for tissue segmentation and quantification. In the first module, we propose to use affinity propagation (AP) (and alternatively morphology based) seed generation approaches to identify sample locations for fat and muscle tissues as well as the background.
The second module describes our efforts for segmenting fat and muscle tissues based on sampled seeds in the first module to make the segmentation algorithm fully automated. We utilize the clean volumetric MRI in three different image contrasts jointly, and apply a novel affinity function to drive multi-object fuzzy connectivity (FC) image segmentation. This new affinity function adaptively combines individual affinities pertaining to each MRI contrast.
Our contributions
Our proposed method is completely data-driven and fully automated.
Unlike the most previous works, we consider multiple MRI contrasts in a unified segmentation platform to optimize complementary information from those individual contrasts. Compared to other studies that use multiple MRI contrast, our proposed study is different and unique since we perform the segmentation by combining strengths of other contrast images through a multi-object affinity function within the FC image segmentation algorithm.
Seed-based segmentation algorithms use different strategies spanning from manual to automated approaches. In our implementation, we optimize the seed selection procedure without the need for extensive search. Instead, we adapt an unsupervised clustering method (i.e., AP) to roughly and efficiently identify objects of interest prior to the delineation procedure.
The remainder of this manuscript is organized as follows. Section II describes the details of the proposed method along with potential alternatives. Section III introduces experiments and performance evaluations. The manuscript ends with a discussion and concluding remarks in Section IV.
II. MATERIALS AND METHODS
A. Data and reference standards
Muscle/Fat separation application:
MRI were acquired using a 3T Philips Achieva MRI scanner (Philips Healthcare, Best, The Netherlands) equipped with a Q-body radiofrequency coil for transmission and reception. Three separate image volumes were obtained using a spoiled gradient echo for readout, with coverage from the proximal to distal ends of the femur using 80 slices in the foot to head direction, a field of view (FOV) of 440 × 296 × 400 mm3 and a voxel size of 1 × 1 mm2 in-plane, and slice thickness varies from 1 mm to 3 mm in different scans (one particular scan was with 5 mm slice thickness). Acquisition parameters included: repetition time (TR) = 7.7 ms, echo time (TE) = 2.4ms, number of signal averages (NSA) = 2, flip angle (FA) = 25 degrees, and the bandwidth of 452 Hz. The three image acquisitions were tailored to yield images containing signal from water-and-fat, fat-only, and water-only, respectively. The first image containing water-and-fat was obtained with the FFE sequence described above. For fat-only and water-only images, water and fat suppression was obtained using spectral pre-saturation with inversion recovery (SPIR). Fifty subjects’ 150 MR images (three contrast images for each subject) were retrospectively collected from BLSA (Baltimore Longitudinal Study of Aging) database [17]. Experiments were conducted on a regular PC with Intel Core i5 at 2.9 GHz configuration with 16 GB RAM at 1867 MHz speed. The details about the experimental cohort including age, weight, height, and BMI of the subjects are included in Table I.
TABLE I:
Experimental Cohort | ||
---|---|---|
Characteristics | Range | mean ± std. |
Age | [44, 89] years | 71 ± 11 years |
Weight | [47.7, 121] kg | 77.5 ± 17 kg |
Height | [148, 187.1] cm | 169.32 ± 9.26 cm |
BMI | [18.67, 45.68] | 26.87 ± 5.02 |
Whole-Body tissue quantification:
Whole-body MR imaging was conducted using 3-D dual-echo, spoiled gradient echo sequences with Dixon-type fat-water separation (Wollenweber et al. [18]). The following protocol was used for determining the images. FOV was 48 cm × 48 cm, TE1/TE2 was 1.15 ms / 2.3 ms, FA/TR=12°/4.33 ms, acquisition matrix = 256 × 128 with 3.8 mm slices, and volumetric images of in-phase, out-phase, fat, and water were reconstructed using Dixon/IDEAL processing on the GE MR systems with a breath hold acquisition with a target of under 20 s. We have used 3 whole body MRI scans to do our segmentation feasibility study.
Brain tissue delineation:
For brain tissue segmentation, the publicly available BRAINWEB dataset was used in our experiments. The dataset comprises simulated phantom images which were constructed from a high-resolution (1-mm isotropic voxels) data which was generated after registering T1-weighted scans having gradient-echo acquisitions with TR/TE/FA = 18 ms/10 ms/30°. The specificity and sensitivity results of the whole body and brain tissue segmentations are given in Table IV in results section.
TABLE IV:
Tissues | Number of Images | Dice Score | Sensitivity | Specificity |
---|---|---|---|---|
Brain Tissue | 20 | - | - | - |
Gray Matter | - | 0.93 | 0.92 | 0.94 |
White Matter | - | 0.93 | 0.95 | 0.91 |
Whole-Body | 3 | - | - | - |
Fat | - | 0.90 | 0.89 | 0.99 |
Muscle | - | 0.88 | 0.87 | 0.99 |
B. Pre-processing MR images
MR images exhibit (1) acquisition-to-acquisition signal intensity variations (called intensity non-standardness), (2) inherent noise, and (3) intensity non-uniformity (i.e., bias-field or inhomogeneity). In our earlier work, we have shown that for precise quantification, all three of these signal variations should be corrected [9], [15], [19]–[21]. As the bias correction promotes noise, denoising filters should be used after the inhomogeneity correction step. Hence, the last step is the intensity standardization for correcting non-standardness among images [9].
1). Inhomogeneity correction:
Inhomogeneity occurs due to low spatial frequency intensity variations. Many MRI vendors include simple inhomogeneity correction methods during image acquisition process [15], [20]–[22]. However, those methods are often based on physical phantoms that are not truly representative of patient anatomy and position. In our experiments, we use post-processing filtering method based on nonparametric nonuniform intensity normalization, called N4ITK [23], which maximizes the high-frequency content of the tissue intensity distribution. We use the following parameters in N4ITK for reproducible results: order of spline (for fitting)=3, number of histogram bins=200, full width at half maximum for Gaussian convolution parameter=0.15.
2). Denoising:
Noise is ubiquitous in MRI scans. In addition to inherent MRI noise, image enhancement algorithms such as inhomogeneity correction can further intensify the noise. To this end, we have used the ball-scale based diffusive filtering method that preserves boundary sharpness and fine structures as reported in [9], [14], [19], which is found to have superior performance than anisotropic diffusion based filtering methods [14], [24].
3). Intensity standardization:
Due to acquisition-to-acquisition signal intensity variations, MR image intensities do not possess a tissue-specific numeric meaning between subsequent scans even for images acquired from the same subject, scanner, body region, or pulse sequence [16]. This phenomenon is called non-standardness. Non-standardness has been largely ignored in the literature until recently since many vendors have started to offer various intensity standardization algorithms for their image reconstruction processes [25]–[27]. Basically, these algorithms (called intensity standardization) are preprocessing techniques, which map image intensities into a standard intensity scale. The mapping is done through a piece-wise linear function which is obtained from a training step. In the current study, we have followed the intensity standardization approach as reported in [27]. For reproducible results, we set the min and max intensity values into 1 and 4095, respectively, as suggested by [16]. In addition, we set only a single foreground histogram mode to be determined automatically from the imaging data.
Figure 3 shows water-only, water-fat, and fat-only MR images obtained directly from the scanner in the first row. Images include inhomogeneity and noise distortion as expected. In the second row, MR images have been cleaned through the pre-processing steps in this order: inhomogeneity correction, denoising, and intensity standardization. Substantial improvements in the images can be readily observed in clean images. Yellow arrows show inhomogeneity and noisy regions prior to pre-processing step. As a quantitative example, we have also provided the coefficient of variance (CoV) values for each image before and after pre-processing steps in Figure 7 of the experiments and results section. As noted, the pre-processing approach improves not only visual interpretation but also reduces intensity variation for each tissue.
C. Background on FC image segmentation
FC defines the “hanging togetherness” between any two voxels p and q within an image [28]. A binary adjacency relationship (μα) determines adjacent voxels in α-adjacency. For any two adjacent voxels pi and pi+1, their local hanging togetherness is generally defined using an affinity function μk (pi, pi+1). This allows us to create probability maps when one of the voxel’s label is known. When two voxels are more closely related, their affinity becomes greater and this can be used for labeling purposes (i.e., segmentation). To generalize this concept from adjacent two voxels to a set of voxels, FC considers a “path”. A path π between p and q is identified as a sequence of adjacent voxels π = 〈p0 = p, p1,…, pl = q〉, where for an arbitrary path π, the strength of the path is defined as the “minimum affinity” along that path:
(1) |
Assuming is the set of all possible paths between p and q, then FC between them is the strength of the strongest path (indicating closely connected voxels with strong affinities as belonging to the same object):
(2) |
Defining a proper affinity function for two adjacent voxels is the key for a successful FC operation. Generally, the affinity function consists of three components: adjacency-based affinity μd, homogeneity-based affinity (μψ), and object-based affinity (μϕ) which are defined as:
(3) |
A wide range of mathematical functions can be used for affinities [29]. Herein, we use Euclidean distance for μd, and adopt the following form of μψ and μϕ for representing tissue class intensity properties:
(4) |
And
(5) |
where σψ and σϕ control the variance, and m controls the expected mean intensity of the target object. In our implementation, these values are all determined in the training step.
FC segmentation is obtained by generating a fuzzy object corresponding to a set of seed points . The fuzzy object membership value at a voxel p is then determined as the maximum FC value of all seed points:
(6) |
The final object is derived from a probability map (i.e., the fuzzy object ) by thresholding.
D. Novel Affinity Function for FC
The FC algorithm is based on affinity functions. The most prominent affinities explained so far in the literature are adjacency-based (μd), homogeneity-based (μψ), and object-feature based (μϕ). In this work, we propose to combine different affinities of different MRI contrasts and utilize the combined affinity function set to conduct segmentation. Figure 4 shows fat and muscle intensity distribution (after images are corrected for noise and inhomogeneity) for water-only (MRI1), water-fat (MRI2), and fat-only (MRI3) images. As it is depicted, the standard deviations of intensities pertaining to muscle are quite different across different contrasts. Similarly, fat intensity patterns have different mean and variance across different image contrasts. It is also worth mentioning that fat-only and water-fat contrasts show large overlaps in fat histograms as would be expected for non-fat suppressed image types. However, infiltrated fat inside the muscle tissue can still be observed in the Figure 4 (first row, the difference between blue and green density distributions indicates infiltrated fat). For precise muscle analysis, it would be more appropriate to choose MRI2 as it follows a balanced Gaussian distribution which is easier to analyze and incorporate into segmentation and/or tissue characterization algorithms (Figure 4, second row).
Based on all these qualitative observations and density histograms where intensity patterns follow distinct distributions, we propose to combine the affinity functions derived from each MRI contrast to quantify muscle and fat tissues jointly. Moreover, we also test the alternative hypothesis where individual segmentation from each image contrast will be combined as a decision fusion process. Here in this subsection, we confine ourselves to unifying affinity functions. In multi-contrast MRI of the thigh, let us have a fuzzy object membership affinity for each contrast: . Assume that there is a fuzzy object membership function that combines the complementary strength of muscle and fat tissue intensity distributions from different MRI contrasts. One straightforward way to combine multiple affinities is the weighted summation approach as:
where each affinity is weighted based on the effects of that particular MRI contrast on the final segmentation. Assuming that the segmentation accuracy is measured by dice similarity coefficient (DSC), then w can be set as:
This step is conducted off-line, as once the weights are determined from each contrast based on the goodness of segmentation results; the system uses the same set of weights for the segmenting different images.
E. Fusion of Multiple Segmentations (FC-Fusion)
As mentioned in the previous section, we also explore an alternative segmentation method based on fusion of multiple segmentations (namely, label or decision fusion). In decision fusion process, for designing FC fusion segmentation, we use a truth table for different input combinations. Figure 5 illustrates this process by indicating 0 or 1 for each entry of the truth label where 1 in output column indicates that both tissues (muscle and fat) can be recovered from the quantification process and 1 in left columns simply indicates if that particular image is used in fusion process or not. As can be seen from output column, to get fusion segmentation results, we need to use either MRI1 and MRI3 or sole MRI2 so as to maximize benefits of multi-contrast MRI. This indicates that, in the absence of MRI2, we need to use both MRI1 and MRI3 to generate the segmentation output. The truth table is simplified using a Karnaugh map (K-Map) to formulate the decision fusion and the resulting output map is determined as (MRI2 + MRI1.MRI3). The main reason for fusion of FC segmentation results is to compensate for any segmentation errors while performing segmentation over each modality separately.
F. Automatic seed selection
The proposed algorithm needs foreground and background seeds to initiate segmentation. Seed selection is critical in many medical image segmentation problems where literature is vast in proposing automatic seed selection procedures [30]. In this study, we devise three strategies for background and foreground seed sampling: manual, morphological filtering, and affinity propagation based seeding. We also evaluate these strategies’ comparative performances in the proposed delineation framework.
1). Manual seed selection:
Although our proposed method is fully automated, sometimes user interaction may be necessary for very challenging cases. In addition, it is often desirable to compare the fully automated method with interactive methods for efficiency purposes. Manual interaction is the most popular and perhaps the most reliable foreground and background seed sampling method. The strength of the manual seed selection comes from the superior recognition performance of human observers compared to automated methods. Sensitivity and specificity of manual seed selection in our experiments are 100% as observers are not required to differentiate vague boundaries between fat and muscle. However, it is also important to note that manual seed selection is the localization/detection part of the segmentation process and it is completely different from manual annotation of the objects, which is often called “manual segmentation”. The drawback of this strategy is the computational overhead when multiple seeds are required to complete segmentation. As summarized in Table V, manual seeding takes around 27 seconds, which is more than three times of the proposed segmentation’s processing time. This duration may increase substantially for other methods, such as graph-cut segmentation where the user may need to put several seeds at locations where the segmentation algorithm is underachieving.
TABLE V:
Method | Fat | Muscle | Total |
---|---|---|---|
FC Segmentation | 3 sec | 5 sec | 8 sec |
FC-Fusion Segmentation | 15 sec | 15 sec | 30 sec |
Proposed Segmentation | 3 sec | 5 sec | 8 sec |
Manual-Seeding | 12 sec | 15 sec | 27 sec |
Morphology-Seeding | 1 sec | 2 sec | 3 sec |
Proposed AP-Seeding | 1.17 sec | 1.16 sec | 2.33 sec |
2). Morphological image filtering for seed generation:
Clustering and morphology filtering based methods have been shown successful in segmenting tissues. Due to the unique challenges of the problem at hand, it may be desirable to have more advanced methods than manual seeding. In our particular problem, clustering and morphology filtering based methods can identify tissue classes up to a certain level, even if they fail to segment whole tissue labeling. Figure 2 second row shows how this information can be obtained through morphology based filtering methods such as thresholding, erosion, and connected component analysis. Parameters of thresholding and connected component analysis are optimized for fat and muscle tissues only in water-fat images due to a better contrast between fat and muscle tissues. Once fat and muscle tissues are roughly identified by morphological filtering, we detect sample voxels for muscle and fat tissue locations from the largest connected components of the output. We also ensure that the sampled seeds have neighboring voxels which have the same label. This constraint avoids seed selection from uncertain locations (i.e., boundaries) which are vulnerable to mis-segmentations.
3). Affinity propagation for seed generation:
Affinity propagation (AP) is an unsupervised learning task which is used to partition data into meaningful and similar groups while considering the similarity between pairs of data points [31]. Although the AP clustering algorithm is used successfully in many areas like image segmentation [32], auto-detection of genes; automatic seed generation via AP clustering is a novel contribution. Moreover, the proposed seed generation methodology through AP is based on multi-objects (i.e., fat and muscle) on multiple MR image contrasts. Herein, we first estimate the probability density functions of multiple MR image contrasts and roughly localize fat and muscle regions by sampling the most discriminative intensity values (Figure 2). To test the reliability of the sampled intensity locations, we use a connected component analysis to check the size of the object that the sampled voxel may belong to. Among the three MR image contrasts, experimental results and qualitative judgments show that water-fat and fat-only images provide the best contrast for seeding procedure which is described in the experiments and results section.
III. EXPERIMENTS AND RESULTS
A. Evaluation metrics
We have evaluated the experimental results both qualitatively and quantitatively. For qualitative evaluations, two participating expert interpreters use their visual assessment to evaluate the segmentation results. Furthermore, we have used dice similarity coefficient (DSC) to evaluate how well the segmented regions overlap with the ground truth provided by the participating interpreters. Intra- and inter-operator agreement rates between expert interpreters were found to be 85.7% and 83.5%, respectively. We have also used the coefficient of variance (CoV) to determine how effectively intensity variations can be eliminated. Lower CoVs indicate improved quality in images due to pre-processing (i.e., inhomogeneity correction).
B. Model evidence for MRI information content, and effects of preprocessing
To quantitatively study the information content of different MRI image contrasts, we calculate DSC pertaining to each image type. As shown in Figure 6, the highest DSC is obtained when water-fat (MRI2) is used. The lowest segmentation accuracy is obtained with water-only images (MRI1) as they contain the least discriminative information about the fat tissue.
It can be seen in Figure 7 that pre-processing makes CoV of fat and muscle tissues smaller. Since intensity variations are minimized prior to segmentation procedure with the pre-processing framework, segmentation quality is improved. In summary, the pre-processing will be substantially affecting the output not only for FC based segmentation methods but also for any segmentation approach that utilizes image intensities.
C. Evaluation of the Segmentation Accuracy
When segmentation accuracies are compared, the most important property is to have high mean accuracy with low standard deviation. We have compared our two proposed methods with the baseline FC segmentation as well (Table II). Furthermore, we have quantitatively tested the effectiveness of the seeding procedure in terms of accuracy and robustness with respect to different seeding strategies. For automated seeding, we have proposed two procedures: AP and morphology based seeding strategies. The proposed method achieves the best results with high mean and low standard deviation as summarized in Table II. All FC based segmentation results with the proposed AP seed generation are substantially better than other segmentation results, which are either based on manual or morphology based seed generation. In the proposed method, the segmentation with the automated AP seed generation gives the best DSC because the combination of weighted affinity functions further improved the segmentation quality. This is primarily due to the improved nonparametric modeling of the tissue class with AP clustering method.
TABLE II:
Method | Fat (%) | Muscle (%) |
---|---|---|
FC Segmentation | DSC (mean/std) | DSC (mean/std) |
Manual-Seeding | 75.60 / 11.18 | 52.59 / 18.14 |
Morphology-Seeding | 54.90 / 15.98 | 71.52 / 20.09 |
With AP-Seeding | 77.26 / 11.57 | 72.88 / 18.59 |
FC-Fusion Segmentation | DSC (mean/std) | DSC (mean/std) |
Manual-Seeding | 77.51 / 17.46 | 82.07 / 9.27 |
Morphology-Seeding | 77.39 / 18.84 | 84.99 / 17.64 |
With AP-Seeding | 81.42 / 14.27 | 87.33 / 13.08 |
Proposed Segmentation | DSC (mean/std) | DSC (mean/std) |
Manual-Seeding | 81.07 / 14.75 | 81.70 / 11.44 |
Morphology-Seeding | 80.66 / 10.36 | 85.27 / 20.36 |
With AP-Seeding | 84.06 / 12.23 | 87.86 / 11.78 |
Proposed Segmentation | % (mean/std) | % (mean/std) |
Sensitivity of AP-Seeding | 79.27 / 12.46 | 84.18 / 16.75 |
Specificity of AP-Seeding | 98.16 / 8.50 | 96.78 / 10.10 |
Once segmentations are finalized, we have conducted a linear regression analysis for volumetric quantification of computed and true fat/muscle tissues. The correlation between the computed muscle-fat volumes and the ground truth volumes is shown in Figure 8. As it can be observed, the computed volumes are highly correlated with the ground truth volumes (R2 > 0.8 and R2 > 0.95 for muscle and fat volumes, respectively).
D. Comparison to other methods
As briefly summarized in Section I-B, available approaches for segmentation of fat and muscle for thigh MRI can be categorized under boundary based methods [11], region based methods such as Markov random field (MRF) [33], graph-cut [34] and clustering [35], and machine learning based methods such as random (decision) forest (RF) [36], multilayer perceptron (MLP) [37], SVM [38], BayesNET [39], Hoeffding trees [40], Random committee [41], and AdaBoost [42]. We compare our proposed method with several existing methods both qualitatively and quantitatively. Table III lists the quantitative comparisons of the proposed method’s DSCs for fat and muscle segmentation with Boykov’s graph-cut [34], active contour [11], [12] and machine learning based methods of random forest [36] and random committee [41].
TABLE III:
Method | Fat (%) | Muscle (%) |
---|---|---|
Graph-Cut | DSC (mean/std) | DSC (mean/std) |
Boykov’s MaxFlow [34] | 73.50 / 16.75 | 72.20 / 15.01 |
Active Contour [11], [12] | DSC (mean/std) | DSC (mean/std) |
Active contour/snake | 71.41 / 16.13 | 74.73 / 14.86 |
Clustering | DSC (mean/std) | DSC (mean/std) |
Random Forest [36] | 73.41 / 15.38 | 71.19 / 18.41 |
Random Committee [41] | 74.83 / 13.39 | 71.90 / 16.42 |
Proposed Segmentation | DSC (mean/std) | DSC (mean/std) |
With AP-Seeding | 84.06 / 12.23 | 87.86 / 11.78 |
We report statistical validation of FC based (proposed) and other methods’ segmentation using DSC in Table III. Proposed FC based algorithm showed higher DSCs compared to other methods. Among the widely used methods, Boykov’s Max Flow Min Cut works better than active contour for fat segmentation but worse for muscle segmentation. While analyzing the classifier based segmentation results, although random forest and random committee show quite similar DSCs, random committee is a little better as it has higher mean and lower deviation. Among the proposed FC based-methods, the best results were obtained when fusion was conducted at the affinity function level: arranging the weight of the affinity functions for each imaging modality (i.e., image contrast of the MRI) significantly improved the DSC for fat and muscle to around 84% and 87%, respectively (Table II). The results from our proposed and alternative segmentation methods prove the efficacy and feasibility of our proposed solution.
For qualitative comparison with other methods, Figure 9 illustrates an example of the segmentation performance for several different methods. We have selected a representative image slice from a single image contrast of the thigh MRI in Figure 9(a), and segmentation results of region based, boundary based, and machine learning based methods are shown in Figure 9(b-m). The proposed algorithm shows improved segmentation results in Figure 9(i) compared to others. Several of the other methods were not able to capture large fat regions in the subcutaneous compartment while leakages (i.e., over-segmentation to other tissues) were inevitable in most cases.
E. Feasibility Studies: Brain and Whole-Body Tissue Segmentation
To evaluate the generalizability of our proposed approach both for different MRI type and body regions, we performed two additional experiments. In the first experiment, we applied our algorithm into Dixon MRI sequences for whole body tissue segmentation. Briefly, for a given sample whole-body MRI (Dixon, Figure 10(a)-left), its 3-class tissue segmentation (Figure 10(a)-right) was obtained within seconds. In our implementation, we considered fat, muscle, and lung fields in the segmentation. Resulting tissue delineations were visually evaluated (and found to be clinically feasible) by the participating experts of this study.
In the second experiment, our proposed method was validated using 20 simulated brain MRI scans (T1-weighted) [43], with the ground truth known. Various noise and smoothing levels were also used to evaluate the robustness of the proposed method. Three smoothing levels (small, medium, and large) were applied to the images using a Gaussian smoothing filter. Three noise levels (small, medium, and large) were applied to the images as well. With the proposed approach, considering all 20 brain MRI scans, Gray Matter (GM) and White Matter (WM) segmentation accuracies were excellent with the sensitivity and specificity rates above 90%; Cerebrospinal fluid (CSF) segmentation was good with the sensitivity and specificity always above 85%. The literature demonstrates that CSF is challenging to segment, with sensitivities generally in the 80% - 88% range and specificities between 51% to 88% [44]. Figure 10(b) shows segmented brain tissues. These feasibility results (both whole body and brain tissue segmentations) demonstrate the promising generic aspects of our proposed method in segmentation and quantification of different body regions, and type of MRI scans. The segmentation results are shown in Table IV.
F. Computational efficiency
Proposed segmentation and seeding (AP) methods have achieved the lowest computational time while guaranteeing higher accuracy as listed in Table V. The slowest segmentation procedure took only 57 seconds, of which 27 seconds were spent on seeding and 30 seconds were used for fusion strategy. The shortest segmentation duration was 10.33 seconds provided by the proposed AP-seeding based FC segmentation algorithm integrating adjusted affinities. Manual seeding was restricted to a real clinical setting where a few seeds were often desired to initiate the segmentation; otherwise, seeding process itself can be time-consuming. Increasing the number of seeds led to similar segmentation results as that obtained through AP based seeding method but at the expense of an increase in the analysis time.
G. Quantification from a single image slice
In clinical practice, radiologists frequently use a single slice to estimate the whole volume for a region in an object. For instance, in abdominal fat quantification, instead of segmenting the whole abdomen and separating subcutaneous tissue from visceral, which is a challenging problem, a single slice at the umbilical level is often selected and used to quantify abdominal fat volume [1], [46]. The main reasons for this practice are two-fold. First is a lack of software algorithms that can work efficiently and accurately. This is because many algorithms fail to provide exact quantification, and thus user interaction is necessary for a better quantification. It is time-consuming for radiologists to correct each segmentation failure for volumetric images. Second, the anatomy does not change abruptly from slice to slice. Thus, it is a reasonable assumption that the content of the tissue distribution in a single slice will be correlated with the whole volume. Then, based on a known correlation between a single slice and volume, creating a linear regression equation will estimate the volumetric tissue distribution from a single slice. Inspired by this clinical practice and to explore the inter-relationship of thigh MRI slices, we perform a Pearson correlation test between fat and muscle volume extracted from a single slice with its volumetric surrogate truths. Figure 11 shows correlation value for each slice when estimating the volumetric distribution of fat and muscle, respectively. As indicated, each slice contributed more than R = 0.97 correlation value, and the highest correlations R > 0.99 were obtained when mid slice of the thigh region was selected. This indicates that the proposed system can be used for much faster muscle and fat volume estimation from a single slice for practical purposes.
IV. DISCUSSIONS AND CONCLUDING REMARKS
A. Limitations and Future Works
Our study has a few limitations to be noted. With its success in almost all tasks of visual domain and its extensive use for medical image analysis, deep learning can serve as an alternative to our proposed approach. The joint segmentation of multi-contrast images in deep neural network framework can lead to accurate and robust performance. In addition, the parameters in our proposed approach were learned offline, which can be estimated in an online manner for reliable performance. Moreover, in the context of skeletal muscle composition studies, more advanced MRI acquisition methods like Dixon imaging can be used which provide an improved correspondence between the image signal and the fat and water fractions within a given voxel. Lastly, it is worth mentioning that partial volume can be an issue to be solved for separating water and fat signals when their contributions are roughly equal at a given voxel. Especially when the resolution is limited, this effect should be corrected. Although we did not observe large errors due to partial volumes, conventional partial volume correction methods for MR images such as [47] can be applied to minimize such errors in volumetric quantification. Furthermore, Dixon can be used for fat/muscle quantification in sub-voxel level unlike our method where we used voxel level quantification.
Our future study will involve exploring another important imaging marker, called biomechanical muscle quality (BMQ), for a potential association of aging and age-related metabolic diseases. In order to calculate BMQ, we will identify the intra-muscular adipose tissue (IMAT), which is associated with a reduction in force production and suggested to affect BMQ. For this purpose, a unified framework to perform segmentation and quantification of the three tissues (muscle, fat, and IMAT) can be utilized in the clinical evaluation of BMQ. The current work also suggests a potential for application of deep learning based methods for improved multi-contrast image segmentation approaches. Inspired by the recent success of purely machine learning methods and the current methods’ limited achievement on IMAT quantification (i.e., around 70–80% success bands [11], [12], [35]), we aim to explore the joint segmentation of these three tissues using a deep learning framework in our future study. Moreover, given a limited number of images, we performed evaluations on a small dataset for whole-body tissue classification as well as brain tissue segmentation. Although the results are quite promising, with a larger dataset, deep learning based approaches can be explored for improved segmentation of whole-body and brain tissues.
B. Summary
Conventional methods for the tissue delineation and quantification mostly rely on manual or semi-automated algorithms, having low accuracy and efficiency. In the case of multi-contrast MRI scans, conventional approaches use isolated frameworks for each imaging modality; thus, leading to suboptimal results. Given the clinical importance of fat and muscle segmentation and quantification in thigh MRI, as well their utility in other research settings, we proposed a novel approach to address unique challenges associated with these images and showed the generalization ability of the proposed system for the whole-body tissue composition and the brain tissue delineation. We provided a complete quantification framework that automatically minimizes extreme challenges (inhomogeneity, noise, and non-standardness) of the MRI automatically. To maintain the unsupervisory and automated nature of our proposed approach, we performed automatic seed generation using an affinity propagation clustering algorithm. In the delineation step, to the best of our knowledge, the proposed approach is the first to explore the utility of multiple affinity functions to perform a unified segmentation of multiple objects from multi-contrast MRI. Our automated seed selection approach also evades the need for filtering approaches which can be prone to subjectivities.
ACKNOWLEDGMENT
This work was supported in part by the Intramural Research Program of the National Institute on Aging of the National Institutes of Health (NIH).
Footnotes
Figure was created with the BioDigital Human Visualization Platform ©.
Contributor Information
Ismail Irmakci, Ege University, Izmir, Turkey. Irmakci’s research is supported by scientific council of Turkey (TUBITAK-BIDEB 2214/A)..
Sarfaraz Hussein, Center for Research in Computer Vision at University of Central Florida, Orlando, FL..
Aydogan Savran, Ege University, Izmir, Turkey. Irmakci’s research is supported by scientific council of Turkey (TUBITAK-BIDEB 2214/A)..
Rita R. Kalyani, Department of Medicine, Johns Hopkins University School of Medicine, Baltimore, MD.
David Reiter, National Institute on Aging, National Institutes of Health (NIH), Baltimore, MD..
Chee W. Chia, National Institute on Aging, National Institutes of Health (NIH), Baltimore, MD.
Kenneth W. Fishbein, National Institute on Aging, National Institutes of Health (NIH), Baltimore, MD.
Richard G. Spencer, National Institute on Aging, National Institutes of Health (NIH), Baltimore, MD.
Luigi Ferrucci, National Institute on Aging, National Institutes of Health (NIH), Baltimore, MD..
Ulas Bagci, Center for Research in Computer Vision at University of Central Florida, Orlando, FL..
REFERENCES
- [1].Hussein S, Green A, Watane A, Reiter D, Chen X, Papadakis GZ, Wood B, Cypess A, Osman M, and Bagci U, “Automatic segmentation and quantification of white and brown adipose tissues from pet/ct scans,” IEEE Transactions on Medical Imaging, vol. 36, no. 3, pp. 734–744, 2017. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [2].Prescott JW, Best TM, Swanson MS, Haq F, Jackson RD, and Gurcan MN, “Anatomically anchored template-based level set segmentation: application to quadriceps muscles in mr images from the osteoarthritis initiative,” Journal of digital imaging, vol. 24, no. 1, pp. 28–43, 2011. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [3].Okada T, Shimada R, Hori M, Nakamoto M, Chen Y-W, Nakamura H, and Sato Y, “Automated segmentation of the liver from 3d ct images using probabilistic atlas and multilevel statistical shape model,” Academic radiology, vol. 15, no. 11, pp. 1390–1403, 2008. [DOI] [PubMed] [Google Scholar]
- [4].Aljabar P, Heckemann RA, Hammers A, Hajnal JV, and Rueckert D, “Multi-atlas based segmentation of brain images: atlas selection and its effect on accuracy,” Neuroimage, vol. 46, no. 3, pp. 726–738, 2009. [DOI] [PubMed] [Google Scholar]
- [5].Ahmad E, Yap MH, Degens H, and McPhee JS, “Atlas-registration based image segmentation of mri human thigh muscles in 3d space,” in SPIE Medical Imaging. International Society for Optics and Photonics, 2014, pp. 90 371L–90 371L. [Google Scholar]
- [6].Essafi S, Langs G, Deux J-F, Rahmouni A, Bassez G, and Paragios N, “Wavelet-driven knowledge-based mri calf muscle segmentation,” in 2009 IEEE International Symposium on Biomedical Imaging: From Nano to Macro. IEEE, 2009, pp. 225–228. [Google Scholar]
- [7].Wang C, Teboul O, Michel F, Essafi S, and Paragios N, “3d knowledge-based segmentation using pose-invariant higher-order graphs,” in International Conference on Medical Image Computing and Computer-Assisted Intervention. Springer, 2010, pp. 189–196. [DOI] [PubMed] [Google Scholar]
- [8].Chen X, Summers RM, Cho M, Bagci U, and Yao J, “An automatic method for renal cortex segmentation on ct images: evaluation on kidney donors,” Academic radiology, vol. 19, no. 5, pp. 562–570, 2012. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [9].Bagci U, Chen X, and Udupa JK, “Hierarchical scale-based multi-object recognition of 3-d anatomical structures,” IEEE Transactions on Medical Imaging, vol. 31, no. 3, pp. 777–789, 2012. [DOI] [PubMed] [Google Scholar]
- [10].Andrews S and Hamarneh G, “The generalized log-ratio transformation: learning shape and adjacency priors for simultaneous thigh muscle segmentation,” IEEE transactions on medical imaging, vol. 34, no. 9, pp. 1773–1787, 2015. [DOI] [PubMed] [Google Scholar]
- [11].Makrogiannis S, Fishbein K, Moore A, Spencer R, and Ferrucci L, “Image-based tissue distribution modeling for skeletal muscle quality characterization,” IEEE Transactions on Biomedical Engineering, vol. 63, no. 4, pp. 805–813, 2016. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [12].Makrogiannis S, Serai S, Fishbein KW, Schreiber C, Ferrucci L, and Spencer RG, “Automated quantification of muscle and fat in the thigh from water-, fat-, and nonsuppressed mr images,” Journal of Magnetic Resonance Imaging, vol. 35, no. 5, pp. 1152–1161, 2012. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [13].Madabhushi A and Udupa JK, “Interplay between intensity standardization and inhomogeneity correction in mr image processing,” IEEE Transactions on Medical Imaging, vol. 24, no. 5, pp. 561–576, 2005. [DOI] [PubMed] [Google Scholar]
- [14].Saha PK and Udupa JK, “Scale-based diffusive image filtering preserving boundary sharpness and fine structures,” IEEE Transactions on Medical Imaging, vol. 20, no. 11, pp. 1140–1155, 2001. [DOI] [PubMed] [Google Scholar]
- [15].Bagci U, Udupa JK, and Chen X, “Intensity non-standardness affects computer recognition of anatomical structures,” in SPIE Medical Imaging. International Society for Optics and Photonics, 2011, pp. 79 642M–79 642M. [Google Scholar]
- [16].Nyúl LG, Udupa JK et al. , “On standardizing the mr image intensity scale,” image, vol. 1081, 1999. [DOI] [PubMed] [Google Scholar]
- [17].Ferrucci L, “The Baltimore Longitudinal Study of Aging (BLSA): a 50-year-long journey and plans for the future,” 2008. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [18].Wollenweber S, Ambwani S, Delso G, Lonn A, Mullick R, Wiesinger F, Piti Z, Tari A, Novak G, and Fidrich M, “Evaluation of an atlas-based PET head attenuation correction using PET/CT & MR patient data,” IEEE Transactions on Nuclear Science, vol. 60, no. 5, pp. 3383–3390, 2013. [Google Scholar]
- [19].Bagci U, Udupa JK, and Chen X, “Orientation estimation of anatomical structures in medical images for object recognition,” in SPIE Medical Imaging. International Society for Optics and Photonics, 2011, pp. 79 622L–79 622L. [Google Scholar]
- [20].Bagci U, Udupa JK, and Bai L, “The influence of intensity standardization on medical image registration,” in SPIE Medical Imaging. International Society for Optics and Photonics, 2010, pp. 76 251X–76 251X. [Google Scholar]
- [21].Bağcı U, Udupa JK, and Bai L, “The role of intensity standardization in medical image registration,” Pattern Recognition Letters, vol. 31, no. 4, pp. 315–323, 2010. [Google Scholar]
- [22].Axel L, Costantini J, and Listerud J, “Intensity correction in surface-coil mr imaging,” American Journal of Roentgenology, vol. 148, no. 2, pp. 418–420, 1987. [DOI] [PubMed] [Google Scholar]
- [23].Tustison NJ, Avants BB, Cook PA, Zheng Y, Egan A, Yushkevich PA, and Gee JC, “N4itk: improved n3 bias correction,” IEEE transactions on medical imaging, vol. 29, no. 6, pp. 1310–1320, 2010. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [24].Bagci U, Udupa J, and Chen X, “Ball-scale based multi-object recognition in a hierarchical framework,” in Proc. of SPIE Medical Imaging, vol. 7623, 2010, pp. 7 623 451–76 234 512. [Google Scholar]
- [25].Robitaille N, Mouiha A, Crépeault B, Valdivia F, and Duchesne S, “Tissue-based mri intensity standardization: application to multicentric datasets,” Journal of Biomedical Imaging, vol. 2012, p. 4, 2012. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [26].Jäger F, Nyúl L, Frericks B, Wacker F, and Hornegger J, “Whole body mri intensity standardization,” in Bildverarbeitung für die Medizin 2007. Springer, 2007, pp. 459–463. [Google Scholar]
- [27].Nyúl LG, Udupa JK, and Zhang X, “New variants of a method of mri scale standardization,” IEEE transactions on medical imaging, vol. 19, no. 2, pp. 143–150, 2000. [DOI] [PubMed] [Google Scholar]
- [28].Udupa JK and Samarasekera S, “Fuzzy connectedness and object definition: theory, algorithms, and applications in image segmentation,” Graphical models and image processing, vol. 58, no. 3, pp. 246–261, 1996. [Google Scholar]
- [29].Ciesielski KC, Udupa JK, Falcão AX, and Miranda PA, “Fuzzy connectedness image segmentation in graph cut formulation: A linear-time algorithm and a comparative analysis,” Journal of Mathematical Imaging and Vision, vol. 44, no. 3, pp. 375–398, 2012. [Google Scholar]
- [30].Pham DL, Xu C, and Prince JL, “Current methods in medical image segmentation 1,” Annual review of biomedical engineering, vol. 2, no. 1, pp. 315–337, 2000. [DOI] [PubMed] [Google Scholar]
- [31].Frey BJ and Dueck D, “Clustering by passing messages between data points,” science, vol. 315, no. 5814, pp. 972–976, 2007. [DOI] [PubMed] [Google Scholar]
- [32].Foster B, Bagci U, Xu Z, Dey B, Luna B, Bishai W, Jain S, and Mollura DJ, “Segmentation of pet images for computer-aided functional quantification of tuberculosis in small animal models,” IEEE Transactions on Biomedical Engineering, vol. 61, no. 3, pp. 711–724, 2014. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [33].Rue H and Held L, Gaussian Markov random fields: theory and applications. CRC Press, 2005. [Google Scholar]
- [34].Chen X, Udupa JK, Bagci U, Zhuge Y, and Yao J, “Medical image segmentation by combining graph cuts and oriented active appearance models,” IEEE Transactions on Image Processing, vol. 21, no. 4, pp. 2035–2046, 2012. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [35].Tan C, Li K, Yan Z, Yang D, Zhang S, Yu HJ, Engelke K, Miller C, and Metaxas D, “A detection-driven and sparsity-constrained deformable model for fascia lata labeling and thigh inter-muscular adipose quantification,” Computer Vision and Image Understanding, 2016. [Google Scholar]
- [36].Breiman L, “Random forests,” Machine learning, vol. 45, no. 1, pp. 5–32, 2001. [Google Scholar]
- [37].Rumelhart DE, Hinton GE, and Williams RJ, “Learning internal representations by error propagation,” DTIC Document, Tech. Rep, 1985. [Google Scholar]
- [38].Cortes C and Vapnik V, “Support-vector networks,” Machine learning, vol. 20, no. 3, pp. 273–297, 1995. [Google Scholar]
- [39].Nielsen TD and Jensen FV, Bayesian networks and decision graphs. Springer Science & Business Media, 2009. [Google Scholar]
- [40].Hulten G, Spencer L, and Domingos P, “Mining time-changing data streams,” in Proceedings of the seventh ACM SIGKDD international conference on Knowledge discovery and data mining. ACM, 2001, pp. 97–106. [Google Scholar]
- [41].Lira MM, de Aquino RR, Ferreira AA, Carvalho MA, Neto ON, and Santos GS, “Combining multiple artificial neural networks using random committee to decide upon electrical disturbance classification,” in Neural Networks, 2007. IJCNN 2007. International Joint Conference on. IEEE, 2007, pp. 2863–2868. [Google Scholar]
- [42].Freund Y and Schapire RE, “A desicion-theoretic generalization of online learning and an application to boosting,” in European conference on computational learning theory. Springer, 1995, pp. 23–37. [Google Scholar]
- [43].Collins DL, Zijdenbos AP, Kollokian V, Sled JG, Kabani NJ, Holmes CJ, and Evans AC, “Design and construction of a realistic digital brain phantom,” IEEE transactions on medical imaging, vol. 17, no. 3, pp. 463–468, 1998. [DOI] [PubMed] [Google Scholar]
- [44].de Boer R, Vrooman HA, Ikram MA, Vernooij MW, Breteler MM, van der Lugt A, and Niessen WJ, “Accuracy and reproducibility study of automatic mri brain tissue segmentation methods,” Neuroimage, vol. 51, no. 3, pp. 1047–1056, 2010. [DOI] [PubMed] [Google Scholar]
- [45].Lim PH, Bagci U, and Bai L, “Introducing willmore flow into level set segmentation of spinal vertebrae,” IEEE Transactions on Biomedical Engineering, vol. 60, no. 1, pp. 115–122, 2013. [DOI] [PubMed] [Google Scholar]
- [46].Green AL, Bagci U, Hussein S, Kelly PV, Muzaffar R, Neuschwander-Tetri BA, and Osman MM, “Brown adipose tissue detected by pet/ct imaging is associated with less central obesity,” Nuclear Medicine Communications, vol. 38, no. 7, pp. 629–635, 2017. [DOI] [PubMed] [Google Scholar]
- [47].Shattuck DW, Sandor-Leahy SR, Schaper KA, Rottenberg DA, and Leahy RM, “Magnetic resonance image tissue classification using a partial volume model,” NeuroImage, vol. 13, no. 5, pp. 856–876, 2001. [DOI] [PubMed] [Google Scholar]