Abstract
Haralick texture features are common texture descriptors in image analysis. To compute the Haralick features, the image gray-levels are reduced, a process called quantization. The resulting features depend heavily on the quantization step, so Haralick features are not reproducible unless the same quantization is performed. The aim of this work was to develop Haralick features that are invariant to the number of quantization gray-levels. By redefining the gray-level co-occurrence matrix (GLCM) as a discretized probability density function, it becomes asymptotically invariant to the quantization. The invariant and original features were compared using logistic regression classification to separate two classes based on the texture features. Classifiers trained on the invariant features showed higher accuracies, and had similar performance when training and test images had very different quantizations. In conclusion, using the invariant Haralick features, an image pattern will give the same texture feature values independent of image quantization.
Introduction
The easiest and most intuitive image features for most applications in image analysis are first order statistics computed from histograms of the gray-level values in images, like their mean, variance, skewness and kurtosis. Such features involve the values of individual pixels, but ignore the spatial interaction between pixels. Histogram features do not reflect objects or patterns in the image, only the distribution of gray-levels. This inability makes first order statistics a blunt tool for quantifying changes in images, or any change in the spatial distribution of gray values.
Most texture analysis methods use higher-order statistics, and consider the relation between two or more pixels at a time. Methods such as local binary patterns [1], wavelets [2] and Gabor filters [3] can be used to assess texture in images. Haralick et al. [4] proposed using a gray-level co-occurrence matrix (GLCM) as a method of quantifying the spatial relation of neighboring pixels in an image. Haralick texture features, computed from the GLCM, are widely used due to their simplicity and intuitive interpretations, and have successfully been applied in e.g. in the analysis of skin texture [5], in land-use and forest-type classification [6], in automatic pollen detection [7], in fabric defect detection [8], in plant leaf classification [9], in cutting tool condition monitoring [10], and electrophoresis analysis [11]. In recent years there has been a rapid increase in the application of Haralick features in medical image analysis, e.g. in the analysis of ultrasound and MRI images of the liver [12, 13], the heart [14], X-ray mammography [15, 16], MRI images in the study of breast cancer [17, 18], prostate cancer [19–21] and brain cancer [22–24]. It is also used in radiomics [25, 26], which is an emerging technique where a large number of quantitative features are extracted from medical images, and used to build models predicting e.g. tumor phenotype [27], survival [28, 29] and tumor classification [30].
When creating a GLCM from an image or a region of interest (ROI), the image bit depth, i.e. the number of gray-levels, may be reduced in a process called quantization. It is, however, not straight-forward to set an appropriate bit depth, and many projects have been looking at the optimal bit depths for various applications. [28, 31–35] The optimal bit depth depends on the size of the image or ROI, image noise [36] and, of course, the content of the image [33]. To complicate things further, the values of the Haralick features depend strongly on the bit depth and the choice of maximum and minimum values used in the quantization [34, 37]. Clausi et al. [38] proposed a method for normalizing two features with respect to the quantization gray-levels to improve classification, however this method is not generalizable to all Haralick features. Shafiq-ul-Hassan et al. [39] proposes another method for scaling a selection of the Haralick features to make them less sensitive to the quantization gray-levels. The presented approach is empirical, and no general rule for feature normalization is presented. Currently, most Haralick feature values cannot be compared between analyses using different numbers of gray-levels, even when they are computed on images depicting the same texture. This means that Haralick features cannot be used to create general models that can be applied regardless of image size or noise levels, and are not reproducible unless the same quantization is performed.
The aim of this work is to propose a modified set of Haralick texture features that are asymptotically invariant to the image quantization, while preserving most of the interpretations of the original features. The modified features allow statistical models to be constructed from images with varying number of gray-levels, or to apply a model trained on one number of gray-levels to image data using a different number of gray-levels.
Theory
The first step in texture analysis using the Haralick features is to map the gray-levels in the original image with size M × K, I′, from the range [a, b], to a quantized image, I, in the range [1, N], that has the desired number of gray-levels, N. We represent this step with a quantization function
(1) |
that returns the quantized image
(2) |
In this work we have set φ to be an affine function, but there is no constraint on this function for the purpose of this work.
The second step is to construct the non-normalized GLCM, , by counting the number of times each pair of gray-levels occurs as neighbors the image, I, or in an arbitrary region in I. The neighbor of a pixel is defined by a displacement vector, δ = (dx, dy), in two dimensions, where represent the displacement in x and y in units of pixels. Each element, X(i, j), in the GLCM is computed as
(3) |
where I(m, k) is element m, k in the quantized image I. Simply put, the element X(i, j) of the non-normalized GLCM counts how many times gray values i and j occur as neighbors in I, where i, j ∈ [1, N]. It is possible to create a GLCM with several displacement vectors, e.g.
(4) |
for the eight immediate neighbors of a pixel, in two dimensions. This can be extended to three dimensions, in which case 26 displacement vectors would be required. A GLCM created using all of these displacement vectors, e.g. by summation, is said to be direction invariant. If displacement vectors with opposite directions are used to construct the GLCM, the GLCM will be symmetric. Symmetric GLCMs are semi-direction invariant, they represent relations between pixels in e.g. vertical or horizontal directions instead of up or down, left or right. A semi-direction invariant GLCM can also be created by adding the transpose of a GLCM created by one displacement vector to itself. This means that only one of the columns in Eq 4 needs to be calculated to obtain a direction invariant GLCM in two dimensions, or 13 displacement vectors for GLCMs constructed from neighbors in three dimensions.
The third step is to construct a normalized GLCM, , where each element represents the estimated probability of each combination of pairs of neighboring gray-levels in the image. The normalized GLCM is computed as
(5) |
The normalized GLCM can be interpreted as a probability mass function of the gray-level pairs in the image.
The fourth step is to compute the texture features from the normalized GLCM. The Haralick texture features are functions of the elements and their corresponding indices in the normalized GLCM, and can be written in a general form as
(6) |
where ψ is a function of an element of P, g is a vector-valued function of P, and ϕ is a function of the indices and g. Examples of g are found in in Table 1. For instance when computing Correlation (see Table 2), we have
Table 1. Variables and notation used to compute the Haralick texture features.
Notation | Definition | Mod. Notation | Definition |
---|---|---|---|
x(i, j) | Element i, j in the unnormalized GLCM | ||
N | Number of gray-levels | ||
Δ | |||
Δx+y | |||
Δij | |||
p(i, j) | |||
px(i) | |||
py(j) | |||
μx | |||
μy | |||
px+y(k) | |||
px−y(k) | |||
μx+y | |||
μx−y | |||
HX | |||
HY | |||
HXY | |||
HXY 1 | |||
HXY 2 | |||
Q(i, j) |
Table 2. The texture features computed from GLCMs.
Feature | Original expression | Invariant expression |
---|---|---|
Autocorrelation [31] | ||
Cluster prominence [4] | ||
Cluster shade [4] | ||
Contrast [4] | ||
Correlation [4] | ||
Difference entropy [4] | ||
Difference variance [4] | ||
Dissimilarity [31] | ||
Energy [4] | ||
Entropy [4] | ||
Homogeneity [31] | ||
Information measure of correlation 1 [4] | ||
Information measure of correlation 2 [4] | ||
Inverse difference [38] | ||
Maximum probability [31] | ||
Sum average [4], μx+y | ||
Sum entropy [4] | ||
Sum of squares [4] | ||
Sum variance [4] | ||
Maximal Correlation Coefficient [4] |
Fig 1 illustrates Steps 2–4 of computing the Haralick texture features from a 4 × 4 example image, that has already been quantized to three gray-levels.
The number of gray-levels in the quantized image determines the size of the GLCM, and will affect the texture feature values. By increasing the number of gray-levels, N, the sums in Tables 1 and 2 will change. Further, the element values, p(i, j), of the normalized GLCM will decrease as the number of gray-levels increases, since the sum of all elements is unity. These two properties of the GLCM and the Haralick texture features means that the texture feature values depend on the number of gray-levels in the quantized image.
Invariant texture features
To develop asymptotically invariant texture features, we propose an interpretation of the GLCM as a discrete approximation of a probability density function. In this interpretation, the texture features are expressed as integrals over functions of the density, i.e. as
(7) |
where p* is the underlying true probability density function, for which
(8) |
Eq 7 can be approximated by a Riemann sum:
(9) |
where
(10) |
and
(11) |
where Δi and Δj are differentials, and is the invariant GLCM. So, to make the Haralick texture features invariant to the number of gray-levels, i and j are normalized to the half-open interval (0, 1], the sums are multiplied by differentials, and the GLCM is normalized so that its Riemann sum is 1, see Fig 2 and Eq 11. The proposed invariant texture features can be seen in Columns 3 and 4 of Tables 1 and 2. The discrete approximations of the integrals depend on the discretization of the distribution (i.e., the quantization, or the number of gray-levels), but the approximation is equal to the integral in the limit as N → ∞. The original features do not give the same approximated values for different discretizations, they either approach zero, a fixed value or infinity when N → ∞.
Materials and methods
We investigated the properties of the original and invariant texture features using two data sets. Dataset 1 contained T1-weighted MRI volumes of the brain. We computed texture features for the cerebellum and the prefrontal cortex, see Fig 3. We used logistic regression to classify each region based on their texture properties, and compared the result for the two methods. This dataset has limited clinical utility, and is used to compare the performance of the original and invariant texture features only. Dataset 2 contained histology images of colorectal cancer glands from the Warwick-QU dataset [40, 41]. This is an open dataset that was used in the Gland Segmentation in Colon Histology Images Challenge Contest (GlaS), held at MICCAI’2015. We computed texture features for each gland, and used logistic regression to classify each gland as benign or malignant.
Dataset 1 imaging
Dataset 1 contained 81 T1-weighted axial spoiled gradient echo images of 29 subjects. The image field of view was 250 × 250 × 240 mm3, with an in-plane resolution of 1.3 × 1.3 mm2 and a slice thickness of 1 mm. The repetition time and echo time were 7.1 and 2.5 ms, respectively, and the bandwidth was 392 Hz/pixel. A region of the prefrontal cortex (30 × 30 × 10 mm3, 23 × 23 × 10 voxels), and a region of the cerebellum (25 × 25 × 10 mm3, 19 × 19 × 10 voxels), were delineated on a reference image set, see Fig 3. The other images were rigidly registered to the reference, and the corresponding regions were extracted. Twenty-three images from 13 patients where the image registration failed were excluded from the study. The clinical trial was approved by the local Regional Ethical Review Board of Umeå University, and oral and written consent was given by all subjects.
Dataset 2 imaging
Dataset 2 contained 1,518 images of colorectal glandular structures in 165 Hematoxylin and Eosin (H&E) stained slides. The slides were digitally scanned at 20 × magnification, with a resolution of 0.62005 μm/pixel, using a Zeiss MIRAX MIDI Slide Scanner. The image dimensions were 520 × 775 pixels for 151 slide scans, and 430 × 575 pixels for 14 slides. 934 gland structures in 74 slides were classified as benign and 584 structures in 91 slides as malignant. The luminance, calculated from the RGB slides were used in the analysis. The images were dithered using uniform noise, to reduce quantization errors.
Texture analysis
Thirty-two direction invariant GLCMs were created from the immediate voxel neighbors in Eq 4, for each brain region or gland structure, one for every quantization of 8–256 gray-levels in steps of 8. We used fixed upper and lower limits of 1500 and 4500 pixel units for the cerebellum and 1000 and 8000 pixel units for the prefrontal cortex, determined by the limits of the histograms from all regions. The upper and lower limits of the glandular structures were set to 0–256. The 20 original Haralick features and the corresponding invariant features described in Table 2 were computed for each quantization. All texture analysis was done using MICE Toolkit [42] and MATLAB 2016b (MathWorks, Inc., Natick, MA).
Classification
The goal of each logistic regression model was to correctly predict the brain regions (cerebellum or prefrontal cortex) in Dataset 1 or the glandular structures (benign or malignant) in Dataset 2, from the texture values. Two scenarios were explored. In the first scenario, the datasets were split into training (50%) and test dataset (50%). One hundred classifiers were trained using a random quantization of each member in the training data. These models were used to predict the correct class in the test set for all quantization levels. In the second scenario, the datasets were split into training (50%), validation (25%) and test (25%) sets. One classifier was trained for each of the 32 quantizations. These classifiers were used to predict the correct class in the test set for all quantization levels. For each dataset and texture method, the validation set was used to perform variable selection to remove irrelevant or highly correlated features. We used forward selection [43], and maximized the mean accuracy.
Evaluation
The first scenario was evaluated by the average accuracy of the 100 classifiers that was trained on randomly quantized data. We compared the original and invariant Haralick features using a two-sample, two-tailed Welch’s t-test for populations with unequal variances. The second scenario was evaluated using a Mann-Whitney U-test to compare the results of the original and invariant Haralick features.
Results
Fig 4 shows the original GLCMs (left) and the invariant GLCMs (right) for the reference region of the cerebellum for 16, 32, 64, and 128 gray-levels. The standard GLCMs decrease in amplitude with orders of magnitude as the number of gray-levels increase. The invariant features retain the same volume, and the amplitude only increases slightly between 8 and 128 gray-levels, due to noise introduced with increased gray-level resolution.
Fig 5 shows how the texture feature values change with the number of gray-levels of the GLCM, computed from a benign glandular structure in the gland dataset (Dataset 2). In the left column, the upper graph shows features that increase rapidly with the number of gray-levels, the middle graph shows features that increase modestly, or reach a limit as the number of gray-levels increases, and the lower graph shows features that decrease with increasing number of gray-levels. The right column shows the corresponding invariant texture features. Note that the original features are plotted on a logarithmic scale on the vertical axis, whereas the invariant features are plotted on a linear scale.
Fig 6 shows the accuracy of 100 logistic regression models trained on randomly quantized images in the range of 8–256 gray-levels in steps of 8, and tested on images quantized to all gray-levels in the range 8–256 in steps of 8. The error bars show the standard deviation of the accuracy of the 100 classifiers, for each test dataset quantization level. Consistently, classifiers trained on the invariant features outperformed classifiers trained on the original features, and had a lower standard deviation of the accuracy. The average accuracies were 0.77 ± 0.08 and 0.96 ± 0.03 (p < 10−39, t = 20.82, using a two-sample Welch’s t-test with 138 degrees of freedom, estimated using the Satterthwaite-Welch formula [44]) for Dataset 1 (the brain dataset), and 0.69 ± 0.06 and 0.80 ± 0.03 (p < 10−35, t = 17.15, using a two-sample Welch’s t-test with 142 degrees of freedom [44]) for Dataset 2 (the gland dataset) for the original and invariant features, respectively.
Fig 7 shows the accuracy of the logistic regression models trained on one quantization levels, and tested on all quantization levels in the range 8–256 in steps of 8, for Dataset 1 (the brain dataset) and Dataset 2 (the gland dataset). The classifiers that were trained and tested on the original features had a high accuracy only when the test data quantization was close to the training data quantization. The classifiers that were trained on the invariant features had a high accuracy for all combinations of training and test dataset quantizations. The renormalized feature accuracies were significantly larger (p < 10−99 using a Mann-Whitney U-test) for both datasets.
Discussion
We have presented a simple modification of the Haralick texture features that makes them asymptotically invariant to the numbers of gray-levels in the quantizations. This is achieved by viewing the GLCM as a discrete approximation of a probability density function, instead of a probability mass function, over the pairs of gray-levels in the image. We have shown examples of how the standard and invariant GLCMs scale with increasing bit depth, Fig 4, and how most of the modified texture features quickly approach a limit, whereas most of the original features diverge or converge to zero, Fig 5. We have demonstrated the benefit of the proposed modified features by training logistic regression models to separate two brain regions in Dataset 1, and the malignancy of colorectal glands in Dataset 2, based only on texture features. Classifiers based on the invariant features performed better than the original features in all tests performed in this work. Further, the invariant Haralick features are rescaled versions of the original features, which means that they retain their original interpretations in most cases. Finally, the proposed modifications allow texture features to be reproducible regardless of quantization, since the same texture feature will give similar values independent of the quantization.
We tested the texture features in two scenarios. In the first scenario, we trained classifiers on texture features computed from images with different quantizations. In this scenario, the quantization levels might be chosen to enhance features optimally in each image, which is employed by e.g. Leijenenaar et al. [45]. Fig 6 shows that models trained on the invariant features have a higher accuracy and smaller standard deviation than models trained on the original features. In the second scenario, we trained the classifiers on one quantization level and predicted the classes from features computed on each of the other quantization levels. This scenario represents cases in e.g. radiomics, where a predictive model is created using one quantization, and employed to other datasets where the quantization is different due to e.g. resolution constrains, the size of the ROI, or where the optimal quantization is different. The classifiers trained on the invariant features had equal or higher accuracies compared to classifiers trained on the original features for the same combination of training and test dataset quantizations in most cases. The exceptions were some classifiers trained on the original features where the test quantization is slightly higher than the training quantization, see the upper left plot in Fig 7. In these situations, feature differences between the two classes are enhanced by the changes in feature values of the test data due to the different quantization gray-levels, as shown in Fig 5. This effect depends on the inherent texture of the two classes to be separated, and the features that are used in the model.
Maximum probability is very sensitive to noise, which is evident from Figs 4 and 5. For the invariant features, the maximum probability is now interpreted as the maximum probability density, i.e. the density value at the mode.
In the continuous case with the invariant features, the entropy becomes a differential entropy, and no longer has all the properties of the discrete counterpart. For instance, the differential entropy can be negative. This reinterpretation has particularly problematic consequences for the Information Measure of Correlation 1, (see Table 2). The denominator can be both positive and negative, and in particular, it can be close to zero. We note that in our particular case, with integration limits running from zero to one, division by zero can only occur if either of the marginal distributions of P are uniform.
It is important to note that except for maximum probability and the features that involve entropy, the interpretations of the texture features do not change. The invariant features are obtained by renormalizing the GLCM, the indices, and by multiplication with a differential determined by the bin size. It is merely a rescaling of the features to make them independent of the number of gray-levels of the GLCM. This has generally no impact on the interpretations of these features.
We chose to study the behavior of the invariant features between 8 and 256 quantization gray-levels. Too few gray levels will lead to a crude Riemann approximation in Eq 9, which can be seen in e.g. Fig 5 where the feature values stabilize around 16 gray-levels for most features, or in Figs 6 and 7 where the accuracies improve drastically between 8 and 16 gray-levels. Choosing excessively many quantizations gray-levels, e.g. on the order of the pixels present in the image, can produce sparse GLCMs. The invariant features will suffer from the same symptoms as the original features in this regard, i.e. an over-sensitivity to noise; and for small image regions which will produce sparse GLCMs, a failure to properly represent the underlying texture information.
Clausi et al. [38] proposed a normalized version of Homogeneity and Inverse difference by dividing the indices i and j by N to improve classification. The result is equivalent to the invariant versions presented in Table 2, since the effect of the differential Δij and the renormalized GLCM, cancel out in these features. However Clausi’s approach to only normalize the indices will not render all features invariant to the quantization; features such as Energy, Entropy measures, Variance measures, Information measure of correlation 1 and 2 and Maximum probability are not explicitly expressed in terms of the GLCM indices and cannot be normalized in this way. Shafic-ul-Hassan et al. [39] empirically derive normalization factors in terms of the quantization gray-levels for a selection of Haralick features. Of those, only one (Contrast) were identical to our results. A few of their normalized features were similar but not identical to the results presented here. Hence, most empirical derivations presented by Shafic-ul-Hassan do not fit the theory presented in this work. Further, many features presented in Table 2 cannot be reduced to a simple scaling factor of the original features, e.g. any feature containing entropy.
In this study we chose global min-max limits when quantizing the images in each dataset to minimize variations in the feature values due to image noise or other structures in the region of interest. Another approach is to set the limits for each ROI based on the minimum and maximum values inside each ROI [37, 45]. However, one extremely high or low pixel value inside the ROI will shift the center of mass of the GLCM, drastically affecting the Haralick feature values. The global limits approach requires that the image intensities are comparable between all images in the dataset, which requires that the images are acquired with the same hardware and imaging settings, or it can be achieved by e.g. normalizing the intensity of common structures in the images.
The invariant features can be used when analyzing multiple images of different sizes (different number of pixels), and possibly even different amounts of noise, by optimizing the quantization level for each image, while still obtaining comparable texture features. Image noise will affect the texture feature values [36, 37, 46], but the smoothing effect of gray-level quantization could reduce the impact on the resulting features. Some features are more sensitive to noise [37], and a more aggressive gray-level quantization for the invariant version of the features could reduce impact of the noise, while retaining the possibility to compare feature values with those from images analyzed at different quantization levels. This approach requires further research. Finally, there are other methods of extracting features, such as the Gray-level Size Zone Matrix (GLSZM), Gray-level Run Length Matrix (GLRLM) and the Neighborhood Gray-tone Difference Matrix (NGTDM), which feature values are also affected by the gray-level quantization. Shafic-ul-Hassan et al. [39] proposes empirical normalization factors for some features from these methods, and it is be feasible that a similar approach to what is presented in this work could reduce the impact of gray-level quantization to these methods as well. Making more methods independent of the gray-level quantization will increase the applicability and reproducibility of radiomics analyses, so this is also an interesting prospect for future research.
Conclusion
By reinterpreting the GLCM as a discretized probability density function, it is possible to construct a modified set of Haralick texture features that are asymptotically invariant to the image quantization. Except for maximum probability and features involving entropy, the invariant features retain their original interpretations. We show that the invariant features can be used in different classification setups, with results superior to the original definitions. This mean that the invariant Haralick texture features are reproducible even when different gray-level quantizations are used, unlike the original definitions.
Data Availability
Data set 1 can be found at the Swedish National Data Service: https://snd.gu.se/sv/catalogue/study/snd1062. Data set 2 is from the Gland Segmentation in Colon Histology Images Challenge Contest (GlaS), held at MICCAI 2015. The data set is open and freely available at https://warwick.ac.uk/fac/sci/dcs/research/tia/glascontest/download/. Both data sets belong to third parties, and the authors confirm they did not have any special access to the data that others would not have. The source code for the calculation of the invariant Haralick features is provided at https://github.com/patrik-brynolfsson/.
Funding Statement
The author(s) received no specific funding for this work.
References
- 1.Ojala, T., Pietikainen, M. & Harwood, D. Performance evaluation of texture measures with classification based on Kullback discrimination of distributions. Pattern Recognition, 1994. Vol. 1—Conference A: Computer Vision amp; Image Processing., Proceedings of the 12th IAPR International Conference on 1, 582–585 (1994).
- 2.Dyck, D. V. Wavelets for texture analysis, an overview. IET Conference Proceedings 581–585 (1997).
- 3. Clausi D. A. & Ed Jernigan M. Designing Gabor filters for optimal texture separability. Pattern Recognition 33, 1835–1849 (2000). [Google Scholar]
- 4. Haralick R. M., Shanmugam K. & Dinstein I. Textural Features for Image Classification. IEEE Transactions on Systems, Man, and Cybernetics 3, 610–621 (1973). 10.1109/TSMC.1973.4309314 [DOI] [Google Scholar]
- 5. Ou X., Pan W. & Xiao P. In vivo skin capacitive imaging analysis by using grey level co-occurrence matrix (GLCM). International Journal of Pharmaceutics 460, 28–32 (2014). 10.1016/j.ijpharm.2013.10.024 [DOI] [PubMed] [Google Scholar]
- 6. Ulaby F., Kouyate F., Brisco B. & Williams T. Textural Infornation in SAR Images. IEEE Transactions on Geoscience and Remote Sensing GE-24, 235–245 (1986). 10.1109/TGRS.1986.289643 [DOI] [Google Scholar]
- 7. Marcos J. V. et al. Automated pollen identification using microscopic imaging and texture analysis. Micron 68, 36–46 (2015). 10.1016/j.micron.2014.09.002 [DOI] [PubMed] [Google Scholar]
- 8. Raheja J. L., Kumar S. & Chaudhary A. Fabric defect detection based on GLCM and Gabor filter: A comparison. Optik—International Journal for Light and Electron Optics 124, 6469–6474 (2013). 10.1016/j.ijleo.2013.05.004 [DOI] [Google Scholar]
- 9. VijayaLakshmi B. & Mohan V. Kernel-based PSO and FRVM: An automatic plant leaf type detection using texture, shape, and color features. Computers and Electronics in Agriculture 125, 99–112 (2016). 10.1016/j.compag.2016.04.033 [DOI] [Google Scholar]
- 10. Bhat N. N., Dutta S., Pal S. K. & Pal S. Tool condition classification in turning process using hidden Markov model based on texture analysis of machined surface images. Measurement 90, 500–509 (2016). 10.1016/j.measurement.2016.05.022 [DOI] [Google Scholar]
- 11. Fernandez-Lozano C. et al. Texture analysis in gel electrophoresis images using an integrative kernel-based approach. Scientific Reports 6, 19256 (2016). 10.1038/srep19256 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12. Lerski R. et al. Computer analysis of ultrasonic signals in diffuse liver disease. Ultrasound in Medicine & Biology 5, 341–343 (1979). 10.1016/0301-5629(79)90004-8 [DOI] [PubMed] [Google Scholar]
- 13. Mayerhoefer M. E. et al. Texture-based classification of focal liver lesions on MRI at 3.0 Tesla: A feasibility study in cysts and hemangiomas. Journal of Magnetic Resonance Imaging 32, 352–359 (2010). 10.1002/jmri.22268 [DOI] [PubMed] [Google Scholar]
- 14. Skorton D. J., Collins S. M., Woskoff S. D., Bean J. a. & Melton H. E. Range- and azimuth-dependent variability of image texture in two- dimensional echocardiograms. Circulation 68, 834–840 (1983). 10.1161/01.CIR.68.4.834 [DOI] [PubMed] [Google Scholar]
- 15. Chan H. P. et al. Computer-aided classification of mammographic masses and normal tissue: linear discriminant analysis in texture feature space. Physics in medicine and biology 40, 857–876 (1995). 10.1088/0031-9155/40/5/010 [DOI] [PubMed] [Google Scholar]
- 16. Li H. et al. Computerized analysis of mammographic parenchymal patterns on a large clinical dataset of full-field digital mammograms: robustness study with two high-risk datasets. Journal of digital imaging 25, 591–8 (2012). 10.1007/s10278-012-9452-z [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17. Chen W., Giger M. L., Li H., Bick U. & Newstead G. M. Volumetric texture analysis of breast lesions on contrast-enhanced magnetic resonance images. Magnetic Resonance in Medicine 58, 562–571 (2007). 10.1002/mrm.21347 [DOI] [PubMed] [Google Scholar]
- 18. Nie K. et al. Quantitative analysis of lesion morphology and texture features for diagnostic prediction in breast MRI. Academic radiology 15, 1513–25 (2008). 10.1016/j.acra.2008.06.005 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19. Fjeldbo C. S. et al. Integrative analysis of DCE-MRI and gene expression profiles in construction of a gene classifier for assessment of hypoxia-related risk of chemoradiotherapy failure in cervical cancer. Clinical Cancer Research 22, 4067–4076 (2016). 10.1158/1078-0432.CCR-15-2322 [DOI] [PubMed] [Google Scholar]
- 20. Wibmer A. et al. Haralick texture analysis of prostate MRI: utility for differentiating non-cancerous prostate from prostate cancer and differentiating prostate cancers with different Gleason scores. European Radiology 25, 2840–2850 (2015). 10.1007/s00330-015-3701-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21. Vignati a. et al. Texture features on T2-weighted magnetic resonance imaging: new potential biomarkers for prostate cancer aggressiveness. Physics in medicine and biology 60, 2685–701 (2015). 10.1088/0031-9155/60/7/2685 [DOI] [PubMed] [Google Scholar]
- 22. Brynolfsson P. et al. ADC texture—An imaging biomarker for high-grade glioma? Medical Physics 41, 101903 (2014). 10.1118/1.4894812 [DOI] [PubMed] [Google Scholar]
- 23. Ryu Y. J. et al. Glioma: Application of whole-tumor texture analysis of diffusion-weighted imaging for the evaluation of tumor heterogeneity. PLoS ONE 9 (2014). 10.1371/journal.pone.0108335 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24. Assefa D. et al. Robust texture features for response monitoring of glioblastoma multiforme on T1-weighted and T2-FLAIR MR images: A preliminary investigation in terms of identification and segmentation. Medical Physics 37, 1722–1736 (2010). 10.1118/1.3357289 [DOI] [PubMed] [Google Scholar]
- 25. Lambin P. et al. Radiomics: extracting more information from medical images using advanced feature analysis. European journal of cancer (Oxford, England: 1990) 48, 441–6 (2012). 10.1016/j.ejca.2011.11.036 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26. Kumar V. et al. Radiomics: The process and the challenges. Magnetic Resonance Imaging 30, 1234–1248 (2012). 10.1016/j.mri.2012.06.010 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27. Aerts H. J. W. L. et al. Decoding tumour phenotype by noninvasive imaging using a quantitative radiomics approach. Nature Communications 5 (2014). 10.1038/ncomms5644 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28. Li Q. et al. A Fully-Automatic Multiparametric Radiomics Model: Towards Reproducible and Prognostic Imaging Signature for Prediction of Overall Survival in Glioblastoma Multiforme. Scientific Reports 7, 14331 (2017). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29. Lovinfosse P. et al. FDG PET/CT radiomics for predicting the outcome of locally advanced rectal cancer. European Journal of Nuclear Medicine and Molecular Imaging 1–11 (2017). [DOI] [PubMed] [Google Scholar]
- 30.Cho, H.-h. & Park, H. Classification of Low-grade and High-grade Glioma using Multi-modal Image Radiomics Features. 2017 39th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC) 3081–3084 (2017). [DOI] [PubMed]
- 31. Soh L.-K. & Tsatsoulis C. Texture Analysis of SAR Sea Ice Imagery Using Gray Level Co-Occurence Matrices. IEEE Transactions on Geoscience and Remote Sensing 37, 780–795 (1999). 10.1109/36.752194 [DOI] [Google Scholar]
- 32. Gómez W., Pereira W. C. A. & Infantosi A. F. C. Analysis of Co-Occurrence Texture Statistics as a Function of Gray-Level Quantization for Classifying Breast Ultrasound. IEEE Transactions on Medical Imaging 31, 1889–1899 (2012). 10.1109/TMI.2012.2206398 [DOI] [PubMed] [Google Scholar]
- 33. Torheim T. et al. Classification of Dynamic Contrast Enhanced MR Images of Cervical Cancers Using Texture Analysis and Support Vector Machines. Ieee Transactions on Medical Imaging 33, 1648–1656 (2014). 10.1109/TMI.2014.2321024 [DOI] [PubMed] [Google Scholar]
- 34. Ahmed A., Gibbs P., Pickles M. & Turnbull L. Texture analysis in assessment and prediction of chemotherapy response in breast cancer. Journal of magnetic resonance imaging: JMRI 38, 89–101 (2013). 10.1002/jmri.23971 [DOI] [PubMed] [Google Scholar]
- 35. Vallières M., Freeman C. R., Skamene S. R. & El Naqa I. A radiomics model from joint FDG-PET and MRI texture features for the prediction of lung metastases in soft-tissue sarcomas of the extremities. Physics in Medicine and Biology 60, 5471–5496 (2015). 10.1088/0031-9155/60/14/5471 [DOI] [PubMed] [Google Scholar]
- 36. Schad L. R. Problems in texture analysis with magnetic resonance imaging. Dialogues in clinical neuroscience 6, 235–42 (2004). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37. Brynolfsson P. et al. Haralick texture features from apparent diffusion coefficient (ADC) MRI images depend on imaging and pre-processing parameters. Scientific Reports 7, 4041 (2017). 10.1038/s41598-017-04151-4 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 38. Clausi D. A. An analysis of co-occurrence texture statistics as a function of grey level quantization. Canadian Journal of Remote Sensing 28, 45–62 (2002). 10.5589/m02-004 [DOI] [Google Scholar]
- 39. Shafiq-ul-Hassan Muhammad et al. Intrinsic dependencies of CT radiomic features on voxel size and number of gray levels. Medical Physics 44, 1050–1062 (2017). 10.1002/mp.12123 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40. Sirinukunwattana K., Snead D. R. J. & Rajpoot N. M. A Stochastic Polygons Model for Glandular Structures in Colon Histology Images. IEEE Transactions on Medical Imaging 34, 2366–2378 (2015). 10.1109/TMI.2015.2433900 [DOI] [PubMed] [Google Scholar]
- 41. Sirinukunwattana K. et al. Gland segmentation in colon histology images: The GlaS challenge contest. Medical Image Analysis 35, 489–502 (2017). 10.1016/j.media.2016.08.008 [DOI] [PubMed] [Google Scholar]
- 42. Nyholm T., Berglund M., Brynolfsson P. & Jonsson J. EP-1533: ICE-Studio—An Interactive visual research tool for image analysis. Radiotherapy and Oncology 115, S837 (2015). 10.1016/S0167-8140(15)41525-7 [DOI] [Google Scholar]
- 43. Hastie T., Tibshirani R. & Friedman J. The Elements of Statistical Learning Springer Series in Statistics (Springer New York Inc., New York, NY, USA, 2001). [Google Scholar]
- 44. Welch B. L. The generalization of ‘Student’s’ problem when several different population variances are involved. Biometrika 34, 28–35 (1947). 10.2307/2332510 [DOI] [PubMed] [Google Scholar]
- 45. Leijenaar R. T. et al. The effect of SUV discretization in quantitative FDG-PET Radiomics: the need for standardized methodology in tumor texture analysis. Scientific Reports 5, 11075 (2015). 10.1038/srep11075 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 46. Mayerhoefer M. E., Szomolanyi P., Jirak D., Materka A. & Trattnig S. Effects of MRI acquisition parameter variations and protocol heterogeneity on the results of texture analysis and pattern discrimination: an application-oriented study. Medical physics 36, 1236–1243 (2009). 10.1118/1.3081408 [DOI] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
Data set 1 can be found at the Swedish National Data Service: https://snd.gu.se/sv/catalogue/study/snd1062. Data set 2 is from the Gland Segmentation in Colon Histology Images Challenge Contest (GlaS), held at MICCAI 2015. The data set is open and freely available at https://warwick.ac.uk/fac/sci/dcs/research/tia/glascontest/download/. Both data sets belong to third parties, and the authors confirm they did not have any special access to the data that others would not have. The source code for the calculation of the invariant Haralick features is provided at https://github.com/patrik-brynolfsson/.