Skip to main content
Sensors (Basel, Switzerland) logoLink to Sensors (Basel, Switzerland)
. 2020 Oct 27;20(21):6110. doi: 10.3390/s20216110

Thyroid Nodule Classification for Physician Decision Support Using Machine Learning-Evaluated Geometric and Morphological Features

Elmer Jeto Gomes Ataide 1,2,*, Nikhila Ponugoti 2, Alfredo Illanes 2, Simone Schenke 1, Michael Kreissl 1, Michael Friebe 2,3
PMCID: PMC7663034  PMID: 33121054

Abstract

The classification of thyroid nodules using ultrasound (US) imaging is done using the Thyroid Imaging Reporting and Data System (TIRADS) guidelines that classify nodules based on visual and textural characteristics. These are composition, shape, size, echogenicity, calcifications, margins, and vascularity. This work aims to reduce subjectivity in the current diagnostic process by using geometric and morphological (G-M) features that represent the visual characteristics of thyroid nodules to provide physicians with decision support. A total of 27 G-M features were extracted from images obtained from an open-access US thyroid nodule image database. 11 significant features in accordance with TIRADS were selected from this global feature set. Each feature was labeled (0 = benign and 1 = malignant) and the performance of the selected features was evaluated using machine learning (ML). G-M features together with ML resulted in the classification of thyroid nodules with a high accuracy, sensitivity and specificity. The results obtained here were compared against state-of the-art methods and perform significantly well in comparison. Furthermore, this method can act as a computer aided diagnostic (CAD) system for physicians by providing them with a validation of the TIRADS visual characteristics used for the classification of thyroid nodules in US images.

Keywords: thyroid nodules, ultrasound imaging, TIRADS, feature extraction, machine learning, classification, computer aided diagnosis

1. Introduction

The thyroid is one of the largest endocrine glands located below the epiglottis. It is responsible for several physiological functions such as the production of hormones, regulation of brain and nerve cells, and development and functioning of organs like the heart, eyes, hair, skin, and intestines [1]. Irregularities and/or deformations of the thyroid lead to its inability to efficiently carry out these functions. Nodules within the thyroid are mostly benign neoplasms. Ultrasound (US) imaging is typically used as the first point of diagnosis and also for the evaluation of the thyroid nodules, as it effectively images and visualizes soft tissue structures. Additionally, it is free of ionizing radiation and is the most widely available and affordable imaging modality [2].

The assessment and evaluation of thyroid nodules using US imaging is done by the physician based on the visual characteristics observed in the scan. For this purpose, the Thyroid Imaging Reporting and Data System (TIRADS) approach is used for the risk stratification and classification of benign vs. malignant nodules. Multiple versions of TIRADS exist such as that of the American Council of Radiology (ACR) TIRADS [3], Kwak-TIRADS [4], etc. Each of these versions differs, but all consider visual and textural features, such as margin, shape, calcification, composition, size, echogenicity and offer a scoring scheme that enables the physician to assess nodule malignancy.

The use of TIRADS has helped to standardize the evaluation of thyroid nodules found in US images. However, there still exists a considerable amount of inter-observer variability and overall subjectivity. To address the issue of subjective diagnoses several computer aided diagnostic (CAD) methods were proposed. These CAD methods use various feature extraction and classification algorithms to characterize thyroid nodules using US images into benign and malignant.

The aim of our work was to develop a more objective diagnostic approach for thyroid nodules using US images. The related studies significant to our work will be discussed in the following sub-section.

Related Work

Several methods have been proposed for feature extraction and computer-aided classification of thyroid nodules using US images. Apart from texture-based features [5], there are other types of feature extraction methods that can be used, like general shape-based feature extraction/classification using different image modalities, as well as shape-based feature extraction for the classification of thyroid nodules in US images. For that we also considered studies that have used the same database that we accessed.

Jianhua Liu et al. introduced the use of shape features of an image. The boundary and region-based feature extraction methods are explained in [6]. The use of shape features was observed in several applications for medical image analysis. Riti et al. employed shape features for the classification of lung cancer from computed tomography (CT) images and obtained an overall classification accuracy of 85% [7]. The same was seen in the work of Ferreira Junior et.al. where the margin sharpness was used for the classification of lung nodules in CT images [8]. Hiremath et al. suggested a shape feature-based approach for detecting follicles in ovarian ultrasound images. The follicles were classified depending on medical knowledge on parametrically defined measures, such as area, compactness, centroid, etc. [9]. Huang et al. in their research suggested the use of functional morphological characteristics to differentiate between benign and malignant breast tumors efficiently. Nineteen morphological features were extracted from ultrasound images and used for the classification of tumors. They obtained a classification accuracy of 82% and a sensitivity of 94% using a support vector machine (SVM) classifier [10]. Nugroho et al. also made use of shape-based feature analysis and extraction for the classification of breast nodules with a specific focus on the marginal characteristics of uncircumscribed versus circumscribed margins [11].

In the case of thyroid nodules, Gopinathan et al. performed thyroid nodule risk stratification and classification by analyzing the roundness and irregularity of nodule margins while using US and fine-needle biopsy [12]. Zulfanahri et al. suggested a system that can classify thyroid nodules in two groups, i.e., round to oval and irregular shapes using three chosen characteristics. The suggested system achieved an accuracy of 91.52% and specificity of 91.35% [13]. Similarly, Lina Choridah et al. proposed a technique to classify thyroid nodules based on marginal features. The suggested strategy effectively classified the thyroid nodules into two smooth and uneven classes using US images and obtained an accuracy and sensitivity of 92.30% and 91.88% respectively [14]. The image pattern classification technique used by Junying Chen et al. to categorize benign and malignant thyroid nodules proved efficient in the classification process to verify the types of pattern properties that could be used for the classification of thyroid nodules in US images [15].

Ding et al. defined statistical characteristics and texture based on elastography of the thyroid lesion area. The selection of features was then done using the algorithm called minimum redundancy-maximum-relevance. The selected features were then plugged into an SVM classifier [16].

Isa et al. used different multi-layer perceptron (MLP) models for the detection of thyroid diseases [17]. Statistical features coupled with demographic details of a sample were fed into three different classification algorithms such as random forest classifier (RFC), SVM, and logistic regression by Patricio et al. to differentiate between thyroid nodules [18]. Song et al. in their work with thyroid nodule US image classification proposed a hybrid multi-branch convolutional neural network (MBCNN) based on a feature cropping method for feature extraction. This work used an open-access dataset [19] as well as a local dataset and obtained a classification accuracy of 96.13% [20]. Chi et al. by using a fine-tuned deep convolutional neural network (FDCNN) approach obtained a classification accuracy of 98.29% [21]. Using the same dataset, Koundal et al. proposed a complete image-driven thyroid nodule detection approach that was able to detect thyroid nodules with an accuracy of up to 93.88% [22]. Nanda and Martha in their work with cancerous thyroid nodule detection employed local binary pattern variants (LBPV)-based feature extraction to classify between benign and malignant thyroid nodules with an accuracy of 94.5% [23] using [19].

In the presented literature it can be seen that there are several methods for the feature extraction and classification of thyroid nodules. Studies [6,7,8,9,10,11] give a broad understanding of the use of shape-based features in medical image analyses. Using various combinations of these shape-based features, studies [12,13,14,15] were able to classify thyroid nodules in US images with significant outcomes. Deep learning-based methods [20,21] result in an automatic feature extraction and classification method that differs from handcrafted features. The methods seen in the literature perform well, but most of them do not take into account the geometrical and morphological attributes of thyroid nodules following TIRADS. Since physicians exhibit significant levels of trust in TIRADS, it is essential to consider the features put forth by [3,4]. These are the attributes that are visible to the physicians on which their decision is partly based. Even the studies [12,13,14,15] that do take into account shape features do so only to a limited extent.

The purpose of this study is to use geometric and morphological feature extraction that takes into consideration features that are closely related to the visual shape-based (TIRADS) features currently used by physicians. This provides them with additional information and mathematical evidence to support their current TIRADS-based classification with an extra layer of objectivity.

In this work, we mainly focus on shape-based geometric and morphological feature extraction for the classification of thyroid nodules as either benign or malignant. Examining physicians use visual and textural characteristics to classify a nodule. Geometric and morphological features represent the visual aspect. The performance of the features extracted was evaluated using a RFC. The results obtained from the RFC were compared against other feature extraction and classification techniques found in the state of the art that use the same database. Furthermore, our approach was also compared to other shape-based feature extraction and classification methods found in the literature. The rest of the paper is structured as follows: Section 2 details the materials and methods used in this work. Section 3 presents the results and comparisons drawn with other feature extraction and classification methods, and the final Section 4 then provides a conclusion and discusses the next steps.

2. Materials and Methods

This section is divided into three parts: (1) database and data augmentation, (2) feature extraction and selection, and (3) classification. The section on the database and data augmentation gives details of the dataset used and the various data augmentation techniques used to balance and maximize the data usage. The second section presents details on the geometrical and morphological features extracted from the dataset and discusses the feature selection methods tested and employed for the selection of the most optimal features. The classification section highlights the classification method used to evaluate the selected features with subsequent classification into benign and malignant. The workflow diagram consisting of each step is presented in Figure 1.

Figure 1.

Figure 1

Workflow diagram.

2.1. Dataset and Data Augmentation

For this study, we used the Digital Database of Thyroid Ultrasound Images (DDTI) open-access dataset of thyroid nodule ultrasound images from the Instituto de Diagnostico Medico (IDIME) [19]. The dataset consists of a total of 99 cases with 2D ultrasound images from different patients that are annotated and classified based on TIRADS classification. The dataset is divided into JPEG files and XML files. Each image has a resolution of 560 × 315. Each image has a corresponding XML file. The XML files provided a detailed classification for each of the nodules. The ground truth (GT) for each of the nodules was generated by experienced physicians and is available in the form of coordinates in the XML file. An example of benign and malignant nodules along with their GTs can be seen in Figure 2. In this study, we only considered two labels (benign and malignant) rather than all the TIRADS labels.

Figure 2.

Figure 2

Examples of ultrasound images of thyroid nodules (a1) malignant nodule, (a2) its ground truth, (b1) benign nodule and (b2) its ground truth [19].

These cases were divided into 17 benign cases and 82 malignant cases. To correct the data imbalance, data augmentation techniques such as flipping, rotation and blurring were employed. Data augmentation was first used to balance the data and then again to further augment it. Finally, a total of 3188 images were obtained (1594 Malignant + 1594 Benign) from the original 134 images and used in the feature extraction phase of the study. The data augmentation was performed using the Augmentor python package. The percentages of each of the operations (flipping, rotation and blurring) cannot be estimated as they are applied stochastically.

2.2. Feature Extraction

In this work, two types of feature were extracted from the GTs of the US thyroid images. These were in the form of geometric and morphological (G-M) features. A summary of all the extracted features can be seen in Table 1. Figure 3 gives a visual depiction of a few geometric and morphological features.

Table 1.

Overview of 27 extracted geometric and morphological (G-M) features.

Sr. No. Features Type
1 Convex Hull Geometric Features
2 Convexity
3 Solidity
4 Elongation
5 Compactness
6 Rectangularity
7 Orientation
8 Roundness
9 Major Axis Length
10 Minor Axis Length
11 Eccentricity
12 Circular Variance
13 Elliptic Variance
14 Ratio of Major Axis Length to Minor Axis Length
15 Bounding Box
16 Centroid
17 Convex Area
18 Filled Area
19 Convex Perimeter
20 Area Morphological Features
21 Perimeter
22 Aspect Ratio
23 Area Perimeter (AP)Ratio
24 Object Perimeter to Ellipse Perimeter (TEP) Ratio
25 TEP Difference
26 Object Perimeter to Circular Perimeter (TCP) Ratio
27 TCP Difference

Figure 3.

Figure 3

Visual depiction of geometric and morphological features. (a) Convexity, (b) elongation, (c) major and minor axes, (d) bounding box, (e1,e2) different instances of eccentricity, (f) orientation, (g) filled area, (h) convex area, (i) circular variance and (j) elliptical variance.

2.2.1. Geometric Features

Geometric features are those features that are used to construct an object with certain geometric elements such as points, curves, and lines as well as information related to edges that describe the shape or irregularity of a given boundary [24,25]. Note that all instances of the mention of the word “object” refer to thyroid nodules. The geometric features extracted from the images are as follows.

Convex Hull:

The convex hull of an object is the smallest convex structure within which an object is positioned. It is the smallest convex polygon that can contain the object. [26,27]:

Convexity:

Convexity is a function that measures the ratio of the convex hull with respect to the original contour of the shape. In this case, a convex hull is drawn around the original contour of the object. Convexity is calculated using the following equation [28]:

Convexity=PcPn (1)

where Pc = perimeter of the convex hull and Pn = perimeter of the object

Solidity:

Solidity helps describe the extent of a shape’s convexity or concavity. It is given by Equation [29]:

Solidity=AnAc (2)

where An = area of the object and Ac = area of the convex hull. A value of a solid object is 1 and an object with an irregular boundary is defined with a value less than 1.

Elongation:

Elongation is the feature that measures the ratio between the length and width of a bounding box around an object. The result is a nodule’s elongation measurement given as a value of 0 to 1. The object is approximately square or round shaped when the ratio is equal to 1. When the ratio is lower than 1, the object is more elongated. The equation can be seen below [26,30]:

Elongation=WnLn (3)

where Wn = width of the object and Ln = length of the object

Compactness:

Compactness is the ratio between the areas of an object with respect to the area of a circle with a perimeter equal to that of the object. It is given by the equation below [26,30]:

Compactness=4πAnPn2 (4)

where Pn = perimeter of the object and An = Area of the object

Rectangularity:

Rectangularity is defined as the ratio between the object area and the area of the minimum-bounding rectangle [31]. When an object returns a rectangularity of 1 it is said to be a rectangular object.

Rectangularity=AnAr (5)

where An = area of the object and Ar = Area of the rectangle.

Roundness:

The roundness of an object is defined as the ratio between the area of the object and the area of a circle with the same convex perimeter. It can be represented by the following equation [26,32]:

Roundness=4πAnPc2 (6)

where Pc = the convex perimeter of the object and An = the area of the object

Major Axis length:

The major axis of an object is the endpoints (X, Y) of the longest line traced through the object. The endpoints of the major axis (X1, Y1, and X2, Y2) are determined by calculating the pixel distance between every combination of border pixels in the object boundary. This is used to find the pair with the maximum length. The object’s major axis length is the pixel distance between the major axis endpoints and is defined by the equation [26,32]:

Major-axis length=(X2X1)2+(Y2Y1)2 (7)

Minor Axis length:

The minor axis is the (x, y) endpoints of the longest line drawn by the object while still perpendicular to the major axis. The endpoints (x1, y1, and x2, y2) of the minor axis are calculated by computing the pixel distance between the two border pixel endpoints. The minor-axis length of an object is the pixel distance between the minor axis endpoints and is defined by the equation [26,32]:

Minor-axis length=(x2x1)2+(y2y1)2 (8)

Eccentricity:

Eccentricity is defined as the ratio between minor axis length and the major axis length of the object [26]. The result is a measure of the eccentricity of the object, given as a value from 0 to 1.

Eccentricity=LmaLMA (9)

where Lma = length of minor axis and LMA = Length of major axis

Circular Variance:

Circular variance (Cva) is the comparison of an object’s shape with respect to a known shape such as a circle. The circular variance is the object’s proportional mean-squared error with respect to the solid circle [26,32]. This returns a null value for a perfect circle and increases and the shape and complexity change. It is given by the following equation.

Cva=σRμR (10)

where µR and σR are the mean and standard deviation of the radial distance from the centroid (cx, cy) of the shape to the boundary points (xi, yi), i ϵ [0, N − 1].

They are represented by the formulae:

μR=1Ni=1N1di (11)

and

σR=1Ni=1N1(diμR)2 (12)

where di = (xicx)2+(yicy)2

Elliptic Variance:

Elliptic variance (Eva) is measured similarly to the circular variance. An oval is fitted to the shape (rather than a circle), and the mean squared error is estimated [32].

Eva=σRμR (13)
μR=1Ni=1N1di (14)

and

σR=1Ni=1N1(diμR)2 (15)

where

  • di=ViT×Cellipse1×Vi

  • Cellipse1 = Inverse of covariance matrix of the shape (ellipse)

  • Vi=(xicxyicy) and ViT=Transpose of Vi

Ratio of Major Axis Length to Minor Axis Length:

Thus is the ratio between major axis lengths to minor axis length.

Ratio of length of Major axis and minor axis = LMALma (16)

where LMA = length of major axis and Lma = length of minor axis.

Orientation:

The orientation is angle between the x-axis and the major axis of the object. It can also be defined as the direction of the shape [26].

Bounding Box:

The bounding box is the region’s smallest rectangle that envelops the object [25]. Dimensions for the bounding box are those equal to the major and the minor axes.

Area of bounding box = LMA×Lma (17)

where LMA = length of major axis and Lma = length of minor axis.

Centroid:

The centroid is defined as the center of gravity of the object [25].

Convex Area:

The convex area of a nodule is the area surrounded by the convex hull [31].

Filled Area:

Is the total number of pixels within the marked object mask. [26]

Convex Perimeter:

The convex perimeter of an object is the perimeter of the convex hull that encloses the object [25].

2.2.2. Morphological Features

Morphological features are those features that define an object’s structuring elements such as area, perimeter, aspect ratio, etc. [33]. The following morphological features were considered for the classification of thyroid nodules in this work.

Area:

Area is the space occupied by objects on a plane surface. Here area is defined as the number of pixels inside the object region [34].

Perimeter:

The number of pixels within the object border is the perimeter [26,34]. If x1… xN is a list of boundaries, and the perimeter is defined by:

Perimeter=i=1N1di=i=1N1|XiXi+1| (18)

Aspect Ratio:

The aspect ratio is defined as the ratio between the tumor’s depth and width [26]:

Aspect Ratio=DnWn (19)

where Dn = Depth of the object and Wn = Width of the object.

AP Ratio (area to perimeter (AP) ratio):

The AP ratio is the ratio between object area and perimeter of the object, and it is defined as:

AP Ratio=AnPn (20)

where An = Area of the object and Pn = Perimeter of the object

TEP Ratio (object perimeter to ellipse perimeter ratio):

The TEP ratio is the ratio of perimeters of an object to the related ellipse [17].

TEP Ratio=PnPe (21)

where Pn = perimeter of the object and Pe = perimeter of the ellipse

TEP Difference:

TEP is determined by the difference between the object perimeter and the related ellipse [17]:

TCP Ratio (object perimeter to circle perimeter ratio):

The TCP ratio is the ratio of the perimeter of the object to the relevant circle [17]:

TCP Ratio=PnPc (22)

where Pn = perimeter of the object and Pc = perimeter of the circle.

TCP Difference:

The TCP difference is known as the difference between the perimeter of the object and the corresponding circle [17].

2.2.3. Feature Selection

Feature selection was based on the most relevant visual characteristics in accordance with TIRADS. This was determined with the help of the two physicians in the Department of Nuclear Medicine at the University Hospital in Magdeburg, Germany. Clinicians use TIRADS classification that classifies a nodule based on its geometrical attributes such as shape, size, irregularity in margins and orientation. The 11 selected features provide the closest estimation to these attributes used by TIRADS. In order to further validate this claim the authors tested the performance of all the geometric and morphological features (Global) as well as those that were not selected (discounted). The performance metrics for each of these are seen in Table 4 in Section 3.

Table 4.

Performance metrics of selected features versus global and discounted features.

Method Accuracy (%) Sensitivity (%) Specificity (%)
Global 70.18 48.07 92.29
Discounted 61.55 31.65 91.45
G-M 99.33 99.39 99.25

Labels were added to each row of features depending on whether they belonged to 0 = benign or 1 = malignant. An overview of all the extracted features can be seen in Table 1. The feature extraction process was carried out using MATLAB 2018b. The feature sets were then exported as .csv files and used for the classification process. The performance of the features was evaluated using a RFC.

2.3. Classification

An RFC was trained for a binary classification problem where each of the feature rows was labeled as 0 or 1. RFC is a type of ensemble learning that builds a final classifier by using weak individual classifiers i.e., binary decision trees. Each tree is a collection of nodes and features that lead to the final classification result. The aggregation of the results from each of the individual trees is considered the final result of the classifier. The features were used as the independent variables and the labels as the dependent variables. The parameters chosen for the RFC are shown in Table 2. A train-test split of 70–30% was used on the data. The classification using RFC was undertaken in Python 3.7 using the scikit-learn library.

Table 2.

Selected random forest classifier (RFC) parameters.

Parameter Value
Number of Decision Trees 400
Criterion Entropy
Bootstrap True

3. Results and Discussion

A total of 27 features were extracted from the thyroid nodule dataset. The feature selection step led to the selection of the 11 most significant features. We selected three prime metrics to compare the G-M features against the performance of global and discounted features in our study for the feature selection as well as the methods found in the literature. The metrics selected were accuracy, sensitivity, and specificity.

The 11 significant features are given in Table 3. These features were selected based on clinical input and expertise. Further experiments confirmed the validity of the features selected. This is highlighted in Table 4.

Table 3.

Eleven most significant features selected from the global feature list of 27.

Sr. No. Features Type
1 Solidity Geometric Features
2 Orientation
3 Roundness
4 Major Axis Length
5 Minor Axis Length
6 Bounding Box
7 Convex Area
8 Area Morphological Features
9 Perimeter
10 Aspect Ratio
11 AP Ratio

As can be seen the accuracy scores of global features and discounted features are considerably lower (70.18% and 61.55% respectively) than compared to G-M. Additionally, both global and discounted features exhibit high specificities, but the sensitivities were low (48.07% and 31.65% respectively). As this is a case of cancer classification, more focus was given to the sensitivity score due to its clinical relevance (true positive rates). The selected 11 final features resulted in high accuracy, specificity and sensitivity scores. i.e., the classifier was able to identify benignity and malignancy of the nodules much more effectively.

The selected 11 features were fed into the RFC and the results obtained were compared against the results from [20,21,23] that used deep learning-based and LBPV-based methods. These studies used feature extraction and classification techniques that were different from our proposed approach but were all tested on the same open-access dataset [19]. Additionally, the results were further compared against other shape-based feature extraction techniques [13,14] found in the literature.

The same performance metrics were used to present the results obtained from the classifiers. This is presented in Table 5 and Table 6 below. The comparison is done in two steps. The first step involves the comparison between the proposed method and the methods found in the state of the art that used the same dataset as us. The second step draws a comparison between the proposed method and other thyroid nodule feature extraction and classification approaches using shape-based features found in other related studies.

Table 5.

Feature evaluation using RFC compared to the performance of the related approaches using the same dataset.

Method Accuracy (%) Sensitivity (%) Specificity (%)
MBCNN [20] 96.13 97.18 -
FDCNN [21] 98.29 99.10 93.90
LBPV (SVM) [23] 94.5 97.25 94.50
G-M (RFC) 99.33 99.39 99.25

Table 6.

Feature evaluation using RFC compared to the performance of shape-based features found in related studies using different datasets.

Method Accuracy (%) Sensitivity (%) Specificity (%)
Margin Features [13] 91.52 91.80 91.35
Margin Features [14] 92.30 91.88 92.73
G-M 99.33 99.39 99.25

Table 5 shows the classification results obtained from our proposed feature extraction approach, G-M and three different thyroid nodule feature extraction and classification approaches found in the state of the art. i.e., MBCNN [20], FDCNN [21] and LBPV [23]. Each of these methods uses a different feature extraction and classification method from the proposed approach. However, the dataset used in all four cases is the same [19]. Data augmentation was used by [20,21]. It can be seen from the table that G-M based features display a higher accuracy of 99.33% as compared to [20,21,23] while using the same database. G-M features also result in a high specificity score of 99.25%. In the case of sensitivity, G-M features exhibit a score of 99.39%. From the depicted results it can be inferred that the proposed feature extraction method can classify benign and malignant thyroid nodules with high accuracy. Additionally, this approach also displays high true positive (Sensitivity) and true negative (Specificity) rates. This means that the feature extracted can be distinguished well.

Table 6 shows the classification results obtained from the proposed G-M feature extraction method compared to two approaches found in the literature that use shape-based features for the classification of thyroid nodules. Each of the two studies [13,14] compared to G-M uses features that help characterize the nodules based on the extent of the margins being oval. Both the compared methods use different datasets in their work. It can be seen from the table that the G-M features when fed into the RFC classifier exhibit an accuracy, sensitivity, and specificity of 99.33%, 99.39%, and 99.25%, respectively. This is a considerable improvement over the same metrics seen in [13,14].

In both comparison cases, across different feature extraction methods using the same dataset as well as similar feature extraction methods using different input data, the G-M approach outperforms the state-of-the-art. Even though [13,14,20,21,23] show high accuracies, sensitivities, and specificities, the features extracted are often not relevant for the physician performing the examination. To a certain extent, [13,14] take into account some shape features. But, [20,21,23] use features that are not in accordance with TIRADS. While reviewing US images of thyroid nodules physicians take into consideration the visual and textural characteristics of a nodule to classify it as benign or malignant. The geometric and morphological features encompass the visual characteristics of a nodule. These are the closest estimation to the features defined by the gold standard in TIRADS. This can be observed across all three calculated metrics. It is evident that the G-M features adequately emulate the visual characteristics that are defined in TIRADS. These visual aspects of a thyroid nodule such as margins irregularity and shape can be directly attributed to the G-M features extracted in this work.

4. Conclusions and Future Work

This study focused on the geometric and morphological feature extraction techniques for the classification of thyroid nodules from US images. This work used a total of 3188 images. A total of 27 geometric and morphological features were defined and extracted from these images and then the 11 most significant features selected in accordance with the TIRADS-based visual features and labeled based on their class (0 = benign, 1 = malignant). The performance of the selected features was then evaluated by the classification accuracy, sensitivity, and specificity obtained from the RFC.

The consideration of G-M features for a computer-aided diagnostic approach for thyroid nodules is clinically relevant. The 11 selected features in this study proved to be the best combination from the overall feature set of 27 because these selected features provide information such as the shape, irregularity in the boundary, orientation, and size of the US thyroid nodule. However, it must be noted that G-M features are only one part of the features that help in the classification. According to TIRADS, physicians also need to consider texture features found in a nodule, which were not considered for this work. Hence, a notable observation of the proposed approach is that it only considers one classification aspect. When G-M features are combined with texture features the overall accuracy might change, which needs to be studied further. Another limitation of this work is that everything was carried out on a single open-source database and further validation of the approach needs to be carried using additional datasets including different ultrasound scanners. We are currently working on acquiring data at the Department of Nuclear Medicine at the University Hospital in Magdeburg, Germany. This would help to improve the robustness of the features extracted across different datasets.

In future work, we would like to omit the use of data augmentation techniques all together. However, this is dependent on the amount of data that is being currently collected as mentioned. Until large volumes of US thyroid nodule images are available, there are two possible strategies that can be employed additionally to test the relevance of the selected features. The first is to augment the data while being aware of the percentage of each operation (flipping, rotation and blurring) performed. This would help us determine and understand how each of the selected features behaves with respect to each data augmentation operation. Additionally, it would also help us determine the extent to which the data augmentation effects the final classification. The second method would be to test the features extracted from non-augmented images against features extracted from augmented ones and to show the deviation between these two sets.

The addition of texture-based features to the G-M features would provide a larger feature set that would then consider both aspects of a TIRADS classification and provide physicians with better decision support for the classification of thyroid nodules. Hence, a step towards reducing inter-observer variability and overall diagnostic subjectivity.

Author Contributions

Conceptualization, E.J.G.A., A.I., S.S., M.K. and M.F.; methodology, E.J.G.A., A.I. and S.S.; software, E.J.G.A. and N.P.; validation, E.J.G.A., A.I. and S.S.; formal analysis, E.J.G.A., N.P. and A.I.; investigation, E.J.G.A. and S.S.; writing—original draft preparation, E.J.G.A.; writing—review and editing, A.I., S.S., M.K. and M.F.; supervision, M.K. and M.F.; project administration, E.J.G.A.; All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflict of interest.

Footnotes

Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

References

  • 1.InformedHealth.org How Does the Thyroid Gland Work? [(accessed on 26 June 2019)]; Available online: https://www.ncbi.nlm.nih.gov/books/NBK279388/
  • 2.Thomas L.S. Diagnostic Ultrasound Imaging: Inside Out. Elsevier Academic Series; New York, NY, USA: 2004. Chapter 1. [Google Scholar]
  • 3.Tessler F.N., Middleton W.D., Grant E.G., Hoang J.K., Berland L.L., Teefey S.A., Cronan J.J., Beland M.D., Desser T.S., Frates M.C., et al. ACR Thyroid Imaging, Reporting and Data System (TI-RADS): White Paper of the ACR TI-RADS Committee. J. Am. Coll. Radiol. 2017;14:587–595. doi: 10.1016/j.jacr.2017.01.046. [DOI] [PubMed] [Google Scholar]
  • 4.Kwak J.Y., Han K.H., Yoon J.H., Moon H.J., Son E.J., Park S.H., Jung H.K., Choi J.S., Kim B.M., Kim E.-K. Thyroid Imaging Reporting and Data System for US Features of Nodules: A Step in Establishing Better Stratification of Cancer Risk. Radiology. 2011;260:892–899. doi: 10.1148/radiol.11110206. [DOI] [PubMed] [Google Scholar]
  • 5.Sollini M., Cozzi L., Chiti A., Kirienko M. Texture analysis and machine learning to characterize suspected thyroid nodules and differentiated thyroid cancer: Where do we stand? Eur. J. Radiol. 2018;99:1–8. doi: 10.1016/j.ejrad.2017.12.004. [DOI] [PubMed] [Google Scholar]
  • 6.Liu1 J., Shi Y. Image Feature Extraction Method Based on Shape Characteristics and Its Application in Medical Image Analysis. Department of Software Institute, Department of Information Engineering, North China University of Water Resources and Electric Power; Zhengzhou, Henan Province, China: 2011. [Google Scholar]
  • 7.Riti Y.F., Nugroho H.A., Wibirama S., Windarta B., Choridah L. Feature extraction for lesion margin characteristic classification from CT Scan lungs image; Proceedings of the 2016 1st International Conference on Information Technology, Information Systems and Electrical Engineering (ICITISEE); Yogyakarta, Indonesia. 23–24 August 2016; pp. 54–58. [Google Scholar]
  • 8.Junior J.R.F., Oliveira M.C. Evaluating Margin Sharpness Analysis on Similar Pulmonary Nodule Retrieval; Proceedings of the 2015 IEEE 28th International Symposium on Computer-Based Medical Systems; Sao Paulo, Brazil. 22–25 June 2015; pp. 60–65. [Google Scholar]
  • 9.Hiremath P.S., Tegnoor J.R. Recognition of Follicles in Ultrasound Images of Ovaries Using Geometric Features. Department of Studies and Research in Computer Science, Gulbarga University; Gulbarga, Karnataka, India: 2009. [Google Scholar]
  • 10.Huang Y.L., Chen D.R., Jiang Y.R., Kuo S.J., Wu H.K., Moon W.K. Computer-Aided Diagnosis Using Morphological Features for Classifying Breast Lesions on Ultrasound. Wiley InterScience; Hoboken, NJ, USA: 2008. [DOI] [PubMed] [Google Scholar]
  • 11.Nugroho H.A., Triyani Y., Rahmawaty M., Ardiyanto I. Analysis of margin sharpness for breast nodule classification on ultrasound images; Proceedings of the 2017 9th International Conference on Information Technology and Electrical Engineering (ICITEE); Phuket, Thailand. 12–13 October 2017; pp. 1–5. [Google Scholar]
  • 12.Anil G., Hegde A., Chong F.H.V. Thyroid nodules: Risk stratification for malignancy with ultrasound and guided biopsy. Cancer Imaging. 2011;11:209. doi: 10.1102/1470-7330.2011.0030. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Zulfanahri, Nugroho H.A., Nugroho A., Frannita E.L., Ardiyanto I. Classification of Thyroid Ultrasound Images Based on Shape Features Analysis; Proceedings of the 2017 Biomedical Engineering International Conference; Yogyakarta, Indonesia. 6–7 November 2017. [Google Scholar]
  • 14.Nugroho H.A., Frannita E.L., Nugroho A., Zulfanahri, Ardiyanto I., Choridah L. Classification of thyroid nodules based on analysis of margin characteristic; Proceedings of the 2017 International Conference on Computer, Control, Informatics and its Applications (IC3INA); Jakarta, Indonesia. 23–26 October 2017; pp. 47–51. [DOI] [Google Scholar]
  • 15.Chen J., You H. Efficient classification of benign and malignant thyroid tumors based on characteristics of medical ultrasonic images; Proceedings of the 2016 IEEE Advanced Information Management, Communicates, Electronic and Automation Control Conference (IMCEC); Xi’an, Shanxi, China. 3–5 October 2016; pp. 950–954. [Google Scholar]
  • 16.Ding J., Cheng H., Huang J., Zhang Y., Ning C. A novel quantitative measurement for thyroid cancer detection based on elastography; Proceedings of the 2011 4th International Congress on Image and Signal Processing (CISP); Shanghai, China. 15–17 October 2011; pp. 1801–1804. [Google Scholar]
  • 17.Isa I.S., Saad Z., Omar S., Osman M., Ahmad K., Sakim H.M. Suitable MLP network activation functions for breast cancer and thyroid disease detection; Proceedings of the 2010 Second International Conference on Computational Intelligence, Modelling and Simulation (CIMSiM); Bali, Indonesia. 28–30 September 2010; pp. 39–44. [Google Scholar]
  • 18.Patrício M., Oliveira C., Caseiro-Alves F. Differentiating malignant thyroid nodule with statistical classifiers based on demographic and ultrasound features; Proceedings of the 2017 IEEE 5th Portuguese Meeting on bioengineering (ENBENG); Coimbra, Portugal. 16–18 February 2017; pp. 1–4. [Google Scholar]
  • 19.Pedraza L., Vargas C., Narváez F., Durán O., Muñoz E., Romero E. An open-access thyroid ultrasound-image Database; Proceedings of the Conference: 10th International Symposium on Medical Information Processing and Analysis; Cartagena de Indias, Colombia. 14–16 October 2014. [Google Scholar]
  • 20.Song R., Zhang L., Zhu C., Liu J., Yang J., Zhang T. Thyroid Nodule Ultrasound Image Classification Through Hybrid Feature Cropping Network. IEEE Access. 2020;8:64064–64074. doi: 10.1109/ACCESS.2020.2982767. [DOI] [Google Scholar]
  • 21.Chi J., Walia E., Babyn P., Wang J., Groot G., Eramian M. Thyroid Nodule Classification in Ultrasound Images by Fine-Tuning Deep Convolutional Neural Network. J. Digit. Imaging. 2017;30:477–486. doi: 10.1007/s10278-017-9997-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Koundal D., Gupta S., Singh S. Computer aided thyroid nodule detection system using medical ultrasound images. Biomed. Signal Process. Control. 2018;40:117–130. doi: 10.1016/j.bspc.2017.08.025. [DOI] [Google Scholar]
  • 23.Nanda S., Mar M. Identification of Thyroid Cancerous Nodule using Local Binary Pattern Variants in Ultrasound Images. Int. J. Eng. Trends Technol. 2017;49:369–374. doi: 10.14445/22315381/IJETT-V49P256. [DOI] [Google Scholar]
  • 24.Mortenson M.E. Geometric Modeling. John Wiley; New York, NY, USA: 1985. p. 763. [Google Scholar]
  • 25.Yang M., Kpalma K., Ronsin J. A Survey of Shape Feature Extraction Techniques. Peng Yeng Yin Pattern Recognit. 2008;15:43–90. [Google Scholar]
  • 26.Sagar C. Feature selection techniques with R. Data Aspirant. Jan 15, 2018.
  • 27.Moltz J., Bornemann L., Kuhnigk J.-M., Dicken V., Peitgen E., Meier S., Bolte H., Fabel M., Bauknecht H., Hittinger M., et al. Advanced Segmentation Techniques for Lung Nodules, Liver Metastases, and Enlarged Lymph Nodes in CT Scans. IEEE J. Sel. Topics Signal Process. 2009;3 doi: 10.1109/JSTSP.2008.2011107. [DOI] [Google Scholar]
  • 28.Kubota T., Jerebko A.K., Dewan M., Salganicoff M., Krishnan A. Segmentation of pulmonary nodules of various densities with morphological approaches and convexity models. Med. Image Anal. 2011;15:133–154. doi: 10.1016/j.media.2010.08.005. [DOI] [PubMed] [Google Scholar]
  • 29.Borghesi A., Scrimieri A., Michelini S., Calandra G., Golemi S., Tironi A., Maroldi R. Quantitative CT Analysis for Predicting the Behavior of Part-Solid Nodules with Solid Components Less than 6 mm: Size, Density and Shape Descriptors. Appl. Sci. 2019;9:3428. doi: 10.3390/app9163428. [DOI] [Google Scholar]
  • 30.Alilou M., Beig N., Orooji M., Rajiah P., Velcheti V., Rakshit S., Reddy N., Yang M., Jacono F., Gilkeson R., et al. An integrated segmentation and shape based classification scheme for distinguishing adenocarcinomas from granulomas on lung CT. Med. Phys. 2017;44 doi: 10.1002/mp.12208. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31.Sahu B., Dehuri S., Jagadev A. A Study on the Relevance of Feature Selection Methods in Microarray Data. Open Bioinf. J. 2018;11:117–139. doi: 10.2174/1875036201811010117. [DOI] [Google Scholar]
  • 32.Söderman C., Johnsson A.A., Vikgren J., Norrlund R.R., Molnar D., Svalkvist A., Månsson L.G., Båth M. Evaluation of Accuracy and Precision of Manual Size Measurements in Chest Tomosynthesis using Simulated Pulmonary Nodules. Acad. Radiol. 2015;22:496–504. doi: 10.1016/j.acra.2014.11.012. [DOI] [PubMed] [Google Scholar]
  • 33.Serra J. Image Analysis and Mathematical Morphology. Academic Press; Cambridge, MA, USA: 1982. [Google Scholar]
  • 34.Gierada D.S., Politte D.G., Zheng J., Schechtman K.B., Whiting B.R., Smith K.E., Crabtree T., Kreisel D., Krupnick A.S., Patterson A.G., et al. Quantitative Computed Tomography Classification of Lung Nodules: Initial Comparison of 2- and 3-Dimensional Analysis. J. Comput. Assist. Tomogr. 2016;40:589–595. doi: 10.1097/RCT.0000000000000394. [DOI] [PMC free article] [PubMed] [Google Scholar]

Articles from Sensors (Basel, Switzerland) are provided here courtesy of Multidisciplinary Digital Publishing Institute (MDPI)

RESOURCES