Skip to main content
Diagnostics logoLink to Diagnostics
. 2023 May 3;13(9):1621. doi: 10.3390/diagnostics13091621

A Novel Hybrid Runge Kutta Optimizer with Support Vector Machine on Gene Expression Data for Cancer Classification

Essam H Houssein 1, Hager N Hassan 1, Nagwan Abdel Samee 2, Mona M Jamjoom 3,*
Editors: Dechang Chen, Wei Shao
PMCID: PMC10178557  PMID: 37175012

Abstract

It is crucial to accurately categorize cancers using microarray data. Researchers have employed a variety of computational intelligence approaches to analyze gene expression data. It is believed that the most difficult part of the problem of cancer diagnosis is determining which genes are informative. Therefore, selecting genes to study as a starting point for cancer classification is common practice. We offer a novel approach that combines the Runge Kutta optimizer (RUN) with a support vector machine (SVM) as the classifier to select the significant genes in the detection of cancer tissues. As a means of dealing with the high dimensionality that characterizes microarray datasets, the preprocessing stage of the ReliefF method is implemented. The proposed RUN–SVM approach is tested on binary-class microarray datasets (Breast2 and Prostate) and multi-class microarray datasets in order to assess its efficacy (i.e., Brain Tumor1, Brain Tumor2, Breast3, and Lung Cancer). Based on the experimental results obtained from analyzing six different cancer gene expression datasets, the proposed RUN–SVM approach was found to statistically beat the other competing algorithms due to its innovative search technique.

Keywords: feature selection, Runge Kutta optimizer, microarray, gene expression, support vector machines, cancer classification

MSC: 68T99, 68U99

1. Introduction

For the past few years, gene expression profiling has allowed scientists to simultaneously examine thousands of genes. As a result, significant knowledge has been gained regarding the processes that take place within cells. In situations involving cancer, having access to such particular information is helpful for both diagnosing the disease and evaluating the patient’s prognosis. Microarray data features such as low sample size, high noise levels, and high dimensionality persist in making gene selection a challenging task. Identifying the best candidates for gene classification is essential to resolving this problem. This method will not only cut down on computational costs, but it will also enable physicians to focus on a small subset of physiologically significant genes that are linked to specific cancers, allowing for the formulation of more cost-effective studies [1]. In addition, a method with a high degree of accuracy can aid people with cancer by facilitating early detection and the development of new treatments [2].

The “curse of dimensionality”, manifested by small sample sizes and the presence of genes that are unrelated to one another, plagues the microarray data [3,4,5]. Due to the presence of this condition, it is more challenging to classify the typical gene expression profile for any given sample. Gene selection has been applied to microarray gene expression analysis as a means of overcoming the “curse of dimensionality” by determining the subset of features that are most significant [6].

There are a few different ways to go about selecting genes, but they can all be broken down into three broad categories: embedding, wrapping, and filtering. Using statistical methods like minimum redundancy maximum relevance, information gain, and ReliefF, the filter strategy collects intrinsic gene properties in order to differentiate the targeted phenotypic class [7]. This approach is easy to implement, but it does not take into account the complex web of interactions between genes. In order to evaluate potential new feature gene subsets, the “wrapper” strategy [8] attempts to pick a tiny subset of the initial feature set, typically through the use of an induction algorithm. Wrapper methods typically outperform filters because they account for multivariate inter-correlation among genes. The optimum number of feature genes for a given classifier can be calculated automatically using the wrapper approach. The embedded method, like the wrapper strategy, permits the combination of many strategies for doing feature subset selection (Kahavi and John, 1997; [9]). Genetic algorithms (GAs) are commonly employed as the search engine for the feature subset in the embedding approach (GAs) [10], while other classification techniques, such as GA-SVM [11], GA-K closest neighbors (GA-KNN) [12], and so on, are used to pick the smallest feature set.

Selecting the significant genes from gene expression data has been implemented using the wellknown machine learning approaches, including Neural Networks (NN), support vector machines (SVMs), and K-nearest neighbor (KNN). There is some evidence that gene selection can improve the accuracy of sample classification [13]. According to studies, the “no free lunch theory” states that no single optimization strategy is generally accepted as providing a globally optimal solution to any given optimization problem [14]. Different practical engineering applications were used to evaluate the RUN algorithm’s efficacy in comparison to numerous metaheuristic algorithms. The RUN’s outcomes were encouraging and on par with previous competitions [15]. Consequently, the objective of this study is to select the genes in the gene expression profile that are the most predictive and beneficial by making use of RUN, a recently proposed optimization method. A preprocessing stage that uses the ReliefF filtering strategy is used to reduce the dimensionality of gene expression datasets and select the valuable genes from among these datasets. As they perform better than other classification methods [16], SVMs have become the most widely used classification method in the field of cancer research, most notably in the analysis of gene expression data [17,18]. in this study, SVM is used in a hybrid cancer classification algorithm. The article also compares the effectiveness of the RUN algorithm as a generation selection algorithm to the effectiveness of other well-known selection algorithms [19], such as the Whale Optimization Algorithm (WOA) [20], the Artificial Bee Colony (ABC) [21], the Harris Hawks Optimization (HHO) [22], the Hunger Games Search (HGS) [23], the Golden Jackal Optimization (GJO) [24], and Manta Ray Foraging Optimization (MRFO) [25]. To evaluate how well the suggested RUN—SVM strategy performs in comparison to existing selection algorithms, the experimental study makes use of six unique microarray datasets.

To emphasize the significance of this study, the following are the contributions that it makes:

  • Here we provide RUN—SVM, a novel approach to classifying gene expression data that combines the SVM classifier with the RUN feature selection optimizer.

  • Overcomes the “feature dimensionality curse” in order to eliminate unnecessary features and focus on the most informative ones when classifying cancers based on gene expression data.

  • Assesses the effectiveness of the introduced classifier in binary and multiclass classification tasks using six benchmark datasets.

  • Compares the capability and feasibility of the proposed classification algorithm to existing approaches through a comprehensive evaluation.

The remaining sections of the article are presented in the following order: The review of the available literature can be found in Section 2. In Section 3, we will go over the first stages of the classification methods that make use of ReliefF, RUN, and SVM. The RUN–SVM technique that has been suggested is discussed in Section 4. In Section 5, In Section, both the experimental design and the results of the experiments are described. Section 6 contains a summary of the findings as well as recommendations for further research.

2. Literature Review

A number of different gene selection optimization approaches have been presented in the literature as a way to improve the accuracy with which cancers are diagnosed. The following is a list of the most recent studies that have been conducted for the purpose of making suggestions as to which genes should be employed in the classification and diagnosis of cancer.

Li [1], proposed a strategy for gene selection that included a GA and a PSO with support vector machine (SVM) classification. Their proposed methodology was put to the test using data from microarray samples. The authors’ intention with their combined strategy was to steer the proposed method away from the local optimum. Using an improved binary PSO and an SVM classifier.

In order to analyze microarray datasets, it was suggested that combining SVM with the artificial bee colony approach might be beneficial [26]. The proposed model was implemented on the group of genes that were chosen in order to evaluate how well it worked. The high computing cost associated with the large dimensionality of the microarray data was a hindrance for all of the strategies, despite the fact that the proposed model performed better than earlier studies that used this methodology. The authors of this study conducted experiments with both genetic algorithm, and particle swarm optimization [2]. It was determined that the suggested approach of GA augmented with SVM is the best option. Instead, both of the techniques required training a classifier with every gene simultaneously due to the high complexity of gene expression datasets. Comparisons with sundry state-of-the-art techniques revealed outcomes that were competitive when measured against industry standards. In the majority of executions, they achieved results with a classification rate of 100% and few genes per subgroup (3 and 4). By combining IG and SVM, the authors of [27] achieved the best results possible for classifying cancers. The most informative and important genes were selected using IG from the original datasets, and then redundant genes were omitted using SVM. They then used the LIBSVM classifier to evaluate the quality of the collected informative genes [28]. For this investigation, the authors prioritized selecting the fewest genes while ignoring the pursuit of the highest accuracy.

In [29], the authors applied filter-, wrapper-, and correlation-based feature selector (CFS) techniques to the acute leukemia microarray dataset. These methods had been used with a number of ML algorithms, such as naive Bayes, decision tree (DT), and SVM. They demonstrated that by integrating several feature selection and classification techniques, relevant genes can be selected with strong certainty. This was the first paper to discuss the biological and computational evidence for zyxin’s role in leukaemogenesis.

In [30], The researchers used a 1-norm support vector machine with squared loss and a second predictor made up of an SVM and the nearest neighbor to classify cancers rapidly using NN. When it comes to assessment and storage that algorithm just needs a few genes. This, of course, ignores the question of how many genes were actually selected.

Mohamad et al., [31], suggested a near-optimal, small selection of the most beneficial genes relevant to cancer classification. The current rule for updating velocity and particle position was updated by the method of Mohamad et al.

Wrapper strategies combining the SVM classification method and the FireFly algorithm (FF-SVM) have been proposed as a means to classify cancer microarray gene expression profiles [32]. The experimental results showed that the suggested algorithm performed better in terms of classification accuracy and selected fewer genes than the other existing techniques. Also, the top five biologically significant genes associated with cancer across all ten datasets were determined.

The authors of [33] concentrated on the filter-based feature selection approach. The primary goals of the proposed effort were to shorten computation times and improve categorization and prediction accuracy. They suggested work that decreases the complexity of data collection as well as the redundancy between different features in order to accomplish this. They employed a score-based criterion fusion approach for feature selection that increased prediction accuracy and cut down on calculation time.

A hybrid model, SVM-mRMRe, was presented for gene selection by the authors of [34]. The most important genes from high-throughput microarrays have been extracted using the proposed model, which combines the feature selection approach, mRMRe, with the SVM classifier. The introduced algorithm was evaluated using eight benchmarking microarray datasets. Several aspects of the selected data set were analyzed by four different classifiers including NN, KNN, SVM, and random forest (RF). As demonstrated experimentally, the suggested approach enhances the current state of the art in cancer tissue classification while using fewer data.

In [35], a number of combinations of wrapping approaches (using NB and KNN with rank search, greedy stepwise, and best first) and ranking methods (using information gain [IG] with threshold of 1% and 5%) were employed for choosing the most crucial genes for microarray datasets. Brain cancer, CNS, breast cancer, and lung cancer datasets were among these datasets. Comparing the scenario where all features were employed with the experimental findings, all feature selection strategies consistently performed well (no feature selection methods). The NB with IG and wrapper (NB and best first) and the KNN with IG and wrapper (KNN and best first) acquired the best performance and defeated all other approaches out of the several techniques that were utilized.

A new hybrid feature selection technique was introduced in [36]. It involved five filters and a wrapper mechanism. It was discovered that the hybrid collection of individual filters has yielded better performance than the retrieved one of the single filter. The suggested method had been tried out and evaluated against well-known hybrid features. Using 10 benchmark micro-array datasets, the suggested strategy was evaluated and contrasted with cutting-edge approaches. Experimental results showed that the proposed strategy exhibited superior performance to state-of-the-art methods in regards to precision in classification.

3. The Methodology

Here, we provide an overview of the methods that would be employed in the proposed RUN—SVM classification technique. These methods include the SVM classification method, the RUN optimization method, and the ReliefF feature selection method for microarray data.

3.1. ReliefF Filter Method

The first Relief algorithm [37] has been developed further with the addition of ReliefF [38]. Selecting one instance at random from the data, the primary structure of Relief then finds its nearest neighbor from the other class as well as its nearest neighbor from the same class. Sampled instances’ attribute values are compared to those of their nearest neighbors in order to revise the relevance scores for each attribute. According to this line of thinking, a valuable feature should be able to distinguish between instances of several classes while maintaining the same value for instances of the same class. With its enhanced capabilities, ReliefF is better able to handle multiclass situations and input that is illegible or noisy. This methodology has global applicability, low bias, involves feature interaction, and has the potential to uncover local dependencies that are missed by other methodologies.

3.2. Runge Kutta Optimizer (RUN)

The goal of optimization algorithms is to enhance the efficiency with which they solve various optimization problems by striking a balance between exploiting and exploring the problem space. This is done in order to improve the efficiency with which they tackle various optimization problems. Iman Ahmadianfar and her colleagues recently presented the RUN, which is one example of such an optimization method [15]. This algorithm makes use of a specialized search mechanism that is founded on the Runge Kutta method, and RUN is an essential component of this algorithm. The second stage is termed solution quality enhancement, and its purpose is to increase the quality of the solutions produced for the researched optimization problem while simultaneously minimizing the hazards of becoming trapped in a locally optimal solution. The mathematical formulation of RUN is explained with the help of an example that can be found in the next subsection.

3.2.1. Initialization

The decision variables yn,v of the optimization problem are initialized at random to begin the optimization process in RUN by:

yn,v=Rmin+rand.(RmaxRmin) (1)

where the lowest and highest ranges of the decision variable v of dimensions 1,2,…, and M are, respectively, Rmin, and Rmax.

3.2.2. Solution Update

The well-known RUN mechanism is applied to the following equations in order to modify the solutions yn+1 at each iteration:

ifrand<0.5yn+1=(yc.SF.r.g+yc)+SF.SM+μ.randn.(ymyc)elseyn+1=(ym.SF.r.g+ym)+SF.SM+μ.randn.(yr1yr2)end (2)

where yr1, and yr2 are solutions selected at random and μ is a typical distributed random number. randn is a random number in the range of [0, 1]. r is an integer number that may be 1 or −1. g is a random number in the range of [0, 2]. xc and xm are solutions around which the local search is performed in order to explore the promising regions in the search space. SF was used to achieve an appropriate balance between exploration and exploitation.

Additionally, using Δy, which is the position increment, and RUN coefficients k1 through k4, the subsequent equation can be used to calculate SF, an adaptable factor, and SM, a RUN-guided search method:

SM=Δy×yRK6,yRK=k1+2×k2+2×k3+k4 (3)

3.2.3. Enhanced Solution Quality (ESQ)

Using the ESQ technique, RUN incorporates better solution quality, escape from the local optima, and quick convergence. In order to use this approach, three new solutions—ynew1, ynew2, and ynew3—will be created based on the equation shown below:

ynew1=β×yavg+(1β)×ybest,yavg=yr1+yr2+yr33 (4)
ifrand<0.5ifw<1ynew2=w.r.|(ynew1yavg)+randn|+ynew1elseynew2=w.r.|(u.ynew1yavg)+randn|+(ynew1yavg)endend (5)
ifrandn<wynew3=SF.((v.ybynew2)+rand.yRK)+(ynew2rand.ynew2)end (6)

where the value of β is a random number between [0, 1]. ybest is the global best solution throughout iterations. r is an integer that can be either 1, 0, or −1. yb is the best solution per iteration. A random number v is equal to 2 × randn. The random number w gets smaller as the number of iterations rises.

The pseudo-code steps of the RUN method is depicted in Algorithm 1.

Algorithm 1 The RUN algorithm
  • 1:

    Input: Population size, Population initialization, and MaxIter of iterations.

  • 2:

    Output: The best Solution.

  • 3:

    Evaluate Objective Function using Equation (7).

  • 4:

    Calculate yn+1 using Equation (2).

  • 5:

    while Stop condition not met do

  • 6:

         for n = 1 ... N do

  • 7:

               Evaluate Objective Function using Equation (7).

  • 8:

               if <0.5 then

  • 9:

                    calculate ynew2 using Equation (5)

  • 10:

                   if F(yn+1<F(ynew2)) then

  • 11:

                       if rand < w then

  • 12:

                             Calculate ynew3 using Equation (6).

  • 13:

                             if F(yn+1)<F(ynew3) then

  • 14:

                                 yn+1=ynew3

  • 15:

                             end if

  • 16:

                       end if

  • 17:

                   else

  • 18:

                       yn+1=ynew2

  • 19:

                   end if

  • 20:

               end if

  • 21:

         end for

  • 22:

    end while

3.3. Support Vector Machine (SVM)

In the realm of supervised machine learning, SVMs are a helpful tool for dealing with classification and regression issues [39]. Despite this, the majority of their applications are found in classification-related debates. It is difficult to find a linear classifier in the dataset that can differentiate between the different classes. The SVM, widely regarded as the only classification method capable of handling this situation, discovers a linear classifier model for categorizing the input data within an excess margin limit after mapping and modifying the input space to turn it into a high-dimensional space. This model is used to determine where the data falls within the space. When dealing with data that has a lot of dimensions, SVM performs the best as a classifier [40]. In light of this, an SVM classification technique is utilized in this investigation in order to assess how useful RUN is when applied to microarrays classification.

The results of a support vector machine are controlled by two variables. C regulates the compromise between accurate training-data classification and a consistently smooth resolution limit. Meanwhile, Γ represents the effect of a single training session. Algorithm 2 demonstrates the pseudo-code for the SVM model.

Algorithm 2 The SVM algorithm
  • 1:

    Input: Test and training points.

  • 2:

    Output: Best fitness.

  • 3:

    Select the SVM’s expense factor C and Γ.

  • 4:

    while stop condition not satisfied do

  • 5:

        For each data point, Apply the SVM training phase.

  • 6:

        For testing data points, Apply the SVM classifying phase.

  • 7:

    end while

  • 8:

    return fitness

In order to partition the data into K separate sets, a statistical technique known as K-fold cross-validation is utilized. As a result, only the training subsets were employed at first; however, more recently, other subsets have been used to test or validate the training subsets. When testing the model, a pending subsample will be used for validating the data, and K-1 sets would be employed for training points.

4. The Proposed RUN-SVM Approach

The combination of ReliefF, RUN, and SVM into a single framework for the effective classification of microarray datasets is novel, as far as we are aware. Not only is RUN being used for the first time as a gene selection method for cancer gene expression datasets, but it is also being used in a novel way. Here, we offer an overview of the hybrid cancer classification method we’ve proposed for picking the most precise SVM-classified genes from cancer gene expression datasets. In order to offer a set of genes that are predictive as well as informative, the proposed RUN–SVM approach, which is depicted in Figure 1, includes two processes: (1) a preprocessing step, and (2) a feature selection (FS) and classification Step. These steps are displayed in order. It is also possible to evaluate the predictability and informability of the other gene selection strategies by using the same two steps of the RUN–SVM approach that we have presented.

Figure 1.

Figure 1

The proposed RUN-SVM approach.

4.1. Preprocessing Phase

Applying algorithms directly to microarray datasets is a real challenge due to the fact that these datasets are of a high dimension and contain thousands of genes that need to be filtered. As a result of this, using these kinds of datasets without first filtering them makes it difficult to accurately train the SVM classification algorithm. For its excellent performance, great efficiency, and widespread application in the field of resolving problems with huge dimensions, we decided to make use of the ReliefF algorithm in this particular piece of work. As a result, the first stage of the proposed RUN–SVM methodology is called ReliefF. It can be utilized for the purpose of removing redundant genes, which contribute to the background noise and lower the classification accuracy (for more information on ReliefF, see Section 3.1).

4.2. Feature Welection and Classification Phase

To pick the most important genes from the smaller pool of genes generated in the first stage, we use wrapper approaches like the RUN algorithm and the SVM. First, RUN generates a randomly selected initial population, which is then subjected to an evaluation based on a fitness function. When the FS and classification phase procedures have been completed, the criterion has been met, and the ideal solution has been created, one with the highest classification performance and the fewest informative genes. This guarantees a productive outcome from the procedure. The SVM classifier performs an evaluation of the RUN solutions through the repetition phase in order to gauge how effective they are. The fitness function is evaluated based on the percentage of samples that are correctly categorized in comparison to the total number of samples (for more information on RUN and SVM, see Section 3.2 and Section 3.3).

fitness=CT100 (7)

where C indicates samples that were successfully classified, and T stands for the overall number of input samples.

The pseudo-code of the proposed RUN–SVM approach is demonstrated in Algorithm 3.

Algorithm 3 Pseudo code of the proposed RUN-SVM approach
  • 1:

    Input: MaxIter, A Microarray Dataset, Population size, Population initialization

  • 2:

    Output: Best solution.

  • 3:

    Apply ReliefF filter on Dataset and generate a filtered dataset.

  • 4:

    Generate an initial population.

  • 5:

    Evaluate Objective Function using Equation (7).

  • 6:

    Calculate yn+1 using Equation (2).

  • 7:

    while t< MaxIter do

  • 8:

         for n = 1 ... N do

  • 9:

               Evaluate Objective Function using Equation (7).

  • 10:

               if rand<0.5 then

  • 11:

                   calculate ynew2 using Equation (5).

  • 12:

                   if F(yn+1<F(ynew2)) then

  • 13:

                   if rand < w then

  • 14:

                       Calculate ynew3 using Equation (6).

  • 15:

                       if F(yn+1)<F(ynew3) then

  • 16:

                           yn+1=ynew3

  • 17:

                       end if

  • 18:

                   end if

  • 19:

               else

  • 20:

                   yn+1=ynew2

  • 21:

               end if

  • 22:

            end if

  • 23:

        end for

  • 24:

    end while

5. Experimental Evaluation and Discussion

5.1. Datasets and the Running Environment

In this investigation, we obtained the gene expression datasets by downloading them from a public database referred to as the GEMS Database. This database is a repository for gene expression data [41,42]. There are two types of gene expression datasets that have been downloaded: multi-class microarray datasets (i.e., Brain Tumor1, Brain Tumor2, Breast3, and Lung Cancer) and binary-class microarray datasets (Breast2 and Prostate). The dataset has been preprocessed to produce the gene expression matrix—S by F matrix, where S is the number of samples and F is the number of genes present in the given samples—which is used to represent the gene expression dataset. The six cancer gene expression datasets employed to evaluate the effectiveness of the proposed RUN–SVM approach are shown in Table 1. A computer with Microsoft Windows 8.1 with core i-5, 2.50 GHz, 64-bit, and 8 GB main memory was employed for the user testing environment, and Matlab R2015a served as the executable environment. Each of the 30 independent runs of the proposed RUN–SVM approach includes 30 search agents and 100 iterations.

Table 1.

An overview of the datasets acquired by the microarrays.

Datasets Genes Samples Classes
Brain Tumor1 5920 90 5
Brain Tumor2 10,368 50 4
Breast2 4869 77 2
Breast3 4869 95 3
Lung Cancer 12,600 203 5
Prostate 10,509 102 2

5.2. Evaluation Metrics

The proposed RUN—SVM method was compared to the other approaches using four different measures of accuracy: best fitness function, standard deviation (STD), worst fitness function, and mean fitness function. In general, these are the four accuracy measurements that were used. The best fitness score at a run I is fitness(ybest).

  • 1.

    Mean: is calculated by multiplying the number of times the algorithm is run by its fitness value (N).

  • 2.

    Best fitness function: is a value that, after running the algorithm N times, outperforms all other best fitness values.

  • 3.

    Worst fitness function: the smallest value for each maximum fitness metric that can be achieved by repeated iterations of the method N times

  • 4.

    The standard deviation (STD): is a quantity that can be obtained by running the algorithm N times and measuring the amount of change in the fitness values. Data points with a smaller STD value show that they cluster tightly around the mean, while data points with a larger STD value suggest that they may deviate greatly from the mean.

5.3. Experimental Findings

The cross-validation method was implemented in this investigation in order to enhance classification performance, minimize bias in point selection for testing and training, and improve overall accuracy. Therefore, 10-fold cross-validation was utilized during the process of training the SVM classifier. In addition, the ReliefF filter and SVM classifier-improved comparison algorithms were each subjected to 30 separate runs with each run consisting of 100 iterations.

5.3.1. Statistical Results Analysis

According to the results displayed in Table 2, Table 3 and Table 4, RUN ranks first in terms of best fitness values, worst fitness values, and mean in the Brain Tumor1, Brain Tumor2, and Breast3 datasets, respectively. Table 5 displays that the three fitness metrics with the highest rankings in Beast2 are RUN, MRFO, and HGS, with RUN being the gold standard. Table 6 displays that RUN and HGS have the highest fitness values in the Lung Cancer dataset, but RUN also has the lowest standard deviation. While HGS is rated highest for greatest fitness, Table 7 shows that RUN ranks highest for mean, worst, and standard deviation on the Prostate dataset.

Table 2.

Performance evaluation of the proposed algorithm on the Brain Tumor 1 dataset.

Algorithms STD Best Mean Worst
MRFO [25] 1.217 86.667 84.667 83.333
HGS [23] 0.4969 86.667 85.778 85.556
GJO [24] 1.7568 86.667 84.444 82.222
ABC [21] 0.9938 85.556 84 83.333
WOA [20] 0.9296 84.444 83.556 82.222
HHO [22] 1.4487 84.444 83.111 81.111
RUN-SVM 2.3040 90 86.222 84.444
Table 3.

Performance evaluation of the proposed algorithm on the Brain Tumor 2 dataset.

Algorithms STD Best Mean Worst
MRFO [25] 1.7889 88 86.8 84
HGS [23] 1.673 92 90.4 88
GJO [24] 0.8944 90 88.4 88
ABC [21] 2.280 92 88.8 86
WOA [20] 2.1909 92 88.4 86
HHO [22] 2 86 84 82
RUN-SVM 2.280 94 90.8 88
Table 4.

Performance evaluation of the proposed algorithm on the Breast 3 dataset.

Algorithms STD Best Mean Worst
MRFO [25] 1.289 72.632 71.579 69.474
HGS [23] 1.729 73.684 72.421 70.526
GJO [24] 1.3725 72.632 71.368 69.473
ABC [21] 0.471 72.632 71.789 71.579
WOA [20] 1.596 71.579 69.895 68.421
HHO [22] 2.025 74.737 71.789 69.474
RUN-SVM 2.306 76.842 73.053 71.579
Table 5.

Performance evaluation of the proposed algorithm on the Breast 2 dataset.

Algorithms STD Best Mean Worst
MRFO [25] 1.693 90.91 89.351 87.013
HGS [23] 1.162 90.91 90.389 88.312
GJO [24] 1.087 88.117 87.273 85.714
ABC [21] 1.837 89.61 87.013 85.714
WOA [20] 1.299 88.312 87.013 85.714
HHO [22] 1.926 88.312 85.974 83.117
RUN-SVM 0.7113 90.91 90.13 89.61
Table 6.

Performance evaluation of the proposed algorithm on the Lung Cancer dataset.

Algorithms STD Best Mean Worst
MRFO [25] 0.9603 90.148 88.867 88.177
HGS [23] 2.072 91.626 89.557 86.207
GJO [24] 1.155 90.148 88.669 87.685
ABC [21] 1.889 91.133 88.571 86.699
WOA [20] 2.013 89.163 87.586 84.729
HHO [22] 1.419 91.133 88.867 87.685
RUN-SVM 0.6033 91.626 91.133 90.148
Table 7.

Performance evaluation of the proposed algorithm on the Prostate dataset.

Algorithms STD Best Mean Worst
MRFO [25] 0.820 99.02 97.843 97.059
HGS [23] 1.201 100 98.039 97.059
GJO [24] 0.537 99.02 98.431 98.04
ABC [21] 0.877 98.039 96.667 96.078
WOA [20] 0.438 99.02 98.235 98.04
HHO [22] 0.537 99.02 98.627 98.039
RUN-SVM 0 99.02 99.02 99.02

5.3.2. Convergence Behavior Analysis

We compared the proposed RUN–SVM approach to some other popular optimization techniques and showed how it converges after 100 iterations. Based on the results shown in Figure 2a, the proposed RUN–SVM approach and the MRFO algorithm are the only ones that can be recognized to have the best accuracy for the Brain Tumor1 dataset. According to Figure 2c, only the proposed RUN–SVM technique and the GJO algorithm are able to achieve the maximum accuracy for the Breast2 dataset within the 100 iterations.

Figure 2.

Figure 2

Figure 2

Convergence curves for the comparison algorithms. (a) Brain Tumor1 Convergence Curve. (b) Brain Tumor2 Convergence Curve. (c) Breast2 Convergence Curve. (d) Breast3 Convergence Curve. (e) Lung Cancer Convergence Curve. (f) Prostate Convergence Curve.

Figure 2d, shows that the introduced RUN–SVM algorithm, together with the HGS algorithm, are the only algorithms that can be said to attain the best performance in the classification of samples in the Breast3 dataset. In addition, Figure 2e, the proposed RUN–SVM approach is the sole algorithm to achieve the greatest accuracy for the Lung Cancer dataset. Only the WOA algorithm and the suggested RUN—SVM method have been shown to achieve the maximum precision on the Prostate database, as depicted in Figure 2f.

5.3.3. Boxplot Behavior Analysis

In this section, we draw attention to the results of the proposed RUN–SVM approach. The proposed RUN–SVM approach outperformed the competition in this evaluation, as is shown by the boxplot curves in Figure 3a–f.

Figure 3.

Figure 3

Boxplots curves for the comparison algorithms. (a) Brain Tumor1 Boxplot. (b) Brain Tumor2 Boxplot. (c) Breast2 Boxplot. (d) Breast3 Boxplot. (e) Lung Cancer Boxplot. (f) Prostate Boxplot.

5.4. Discussion

The purpose of this section is to examine the numerous systems now in use for classifying cancers. The goal of this research is to propose a search strategy for the FS problem that takes into account the challenges posed by high-dimensional datasets. The research suggested using a ReliefF filtering approach for preprocessing and then combining RUN and SVM for the FS and classification stages. The effectiveness of the proposed RUN–SVM approach is demonstrated by experimental analysis and a comparison study. The following is a list of benefits that come with using our proposed RUN–SVM technique:

  • The datasets that were utilized for the analysis of this study contain feature sizes that range anywhere from 4869 to 12,600 features, which makes for an excellent testing environment for an optimization method.

  • Any chance for improving RUN–SVM can easily be implemented due to its straightforward architecture.

  • Figure 4 shows how RUN–SVM outperforms other optimization algorithms in terms of classifying cancer.

Figure 4.

Figure 4

Average accuracy percentage.

On the other hand, the proposed RUN–SVM method offers certain restrictions, and these are listed below.

  • Due to the fact that RUN–SVM uses a randomization-based optimization technique, the features it uses may vary from run to run. The feature subset picked in one run may not be present in another.

  • As most genes are related to one another either directly or indirectly, microarray data processing is quite complex.

  • The limited number of samples available in microarray datasets, the presence of noise and genes that are not relevant to the study, and the “curse of dimensionality” all add to the difficulty of classifying a particular sample.

6. Conclusions and Future Work

Using the Runge Kutta optimizer (RUN) approach, this investigation presents an improved way for choosing significant genes from cancer microarray patterns. The first primary goal was to pick the genes from the microarray datasets that were the most relevant and helpful with the least amount of effort, and the second primary goal was to select the genes that were the most accurate with the least amount of effort. In this paper, a wrapper gene selection approach called RUN—SVM is proposed. RUN—SVM combines the RUN algorithm with the support vector machine (SVM) classifier in order to identify the most relevant genes from a small dataset after removing redundant genes and selecting related genes from gene expression datasets. Six standard microarray datasets were used to evaluate the proposed RUN–SVM approach, namely Brain Tumor1, Brain Tumor2, Breast2, Breast3, Lung Cancer, and Prostate. The findings of the experiments showed that the RUN-SVM methodology obtains the highest level of accuracy in comparison to other comparative methodologies, such as HHO, MRFO, GJO, HGS, WOA, and ABC. The effectiveness of recent deep learning classifiers, such as transformer-based approaches and self-attention methods, could be examined for cancer classification using microarray datasets as part of future work that is associated with this research. Also, the performance of those classifiers could be evaluated in order to determine whether or not it would be enhanced by making use of the optimizer that was proposed in this study.

Acknowledgments

The authors would like to express their gratitude to Princess Nourah bint Abdulrahman University Researchers Supporting Project number (PNURSP2023R104), Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia.

Author Contributions

Supervision, E.H.H.; conceptualization, N.A.S. and M.M.J.; methodology, E.H.H. and H.N.H.; software, E.H.H. and H.N.H.; validation, N.A.S. and M.M.J.; formal analysis, E.H.H. and H.N.H.; investigation, E.H.H. and H.N.H.; resources, N.A.S. and M.M.J.; data curation, N.A.S. and M.M.J.; writing—original draft preparation, E.H.H., H.N.H., N.A.S. and M.M.J.; writing—review and editing, E.H.H., H.N.H., N.A.S. and M.M.J.; visualization, E.H.H. and H.N.H.; funding acquisition, N.A.S. All authors have read and agreed to the published version of the manuscript.

Institutional Review Board Statement

This article does not contain any studies with human participants or animals performed by any authors.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data sharing is not applicable to this article, as no datasets were generated or analyzed during the current study.

Conflicts of Interest

The authors declare that there are no conflict of interest.

Funding Statement

This research was funded by Princess Nourah bint Abdulrahman University Researchers Supporting Project number (PNURSP2023R104), Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia.

Footnotes

Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

References

  • 1.Li S., Wu X., Tan M. Gene selection using hybrid particle swarm optimization and genetic algorithm. Soft. Comput. 2008;12:1039–1048. doi: 10.1007/s00500-007-0272-x. [DOI] [Google Scholar]
  • 2.Alba E., Garcia-Nieto J., Jourdan L., Talbi E.G. Gene selection in cancer classification using PSO/SVM and GA/SVM hybrid algorithms; Proceedings of the 2007 IEEE Congress on Evolutionary Computation; Singapore. 25–28 September 2007; pp. 284–290. [Google Scholar]
  • 3.Alshamlan H.M., Badr G.H., Alohali Y. A study of cancer microarray gene expression profile: Objectives and approaches; Proceedings of the World Congress on Engineering; London, UK. 3–5 July 2013; pp. 1–6. [Google Scholar]
  • 4.Ghorai S., Mukherjee A., Sengupta S., Dutta P.K. Multicategory cancer classification from gene expression data by multiclass NPPC ensemble; Proceedings of the 2010 International Conference on Systems in Medicine and Biology; Kharagpur, India. 16–18 December 2010; pp. 41–48. [Google Scholar]
  • 5.Guo S.B., Lyu M.R., Lok T.M. Gene selection based on mutual information for the classification of multi-class cancer; Proceedings of the International Conference on Intelligent Computing; Kunming, China. 16–19 August 2006; pp. 454–463. [Google Scholar]
  • 6.Alanni R., Hou J., Azzawi H., Xiang Y. A novel gene selection algorithm for cancer classification using microarray datasets. BMC Med. Genom. 2019;12:10. doi: 10.1186/s12920-018-0447-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Su Y., Murali T., Pavlovic V., Schaffer M., Kasif S. RankGene: Identification of diagnostic genes based on expression data. Bioinformatics. 2003;19:1578–1579. doi: 10.1093/bioinformatics/btg179. [DOI] [PubMed] [Google Scholar]
  • 8.Ron K., George H.J. Wrappers for feature subset selection. Artif. Intell. 1997;97:273–324. [Google Scholar]
  • 9.Li X., Rao S., Wang Y., Gong B. Gene mining: A novel and powerful ensemble decision approach to hunting for disease genes using microarray expression profiling. Nucleic Acids Res. 2004;32:2685–2694. doi: 10.1093/nar/gkh563. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Zhao X.M., Cheung Y.M., Huang D.S. A novel approach to extracting features from motif content and protein composition for protein sequence classification. Neural Netw. 2005;18:1019–1028. doi: 10.1016/j.neunet.2005.07.002. [DOI] [PubMed] [Google Scholar]
  • 11.Li L., Jiang W., Li X., Moser K.L., Guo Z., Du L., Wang Q., Topol E.J., Wang Q., Rao S. A robust hybrid between genetic algorithm and support vector machine for extracting an optimal feature gene subset. Genomics. 2005;85:16–23. doi: 10.1016/j.ygeno.2004.09.007. [DOI] [PubMed] [Google Scholar]
  • 12.Li L., Darden T.A., Weingberg C., Levine A., Pedersen L.G. Gene assessment and sample classification for gene expression data using a genetic algorithm/k-nearest neighbor method. Comb. Chem. High Throughput Screen. 2001;4:727–739. doi: 10.2174/1386207013330733. [DOI] [PubMed] [Google Scholar]
  • 13.Alshamlan H.M., Badr G.H., Alohali Y.A. The performance of bio-inspired evolutionary gene selection methods for cancer classification using microarray dataset. Int. J. Biosci. Biochem. Bioinform. 2014;4:166. doi: 10.7763/IJBBB.2014.V4.332. [DOI] [Google Scholar]
  • 14.Wolpert D.H., Macready W.G. No Free Lunch Theorems for Search. Santa Fe Institute; Santa Fe, NM, USA: 1995. Technical Report SFI-TR-95-02-010. [Google Scholar]
  • 15.Ahmadianfar I., Heidari A.A., Gandomi A.H., Chu X., Chen H. RUN beyond the metaphor: An efficient optimization algorithm based on Runge Kutta method. Expert Syst. Appl. 2021;181:115079. doi: 10.1016/j.eswa.2021.115079. [DOI] [Google Scholar]
  • 16.Alshamlan H., Badr G., Alohali Y. Proceedings of the First International Conference on Advanced Data and Information Engineering (DaEng-2013) Springer; Berlin/Heidelberg, Germany: 2014. A comparative study of cancer classification methods using microarray gene expression profile; pp. 389–398. [Google Scholar]
  • 17.Guyon I., Weston J., Barnhill S., Vapnik V. Gene selection for cancer classification using support vector machines. Mach. Learn. 2002;46:389–422. doi: 10.1023/A:1012487302797. [DOI] [Google Scholar]
  • 18.Huerta E.B., Duval B., Hao J.K. A hybrid GA/SVM approach for gene selection and classification of microarray data; Proceedings of the Workshops on Applications of Evolutionary Computation; Budapest, Hungary. 10–12 April 2006; pp. 34–44. [Google Scholar]
  • 19.Houssein E.H., Sayed A. Dynamic Candidate Solution Boosted Beluga Whale Optimization Algorithm for Biomedical Classification. Mathematics. 2023;11:707. doi: 10.3390/math11030707. [DOI] [Google Scholar]
  • 20.Mirjalili S., Lewis A. The whale optimization algorithm. Adv. Eng. Softw. 2016;95:51–67. doi: 10.1016/j.advengsoft.2016.01.008. [DOI] [Google Scholar]
  • 21.Karaboga D. An Idea Based on Honey Bee Swarm for Numerical Optimization. Department of Computer Engineering, Engineering Faculty, Erciyes University; Kayseri, Turkey: 2005. Technical Report-tr06. [Google Scholar]
  • 22.Heidari A.A., Mirjalili S., Faris H., Aljarah I., Mafarja M., Chen H. Harris hawks optimization: Algorithm and applications. Future Gener. Comput. Syst. 2019;97:849–872. doi: 10.1016/j.future.2019.02.028. [DOI] [Google Scholar]
  • 23.Yang Y., Chen H., Heidari A.A., Gandomi A.H. Hunger games search: Visions, conception, implementation, deep analysis, perspectives, and towards performance shifts. Expert Syst. Appl. 2021;177:114864. doi: 10.1016/j.eswa.2021.114864. [DOI] [Google Scholar]
  • 24.Chopra N., Ansari M.M. Golden jackal optimization: A novel nature-inspired optimizer for engineering applications. Expert Syst. Appl. 2022;198:116924. doi: 10.1016/j.eswa.2022.116924. [DOI] [Google Scholar]
  • 25.Zhao W., Zhang Z., Wang L. Manta ray foraging optimization: An effective bio-inspired optimizer for engineering applications. Eng. Appl. Artif. Intell. 2020;87:103300. doi: 10.1016/j.engappai.2019.103300. [DOI] [Google Scholar]
  • 26.Alshamlan H.M., Badr G.H., Alohali Y.A. Abc-svm: Artificial bee colony and svm method for microarray gene selection and multi class cancer classification. Int. J. Mach. Learn. Comput. 2016;6:184. doi: 10.18178/ijmlc.2016.6.3.596. [DOI] [Google Scholar]
  • 27.Gao L., Ye M., Lu X., Huang D. Hybrid method based on information gain and support vector machine for gene selection in cancer classification. Genom. Proteom. Bioinform. 2017;15:389–395. doi: 10.1016/j.gpb.2017.08.002. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Chang C.C., Lin C.J. LIBSVM: A library for support vector machines. ACM Trans. Intell. Syst. Technol. (TIST) 2011;2:1–27. doi: 10.1145/1961189.1961199. [DOI] [Google Scholar]
  • 29.Wang Y., Tetko I.V., Hall M.A., Frank E., Facius A., Mayer K.F., Mewes H.W. Gene selection from microarray data for cancer classification—A machine learning approach. Comput. Biol. Chem. 2005;29:37–46. doi: 10.1016/j.compbiolchem.2004.11.001. [DOI] [PubMed] [Google Scholar]
  • 30.Zhang L., Zhou W., Wang B., Zhang Z., Li F. Applying 1-norm SVM with squared loss to gene selection for cancer classification. Appl. Intell. 2018;48:1878–1890. doi: 10.1007/s10489-017-1056-3. [DOI] [Google Scholar]
  • 31.Mohamad M.S., Omatu S., Deris S., Yoshioka M. Particle swarm optimization for gene selection in classifying cancer classes. Artif. Life Robot. 2009;14:16–19. doi: 10.1007/s10015-009-0712-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Almugren N., Alshamlan H. FF-SVM: New FireFly-based gene selection algorithm for microarray cancer classification; Proceedings of the 2019 IEEE Conference on Computational Intelligence in Bioinformatics and Computational Biology (CIBCB); Siena, Italy. 9–11 July 2019; pp. 1–6. [Google Scholar]
  • 33.Kavitha K., Prakasan A., Dhrishya P. Score-based feature selection of gene expression data for cancer classification; Proceedings of the 2020 Fourth International Conference on Computing Methodologies and Communication (ICCMC); Erode, India. 11–13 March 2020; pp. 261–266. [Google Scholar]
  • 34.El Kafrawy P., Fathi H., Qaraad M., Kelany A.K., Chen X. An efficient SVM-based feature selection model for cancer classification using high-dimensional microarray data. IEEE Access. 2021;9:155353–155369. doi: 10.1109/ACCESS.2021.3123090. [DOI] [Google Scholar]
  • 35.Qasem S.N., Saeed F. Hybrid feature selection and ensemble learning methods for gene selection and cancer classification. Int. J. Adv. Comput. Sci. Appl. 2021;12 doi: 10.14569/IJACSA.2021.0120225. [DOI] [Google Scholar]
  • 36.Chaudhuri A., Sahu T.P. A hybrid feature selection method based on Binary Jaya algorithm for micro-array data classification. Comput. Electr. Eng. 2021;90:106963. doi: 10.1016/j.compeleceng.2020.106963. [DOI] [Google Scholar]
  • 37.Kira K., Rendell L.A. The feature selection problem: Traditional methods and a new algorithm; Proceedings of the Tenth National Conference on Artificial Intelligence; San Jose, CA, USA. 12–16 July 1992; pp. 129–134. [Google Scholar]
  • 38.Kononenko I. Estimating attributes: Analysis and extensions of RELIEF; Proceedings of the European Conference on Machine Learning; Catania, Italy. 6–8 April 1994; pp. 171–182. [Google Scholar]
  • 39.Rodríguez-Peérez R., Vogt M., Bajorath J. Support vector machine classification and regression prioritize different structural features for binary compound activity and potency value prediction. ACS Omega. 2017;2:6371–6379. doi: 10.1021/acsomega.7b01079. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40.Wang X., Gotoh O. Microarray-based cancer prediction using soft computing approach. Cancer Inform. 2009;7:CIN-S2655. doi: 10.4137/CIN.S2655. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41.Statnikov A., Tsamardinos I., Dosbayev Y., Aliferis C.F. GEMS: A system for automated cancer diagnosis and biomarker discovery from microarray gene expression data. Int. J. Med. Inform. 2005;74:491–503. doi: 10.1016/j.ijmedinf.2005.05.002. [DOI] [PubMed] [Google Scholar]
  • 42.Bolón-Canedo V., Sánchez-Marono N., Alonso-Betanzos A., Benítez J.M., Herrera F. A review of microarray datasets and applied feature selection methods. Inf. Sci. 2014;282:111–135. doi: 10.1016/j.ins.2014.05.042. [DOI] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

Data sharing is not applicable to this article, as no datasets were generated or analyzed during the current study.


Articles from Diagnostics are provided here courtesy of Multidisciplinary Digital Publishing Institute (MDPI)

RESOURCES