ABSTRACT
In recent years, tumor classification based on gene expression profiles has drawn great attention, and related research results have been widely applied to the clinical diagnosis of major gene diseases. These studies are of tremendous importance for accurate cancer diagnosis and subtype recognition. However, the microarray data of gene expression profiles have small samples, high dimensionality, large noise and data redundancy. To further improve the classification performance of microarray data, a gene selection approach based on the Fisher linear discriminant (FLD) and the neighborhood rough set (NRS) is proposed. First, the FLD method is employed to reduce the preliminarily genetic data to obtain features with a strong classification ability, which can form a candidate gene subset. Then, neighborhood precision and neighborhood roughness are defined in a neighborhood decision system, and the calculation approaches for neighborhood dependency and the significance of an attribute are given. A reduction model of neighborhood decision systems is presented. Thus, a gene selection algorithm based on FLD and NRS is proposed. Finally, four public gene datasets are used in the simulation experiments. Experimental results under the SVM classifier demonstrate that the proposed algorithm is effective, and it can select a smaller and more well-classified gene subset, as well as obtain better classification performance.
KEYWORDS: Gene selection, Fisher linear discriminant, neighborhood rough set, reduction
Introduction
With the development of gene expression profiles, the analysis and modelling of gene expression profiles has become an important topic in the field of bioinformatics research [1–3]. However, the high dimension of tumor gene expression data, which is often in the thousands or even tens of thousands, increases the learning cost and deteriorates learning performance. This is widely known as the ‘‘Curse of Dimensionality’’, which costs time and reduces the effectiveness of classification when using a classifier to forecast new samples [4,5]. Thus, the dimensionality reduction has been a research hotspot in different fields as an important step in pattern recognition, machine learning and data mining [6-14].
In general, dimensionality reduction algorithms can be categorized as feature extraction and feature selection. Feature extraction constructs a new low-dimensional space out of the original high-dimensional data through projection or transformation, while the aim of feature selection is to reduce the dimensionality of microarray data[15,16] and to enhance classification accuracy [17,18]. The existing feature selection methods can be broadly categorized into the following three classes: filter, wrapper, and hybrid [19]. A good feature selection algorithm should be reasonable and efficient; the algorithm should be able to find a typical genome containing fewer genes [20].
Many scholars have conducted research on gene selection and have generated many results. FLD is a classical technique in pattern recognition; Robert Fisher first developed FLD in 1936 for taxonomic classification [21]. FLD can be used to select the characteristics possessed by classified information, eliminate redundant attributes, and achieve the dimensionality reduction processing of gene data. Since Pawlak in the early 1980s proposed rough set theory, it has been widely used in various fields [22]. However, the classical rough set theory is only applicable to discrete-valued information systems, and it is not suitable for real-valued datasets. To overcome this weakness, Dai and Xu presented a gene selection method based on fuzzy rough sets and a fuzzy gain ratio [23]. Hu et al. proposed a neighborhood rough set model to address both discrete and continuous data sets with a δ-neighborhood parameter, which can maintain the rich information for classifying the data sets [24].
To further improve the classification performance of microarray data, effectively remove the redundant gene, and reduce the computational time complexity of the gene selection algorithm, the FLD method is employed to conduct the preliminary dimensionality reduction for microarray gene data. FLD effectively removes genes that do not contribute to classification. The neighborhood rough set can process continuous data sets and avoid the loss of information caused by discretizing. Then, a new neighborhood dependency and its attribute significance are given, and an attribute reduction method of neighborhood decision systems is presented. A gene selection approach based on FLD and NRS is proposed. A number of simulation experiments were conducted on public gene data sets, and the best parameters were determined according to the experimental results. Therefore, high classification accuracy can be obtained using the selected gene subset under the support vector machine (SVM) classifier [25].
The remainder of this paper is structured as follows: Section 2 introduces related concepts of FLD. An effective and efficient feature selection method based on FLD and NRS is given in Section 3. To evaluate the performance of the proposed algorithm, five related algorithms are employed to compare four public gene expression data sets. The experimental results are described in Section 4. Finally, the conclusion is drawn in Section 5.
Fisher linear discriminant model
The Fisher linear discriminant is a classical algorithm introduced by Belhumeur in the field of pattern recognition and artificial intelligence [21]. The basic idea of the FLD model is to project the sample onto a straight line by transforming the sample so the projection of the sample can be best divided. That is, the dispersion degree between the transformed sample classes reaches the highest level, and the sample dispersion within the classes reaches the minimum, which increases the distinction among the categories. Therefore, FLD can be used to select the characteristics with the possessed information classified, eliminate redundant attributes, and achieve the processing of dimensionality reduction for gene data. The method is an effective, supervised dimensionality reduction technology. The related concepts of FLD are described as follows.
Let c be the number of classes of the sample matrix , where ni is the number of samples belonging to the i-th class ωi, and . The centre point of each sample is , and the centre point of all samples is , where xj is the j-th sample. The between-class scatter matrix SB and the within-class matrix SW can be expressed, respectively, as
(1.1) |
(1.2) |
On the basis of Formulas (1.1) and (1.2), the between-class scatter JB and the within-class scatter JW of the samples after projection are expressed, respectively, as
(1.3) |
(1.4) |
The objective function established by the Fisher discriminant criterion is described by
(1.5) |
If the k-th column wk of W is considered, the objective function can be transformed into
(1.6) |
A Lagrangian equation is established as
(1.7) |
Take the derivative of wk, and make it equal to 0, to obtain the following formula:
(1.8) |
To maximize the value of , the projection matrix W can be constructed by simply taking the eigenvectors corresponding to the k largest eigenvalues.
Fisher linear discriminant and neighborhood rough set based gene selection method
When using classical rough sets to solve continuous data problems, the data set must be discretized; however, processing the original properties of the data will change, and some useful information will be lost [26]. The neighborhood rough set is proposed to solve the problem where the classical rough set cannot handle the numerical attributes [27,28]. In addition, the effect of the classical neighborhood rough set model is not obvious. Then, to resolve this issue, this paper proposes a feature selection method based on FLD and NRS, which is applied to gene selection of a cancer data set.
There are N dimensions in a determined real space U. Let Δ = RN×RN → R. Δ is called as a measure on RN, and (Δ, U) is called as a measure space, when Δ meets the following three conditions:
-
(1)
Δ(x1, x2) = Δ(x2, x1),
-
(2)
Δ(x1, x2) ≥ 0, where the equation holds if and only if x1 = x2,
-
(3)
Δ(x1, x3) ≤ Δ(x1, x2) + Δ(x2, x3),
where Δ(x1, x2) is a distance function between two elements x1 and x2. The distance functions used always include a Manhattan distance function, a Euclidean distance function, and a p-normal form distance function. Since the Euclidean distance function can reflect the basic situation of unknown data [29]. the Euclidean distance function is used in this paper. The formula is described as follows:
(2.1) |
Let U = {x1, x2, x3,…, xm} be a nonempty finite set on a given real space Ω, then the β-neighborhood of any xi (1 ≤ i ≤ m) is defined as
(2.2) |
Let U = {x1, x2, x3,…, xm} be a nonempty finite set on a given real space Ω, and its neighborhood relationship N on the real field Ω is expressed as a binary group NA = (U, N). For any XU, the upper approximation and the lower approximations of X in a neighborhood approximate space NA = (U, N) can be defined respectively as
(2.3) |
(2.4) |
The approximate boundary region of X is defined as
(2.5) |
Suppose NDS = (U, A∪D) is a neighborhood decision system, A is a conditional attribute set, D is a decision attribute, and U/D = {X1, X2, X3, …, Xn}. For any conditional attribute subset B A, the upper approximation and the lower approximation of decision attribute D with respect to B are expressed, respectively, as
(2.6) |
(2.7) |
It follows that the boundary region of the decision system can be expressed as
(2.8) |
where is a positive domain of the decision system, is a negative domain of the decision system.
The existence of the boundary domain causes the uncertainty of the set. Greater uncertainty occurs with larger boundary domain sets. This paper studies the boundary domain of the neighborhood decision system and investigates various uncertainty measures.
The roughness measure, a quantitative index for processing uncertain information by using the rough set theory, is the basis of resource management, system optimization, and many other decision-making problems [30].
Suppose NDS = (U, A∪D) is a neighborhood decision system, U/D = {X1, X2, X3,…, Xn}. Then for any conditional attribute subset B A, the neighborhood precision of U/D with respect to B is described by
(2.9) |
The neighborhood roughness of U/D with respect to B is expressed as
(2.10) |
Definition 1.
Suppose that NDS = (U, A∪D) is a neighborhood decision system and any conditional attribute subset BA. Then, a dependency of decision attribute D with respect to B is defined as
(2.11)
Definition 2.
Suppose that NDS = (U, A∪D) is a neighborhood decision system, any conditional attribute subset BA, and a∈B. Then, an internal significance of attribute a with respect to B is defined as
(2.12)
Definition 3.
Suppose that NDS = (U, A∪D) is a neighborhood decision system, any conditional attribute subset BA, and a∈A−B. Then, an external significance of attribute a with respect to B is defined as
(2.13)
Definition 4.
Suppose that NDS = (U, A∪D) is a neighborhood decision system, any conditional attribute subset BA, and a∈B. Then, B is a reduction set of A if and only if it is satisfied with the following conditions:
- (1)
K(B, D) = K(A, D),
- (2)
K(B, D) > K(B−{a}, D),
where K(B, D) is the dependency of the decision attribute D with respect to the conditional attribute subset B.
Based on the dimensionality reduction technology FLD and the feature reduction of the neighborhood decision system, a feature selection algorithm based on FLD and NRS (FLD-NRS) is designed. The detailed steps are described as follows.
Input: A neighborhood decision system NDS = (U, A∪D), and a neighborhood radius β
Output: A reduction set red
Step 1: Calculate a centre point μi of the various samples for the conditional attribute set A and the centre point μ of all the samples.
Step 2: Calculate the within-class scatter matrix SW and the between-class scatter matrix SB according to Formulas (1.3) and (1.4).
Step 3: Decompose the eigenvalues of obtained from Formula (1.8) and sort the eigenvalues with descending order.
Step 4: Take the eigenvector corresponding to the first k eigenvalues to form the projection matrix W.
Step 5: Calculate X′ = WTX, X′∈ and obtain the conditional attribute subset C after dimensionality reduction that is NDS = (U, C∪D), where d' is the number of attributes of C, and n is the number of samples.
Step 6: Let Ø → red, and B = C.
Step 7: Calculate SIGinner(a, B, D) > 0 with Formula (2.12) for any attribute a∈B and ared, get the indispensable attribute a, and let red = red∪{a}.
Step 8: Calculate SIGouter(ak, red, D) with Formula (2.13) for any attribute ak∈C−red, get the most important attribute ak according to the size of the order and add it to the reduction set red = red∪{ak}.
Step 9: Calculate K(red, D) and K(C, D) according to Formula (2.11).
Step 10: If K(red, D) ≠ K(C, D), update the conditional attribute set C = C−{ak}, and perform Step 8.
Step 11: Output the reduction set red.
Step 12: End.
For a group of gene expression data, it is assumed that the number of samples is K, and the number of attributes is T. After the dimensionality reduction according to the FLD algorithm, the M genes can be obtained. To select a gene, it is necessary to add samples to the positive domain set, and the neighborhood computational time complexity of the gene data set is O(KlogK). Since the computational time complexity of the first gene is TKlogK, and the computational time complexity of the second gene is (T−1)(K−)log(K−), the computational time complexity of the M-th gene is (T−M+1)loglog. After the above analysis, the worst computational time complexity of the FLD-NRS algorithm is MTKlogK. Due to M ≪ T, the computational time complexity of the proposed algorithm is less than O(T2KlogK).
Experimental results and analysis
To verify the effectiveness of the proposed FLD-NRS algorithm, simulation experiments are performed on four public gene expression profile data sets, which include colon, leukaemia, lung, and prostate cancer data downloaded from http://bioinformatics.rutgers.ed/Static/ Supplemens/CompCancer/datasets. The specific description of the datasets is shown in Table 1. The computer system used in this experiment is Windows 7 64-bit operating system, Intel(R) Core(TM) i5-3470 CPU @ 3.20 GHz, Memory (RAM) 4.00 GB. All simulation experiments are implemented in Matlab R2012b (The MathWorks, Inc., 1 Apple Hill Drive Natick, United States).
Table 1.
Data set | Feature size | Sample size (normal/tumor) | Class size |
---|---|---|---|
Colon | 2000 | 62 (40/20) | 2 |
Leukemia | 7129 | 72 (25/47) | 2 |
Lung | 12533 | 181 (31/150) | 2 |
Prostate | 12600 | 136 (77/59) | 2 |
It is noted that the values of the partial gene columns in the lung and prostate data sets are all zero. Thus, the 121 columns of noise gene data from the lung cancer set and the 394 columns of noise gene data from the prostate set should be eliminated. Finally, the gene number of the lung cancer data set is 12412, and the gene number of the prostate cancer data set is 12206.
In Table 1, the four data sets have two categories, namely, belonging to two classification problems. Taking the colon cancer data set with a high dimension and a small sample as an example, there are 2000 conditional attributes and 62 samples. The number of positive samples is 40, and the number of negative ones is 20. Since the external manifestation of the gene data is a numerical matrix, the model described needs to name the gene dataset as a high-dimensional data matrix, and then the dimensionality needs to be reduced. To ensure that the gene data cannot lose its characteristics, the data of each gene is then marked. That is, the numbers between genes and markers in the data set are corresponding to each other.
In this paper, FLD is employed to preliminary dimensionality reduction, and the neighborhood rough set algorithm is used to further reduce the attributes, which can remove the redundant data of the original data set. The effect of dimensionality reduction then becomes obvious. The neighborhood radius parameter λ is set for each data set, and the lower limit of importance is 0.00001. The FLD-NRS algorithm is used to reduce the attributes for the four data sets respectively in Table 1. The experimental results of the selected gene subsets are shown in Table 2.
Table 2.
Data set | Gene subset after reduction |
---|---|
Colon | {1423, 765, 822, 66, 1870, 590} |
Leukemia | {1834, 2354, 2642, 1685, 758} |
Lung | {2549, 7200, 6139} |
Prostate | {8986, 11052, 6392, 4050} |
To verify the classification performance of the selected gene subsets, four classifiers are employed to do this experiment on each data set. The results are indicated in Figure 1.
According to Fig. 1, SVM has the best classification performance on the four data sets when compared with the other three. Then, to verify the validity of the proposed algorithm for selecting a gene subset with strong classification, the classification accuracy of the gene subset after reduction is evaluated on SVM. The FLD-NRS algorithm is compared with the other three related algorithms on four gene data sets, where the original data processing (ODP) algorithm is used to classify the original data set directly. The Lasso[31,32] algorithm is a feature selection method by coefficient compression estimation, and the NRS [24]. algorithm is a feature selection method using the neighborhood rough set theory. The experimental results are illustrated in Table 3, where m describes the gene number after gene selection, and Acc describes the optimal classification accuracy. Meanwhile, the time complexities of these algorithms are given in Table 3.
Table 3.
ODP |
Lasso |
NRS |
FLD-NRS |
|||||
---|---|---|---|---|---|---|---|---|
Data set | m | Acc | M | Acc | m | Acc | m | Acc |
Colon | 2000 | 0.811 | 5 | 0.887 | 4 | 0.611 | 6 | 0.880 |
Leukemia | 7129 | 0.944 | 23 | 0.986 | 5 | 0.645 | 5 | 0.828 |
Lung | 12412 | 0.903 | 8 | 0.995 | 3 | 0.641 | 3 | 0.889 |
Prostate | 12206 | 0.619 | 63 | 0.961 | 4 | 0.647 | 4 | 0.800 |
Time complexity | — | O(PT3) | O(T2KlogK) | O(MTKlogK) |
It can be seen from Table 3 that, although the classification accuracy of the leukemia data set is 94.4% with the ODP algorithm, the original data is directly classified by using ODP, and the size of the selected gene subset is very large. The NRS algorithm can effectively remove irrelevant genes to obtain a smaller gene subset. However, some genes with strong classifications have been removed. This leads to the lower classification accuracy of the selected gene subset. For example, the classification accuracy of leukemia with NRS is reduced to 64.5%. In the process of gene selection, it is known that the scale of the selected gene subset and its classification accuracy are two important aspects. The FLD-NRS algorithm presented in this paper can select a smaller gene subset, and the classification accuracy has clearly been improved. For the colon data set, our algorithm can also select fewer genes with higher classification accuracy than the other two algorithms. For the leukemia and lung data sets, although our accuracy is lower than those of the ODP and Lasso algorithms, the selected gene subset is much smaller than those of the above two algorithms. Meanwhile, for the prostate data set, the classification accuracy with FLD-NRS is higher than those of the ODP and NRS algorithms, and the selected gene number is smaller than those of the ODP and Lasso algorithms. These results prove the effectiveness of the proposed algorithm for gene selection.
To further investigate the performance of the proposed algorithm, the FLD-NRS algorithm is compared with two random forest algorithms, where RF represents the classical random forest algorithm [33]. and SNRRF [34]. represents an improved random forest algorithm. The time complexity of the random forest algorithms can be approximated as O(kTK(logK)2), where k is the number of random classifiers in a random forest. The experimental results are shown in Table 4. The time complexities of these algorithms can be found in Table 4.
Table 4.
RF |
SNRRF |
FLD-NRS |
||||
---|---|---|---|---|---|---|
Data set | m | Acc | m | Acc | m | Acc |
Colon | 2000 | 0.848 | 72 | 0.875 | 6 | 0.880 |
Leukemia | 7129 | 0.902 | 26 | 0.948 | 5 | 0.828 |
Lung | 2880 | 0.864 | 10 | 0.899 | 3 | 0.889 |
Prostate | 12600 | 0.925 | 49 | 0.931 | 4 | 0.800 |
Time complexity | O(kTK(logK)2) | O(kTK(logK)2) | O(MTKlogK) |
According to Table 4, the classification accuracy of the colon data set with the FLD-NRS algorithm is 88%, which is higher than those of the RF and SNRRF algorithms. For the lung data set, the accuracy of the proposed algorithm is 88.9%, which is basically equivalent to those of the two random forest algorithms, but the selected gene number is very small. These results demonstrate the validity of the proposed algorithm. However, for the leukemia and prostate data sets, the classification accuracy of this algorithm is slightly lower than those of the two random forest algorithms. These results explain that when using FLD to filter irrelevant genes, the genes with large influence on classification are mistakenly filtered out; therefore, the classification accuracy will be affected and reduced.
Through the time complexity analyses presented in Tables 3 and 4, it is obvious that the Lasso algorithm costs significantly more time, which is higher than those of the other four algorithms; although, the classification accuracy of the selected gene subset is high. The gene number of the original data set is usually much larger than that of the selected gene subset, so the time complexity of the proposed algorithm is obviously lower than those of the other five algorithms.
The above experimental results show that the FLD-NRS algorithm can solve the high-dimensional and high-redundancy problem of gene expression profile data well. The selected gene subset is smaller, and the dimensionality reduction effect is obvious. Hence, the FLD-NRS algorithm is superior to the other four algorithms mentioned in this paper under the overall situation of the three indicators, including selected gene number, classification accuracy, and computational time complexity. Therefore, the FLD-NRS algorithm can accomplish dimensionality reduction processing well, and the selected gene subset has strong classification abilities.
Conclusion
The challenge of selecting genes with important classification information from tens of thousands of gene expression profiles is an important problem in the field of bioinformatics. In this paper, a genetic selection method based on the FLD and NRS is proposed in view of poor stability, large feature subset size, and time-consuming calculations of various gene selection algorithms. The FLD approach is applied into the preliminary dimensionality reduction of gene data to obtain the candidate gene subset. A novel feature reduction algorithm in neighborhood decision systems is proposed to optimize the features after dimensionality reduction. Then, a gene subset with strong classification ability is selected. The experimental results all show that the FLD-SNR algorithm can select a gene subset with smaller scale and stronger classification ability. The proposed algorithm is of great practical significance for the future study of cancer clinical diagnosis.
Funding Statement
China Postdoctoral Science Foundation, 2016M602247, Founf for Youth Key Teachers of Henan Normal University of China, qd15132, Key Research Project of High School of Henan Province of China, 14A520069, Key Scientific and Technological Project of Xinxiang City of China, CXGG17002, National Natural Science Foundation of China (NSFC), 61772176, 61402153, 61370169, 61502319,U1604154 Key Project of Science and Technology Development of Henan Provience of China, 162102210261
Disclosure of potential conflicts of interest
No potential conflicts of interest were disclosed.
References
- [1].Liu F, Wang LB. Biclustering of time-lagged gene expression data using real number. J Biomed Sci Eng. 2010;3:217–220. doi: 10.4236/jbise.2010.32029. [DOI] [Google Scholar]
- [2].Liu F. Time-lagged co-expression gene analysis based on biclustering technology. Biotechnol Biotec Eq. 2013;27:4031–4039. doi: 10.5504/BBEQ.2013.0058. [DOI] [Google Scholar]
- [3].Sun L, Xu JC, Yin Y. Principal component-based feature selection for tumor classification. Bio-Med Mater Eng. 2015;26:S2011–S2017. [DOI] [PubMed] [Google Scholar]
- [4].Liu HQ, Li JY, Wong L. A comparative study on feature selection and classification methods using gene expression pro-files and proteomic patterns. Genome Inform. 2002;13:51–60. [PubMed] [Google Scholar]
- [5].Xu JC, Sun L, Gao YP. An ensemble feature selection technique for cancer recognition. Bio-Med Mater Eng. 2014;24:1001–1008. [DOI] [PubMed] [Google Scholar]
- [6].Chakraborty B, Chakraborty G. A new feature extraction technique for on-line recognition of handwritten alphanumeric characters. Inform Sciences. 2002;148:55–70. doi: 10.1016/S0020-0255(02)00276-1. [DOI] [Google Scholar]
- [7].Lotlikar R, Kothari R. Adaptive linear dimensionality reduction for classification. Pattern Recogn. 2000;33:185–194. doi: 10.1016/S0031-3203(99)00053-9. [DOI] [Google Scholar]
- [8].Sun YJ. Iterative RELIEF for feature weighting: algorithms, theories, and applications. IEEE T Pattern Anal. 2007, 29:1035–1051. doi: 10.1109/TPAMI.2007.1093. [DOI] [PubMed] [Google Scholar]
- [9].Chen HM, Li TR, Cai Y, et al.. Parallel attribute reduction in dominance-based neighborhood rough set. Inform Sciences. 2016;373:351–368. doi: 10.1016/j.ins.2016.09.012. [DOI] [Google Scholar]
- [10].Banka H, Dara S. A Hamming distance based binary particle swarm optimization (HDBPSO) algorithm for high dimensional feature selection, classification and validation. Pattern Recogn Lett. 2015;52:94–100. doi: 10.1016/j.patrec.2014.10.007. [DOI] [Google Scholar]
- [11].Saeed M, Kashif J, Babri HA. Machine learning using Bernoulli mixture models: Clustering, rule extraction and dimensionality reduction. Neurocomputing. 2013;119:366–374. doi: 10.1016/j.neucom.2013.03.021. [DOI] [Google Scholar]
- [12].Wang CZ, Hu QH, Wang XZ, et al.. Feature selection based on neighborhood discrimination index. IEEE T Neural Networ Learn Syst. 2017. doi: 10.1109/TNNLS.2017.2710422 [DOI] [PubMed] [Google Scholar]
- [13].Bhuyan HK, Kamila NK. Privacy preserving sub-feature selection in distributed data mining. Appl Soft Comput. 2015;36:552–569. doi: 10.1016/j.asoc.2015.06.060. [DOI] [Google Scholar]
- [14].Yu L, Liu H. Efficient feature selection via analysis of relevance and redundancy. J Mach Learn Res. 2004;5:1205–1224. [Google Scholar]
- [15].Kuo BC, Li CH, Yang JM. Kernel nonparametric weighted feature extraction for hyperspectral image classification. IEEE T Geosci Remote. 2009;47:1139–1155. [Google Scholar]
- [16].Feng J, Jiao LC, Liu F, et al.. Unsupervised feature selection based on maximum information and minimum redundancy for hyperspectral images. Pattern Recogn. 2016;51:295–309. doi: 10.1016/j.patcog.2015.08.018. [DOI] [Google Scholar]
- [17].Tabakhi S, Najafi A, Ranjbar R, et al.. Gene selection for microarray data classification using a novel ant colony optimization. Neurocomputing. 2015;168:1024–1036. doi: 10.1016/j.neucom.2015.05.022. [DOI] [Google Scholar]
- [18].Sun L, Xu JC, Tian Y. Feature selection using rough entropy-based uncertainty measures in incomplete decision systems. Knowl-Based Syst. 2012;36:206–216. doi: 10.1016/j.knosys.2012.06.010. [DOI] [Google Scholar]
- [19].Senawi A, Wei HL, Billings SA. A new maximum relevance-minimum multicollinearity (MRmMC) method for feature selection and ranking. Pattern Recogn. 2017;67:47–61. doi: 10.1016/j.patcog.2017.01.026. [DOI] [Google Scholar]
- [20].Lu HJ, Chen JY, Yan K, Jin Q, et al.. A hybrid feature selection algorithm for gene expression data classification. Neurocomputing. 2017;256:56–62. doi: 10.1016/j.neucom.2016.07.080. [DOI] [Google Scholar]
- [21].Belhumeur PN, Hespanha JP, Kriegman DJ. Eigenfaces vs. fisherfaces: Recognition using class specific linear projection. IEEE T Pattern Anal. 2002;19:711–720. doi: 10.1109/34.598228. [DOI] [Google Scholar]
- [22].Pawlak Z. Rough sets: Theoretical aspects of reasoning about data. Dordrecht & Boston: Kluwer Academic Publishers; 1991. [Google Scholar]
- [23].Dai JH, Xu Q. Attribute selection based on information gain ratio in fuzzy rough set theory with application to tumor classification. Appl Soft Comput. 2013;13:211–221. doi: 10.1016/j.asoc.2012.07.029. [DOI] [Google Scholar]
- [24].Hu QH, Yu DR, Liu JF, et al.. Neighborhood rough set based heterogeneous feature subset selection. Inform Sciences. 2008;178:3577–3594. doi: 10.1016/j.ins.2008.05.024. [DOI] [Google Scholar]
- [25].Maldonado S, Pérez J, Bravo C. Cost-based feature selection for Support Vector Machines: An application in credit scoring. Eur J Oper Res. 2017;261:656–665. doi: 10.1016/j.ejor.2017.02.037. [DOI] [Google Scholar]
- [26].Wang M, Min F, Wu YX, et al.. Active learning through density clustering. Expert Syst Appl. 2017;85:305–317. doi: 10.1016/j.eswa.2017.05.046. [DOI] [Google Scholar]
- [27].Sun L, Xu JC. Feature selection using mutual information based uncertainty measures for tumor classification. Bio-Med Mater Eng. 2014;24:763–770. [DOI] [PubMed] [Google Scholar]
- [28].Xu JC, Li T, Sun L, Li YH. Feature gene selection based on SNR and neighborhood rough set. J. Data Acquis and Proc. 2015;30:973–981. [Google Scholar]
- [29].Liu Y, Huang WL, Jiang YL, et al.. Quick attribute reduct algorithm for neighborhood rough set model. Inform Sciences. 2014;271:65–81. [Google Scholar]
- [30].Li FC, Yang JN, Jin CX, et al.. A new effect-based roughness measure for attribute reduction in information system. Inform Sciences. 2017;378:348–362. doi: 10.1016/j.ins.2016.08.056. [DOI] [Google Scholar]
- [31].Ma SG, Song X, Huang J. Supervised group Lasso with applications to microarray data analysis. BMC Bioinformatics. 2007;8:60. doi: 10.1186/1471-2105-8-60. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [32].Zhang J, Hu XG, Li PP, et al.. Informative Gene Selection for Tumor Classification Based on Iterative Lasso. Pattern Recogn Artif Intell. 2014;27:49–59. [Google Scholar]
- [32].Strobl C, Boulesteix AL, Kneib T, et al.. Conditional variable importance for random forests. BMC Bioinformatics. 2008;9:307. doi: 10.1186/1471-2105-9-307. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [34].Xu JC, Feng S, Mu HY. Tumor feature gene selection based on SNR and random forest. J. Henan Normal Univ (Nat Sci Ed). 2017;45:87–92. [Google Scholar]