Abstract
Background
Lung cancer is one of the most common types of cancer, among which lung adenocarcinoma accounts for the largest proportion. Currently, accurate staging is a prerequisite for effective diagnosis and treatment of lung adenocarcinoma. Previous research has used mainly single-modal data, such as gene expression data, for classification and prediction. Integrating multi-modal genetic data (gene expression RNA-seq, methylation data and copy number variation) from the same patient provides the possibility of using multi-modal genetic data for cancer prediction. A new machine learning method called gcForest has recently been proposed. This method has been proven to be suitable for classification in some fields. However, the model may face challenges when applied to small samples and high-dimensional genetic data.
Results
In this paper, we propose a multi-weighted gcForest algorithm (MLW-gcForest) to construct a lung adenocarcinoma staging model using multi-modal genetic data. The new algorithm is based on the standard gcForest algorithm. First, different weights are assigned to different random forests according to the classification performance of these forests in the standard gcForest model. Second, because the feature vectors generated under different scanning granularities have a diverse influence on the final classification result, the feature vectors are given weights according to the proposed sorting optimization algorithm. Then, we train three MLW-gcForest models based on three single-modal datasets (gene expression RNA-seq, methylation data, and copy number variation) and then perform decision fusion to stage lung adenocarcinoma. Experimental results suggest that the MLW-gcForest model is superior to the standard gcForest model in constructing a staging model of lung adenocarcinoma and is better than the traditional classification methods. The accuracy, precision, recall, and AUC reached 0.908, 0.896, 0.882, and 0.96, respectively.
Conclusions
The MLW-gcForest model has great potential in lung adenocarcinoma staging, which is helpful for the diagnosis and personalized treatment of lung adenocarcinoma. The results suggest that the MLW-gcForest algorithm is effective on multi-modal genetic data, which consist of small samples and are high dimensional.
Keywords: MLW-gcForest staging model multi-modal genetic data lung adenocarcinoma
Introduction
Lung cancer is one of the most common cancers and possesses the highest morbidity and mortality, causing more than 1.4 million deaths each year. Lung cancer can be classified into non-small cell lung cancer (NSCLC) and small cell carcinoma. Lung adenocarcinoma and lung squamous cell carcinoma are common types of NSCLC [1, 2], with lung adenocarcinoma accounting for approximately 70% of NSCLC. Therefore, the study of lung adenocarcinoma is crucial in the study of lung cancer. The 5-year survival rate of lung adenocarcinoma does not exceed 5% [3]. Different treatments are needed during different stages of lung adenocarcinoma to improve the patient’s survival rate. Therefore, the accurate staging of lung adenocarcinoma is the first step in clinical diagnosis and targeted treatment.
With the development of high-throughput sequencing technology, a large number of microarrays and genetic data have been produced. An increasing number of researchers have engaged in the analysis of genetic data. As an important branch of artificial intelligence, machine learning methods are favored by many researchers. Various machine learning methods based on cancer gene data have been widely used in disease prognosis and prediction. Based on the Cancer Genome Atlas (TCGA) and Stanford Tissue Microarray Database, Yu et al. [4] used regularized machine learning methods to select the top quantitative image features and classified patients as having lung adenocarcinoma and squamous cell carcinoma. Cai et al. [5] used machine learning methods to capture unbiased and compact molecular features to classify lung adenocarcinoma, small cell lung cancer, and NSCLC. Li et al. [6] proposed a method that combines support vector machine (SVM) and random forest to predict lung cancer adenocarcinoma stages. Nguyen et al. [7] proposed a multi-class machine learning technique using SVM to classify tumor node metastasis (TNM) staging of lung cancer patients by analyzing their free-text histology reports. Singh et al. [8] used machine learning methods to identify biomarkers and constructed a model to distinguish early and late stages of papillary renal cell carcinoma based on gene expression profiles. Xiao et al. [9] used an ensemble deep neural method comprising five machine learning models to predict cancer. The proposed deep learning-based multi-modal ensemble method achieved better predictive performance than that of any single model. Many machine learning or deep learning algorithms classify or predict cancer by analyzing different types of cancer gene data [10–22]. Many scholars have analyzed and studied lung cancer gene data using other methods [23–27]; however, to the best of our knowledge, few studies have applied machine learning to the staging of lung adenocarcinoma based on multi-modal genetic data.
Multi-modal genetic data mainly include gene expression RNAseq (RNA-seq), methylation data, and copy number variation (CNV) from the same patients, which are usually characterized by small sample sizes and high dimensionality. The construction of deep neural networks typically relies on a large amount of data: small samples and high-dimensional genetic data increase the risk of overfitting deep neural networks during training. A deep forest model called gcForest [28] was recently proposed as an alternative method of deep learning to alleviate the overfitting problem of deep neural networks for small samples.
The gcForest model is a new decision tree integration of the deep forest method. The algorithm is a combination of traditional machine learning algorithms and deep learning ideas. The gcForest model implements multi-grained scanning to further enhance the learning ability and can achieve good performance in high-dimensional, small-scale data. The gcForest model adopts a cascade structure in which each layer receives information processed by the previous layer and transmits information to the next layer. The standard gcForest algorithm is expected to deliver better predictions than those of traditional machine learning methods, even in cases of small-scale training data [28]. However, the gcForest algorithm still has the following shortcomings in the analysis of cancer genetic data: (1) the multi-grained scanning of the gcForest algorithm does not account for the different effects of each random forest on the final prediction, which is not conducive to capturing diverse features, especially in small-sample data; and (2) the class vectors obtained under different scanning granularities have different effects on the final classification decision-making ability, but the standard gcForest algorithm simply concatenates the class vectors from different granularity sliding windows, which potentially weakens the final classification ability.
In our previous work, an improved gcForest algorithm based on methylation data [29] was proposed and successfully applied to the subtype classification of cancer. In this paper, we have substantially revised and applied to the staging model of lung adenocarcinoma.
The main contribution of our approach is the proposal of MLW-gcForest to construct a staging model of lung adenocarcinoma based on multi-modal genetic data. Specifically, (1) we set dynamic weights for different random forests in the multi-grained scanning according to the classification performance of each random forest; (2) we propose a sorting optimization algorithm to set different weights for each sliding window, as the class vectors generated by each sliding window have varying effects on the final prediction results; and (3) we adopt decision-level fusion to construct a staging model of lung adenocarcinoma based on multi-modal genetic data.
Method
Feature selection
In our experiment, lasso regression is used for feature selection [30]. The lasso method belongs to a class of embedded feature selection methods that can overcome the problems of efficiency and computational cost in traditional feature selection and has been successfully applied to microarray classification and gene selection [31].
The lasso method uses a paradigm penalty-based regression to find the optimal solution for formulas (1) and (2).
1 |
2 |
where r is the number of features of the data, n is the number of samples, δj is the regression coefficient of the g th variable, and z is the constraint value, which is a paradigm penalty for the regression coefficient δj. The value of z can vary from 0 to infinity. When z is small, some variable coefficients with small effects are compressed to 0 so that these variables are deleted to achieve feature selection. When z is sufficiently large, it no longer constitutes an actual constraint, and all the attributes are selected.
GcForest
The gcForest model comprises two components [28], as shown in Fig. 1. (1) The first component is multi-grained scanning, which adapts sliding windows to cut raw features into feature vectors. After feeding these feature vectors into different types of random forest, the model outputs class vectors. Then, all the class vectors are concatenated and output as the result of the multi-grained scanning. (2) The second component is the cascade forest. Each cascade layer is composed of multiple random forests, which comprise decision trees. The input is composed of the class vectors from the output of the first component. Each cascade layer outputs a new class vector that is concatenated with the original class vector to form a new class vector as the input of the next layer (detail in [28]). With multiple random forests in each cascade, more discriminative features can be learned from the input vector to that cascade. A more accurate prediction is finally obtained through the layer-by-layer transfer of each cascade layer. k-fold cross-validation is used to reduce the risk of overfitting when extending a new layer. Specifically, the training data are divided into k folds. k-1 folds are selected as the training data in turn, and the remaining fold is used as the validation data. After extending the new layer, the performance of the entire cascade is estimated on the validation data, and if no significant performance gain is observed, the training process is terminated. Finally, the average of each class probability is calculated from the class vectors of the last cascade layer’s outputs: the class with the maximum probability value is used as the prediction result.
Figure 1 shows the standard gcForest model [28], which is composed of multi-grained scanning and a cascade forest. Assume that the input feature vector has 400 dimensions and that three sizes of sliding windows (100, 200, 300) are used to cut the input feature vector. The first sliding window size is 100, and the sliding stride is 1. Thus, a total of 301 scans are required, and 301*100-dimensional(dim) feature vectors are generated. These feature vectors are the input to a completely random forest and a random forest. Suppose the samples have three classes; each sample is trained using completely random forest and random forest, and 1806-dim (2*301*3 dim) class vectors are output. Similarly, when the sliding window size is 200 and 300, respectively, we obtain class vectors of 1206-dim (2*201*3 dim) and 606-dim (2*101*3 dim).
The second component is the cascade structure. The class vector output by the first component is used as input to the cascading forest component. First, the 1806-dim class vector obtained from the 100-dim sliding window is used as the input to train the first cascade layer. Notably, the diversity of forests plays an important role in constructing the model. After training four forests (two random forests [32] and two completely random forests [33]), a 12-dim class vector (4 forests, 3 classes) is generated. The 12-dim class vector is then concatenated with the original 1806-dim class vector (as shown in Fig. 1) to obtain an 1818-dim vector as the input of the second cascade layer. Similarly, the second cascade layer’s random forests are trained to generate a 12-dim class vector, which is concatenated with the class vector (1206-dim) obtained from the 200-dim sliding window in the first component. Therefore, we obtain a 1218-dim class vector as the input of the third cascade layer. Similarly, we train the third cascade layer’s random forest and obtain a 12-dim class vector, which is concatenated with the class vectors (606 dim) from the 300-dim sliding window in the first component. Therefore, we obtain a 618-dim class vector as the input of the next cascade layer. We repeat the above process each time a new layer is generated. The performance of the entire cascade is estimated on the validation set whenever a new layer is extended: if no significant performance gain is observed, the training process is terminated [28].
Multi-weighted gcForest (MLW-gcForest)
Two challenges may limit gcForest’s application to small-scale genetic data. 1) Each forest in the original gcForest has the same impact on the final prediction, but in reality, the classification ability of each forest is different. Therefore, we assign different weights αto the random forests according to the classification performance of each random forest. 2) In the original gcForest, for the same raw data, scanning with multiple scale windows can generate different-dimension feature vectors. Multi-grained scanning enriches the diversity of generated features. The above method of generating features can capture more comprehensive features, and the different grains of feature vectors generated under different sliding windows have different effects on the final classification results. We believe giving equal attention to different sliding windows in the original gcForest algorithm is unreasonable; therefore, we consider assigning corresponding weights β to different sliding windows. A structure diagram of MLW-gcForest structure is shown in Fig. 2.
Determination of weights
The weight assignment is the first improvement to the algorithm. We use one random forest and one completely random forest for each sliding window and compute the weights of each forest, α1 and α2. The specific method is as follows.
The performance of each random forest must be evaluated objectively; therefore, evaluation criteria must be introduced. The receiver operating characteristics (ROC) [34] curve is a common indicator to measure the performance of a model and indicate the performance of a classifier. In the case of a binary classification task, the area under the curve (AUC) is shown in formula (3).
3 |
However, this metric is not intuitive for multi-class tasks because the AUC is usually used to measure the classification ability of binary classification tasks.
For multi-class tasks, Scurfiled et al. [35] proposed the concept of multiple ROC [36] analysis and the measurement of hypervolume under multi-flow (HUM) to evaluate the identification ability of the corresponding biomarkers. The classification of lung adenocarcinoma is a three-class task, so a double integral is used, as shown in formula (4).
4 |
In this paper, we use one random forest and one completely random forest to obtain HUM1 and HUM2 as the evaluation indicators for the classifier; the HUM values are then normalized to calculate the weight of each forest, as shown in formulas (5) and (6).
5 |
6 |
The values α1 and α2 are then used to assign weights to the class vectors produced by the different forests.
Sorting optimization algorithm
As using different sliding windows to extract class vectors greatly influences the final classification results, different weights are assigned to the class vectors generated by different sliding windows; this algorithm is called the sorting optimization algorithm (as shown in Fig. 3). The basic structure of the algorithm is as follows.
Suppose the number of samples is N, the size of the original features is M, and the number of class labels for the sample is C. For each sample i, we use a sliding window t to cut features (the index of the current window is t (1≤ t ≤ T). The size of the sliding window is L (1≤ L ≤ M). The stride of the scan is S (default S =1). We designate the number of feature vectors after scanning as V and calculate this value with formula (7).
7 |
-
(2)
Each L-dim feature vector is input into one random forest and one completely random forest. The two forests each output a C-dim class vector. Each random forest uses V L-dim feature vectors as input and outputs V C-dim class vectors. Then, these V C-dim class vectors are concatenated into V * C-dim class vectors (called RF-vec). Similarly, we concatenate the class vectors from the output of the completely random forest and call it cRF-vec.
-
(3)
RF-vec and cRF-vec are multiplied by their respective weights α1 and α2 (the weights of the different random forests, calculated in the previous section) and concatenated as 2 ∗V ∗ C-dim class vectors (length L′ = 2 ∗ V ∗ C).
-
(4)
First, the L′-dim class vector obtained in step (3) is sorted in descending order. Then, the average of the top 1/C of the sorted class vector values is calculated. This indicator can be viewed as the approximate ability of the current window to predict the current sample i.We name it t _ abilityi, and it is calculated as shown in formula (8).
8 |
where conc represents the concatenate operation.
-
(5)
Steps (1)–(4) are repeated for the N samples, and we obtain the prediction abilities for the current window for the N samples (t _ ability1, t _ ability2, …, t _ abilityN). This indicator approximates the prediction performance of the current window t.
-
(6)
The prediction ability W _ abilityt of the sliding window t is obtained by averaging the prediction abilities of the current window for the N samples, as shown in formula (9).
9 |
-
(7)
Steps (1)–(6) are repeated to obtain the prediction ability of each window (W _ ability1…W _ abilityt…W _ abilityT). The individual W _ ability values are normalized to obtain the predictive weights βt of each sliding window, as shown in formula (10), which is used to obtain the weights of each window β1 β2 … βt … βT.
10 |
The class vector obtained from each window is multi-plied by its corresponding β value and then concatenated with each other as the output of the multi-grained scanning component.
In step (4) of the algorithm, we calculate the average of the top 1/C class vector values to approximate the current scan window prediction ability because the random forest outputs the confidence probabilities that the samples belong to a certain class. If the maximum value of the confidence probabilities is closer to 1, the random forest has a stronger ability to distinguish the sample categories. Therefore, we take the average of the top 1/C class vector values to approximate the current scan window prediction ability.
After assigning various weights to the gcForest algorithm, we obtain the MLW-gcForest classification model.
MLW-gcForest decision fusion of multi-modal data
The pathogenesis of lung adenocarcinoma is complex, and satisfactory staging results are often difficult to obtain using only single-modal data. A more accurate diagnosis of lung adenocarcinoma is achieved by combining multi-modal data (methylation data, RNA-seq data and CNV data) and taking full advantage of the complementarity between the advanced features of the different modal datasets. Therefore, multi-modal lung adenocarcinoma genetic data are used to train different MLW-gcForest models, and decision-level fusion is performed.
The basic idea of decision-level fusion is to determine which class a sample belongs to by considering the classification results of multiple models. In our algorithm, we obtain the final classification results via weighted voting of multiple models.
Each classification models hp predicts a label from the category label set {class1, class2, …, classC}. The forecast output is represented as a C-dim class vector , such as (0.12, 0.33, …, 0.45), where is the output of hp on classq. Different types of individual models can produce different types of values (). The weighted voting method used in this paper is shown in the following formula:
11 |
m is the number of modalities of the data, and γp is the degree of influence of modality p on the classification results based on the experimental results, where γMethylation + γRNA + γCNV = 1. γMethylation, γRNA, and γCNV are normalized by the accuracies of each type of data’s MLW-gcForest model, as shown in formula (12).
12 |
mp is the model trained by modality p.
Finally, weighted voting is performed on the MLW-gcForest model trained using three different modal datasets to obtain the final classification result (as shown in Fig. 4).
Results
Materials
To evaluate the performance of the MLW-gcForest algorithm, methylation data, RNA-seq data, CNV data and corresponding clinical data of lung adenocarcinoma are downloaded from the TCGA [37] (https://portal.gdc.cancer.gov/). The data include 492 methylation samples, 576 RNA-seq samples, and 516 CNV samples. After excluding samples without clinical staging and feature values that were missing more than 50% of information, we obtained 155 cases of stage I data, 243 cases of stage II data, 41 cases of stage III data, and 16 cases of stage IV data. The small number of samples in stage IV (less than 20 samples) would result in a very unbalanced training sample; therefore, the data for stage IV were excluded. Then, we excluded samples that did not have complete multi-modal data and obtained 369 samples of multi-modal data as our final dataset.
Each sample has 485,577 columns of feature values for the methylation data, 60,483 columns of feature values for the RNA-seq data, and 39 columns of feature values for the CNV data (all corresponding to valid data columns after null values are deleted). After feature selection, the methylation data retained 340 columns, RNA-seq retained 320 columns and CNV had too few columns to perform feature selection.
Experiment
To evaluate the performance of the proposed algorithm, we use nested cross-validation [38] to train and test the model. Compared with standard cross-validation, nested cross-validation can achieve an almost unbiased estimation of model performance [38]. The process of nested cross-validation is divided into an outer loop and an inner loop, as shown in Fig. 5. The inner loop is used to perform parameter adjustments, while the outer loop is used to calculate the final error estimate for model performance.
As shown in Fig. 5, we divide the dataset into ten folds in a mutually exclusive manner. Each time we select nine folds to execute the inner loop (the inner loop performs standard 10-fold cross-validation), and the remaining fold is used for testing. The above process is the outer loop. The outer loop is repeated 10 times until each fold is used as a test set. Therefore, results are obtained for 10 test sets. We calculated the average of the ten test results from the nested 10-fold cross-validation.
It is worth noting that the inner loop performs standard 10-fold cross-validation; that is, the data used to execute the inner loop are divided into ten folds, nine folds are selected for training, and the other fold is used for validation. This process is repeated 10 times until each fold is used as the validation set. Moreover, when performing the inner loops, if the error on the training data continues to decrease but the error on the validation data stops decreasing, the training process terminates early, even before reaching the maximum number of epochs to avoid overfitting.
As described in section 2.4, we trained different MLW-gcForest models using multi-modal data with nested 10-fold cross-validation and then performed decision-level fusion.
In our algorithm, one random forest and one completely random forest were set up in the multi-grained scanning, and 300 decision trees were used in each forest. In the cascade forest layer, we used 300 decision trees for two completely random forests and two random forests. For better comparison with existing algorithms, we used different machine learning algorithms to construct lung adenocarcinoma staging models based on single and multi-modal data (methylation, RNA, CNV): SVM, K-nearest neighbors (KNN), logistic regression (LR), random forest (RF), gcForest and the proposed MLW-gcForest. We considered commonly used evaluation indexes, namely, AUC, accuracy, precision, recall and F1 score, to evaluate the performance of the algorithm.
Result of lung adenocarcinoma staging models based on single-modal data
To evaluate the performance of the MLW-gcForest algorithm, MLW-gcForest, gcForest, and traditional machine learning methods were used to build lung adenocarcinoma staging models from different single-modal data. Based on the different single-modal datasets, we trained different methods with nested 10-fold cross-validation to evaluate the classification ability (Fig. 6). The three rows of Fig. 6 show the classification performance of different algorithms based on methylation, RNA, and CNV measured through nested 10-fold cross-validation. The three columns in Fig. 6 show the classification performance of different algorithms based on different modal data under the same evaluation metric. As shown in Fig. 6a(1), b(1) and c(1), the AUC values of the proposed MLW-gcForest algorithm are higher than those of the remaining algorithms on all three modal datasets.
The accuracy of the different algorithms on each data modality is shown in Fig. 6a (2), b(2), and c(2). The accuracy of the MLW-gcForest algorithm is higher than that of gcForest, SVM, KNN, LR and RF.
In addition, we also compare the precision and F1 scores of MLW-gcForest with those of the other algorithms in Table 1. Table 1 shows that our algorithm achieves a precision of 0.771 and an F1 score of 0.767, which are higher than those of the standard gcForest (precision and F1 score of 0.715 and 0.709) in methylation data. MLW-gcForest always outperforms gcForest, and their performances are superior to those of traditional machine methods on the other two single-modal data (RNA and CNV). This result demonstrates that the deep forest structure can capture more complex and diverse features, making it more suitable for small-sample genetic data. Furthermore, the proposed multi-level weighting strategy can help deep forests extract more valuable multi-level features, thus effectively improving the classification ability of the standard gcForest on small-sample genetic data.
Table 1.
Classification algorithm | Methylation | RNA | CNV | |||
---|---|---|---|---|---|---|
Precision | F1 | Precision | F1 | Precision | F1 | |
SVM | 0.524 | 0.519 | 0.552 | 0.558 | 0.427 | 0.434 |
KNN | 0.584 | 0.605 | 0.533 | 0.528 | 0.460 | 0.466 |
LR | 0.575 | 0.572 | 0.609 | 0.603 | 0.446 | 0.486 |
RF | 0.606 | 0.618 | 0.611 | 0.602 | 0.512 | 0.557 |
gcForest | 0.715 | 0.709 | 0.634 | 0.643 | 0.616 | 0.628 |
MLW-gcForest | 0.771 | 0.767 | 0.659 | 0.669 | 0.675 | 0.677 |
Comparison of the staging of lung adenocarcinoma based on different algorithms using multi-modal data
We used different classification algorithms to construct staging models of lung adenocarcinoma based on multi-modal data to demonstrate the performance of the proposed MLW-gcForest algorithm in integrating multi-modal data for lung adenocarcinoma staging. In addition to plotting the ROC curves and calculating the AUC, the accuracy for each method under the multi-modal data measured through nested 10-fold cross-validation was calculated and is shown in Fig. 7. As shown in Fig. 7, the MLW-gcForest algorithm achieves better classification results on the staging of lung adenocarcinoma. The AUC and accuracy values of MLW-gcForest are higher than those of the standard gcForest algorithm and traditional machine learning algorithms.
The precision, recall and F1 score of the proposed MLW-gcForest and the other algorithms are shown in Table 2. The precision, recall and F1 score of the MLW-gcForest classification reached 0.896, 0.882 and 0.889, respectively, which are higher than those of the standard gcForest (0.764, 0.795, and 0.779). The comparison results indicate that MLW-gcForest and gcForest perform significantly better than the other traditional machine learning algorithms when using multi-modal data for lung adenocarcinoma staging. The results suggest that the deep forest algorithms (MLW-gcForest and gcForest) are more effective in lung adenocarcinoma staging because more complex and diverse features can be captured to distinguish different classes.
Table 2.
Algorithm | Precision | Recall | F1 |
---|---|---|---|
SVM | 0.674 | 0.664 | 0.669 |
KNN | 0.664 | 0.646 | 0.655 |
LR | 0.675 | 0.669 | 0.672 |
RF | 0.706 | 0.730 | 0.718 |
gcForest | 0.764 | 0.795 | 0.779 |
MLW-gcForest | 0.896 | 0.882 | 0.889 |
Furthermore, our result indicates that comprehensive multi-modal genetic data and the multiple decision-level fusion strategies of the MLW-gcForest model effectively improve the accuracy of lung cancer staging because the proposed algorithm not only enables the deep forests to extract more valuable and multi-level features through improved multi-level weighting strategies but also effectively utilizes the complementarity of multi-modal genetic data.
Comparison of multi-modal data and single-modal data in the staging of lung adenocarcinoma with the MLW-gcForest algorithm
To confirm the effectiveness of the multi-modal data, we compared the classification performance of MLW-gcForest using single-modal (methylation, RNA-seq, CNV) and multi-modal data. We plotted the ROC curves and calculated the AUC value between different modal data measured through nested 10-fold cross-validation, as shown in Fig. 8. The results indicate that the MLW-gcForest algorithm achieves better classification performance when using multi-modal data than when using single-modal data. In addition, comparing the accuracy in the multi-modal scenario in Fig. 7b with that in the single-modal scenario in Fig. 6a(2), b(2) and c(2), it can be seen that the integration of multi-modal data effectively improves the accuracy of lung adenocarcinoma stage prediction. Table 3 shows the accuracy, precision, recall, and F1 score of the lung adenocarcinoma staging model measured through nested 10-fold cross-validation of different modal datasets. The results indicate that MLW-gcForest achieves better performance (accuracy 0.908, precision 0.896, recall 0.882, F1 0.889) with multi-modal data than with single-modal data. Using multi-modal data in the proposed MLW-gcForest significantly improved the accuracy of lung adenocarcinoma staging, which suggests that integrating multi-modal genetic data can effectively improve the accuracy of lung adenocarcinoma staging compared to using only single-modal data.
Table 3.
Modality | Accuracy | Precision | Recall | F1 |
---|---|---|---|---|
Methylation | 0.751 | 0.771 | 0.763 | 0.767 |
RNA | 0.689 | 0.659 | 0.679 | 0.669 |
CNV | 0.645 | 0.675 | 0.677 | 0.677 |
Multi-modal | 0.908 | 0.896 | 0.882 | 0.889 |
Table 3 also shows that the methylation data have a higher staging classification ability for lung adenocarcinoma than do RNA-seq and CNV data.
Discussion
Methylation data are found to be the most discriminative in building lung adenocarcinoma staging models using single-modal data. MLW-gcForest based on multi-modal genetic data achieves better classification performance than that with single-modal data. These results indicate that combining multi-modal genetic data is an efficient way to improve the classification ability of lung adenocarcinoma staging. We also found that in the process of staging lung adenocarcinoma, the MLW-gcForest model and the gcForest model are superior to the traditional machine learning algorithms. The most likely reason is that deep forests (MLW-gcForest model, gcForest) can learn more valuable and advanced features with multi-grained and cascade layers. In addition, MLW-gcForest outperformed the standard gcForest on most lung adenocarcinoma datasets. The results suggest that our multi-level weighting strategy effectively improves the classification ability of the standard gcForest model on small-sample cancer datasets.
Now, we explain why 300 decision trees are chosen in our algorithm. We conducted comparative experiments to determine the required number of decision trees in the RF for the algorithm to achieve the best results. Figure 9 shows the results of these comparative experiments.
As shown in Fig. 9, changing the number of decision trees has little effect on the results when using RNA-seq data (Fig. 9b) but a larger effect on the results when using methylation (Fig. 9a) and CNV (Fig. 9c) data. For the methylation and CNV data, the worst performance is obtained when the number of decision trees in the algorithm is set to [30, 101] (i.e., the number of decision trees in the RF at the multi-grained scanning step is set to 30 and that of the cascade forest is set to 101). The algorithm performs best when the number of trees is set to [300, 300] or [500, 500].
Based on the above comparative results, [300, 300] was selected as the final experimental parameters: although [300, 300] and [500, 500] yielded similar experimental results, the time and calculation costs of using [300, 300] are lower.
Other scholars have used machine learning algorithms to stage lung adenocarcinoma. Li et al. [6] provided a staging model of lung cancer with an accuracy of 0.71. Nicolas Anthony Nguyen et al. [7] used SVM to classify TNM staging in lung cancer patients, with overall accuracies of 0.64 and 0.82 across T and N stages, respectively. The comparison of the experimental results shows that our proposed algorithm achieves higher AUC values and accuracy for lung adenocarcinoma classification. Moreover, MLW-gcForest based on multi-modal genetic data is an effective method to improve the classification model of lung adenocarcinoma compared to traditional machine learning algorithms.
Conclusion
In this paper, we propose an improved gcForest model called MLW-gcForest and implement decision-level fusion to address the challenge of staging lung adenocarcinoma using small-sample multi-modal genetic data. The experimental results show that the MLW-gcForest algorithm has an AUC of 0.96 and an accuracy of 0.908 for lung adenocarcinoma staging, which are better than those achieved by the standard gcForest and traditional machine learning algorithms. Therefore, the proposed MLW-gcForest algorithm is more suitable for small-sample genetic data, and the integration of multi-modal genetic data can effectively improve the accuracy of lung adenocarcinoma staging compared to that achieved with single-modal data.
Although the experimental results show that the proposed combination of MLW-gcForest and multi-modal genetic data has the potential to improve the staging of lung adenocarcinoma, some limitations remain in our research. First, the amount of multi-modal genetic data collected in the experiment is relatively small, which may limit the training of more powerful lung adenocarcinoma staging models. Second, our experiments integrate only three types of genetic data, namely, methylation data, RNA-seq, and CNV, which may ignore the value of other types of genetic data for the staging of lung adenocarcinoma. In addition, pathological images, another class of valuable data found in the TCGA, are not considered in our study. In our future work, we will collect more types of genetic data to train and test the proposed model and explore the possibility of combining pathological images with multi-modal genetic data for the staging of lung adenocarcinoma. In addition, we intend to extend the proposed algorithm to the classification task of cancer subtypes.
Acknowledgments
We would like to thank the National Natural Science Foundation of China for financial support.
Abbreviations
- AUC
Area under the curve
- CNV
Copy number variation
- Dim
Dimensional
- HUM
Hypervolume under multi-flow
- KNN
K-nearest neighbors
- LR
Logistic regression
- MLW-gcForest
Multi-weighted gcForest
- NSCLC
Non-small cell lung cancer
- RF
Random forest
- ROC
Receiver operating characteristics
- SVM
Support vector machine
- TCGA
The Cancer Genome Atlas
- TNM
Tumor node metastasis
Authors’ contributions
YYD, JJZ and YQ designed the research; WKY and JWW performed the experiments; YYD, WKY, JWW, ZJZ, JJZ and YQ wrote and edited the manuscript. YFC and XTY provided clinical guidance. NGFK and SYL polished the article. All authors read and approved the final manuscript.
Funding
This research was funded in part by National Natural Science Foundation of China (Grant number 61872261), in part by the open funding project of State Key Laboratory of Virtual Reality Technology and Systems, Beihang University (Grant No. 2018-VRLAB2018B07) and in part by Research Project Supported by Shanxi Scholarship Council of China (201801D121139).The funders had no rule in the design of the study and collection, analysis, and interpretation of data and in writing the manuscript.
Availability of data and materials
The datasets used and/or analyzed during the current study are available from the corresponding author on reasonable request.
Ethics approval and consent to participate
Not applicable.
Consent for publication
Not applicable.
Competing interests
The authors declare that they have no competing interests.
Footnotes
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
References
- 1.Motono N, Funasaki A, Sekimura A, et al. Prognostic value of epidermal growth factor receptor mutations and histologic subtypes with lung adenocarcinoma[J] Med Oncol. 2018;35(3):22. doi: 10.1007/s12032-018-1082-y. [DOI] [PubMed] [Google Scholar]
- 2.Lin CY, Wu YM, Hsieh MH, et al. Prognostic implication of EGFR gene mutations and histological classification in patients with resected stage I lung adenocarcinoma. [J]. Plos One. 2017;12(10):e0186567. doi: 10.1371/journal.pone.0186567. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Perezmoreno P, Brambilla E, Thomas R, et al. Squamous cell carcinoma of the lung: molecular subtypes and therapeutic opportunities. [J]. Clinical Cancer Research An Official Journal of the American Association for Cancer Research. 2012;18(9):2443. doi: 10.1158/1078-0432.CCR-11-2370. [DOI] [PubMed] [Google Scholar]
- 4.Yu KH, Zhang C, Berry GJ, et al. Predicting non-small cell lung cancer prognosis by fully automated microscopic pathology image features:[J] Nat Commun. 2016;7:12474. doi: 10.1038/ncomms12474. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.Cai Z, Xu D, Zhang Q, et al. Classification of lung cancer using ensemble-based feature selection and machine learning methods[J] Mol BioSyst. 2015;11(3):791–800. doi: 10.1039/C4MB00659C. [DOI] [PubMed] [Google Scholar]
- 6.Li X, Scheich B. Predicting tumour stages of lung cancer adenocarcinoma tumours from pooled microarray data using machine learning methods[J] International Journal of Computational Biology and Drug Design. 2015;8(3):275–292. doi: 10.1504/IJCBDD.2015.072109. [DOI] [Google Scholar]
- 7.Nguyen A, Moore D, McCowan I, et al. Multi-class classification of cancer stages from free-text histology reports using support vector machines[C]//2007 29th annual international conference of the IEEE engineering in medicine and biology society. IEEE. 2007:5140–3. [DOI] [PubMed]
- 8.Singh NP, Bapi RS, Vinod PK. Machine learning models to predict the progression from early to late stages of papillary renal cell carcinoma. [J]. Computers in Biology & Medicine. 2018:100. [DOI] [PubMed]
- 9.Xiao Y, Wu J, Lin Z, et al. A deep learning-based multi--model ensemble method for cancer prediction[J] Computer Methods & Programs in Biomedicine. 2018;153:1. doi: 10.1016/j.cmpb.2017.09.005. [DOI] [PubMed] [Google Scholar]
- 10.Statnikov A, Wang L, Aliferis CF. A comprehensive comparison of random forests and support vector machines for microarray-based cancer classification[J] BMC bioinformatics. 2008;9(1):319. doi: 10.1186/1471-2105-9-319. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Chaudhary K, Poirion OB, Lu L, et al. Deep learning–based multi-omics integration robustly predicts survival in liver cancer[J] Clin Cancer Res. 2018;24(6):1248–1259. doi: 10.1158/1078-0432.CCR-17-0853. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Wang Haifeng, Zheng Bichen, Yoon Sang Won, Ko Hoo Sang. A support vector machine-based ensemble algorithm for breast cancer diagnosis. European Journal of Operational Research. 2018;267(2):687–699. doi: 10.1016/j.ejor.2017.12.001. [DOI] [Google Scholar]
- 13.Tan MS, Chang SW, Cheah PL, et al. Integrative machine learning analysis of multiple gene expression profiles in cervical cancer[J] PeerJ. 2018;6:e5285. doi: 10.7717/peerj.5285. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Alkawaa FM, Chaudhary K, Garmire LX. Deep learning accurately predicts estrogen receptor status in breast cancer metabolomics data[J] J Proteome Res. 2018;17(1):337–347. doi: 10.1021/acs.jproteome.7b00595. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Coudray N. Et al. classification and mutation prediction from non–small cell lung cancer histopathology images using deep learning. Nat Med. 2018;1. [DOI] [PMC free article] [PubMed]
- 16.Litjens G, Sánchez CI, Timofeeva N, et al. Deep learning as a tool for increased accuracy and efficiency of histopathological diagnosis[J] Sci Rep. 2016;6:26286. doi: 10.1038/srep26286. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Sun D, Wang M, Li A. A multimodal deep neural network for human breast cancer prognosis prediction by integrating multi-dimensional data[J] IEEE/ACM Transactions on Computational Biology and Bioinformatics (TCBB) 2019;16(3):841–850. doi: 10.1109/TCBB.2018.2806438. [DOI] [PubMed] [Google Scholar]
- 18.Cruz-Roa Angel Alfonso, Arevalo Ovalle John Edison, Madabhushi Anant, González Osorio Fabio Augusto. Advanced Information Systems Engineering. Berlin, Heidelberg: Springer Berlin Heidelberg; 2013. A Deep Learning Architecture for Image Representation, Visual Interpretability and Automated Basal-Cell Carcinoma Cancer Detection; pp. 403–410. [DOI] [PubMed] [Google Scholar]
- 19.Wilson CM, Li K, Yu X, et al. Multiple-kernel learning for genomic data mining and prediction[J] BMC bioinformatics. 2019;20(1):426. doi: 10.1186/s12859-019-2992-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Feng PH, Chen TT, Lin YT, et al. Classification of lung cancer subtypes based on autofluorescence bronchoscopic pattern recognition: a preliminary study[J] Comput Methods Prog Biomed. 2018;163:33–38. doi: 10.1016/j.cmpb.2018.05.016. [DOI] [PubMed] [Google Scholar]
- 21.Liao Z, Li D, Wang X, et al. Cancer diagnosis through IsomiR expression with machine learning method[J] Curr Bioinforma. 2018;13(1):57–63. doi: 10.2174/1574893611666160609081155. [DOI] [Google Scholar]
- 22.Zhu Xiao-Feng, Zhu Bi-Sheng, Wu Fei-Ma, Hu Hai-Bo. DNA methylation biomarkers for the occurrence of lung adenocarcinoma from TCGA data mining. Journal of Cellular Physiology. 2018;233(10):6777–6784. doi: 10.1002/jcp.26531. [DOI] [PubMed] [Google Scholar]
- 23.Tran N, Abhyankar V, Nguyen KT, et al. Microrna dysregulational synergistic network: learning context-specific microrna dysregulations in lung cancer subtypes[C]//2017 IEEE international conference on bioinformatics and biomedicine (BIBM). IEEE. 2017:142–5.
- 24.Molina-Romero C, Rangel-Escareño C, Ortega-Gómez A, et al. Differential gene expression profiles according to the Association for the Study of lung Cancer/American Thoracic Society/European Respiratory Society histopathological classification in lung adenocarcinoma subtypes[J] Hum Pathol. 2017;66:188–199. doi: 10.1016/j.humpath.2017.06.002. [DOI] [PubMed] [Google Scholar]
- 25.Shin J, Keam B, Kim M, et al. Prognostic impact of newly proposed M descriptors in TNM classification of non-small cell lung Cancer[J] J Thorac Oncol. 2017;12(3):520–528. doi: 10.1016/j.jtho.2016.11.2216. [DOI] [PubMed] [Google Scholar]
- 26.Cristescu R, Lee J, Nebozhyn M, et al. Molecular analysis of gastric cancer identifies subtypes associated with distinct clinical outcomes[J] Nat Med. 2015;21(5):449. doi: 10.1038/nm.3850. [DOI] [PubMed] [Google Scholar]
- 27.Xiong JF, Jia TY, Li XY, et al. Identifying epidermal growth factor receptor mutation status in patients with lung adenocarcinoma by three-dimensional convolutional neural networks[J] Br J Radiol. 2018;91(1092):20180334. doi: 10.1259/bjr.20180334. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28.Zhou Z H, Feng J. Deep forest: Towards an alternative to deep neural networks. arXiv 2017[J]. arXiv preprint arXiv:1702.08835.
- 29.Dong Y, Yang W, Wang J, et al. MLW-gcForest: a multi-weighted gcForest model for Cancer subtype classification by methylation data[J] Appl Sci. 2019;9(17):3589. doi: 10.3390/app9173589. [DOI] [Google Scholar]
- 30.Tibshirani R. The lasso method for variable selection in the cox model. Stat Med. 1997;16(4):385–395. doi: 10.1002/(SICI)1097-0258(19970228)16:4<385::AID-SIM380>3.0.CO;2-3. [DOI] [PubMed] [Google Scholar]
- 31.Lin Y, Liu X, Hao M. Model-free feature screening for high-dimensional survival data[J] SCIENCE CHINA Math. 2018;61(9):1617–1636. doi: 10.1007/s11425-016-9116-6. [DOI] [Google Scholar]
- 32.Breiman L. Random forests[J] Mach Learn. 2001;45(1):5–32. doi: 10.1023/A:1010933404324. [DOI] [Google Scholar]
- 33.Fan W, Wang H, Yu PS, et al. Is random model better? On its accuracy and efficiency[C]//third IEEE international conference on data mining. IEEE. 2003:51–8.
- 34.Kourou K, Exarchos TP, Exarchos KP, et al. Machine learning applications in cancer prognosis and prediction[J] Computational and structural biotechnology journal. 2015;13:8–17. doi: 10.1016/j.csbj.2014.11.005. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35.Scurfield BK. Generalization of the theory of signal detectability ton-Eventm-dimensional forced-choice tasks[J] J Math Psychol. 1998;42(1):5–31. doi: 10.1006/jmps.1997.1183. [DOI] [PubMed] [Google Scholar]
- 36.Li J, Fine JP. ROC analysis with multiple classes and multiple tests: methodology and its application in microarray studies[J] Biostatistics. 2008;9(3):566–576. doi: 10.1093/biostatistics/kxm050. [DOI] [PubMed] [Google Scholar]
- 37.Network TCGA. Comprehensive molecular profiling of lung adenocarcinoma[J] Nature. 2014;511(7511):543–550. doi: 10.1038/nature13385. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 38.Varma S, Simon R. Bias in error estimation when using cross-validation for model selection[J] BMC bioinformatics. 2006;7(1):91. doi: 10.1186/1471-2105-7-91. [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
The datasets used and/or analyzed during the current study are available from the corresponding author on reasonable request.