Abstract
Cancer is a complex disease that deregulates cellular functions at various molecular levels (e.g., DNA, RNA, and proteins). Integrated multi-omics analysis of data from these levels is necessary to understand the aberrant cellular functions accountable for cancer and its development. In recent years, Deep Learning (DL) approaches have become a useful tool in integrated multi-omics analysis of cancer data. However, high dimensional multi-omics data are generally imbalanced with too many molecular features and relatively few patient samples. This imbalance makes a DL based integrated multi-omics analysis difficult. DL-based dimensionality reduction technique, including variational autoencoder (VAE), is a potential solution to balance high dimensional multi-omics data. However, there are few VAE-based integrated multi-omics analyses, and they are limited to pancancer. In this work, we did an integrated multi-omics analysis of ovarian cancer using the compressed features learned through VAE and an improved version of VAE, namely Maximum Mean Discrepancy VAE (MMD-VAE). First, we designed and developed a DL architecture for VAE and MMD-VAE. Then we used the architecture for mono-omics, integrated di-omics and tri-omics data analysis of ovarian cancer through cancer samples identification, molecular subtypes clustering and classification, and survival analysis. The results show that MMD-VAE and VAE-based compressed features can respectively classify the transcriptional subtypes of the TCGA datasets with an accuracy in the range of 93.2-95.5% and 87.1-95.7%. Also, survival analysis results show that VAE and MMD-VAE based compressed representation of omics data can be used in cancer prognosis. Based on the results, we can conclude that (i) VAE and MMD-VAE outperform existing dimensionality reduction techniques, (ii) integrated multi-omics analyses perform better or similar compared to their mono-omics counterparts, and (iii) MMD-VAE performs better than VAE in most omics dataset.
Subject terms: Cancer, Mathematics and computing
Introduction
Ovarian cancer is a common and deadly gynaecological cancer with a high mortality rate in developed countries. It accounts for 5% of all cancer deaths in females in the UK1 and USA2. Ovarian cancers are generally diagnosed at an advanced age as the early-stage disease is usually asymptomatic, and symptoms of the late-stage disease are nonspecific3. The anatomical location and the ovaries’ position are mainly responsible for asymptomatic and nonspecific nature of the disease. Since the ovaries have limited interference with the surrounding structures, ovarian cancer is hard to detect until the ovarian mass is significant, or metastatic disease supervenes. Due to the symptoms’ nonspecific nature, it often requires multiple consultations with a primary care physician and several investigations in finding the disease and for an appropriate therapy/treatment. In this context, the development of useful tools to better understand the complex pathogenesis is needed for effective cancer management and prognosis3–5.
Cancer is a complex and heterogeneous disease that deregulates cellular functions in different molecular levels, including DNA, RNA, proteins and metabolites. Importantly, molecules from different levels are mutually associated in reprogramming the cellular functions6–8. Any study limited to any of these levels is insufficient to understand the complex pathogenesis of cancer. Integrated multi-omics analysis of data from these levels is essential to understand cellular malfunctions responsible for cancer and its progression holistically. Importantly, integrated multi-omics analysis, taking the advantage of various omic technologies (i.e., genomics, transcriptomics, epigenomics and proteomics), can identify reliable and precise biomarkers for diagnosis, treatment stratification and prognosis5,9. Recent advancements of omics and computational technologies, including deep learning, have boosted the research in integrated multi-omics analysis for precision medicine and cancer. In recent years, many research works10–12 have been published on integrated multi-omics analysis of cancer. These works are either on individual cancer13–18 or pancancer19–22. Most of these works integrated di-omics15,16, few of them integrated tri-omics23, and very few of them integrated tetra-omics24 data. There are a few integrated multi-omics analyses of ovarian cancer, including di-omics15,25 and tri-omics13,25–27 based analyses.
Due to the increasing availability of large-scale multi-omics cancer data, machine learning approaches, especially DL approaches16,18,28,29 are becoming very useful and effective in integrated multi-omics analysis of cancer data. However, high dimensional omics data are normally imbalanced with a large number of molecular features and a relatively small number of available samples with clinical labels29. For example, DNA methylation and mRNA integrated dataset (TCGA) for ovarian cancer has 39,622 features (27,579 features for DNA methylation and 12,043 features for mRNA) for common 481 samples with clinical labels. This imbalanced dimensionality of cancer datasets makes it challenging to use a machine learning (ML) or DL in integrated multi-omics analysis, especially in individual cancers as they have few samples. For instance, TCGA ovarian cancer RNAseq dataset has only 308 samples, whereas pancancer RNAseq dataset has 9081 samples 29. Algorithms for dimensionality reduction, such as autoencoder-based DL algorithms16,18, together with conventional solutions, such as principal component analysis (PCA)30, are possible solutions to the dimensionality problem. Importantly, considering the discontinuous and non-generative nature of traditional autoencoders, VAEs31 have emerged as DL-based generative models for compressed features learning or dimensionality reduction. There are many works which have used VAE in their studies. However, they are mostly mono-omics studies of individual cancer28,32,33 or pancancer29,34,35. OmiVAE29 is the only work that considered VAE for integrated multi-omics (di-omics) analysis of pancancer. Also, VAE may suffer in representing the input features due to uninformative compressed features and variance over-estimation in feature space36. None of the existing works has used MMD-VAE, and also their analyses are limited to cancer molecular subtypes classification only.
Moreover, most of the existing works28,32,33,35 use unsupervised dimensionality reduction methods, separating the downstream analysis from the reduction method. However, dimensionality reduction in cancer multi-omics analysis is an intermediate step toward the downstream analysis, like classification (e.g., cancer vs normal cell). Separating the dimensionality reduction and model (e.g., a classifier) learning may not be optimal for classification as datasets are not always suitable for a classification task. For example, a DNA methylation dataset that includes cancer and normal samples is readily applicable for classification due to their discriminative features. However, a DNA methylation dataset that only includes cancer samples may not be useful in classification. Due to lack of supervision during a dimensionality reduction process, some key features can be filtered before training the classifier, affecting the final performance29,37,38. In this context, supervised dimensionality reduction methods can be more useful in balancing multi-omics datasets and their integrated downstream analysis.
In this work, we did an integrated multi-omics analysis of ovarian cancer using VAE and MMD-VAE36. To the best of our knowledge, this is the first work that does a comprehensive mono-omics and integrated multi-omics (i.e., di- and tri-omics) analysis of individual cancer (ovarian) using VAE or MMD-VAE, or both. The objectives of this work are three-fold. First, we have designed and developed a DL architecture of VAE and MMD-VAE that supports unsupervised and supervised learning of latent features from mono-omics, di-omics and tri-omics data. Second, we did a dimensionality reduction performance analysis of the developed DL architecture on ovarian cancer by cancer samples identification, molecular subtypes clustering and classification. As a dimensionality reduction technique, the performance of MMD-VAE or VAE depends on input features and sample size, not on the cancer type. Hence, we have tested the developed MMD-VAE and VAE using datasets with three different sample sizes (i.e., 292, 459 and 481) to demonstrate our findings’ robustness. Finally, a survival analysis of an existing ovarian cancer dataset has been carried out using the reduced or latent features sets.
Methods
In the following, we briefly discuss the datasets used, data preprocessing, VAE/MMD-VAE architecture, dimensionality reduction and survival analysis methods.
Datasets used
We used mono-omics and multi-omics (i.e., di- and tri-omics) data for the study. We generated multi-omics data using different combinations of high dimensional mono-omics data. Table 1 summarises the datasets used in this study in terms of their key features (i) omic-count (mono/di/tri), (ii) omic type (e.g.,genomics, epigenomics, transcriptomics, and their combinations) (iii) omic data (i.e., mRNA, CNV/CNA, DNA methylation, RNAseq, and miRNA) (iv) input features dimension, (v) sample size (after processing), and (vi) the unit used for data values (e.g., beta value for DNA methylation). Here, CNV/CNA means copy number variation/alteration, mRNA means gene expression array, DNA methylation means methylation of CPG islands, and RNAseq means gene expression by RNAseq. We have downloaded four mono-omics TCGA datasets from UCSC Xena data portal39, one for mRNA, CNV/CNA and RNAseq, and two for DNA methylation. All the mono-omics datasets except the second DNA methylation dataset are from the TCGA Ovarian Cancer (OV) cohort40–43. The second DNA methylation dataset is from GDC TCGA Ovarian Cancer (OV) cohort44. It includes cancer and normal samples. We have concatenated these mono-omics data to form the di-omics and tri-omics datasets. The table’s ’Feature dimension’ and ’Sample size’ columns demonstrate that all the datasets are imbalanced with too many input features and relatively too few numbers of samples with clinical labels. For example, one tri-omics (CNV + DNA methylation + RNAseq) dataset has 72,885 input features with only 292 samples.
Table 1.
Omic count | Omic type | Omic data | Feature size | Sample size () | Feature & Data values (unit) | Source |
---|---|---|---|---|---|---|
mono-omics | Genomics (G) | CNV | 24,776 | 481 | Gene name & Gistic2 copy number | 41 |
Transcriptomics (T) | mRNA | 12,043 | 481 | Gene name & log2(affy RMA) | 40 | |
RNAseq | 20,530 | 292 | Gene name & pan-cancer normalized log2(norm_count+1) | 42 | ||
Epigenomics (E) | DNA methylation | 27,579 / 21,675 | 481/886 | CPG probe identifier or CG number & Beta value | 43,44 | |
Di-omics | G + T | CNV + mRNA | 36,819 | 481 | A combination of the respective mono-omics values | |
CNV+ RNAseq | 45,306 | 292 | ||||
E + T | DNA methylation + mRNA | 39,622 | 481 | |||
DNA methylation + RNAseq | 48,109 | 292 | ||||
G + E | CNV+ DNA methylation | 52,355 | 481 | |||
Tri-omics | G + E + T | CNV+ DNA methylation + mRNA | 64,398 | 481 | ||
CNV+ DNA methylation + RNAseq | 72,885 | 292 |
Sample size after intersection.
Generated through the concatenation of respective mono-omics datasets.
Data prepossessing
The downloaded datasets are not ready (e.g., sample sizes are not equal- CNV dataset has 579 and mRNA dataset has 593 samples) to be used in dimensionality reduction and integrated multi-omics analysis. They need to be preprocessed, such as the datasets’ sample sizes need to be same to integrate and generate di- and tri-omics datasets. We have preprocessed the downloaded datasets (TCGA Ovarian Cancer cohort) using a four steps method (see Fig. 6 in Supplementary documents).
Step 1 First, we intersected the mono-omics datasets to find the common and same size samples. We did two different intersections of the datasets (Fig. 6 (step: 1) in Supplementary documents) using common sample IDs (also represent the patient IDs) to keep the maximum number of samples for the study. The intersection of CNVs, mRNA and DNA methylation datasets has found 481 samples, and the intersection of CNVs, DNA methylation and RNAseq has found 292 samples in common within the datasets.
Step 2 We identified and removed the missing/zero/NA values in the four downloaded omics files. All the data files, except the RNAseq, had no missing/zero/NA values, and 212 input features or genes (particularly small nucleolar RNA/SNORD) with zero expression values were removed from the RNAseq dataset.
Step 3 Non-normalised datasets, such as CNVs and RNAseq datasets, were normalised using the min-max technique. We used the min-max normalisation as unlike other techniques (i.e., Z-score normalisation) it guarantees multi-omics features will have the same scale45. Thus, all the features will have equal importance in the multi-omics analysis.
Step 4 Finally, we concatenated the normalised mono-omics datasets to form the di- and tri-omics datasets. Concatenations involving RNAseq, such as and datasets have 292 samples and others have 481 samples.
The second DNA methylation44 dataset is highly imbalanced (Fig. 8a in Supplementary documents) as it has only 10 normal samples compared to the 603 cancer samples (class ratio: 1.36:98.64). We used the Borderline-SMOTE SVM46 to reduce the class imbalance (Fig. 8b in Supplementary documents) by re-sampling of normal samples (10 to 283). After the re-sampling, we have 886 samples compared to the original 613 samples. This has increased class ratio (31.94:68.06) between the normal and cancer samples.
VAE/MMD-VAE architecture
Standard VAE
A VAE31 is a deep generative model, which can learn meaningful data manifold from high dimensional input data. Unlike, standard autoencoders (Fig. 7 in Supplementary documents), a VAE encodes an input () as a distribution over a latent space instead of as a single point. Given an omic/multi-omics dataset D with N samples with d dimensional omic or multi-omics features, a VAE/MMD-VAE assumes each sample is generated from a latent vector , where . A DL model for VAE follows a four-step process:
Step 1 (encoding): an encoder encodes or generate each latent variable from a prior distribution or latent distribution . Importantly, the encoder introduces a variational distribution (also known as encoding distribution) to estimate the posterior and address the intractability of the true posterior in calculating the distribution of X or 31. Here, is the set of learnable parameters of the encoder.
Step 2 (sampling): a sampler samples points from the latent space by sampling from the encoded or encoding distribution .
- Step 3 (decoding) : a decoder decodes the sampled points from a conditional distribution and reconstructs the inputs , where is the set of learnable parameters of the decoder. In this step, VAE also calculates the loss or error using loss function that is composed of a reconstruction term and a regularisation term. The reconstruction term calculates the reconstruction loss and the regularisation term quantifies the distance between the estimated posterior and true posterior to regularise the latent space. A standard VAE uses Kullback-Leibler divergence47 for the regularisation term and jointly optimises the encoder and decoder using the following loss function that rely on the traditional evidence lower bound (ELBO) criterion:
where is the Kullback-Leibler (KL) divergence between two probability distributions.1 Step 4 (backpropagation): finally, the calculated loss is backpropagated through the network to update the model accordingly.
MMD-VAE
VAE using ELBO-based loss function (Eq. 1) may suffer from the following two issues36:
Uninformative latent code/feature: The regularisation term ( used in the loss function (Eq. 1) might be too restrictive48,49. KL divergence naturally encourages the latent code to be a random sample from for each x, making the code uninformative/unaware about the input. In this context, the encoder could fail to learn any meaningful latent representation of the input.
Overestimation of variance in feature space: The ELBO-based VAE tends to over-fit data. Due to the over-fitting, it could learn a that has variance tending to infinity36. For example, training ELBO-based VAE on a dataset with two data points , and both encoder () and decoder () output Gaussian distributions with non-zero variance.
Use of Maximum Mean Discrepancy (MMD) in loss function instead of the 50 can address the above issues. According to the MMD, two distributions are identical if and only if all of their moments are same. Unlike , MMD-based regularisation term estimate divergence by how “different” the moments of two distributions p(z) and q(z) are. We can use the kernel embedding trick to estimate MMD for two distributions as Eq. (2).
2 |
where is any universal kernel, including Gaussian kernel . The VAE using the MMD-based loss function is known as MMD-VAE, and the corresponding loss function for can be expressed as Eq. (3).
3 |
VAE/MMD-VAE architecture
As VAE and MMD-VAE differ only in the loss function, we can implement them using the same architecture. The implementation presented in Fig. 1 can support unsupervised and supervised dimensionality reductions. The architecture includes three main components: an encoder, a decoder and a classifier. For the unsupervised dimensionality reduction, the encoder and the decoder learn latent features from the input without the classifier’s support. However, we need the classifier for supervised learning of latent features.
Like other deep neural network architectures, VAE has two main hyperparameters: the number of layers and the number of nodes in each hidden layer. Systematic experimentation is the most reliable way to configure this hyperparameters51. We used the configurations from an existing and related work OmiVAE29 to avoid the experimentation from scratch. We used the same number of hidden layers as the OmiVAE and ran a few experiments to identify the suitable nodes for the hidden and bottleneck layers. For example, we experimented with the hidden layer one of the encoder and decoder with 4096 and 2048 nodes. We selected 2048 due to insignificant performance difference between the two sizes and shorter processing time for 2048 nodes.
The encoder network comprised of an input layer and three hidden layers. The decoder network structure is the mirror image of the encoder structure. Notably, the encoder and decoder share the necessary bottleneck layers. We used the architecture for mono-omics, di- and tri-omics data, and the size of the bottleneck layers is same for all the datasets. However, the other layers’ sizes varied according to the omic count (i.e., mono, di and tri) and omic data. For example, for mono-omics data, such as mRNA data, the input and output layers are 12,043, and hidden layers sizes are 2048 and 1024. As shown in Fig. 1, multi-omics data were integrated using an unsupervised parallel integration method12. The classifier used a 3-layered fully connected artificial neural network (ANN) with an input layer with nodes equal to the LFs (32/64/128), a hidden layer with nodes equal to the half of LFs, and an output layer with nodes equal to the class numbers (e.g., 2 for cancer vs normal samples, 4 for molecular subtypes).
The VAE/MMD-VAE architecture does all the activities illustrated in Fig. 1. In the following, using omic(s) data, we briefly discuss these activities in the perspective of the encoder, decoder and classifier.
Encoder: The encoder network using two hidden layers encodes mono-omics data into a 1024 dimensional vector, di-omics data into two 1024 dimensional vectors and tri-omics data into three 1024 dimensional vectors. The encoding network for the DNA methylation data is different from the other omics data. For example, in the first hidden layer, each chromosome related DNA methylation data are encoded into corresponding vectors with 256 dimensions wheres for the others, input data are encoded into a 2048 dimensional vector. This encoding is to capture the intra-chromosome relationships, and second hidden layer for the DNA methylation data captures the inter-chromosome relationships. For di- and tri-omics data, the second hidden layer respectively concatenates two and three 1024 dimensional vectors and produces an encoded 512-dimensional vector. The encoder’s final hidden layer fully connects to two output layers. These two layers of the size of latent code or features (32/64/128) are part of the bottleneck layers and represent the mean and the standard deviation in the Gaussian distribution of the latent variable or feature z given input sample x or simply . As illustrated in Figs. 1 and 7 in Supplementary documents, a reparameterisation trick is applied (, where is a random variable sampled from unit normal distribution N(0, I)) in the bottleneck layer to make the sampling process differentiable and suitable for backpropagation. The sampled latent features vector (z/LFs) is the compressed lower-dimensional representation of omics or integrated multi-omics data.
- Decoder: The decoder network takes the latent feature vector z as the input and passes through three hidden layers, and finally outputs the reconstructed vector of the input omics data. The decoder is also responsible for estimating the overall loss using Eqs. (4) and (6) respectively for VAE and MMD-VAE.
where k is a binary variable set to 1, if there is any DNA methylation data in the input otherwise set to 0, M is the number of chomosomes, CE is the binary cross-entropy between input data (i.e., - DNA methylation, - other omic data) and reconstructed data (i.e., - DNA methylation and - other omic data), - other omic data count, and is the KL divergence between the learned distribution and a unit normal distribution N(0, I), which is:4 5
where nll- negative log likelihood which can be calculated as mean of for DNA methylation data and mean of for other omics data, is MMD (Eq. 2) between the learned distribution and a unit normal distribution N(0, I), which is:6
Classifier: In an unsupervised VAE, the bottleneck layer tends to extract the essential features to reconstruct input samples as closely as possible. However, these extracted features may not be related to a specific task, such as a molecular subtype classification. The classifier works as an additional regularisation on top of the bottleneck layer. With this additional regularisation, the classifier encourages the VAE or MMD-VAE network to learn LFs that can not only accurately reconstruct the input sample but also, identify cancer and classify molecular subtypes29. The binary cross-entropy based classification loss () can be added to or to estimate the total loss using the following Equation:7
where and are weights of the two losses in the total loss. Equation (8) can be used for the total loss of MMD-VAE () by replacing with . The supervised and unsupervised learning of LFs depends on the value of . We use for the unsupervised and or any positive value for the supervised learning of LFs.8
We used a batch normalisation technique in each fully connected block to implement the VAE/MMD-VAE DL architecture. This is to address the internal covariate shift (The distribution of the inputs of each layer changes during training, when previous layers’ parameters change, which slows down the training process.) by normalising layer inputs52. Thus, it stabilises the learning process and significantly improves the learning speed. As the activation function, we used the rectified linear units (ReLU) for the hidden layers, the sigmoid for the decoder’s output layer and the softmax for classifiers’ output layer. We built the model using PyTorch (version 1.5.0). The implementations of the models used in this paper are available on GitHub (https://github.com/hiraz/MMD-VAE4Omics).
Clustering and classification in cancer
Cancer samples identification and molecular subtypes are useful in prognostic and therapeutic stratification of patients and improved management of cancers13,25,53. Hence, correct clustering and classification of ovarian cancer samples and molecular subtypes are important for improved disease management. Authors in13,25,53 have identified four ovarian cancer transcriptional (one molecular subtype (We will use transcriptional subtypes and molecular subtypes interchangeably.)) subtypes, which may have clinical significance. These four subtypes of high grade serous ovarian cancer (HGS-OvCa) are named as Immunoreactive, Differentiated, Proliferative and Mesenchymal13,54. The datasets used in this work are about HGS-OvCa, and the clinical data include these molecular subtypes for most of the samples. Although these molecular subtypes are transcriptional (e.g., mRNA), they can be used for other omics data analysis due to their correlation or association with transcriptional data55–58. For example, authors in55 have reported that DNA methylation is often negatively associated with gene expression in promoter regions, while DNA methylation is often positively associated with gene expression in gene bodies.
VAE or MMD-VAE generated latent and compressed features (z or LFs) can be used to cluster and classify cancer samples, subtypes, including existing transcriptional or molecular subtypes of ovarian cancer. The performance of clustering and classification exploiting z can demonstrate the dimensionality reduction capability of VAE or MMD-VAE. We demonstrated the dimensionality reduction capability of VAE and MMD-VAE using the latent features learned from the mono-omics, integrated di-and tri-omics data of ovarian cancer, and used for the followings:
Clustering We can use the LFs learned (unsupervised and supervised) by the VAE/MMD-VAE models to cluster samples into cancer vs normal and molecular subtypes. We used a two- and three-dimensional embedding of the mono-omics, di- and tri-omics features for the selected samples, and visualised the clustered samples using scatter plots. Two dimensional (2D) and three dimensional (3D) embedding of the omic(s) features for the selected samples are accomplished by selecting first 2 and 3 LFs from the learned LFs (Fig. 1B (left-side)). We then used the embedded features to cluster the samples into two groups for cancer identification (cancer and normal samples), and four groups (4 molecular subtypes) for molecular subtypes using 2D and 3D scatter plots.
Classification We used an ANN-based classifier to classify cancer samples and molecular subtypes using the LFs learned through the unsupervised process. For all the omics data, we selected the first two and all LFs learned by VAE/MMD-VAE to classify the samples (Fig. 1B right-side). For the LFs learned through the supervised process, we used the VAE/MMD-VAE architecture’s classifier to classify the molecular subtypes. All the classification experiments were validated using a 5-fold cross-validation. In each round of the validation, 80% data were used for the training, and the rest 20% were left out from the training and used for separate testing. We presented the classification performances for both classifiers in terms of accuracy, precision, recall, and f1 score. We also presented a confusion matrix for each classification task done using the LFs learned through supervised VAE/MMD-VAE models. We have selected the first 2 LFs for the clustering and classification for simplicity reason. However, one can select any 2 LFs from the learned LFs, and the performance will be similar to the presented ones.
For the LFs learned using the unsupervised VAE/MMD-VAE models, we compared the clustering and classification performances with two popular traditional dimensionality reduction methods, namely PCA and t-SNE59. We also illustrated how a combination of a traditional method (e.g., t-SNE) and MMD-VAE/VAE performs in molecular subtypes clustering.
Survival analysis
Identification of robust survival subgroups of ovarian cancer (HGS-OvCa) can significantly improve patient care. Existing molecular subtypes of HGS-OvCa, such as transcriptional molecular subtypes13 may not be useful in survival subgroups prediction as most of these studies do the subtyping without relying on survival data. In this study, first, we used existing transcriptional subtypes for survival analysis and then used the learned (supervised) LFs inferring and predicting survival subgroups of HGS-OvCa. We followed a 3-step process (Fig. 1C) as below to do the subgrouping and their corresponding survival analysis:
Inferring survival subgroup: We built a univariate Cox proportional hazards (Cox-PH) model for each of the LFs produced by the VAE/MMD-VAE (Fig. 1C(i)). Then, we identified clinically relevant LFs for which a significant Cox-PH model was found (log-rank ). Next, we used these reduced and clinically relevant LFs (CRLFs) to cluster the samples using a K-means clustering algorithm. We used the R package NbClust60 to determine the optimal K value (number of clusters). NbClust can calculate up to 30 indices or metrics to determine the optimal number of clusters in a data set. It also identifies the best value for K by the majority rule. In our all 11 datasets (see Table 1), optimal values were between 4 and 2. Considering the small sample sizes 481 and 292 with a low number of events, we chose , which means we identified/inferred two survival subgroups.
-
Predicting survival group labels for new samples: After having the survival subgroups labels from K-means clustering, we used an SVM-based classifier (Fig. 1C(ii)) to predict survival subgroup labels for new samples. We used a 60%/40% (training/test sets) of all the datasets to have sufficient test samples in most cases that generate evaluation metrics. We used the tune function of R package e107161 to train the SVM model as it tunes the model parameters through cross-validation (5-fold) and identify the best model for a training dataset. In each round of the validation, 60% data were used for training and rest 40% were left out from the training as the test dataset. Finally, we used the test dataset to predict the survival subgroup or risk labels. We used the Cox-PH model and Kaplan-Meier (KM) survival curves to evaluate survival prediction performance. We used the following three metrics for the evaluation:
Concordance index The concordance index or C-index is a metric to evaluate the predictions made by an algorithm. Based on Harrell C statistics62, C-index can be defined as the fraction of all pairs of individuals whose predicted survival times are correctly ordered63. The C-index score range between 0 and 1 and a score around 0.70 indicates a good model, whereas a score around 0.50 means predictions are no better than a coin flip in determining which patient will live longer. To calculate the C-index, we first built a multi-variate Cox-PH model using the training set (including the inferred survival subgroup labels and clinical features). We then predicted survival using the labels of the test set. We then computed the C-index using concordance function of R’s survival package64. Similarly, we calculated the C-index only considering the clinical features (i.e., status, grade).
P value of Cox-PH regression The Cox-PH models built on training datasets compute log-rank p values for the models. Also, we plotted the Kaplan-Meier survival curves of the two survival subgroups (predicted) and calculated the log-rank p-value of the survival difference between them.
Brier score Brier score function measures the accuracy of probabilistic prediction65. In survival analysis, it measures the mean of the difference between the observed and the estimated survival beyond a certain time66. The score ranges between 0 and 1, and a smaller score indicates better accuracy of the prediction. We used the R Package survAUC67 to compute the Brier score.
Identifying prognostic biomarkers from LFs: We did all the above analyses using the LFs as their compressed representations simplify the survival analysis and molecular subtyping. However, we need to map these LFs back to their corresponding input features to identify potential molecular biomarkers. We mapped the associated input features for each clinically relevant LF using a linear model and filtered the features with zero or insignificant input feature values. Next, we estimated the correlation between the CRLFs and their corresponding input features. Finally, we used the filtered correlation data for hierarchical clustering (colour map) of LFs and their input features.
Results
We used the developed DL architecture of VAE/MMD-VAE for cancer samples identification, molecular subtypes clustering and classification, and survival analysis using the TCGA ovarian cancer datasets. The results demonstrate the performance of the VAE and MMD-VAE in dimensionality reduction and survival analysis.
We trained and tested the developed VAE and MMD-VAE models with three different bottleneck layers () on the preprocessed omics datasets to demonstrate the integrated multi-omics data analysis capability. We implemented the DL model of VAE/MMD-VAE using the network architecture presented in Fig. 1. We tested the model in unsupervised and supervised settings. We used the Adam optimiser with learning rate due to its superior performance compared to other stochastic optimisation methods68. We reported the results only for due to space limitation and a similar performance pattern. All the classification performances were cross-validated. Two sets of results were generated, one on cancer samples identification and molecular subtypes clustering and classification, and another on survival analysis. Importantly, we ran the experiments on four mono omics, five di-omics and two tri-omics datasets’. We presented the results for only one for each omics data due to space limitation.
Dimensionality reduction
We have demonstrated the dimensionality reduction capability of the developed VAE/MMD-VAE by ovarian cancer samples identification, and molecular subtypes clustering and classification. We also carried out a survival analysis of the TCGA ovarian cancer dataset with the latent features set.
Clustering
Cancer vs Normal samples: We used the unsupervised setting of the VAE and MMD-VAE to learn the LFs of the DNA methylation data of 886 samples (GDC cohort). We have selected the first 2 LFs of the 128 LFs to cluster the samples into two groups (cancer and normal). The two-dimensional embedding of the DNA methylation dataset’s input features was plotted on scatter plots for PCA, t-SNE, VAE and MMD-VAE. As illustrated in Fig. 2, even with the unsupervised setting, all the dimensionality reduction methods demonstrate clustering accuracy over 95%, thanks to the discriminative nature of the input features. MMD-VAE outperforms others by correctly clustering 883 samples out of 886. However, the distance between the clusters is an issue, especially in MMD-VAE, which was improved (shown in Fig. 2e,f) for VAE and MMD-VAE by combining t-SNE with them. The cancer samples are compact within the cluster (orange dots) compared to the normal samples. The sub-clusters within the normal samples could be due to the variances within the samples.
Molecular subtypes clustering: We clustered the transcriptional subtypes using the LFs learned through unsupervised and supervised VAE and MMD-VAE models. For the LFs learned via unsupervised model, we have selected the first 2 LFs of the learned 128 LFs to cluster the molecular subtypes. The two-dimensional embedding of the mono omic, di-omics and tri-omics datasets’ input features were plotted on scatter plots for PCA, t-SNE, VAE and MMD-VAE. Figure 9 in Supplementary documents presents the results of 2 LFs-based molecular subtypes clustering. As seen in Fig. 9b–i in Supplementary documents, all the dimensionality reduction methods poorly clustered the samples into four subtypes using the mono- and tri-omics datasets. This result is expected as the original omics datasets are not discriminative or well representative of the transcriptional subtypes. As Fig. 9a in Supplementary documents illustrates, even the most relevant transcriptional dataset (mRNA) do not represent the transcriptional subtypes. Hence, the unsupervised PCA, t-SNE, VAE and MMD-VAE models struggle to cluster the transcriptional subtypes. In this context, we can use the supervised versions of these models, especially VAE and MMD-VAE. We used the supervised VAE and MMD-VAE models to learn the task-oriented (i.e., the transcriptional subtypes) or guided LFs from the mono-, di- and tri-omics datasets. We have selected the first 2 LFs of the learned 128 LFs to cluster the molecular subtypes. Figure 3 presents a part of the clustering results for the supervised VAE and MMD-VAE. As Fig. 3a–j illustrates, the supervised VAE and MMD-VAE have significantly improved their clustering performance compared to their unsupervised counterparts (Fig. 9 in Supplementary documents) in all omics datasets. As illustrated in the Figure, the transcriptional (mRNA -mono-omics) dataset is outperforming other datasets, mainly other mono- omics (i.e., methylation and CNV) datasets, and MMD-VAE outperforms VAE in most datasets. Also, we have combined the t-SNE with VAE and MMD-VAE, which improve the performance (shown in Fig. 3k,l compared to their implementations without t-SNE.
Classification
Cancer samples identification: We used an SVM-based classifier to identify the cancer samples from the normal samples using the LFs learned through the unsupervised PCA, t-SNE, VAE and MMD-VAE. Table 3 in Supplementary documents presents the classification performances for the DNA methylation dataset of 886 samples (GDC cohort). All the models except t-SNE have more than 99% classification accuracy with very high precision (0.99), recall (0.99) and f1 score (0.99). The discriminative features (cancer vs normal) of the DNA methylation data is the main reason for this classification performance. Molecular subtypes classification: Like the transcriptional subtypes clustering, we used the LFs learned through the unsupervised and supervised VAE and MMD-VAE models in molecular subtypes classification. For the unsupervised setting, we also compared the classification performance of the LFs learned through VAE and MMD-VAE with of the LFs learned through PCA and t-SNE. Table 4 in Supplementary documents presents the classification performance of an ANN-based classifier utilising the LFs learned via these unsupervised models from mono-, di- and tri-omics datasets. As we can see from the table, the classifier using the PCA and t-SNE generated LFs poorly classify the existing transcriptional subtypes in all omics datasets. On the other hand, the classifier using the VAE and MMD-VAE generated LFs can classify the transcriptional subtypes for mono-omics (mainly mRNA), di- and tri-omics data with higher accuracies in the range of 73.2–81.44%. However, the performances may not be acceptable in many real-life applications. Lack of discriminative features within the omics datasets for the transcriptional subtypes is the main reason for the low accuracies. Supervised learning of the LFs can improve the classification performance. In the supervised setting, VAE or MMD-VAE and the classifier jointly learn the LFs using the transcriptional subtypes as the supervisory guidance. We trained the joint models on the mono-, di- and tri-omics datasets and tested the models separately. Table 2 presents the performances of the molecular subtypes classification in terms of accuracy, precision, recall and f1 score. Figure 10 in Supplementary documents presents the confusion matrices for few of these classification tasks. As presented in the table, the molecular subtypes classification performances have significantly improved in all matrices (i.e., accuracy, precision, recall and f1 score) compared to the unsupervised VAE/MMD-VAE (Table 4 in Supplementary documents). For example, except the CNV and methylation datasets, MMD-VAE and VAE respectively show accuracies in the range of 93.2-95.5%, and 87.1-95.7% with high precision, recall and f1 scores. The performances of the CNV and DNA methylation are not satisfactory as they are not transcriptional omics data. Even these non-transcriptional datasets, especially the DNA methylation dataset, show a good classification performance with an accuracy range 72.3–75.2%. This performance could be due to the association or correlation between the omics datasets55–58. For the same reason, the use of integrated non-transcriptional and transcriptional data helps to maintain a similar performance or improve the performances of the transcriptional subtypes clustering and classification. For example, the accuracy of MMD-VAE using mRNA is 93.8%, which has been maintained (93.7%) in case of the integrated CNV-mRNA, and improved to 95.5% in case of the integrated CNV-mRNA-methylation datasets. The confusion matrices in Fig. 10b,d,f in Supplementary documents for MMD-VAE illustrate similar results. As a dimensionality reduction algorithm, in majority datasets, MMD-VAE shows better performance compared to VAE. This performance difference could be due to the MMD-based loss function. We tested the supervised MMD-VAE and VAE on omics datasets with three different sample sizes (i.e., 292, 459, and 481) to demonstrate our findings’ robustness. We used the learned LFs to classify the transcriptional subtypes and presented the classification accuracies in Table 5 in Supplementary documents. As seen in the Table, both VAE and MMD-VAE show robust classification accuracy for the same omics data with different sample sizes.
The dimensionality reduction performance results of VAE and MMD-VAE in clustering and classification demonstrate the followings:
in any downstream analysis (e.g., classification) unsupervised dimensionality reduction is useful if the input dataset is discriminative (e.g., cancer vs normal samples), otherwise supervised dimensionality reduction is necessary, and
integrated dimensionality reduction and multi-omics analysis of data may improve or maintain the similar performance of their mono-omics counterparts exploiting their association, without confounding each other.
Table 2.
Method | Omics_data | Accuracy | Precision | Recall | f1 score |
---|---|---|---|---|---|
V-VAE | CNV | 58.3 ± 0.3 | 0.63 ± 0.01 | 0.58 ± 0.03 | 0.579 ± 0.03 |
MMD-VAE | CNV | 54.3 ± 0.31 | 0.58 ± 0.02 | 0.54 ± 0.02 | 0.53 ± 0.03 |
V-VAE | mRNA | 95.7 ± .5 | 0.95 ± 0.008 | 0.95 ± 0.05 | 0.95 ± 0.006 |
MMD-VAE | mRNA | 93.8 ± .97 | 0.93 ± 0.006 | 0.93 ± 0.005 | 0.93 ± 0.006 |
V-VAE | Methylation | 72.3 ± .8 | 0.73 ± 0.02 | 0.72 ± 0.009 | 0.71 ± 0.006 |
MMD-VAE | Methylation | 75.2 ± .9 | 0.75 ± 0.019 | 0.75 ± 0.018 | 0.75 ± 0.015 |
V-VAE | CNV_mRNA | 93.7 ± .27 | 0.93 ± 0.01 | 0.93 ± 0.008 | 0.93 ± 0.007 |
MMD-VAE | CNV_mRNA | 93.7 ± .37 | 0.94 ± 0.006 | 0.93 ± 0.007 | 0.93 ± 0.007 |
V-VAE | mRNA_methylation | 87.1 ± 1.1 | 0.87 ± 0.009 | 0.87 ± 0.008 | 0.87 ± 0.005 |
MMD-VAE | mRNA_methylation | 93.2 ± .97 | 0.93 ± 0.02 | 0.93 ± 0.008 | 0.93 ± 0.005 |
V-VAE | CNV_mRNA_methylation | 89.4 ± .6 | 0.89 ± 0.02 | 0.89 ± 0.006 | 0.89 ± 0.004 |
MMD-VAE | CNV_mRNA_methylation | 95.5 ± .37 | 0.95 ± 0.02 | 0.95 ± 0.008 | 0.95 ± 0.009 |
Survival analysis
We did a comprehensive survival analysis using eleven datasets, including mono-omics and multi-omics data, particularly for the samples with existing transcriptional subtypes and inferred survival/risk groups. Considering the space limitation, we presented a subset but enough of the results (Fig. 4 and Fig. 12, and Table 7 in Supplementary documents) that significantly represent the performance of VAE and MMD-VAE in survival analysis. Figure 4a (for 481 samples) and Fig. 11 in Supplementary documents (for 292 samples) present the Kaplan-Meier survival curves for existing transcriptional subtypes. The subtypes are not clinically significant or associate with survival of patients/samples (log-rank ) (Fig. 4a and Fig. 11 in Supplementary documents).
For LFs-based survival analysis, we conducted a univariate Cox-PH regression on each of the 128 LFs from each dataset. We identified 5-22 CRLFs associated with survival. The number of CRLFs is different for each omics dataset. For example, we found 22 LFs for CNV dataset and only 5 LFs for the integrated CNV, DNA methylation and mRNA dataset). We did a two-stage survival analysis of the samples (481 and 292) using the two inferred subgroups. In the first stage, we plotted Kaplan-Meier survival curves for all the samples. As seen in the Kaplan-Meier survival curves (Fig. 4b–f) of the inferred groups by VAE and MMD-VAE, there is a significant survival differences (log-rank ) for all the omic(s) data accept the tri-omics (log-rank is higher than threshold ), especially for the VAE. This results could be due to the uninformative LFs learned by the VAE.
In the second stage, we predicted survival subgroup labels using an SVM-based classifier splitting the samples into training and test data using a 60/40 split ratio. After predicting survival groups for the test datasets, we ran two multivariate Cox-PH regressions (one for clinical and one for combined = subgroup + clinical co-variates) on the training samples, then predicted survival using the labels of the test datasets. For the clinical co-variates, we considered three clinicopathological characteristics of the considered patients: (i) age at diagnosis, (ii) clinical or FIGO stage, and (iii) grade. We calculated C-indexes, Brier scores and models’ p-values for the training and held-out test samples for the multivariate Cox-PH regressions. As seen in Table 7 in Supplementary documents, the training samples generated moderately high C-indexes in between , low Brier scores in between with significant log-rank p-values of the Cox-PH model. A similar trend is observed for the held-out datasets with little lower C-indexes () and little higher Brier scores () with significant log-rank p-values of the Cox-PH model. Importantly, as seen from the Table 7 in Supplementary documents the performances of VAE and MMD-VAE have been improved in case of combined survival analysis compared to only clinical variables. This confirm that identified survival subgroup does not confound with clinicopathological variables, rather it improves the prognosis. Finally, we have plotted Kaplan-Meier survival curves for the predicted survival group labels. As shown in Fig. 12a–f in Supplementary documents, similar to Fig. 4b–f, there is a significant survival differences (log-rank ) for all the omic(s) between predicted survival groups for all (presented) omics data. However, few datasets’ p values are higher than the threshold (0.05). Potential reasons for the higher p values or insignificant differences between the predicted survival groups for the datasets could be (i) the smaller sample size with few events to identify the differences and (ii) too much compression may have obscured the clinically relevant features.
From Fig. 4b–f and Fig. 12 and Table 7 in Supplementary documents we have the following two key observations:
Impact of inferred subgroups on survival: In case of mono-omics and multi-omics data, inferred subgroups combined with the clinical covariates (i.e., stage, grade and age) improve the survival prediction. For example, subgroups learned using the MMD-VAE on CNV mono-omics data combined with clinical features shows higher C-index value (0.63) than the C-index value (0.62) for the clinical features. However, the improvement is not that significant. In summary, inferred subgroups, combined with the clinical features, demonstrate survival prediction performance higher than or similar to the clinical features.
mono-omics vs multi-omics based LFs in survival: Similar to molecular subtypes classification, multi-omics based LFs may incrementally (e.g., di- and tri-omics) improve the survival subgroups classification accuracy compared to mono-omics based LFs (shown in Table 6 in Supplementary documents). However, multi-omics (i.e., CNV_mRNA, CNV_mRNA_methylation) based LFs and subgroups predicted based on them do not improve the survival prediction performances than their mono-omics counterparts. Potential reasons for the lower performance using multi-omics based LFs could be due to (i) smaller sample size with few events (e.g., 143 out of 292 samples) to identify the differences, (ii) too much compression (e.g., for CNV + DNA methylation + mRNA data 64,398 input features to 128 LFs) may have obscured the clinically relevant features.
Finally, we presented a simple method to identify potential prognostic biomarkers from the clinically relevant LFs using a linear model. Figure 5 presents the association between the CRLFs and the input features for tri-omics (CNV_mRNA_methylation) data. For example, our mapping has identified the NDRG2 gene as a potential biomarker for ovarian cancer69. Thus, we can identify the other responsible gene or cgp island that could jointly work as a multi-omics prognostic biomarker. However, the linear mapping is not the best solution to identify a set of biomarkers from the learned CRLFs as it is only identifying correlations values within . We need further research in this direction.
Conclusion
Integrated multi-omics data analysis of cancer is vital for a comprehensive understanding of cellular functions, responsible for cancer and its development. DL algorithms have become a useful and popular tool for integrated multi-omics analysis of cancer data in recent years. However, imbalanced high dimensional multi-omics data makes a DL based integrated multi-omics analysis difficult, especially in individual cancer. DL based dimensionality reduction techniques, such as variational autoencoder is a potential solution to this dimensionality issue. In this work, we have designed and developed VAE and MMD-VAE based dimensionality reduction techniques for mono-omics and multi-omics data analysis of cancer. These techniques can learn latent features from any omics data using unsupervised and supervised training process. Then, we used the developed techniques to learn latent features from the mono-omics, di-omics and tri-omics datasets of ovarian cancer. Finally, we used the learned latent features to analyse ovarian cancer through cancer samples identification, molecular subtypes clustering and classification and survival analysis. The results, especially subtype clustering and classification and survival subgroups prediction demonstrate that integrated di- and tri-omics based LFs can perform better or similar to their mono-omics counterparts based LFs. Also, molecular subtypes clustering and classification results show that MMD-VAE is outperforming VAE in most datasets. Notably, the results demonstrate that unsupervised dimensionality reduction is useful in the downstream analysis (e.g., classification) when the input dataset is discriminatory, otherwise supervised dimensionality reduction is needed. Finally, inferred and predicted survival subgroups results show a significant survival difference between the two subgroups. However, multi-omics based LFs and subgroups predicted based on them do not improve the survival prediction performances than their mono-omics counterparts. One of the potential reasons for this could be a sub-optimal number of LFs or dimension of the selected dimension space. Moreover, the straightforward integration (concatenation based) strategy, treating omics measurements from different platforms equally and performing integration in a parallel fashion may not always be useful. The linear model used for mapping between CRLFs and input features may not be the best solution for biomarkers identification. There is significant scope for future work in these areas.
The residual disease is one of the strongest predictors for outcome estimation after any cancer therapy or treatment. Hence, VAE or MMD-VAE based integrated multi-omics analysis on predicting residual disease for ovarian cancer will be one of our future works.
Supplementary Information
Acknowledgements
This research was partially funded by the Newton Fund—Institutional Links from British Council and National Research Council of Thailand with the grant ID is 527643161.
Author contributions
M.H. and M.A.R. performed all experiments, analysed the results and prepared the manuscript. M.S., C.A., J.S., and S.S. as the supervisory (PhD) team of M.H., reviewed and helped to improve the manuscript.
Competing interests
The authors declare no competing interests.
Footnotes
The original online version of this Article was revised: The original version of this Article contained an error in the spelling of the author Mosharraf Sarker which was incorrectly given as Mosharraf Sarkar.
Publisher's note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
These authors contributed equally: Muta Tah Hira and M. A. Razzaque.
Change history
8/11/2021
A Correction to this paper has been published: 10.1038/s41598-021-95882-y
Supplementary Information
The online version contains supplementary material available at 10.1038/s41598-021-85285-4.
References
- 1.UK. Cancer Research, Ovarian cancer statistics. https://www.cancerresearchuk.org/health-professional/cancer-statistics/statistics-by-cancer-type/ovarian-cancer#heading-One.
- 2.Torre, L. A. et al. Ovarian cancer statistics, 2018. CA68, 284–296 (2018). [DOI] [PMC free article] [PubMed]
- 3.Doubeni CA, Doubeni AR, Myers AE. Diagnosis and management of ovarian cancer. Am. Fam. Physician. 2016;93:937–944. [PubMed] [Google Scholar]
- 4.Rosenthal AN, Menon U, Jacobs IJ. Screening for ovarian cancer. Clin. Ostet. Gynecol. 2006;49:433–447. doi: 10.1097/00003081-200609000-00004. [DOI] [PubMed] [Google Scholar]
- 5.Lu M, Zhan X. The crucial role of multiomic approach in cancer research and clinically relevant outcomes. EPMA J. 2018;9:77–102. doi: 10.1007/s13167-018-0128-8. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.Hu R, Wang X, Zhan X. Multi-parameter systematic strategies for predictive, preventive and personalised medicine in cancer. EPMA J. 2013;4:2. doi: 10.1186/1878-5085-4-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Cheng T, Zhan X. Pattern recognition for predictive, preventive, and personalized medicine in cancer. EPMA J. 2017;8:51–60. doi: 10.1007/s13167-017-0083-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Zhan, X., Long, Y. & Lu, M. Exploration of variations in proteome and metabolome for predictive diagnostics and personalized treatment algorithms: Innovative approach and examples for potential clinical application. J. Proteom.188, 30 – 40, 10.1016/j.jprot.2017.08.020 (2018). The key-role of multi-omics in predictive, preventive and personalised medicine as the medicine of the future. [DOI] [PubMed]
- 9.Hasin Y, Seldin M, Lusis A. Multi-omics approaches to disease. Genome Biol. 2017;18:83. doi: 10.1186/s13059-017-1215-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Chakraborty S, Hosen MI, Ahmed M, Shekhar HU. Onco-multi-omics approach: A new frontier in cancer research. BioMed Res. Int. 2018;2018:9836256–9836256. doi: 10.1155/2018/9836256. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Olivier M, Asmis R, Hawkins GA, Howard TD, Cox LA. The need for multi-omics biomarker signatures in precision medicine. Int. J. Mol. Sci. 2019;20:4781. doi: 10.3390/ijms20194781. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Wu C, et al. A selective review of multi-level omics data integration using variable selection. High-Throughput. 2019;8:4. doi: 10.3390/ht8010004. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Bell D, et al. Integrated genomic analyses of ovarian carcinoma. Nature. 2011;474:609–615. doi: 10.1038/nature10166. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Kwon M-S, et al. Integrative analysis of multi-omics data for identifying multi-markers for diagnosing pancreatic cancer. BMC Genom. 2015;16(Suppl 9):S4–S4. doi: 10.1186/1471-2164-16-S9-S4. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Zhang H, et al. Integrated proteogenomic characterization of human high-grade serous ovarian cancer. Cell. 2016;166:755–765. doi: 10.1016/j.cell.2016.05.069. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Chaudhary K, Poirion OB, Lu L, Garmire LX. Deep learning-based multi-omics integration robustly predicts survival in liver cancer. Clin. Cancer Res. 2018;24:1248–1259. doi: 10.1158/1078-0432.CCR-17-0853. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Zhang, Q. et al. Integrated multiomic analysis reveals comprehensive tumour heterogeneity and novel immunophenotypic classification in hepatocellular carcinomas. Gut68, 2019–2031, 10.1136/gutjnl-2019-318912 (2019). https://gut.bmj.com/content/68/11/2019.full.pdf. [DOI] [PMC free article] [PubMed]
- 18.Liu Q, Hu P. Association analysis of deep genomic features extracted by denoising autoencoders in breast cancer. Cancers. 2019;11:494. doi: 10.3390/cancers11040494. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.Weinstein JN, et al. The cancer genome atlas pan-cancer analysis project. Nat. Genet. 2013;45:1113. doi: 10.1038/ng.2764. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Champion M, et al. Module analysis captures pancancer genetically and epigenetically deregulated cancer driver genes for smoking and antiviral response. EBioMedicine. 2018;27:156–166. doi: 10.1016/j.ebiom.2017.11.028. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Zhang, X. Integrated. et al. IEEE Int. Conf. on Bioinforma. Biomed. (BIBM)2019. 10.1109/bibm47256.2019.8983228 (2019).
- 22.Gonzalez-Reymundez, A. & Vazquez, A. I. Multi-omic signatures identify pan-cancer classes of tumors beyond tissue of origin. bioRxiv10.1101/806323 (2019). [DOI] [PMC free article] [PubMed]
- 23.Singh A, et al. DIABLO: An integrative approach for identifying key molecular drivers from multi-omics assays. Bioinformatics. 2019;35:3055–3062. doi: 10.1093/bioinformatics/bty1054. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24.Terunuma A, et al. Myc-driven accumulation of 2-hydroxyglutarate is associated with breast cancer prognosis. J. Clin. Investig. 2014;124:398–412. doi: 10.1172/JCI71180. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.Zhang Z, et al. Molecular subtyping of serous ovarian cancer based on multi-omics data. Sci. Rep. 2016;6:26001. doi: 10.1038/srep26001. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Fu A, Chang HR, Zhang Z-F. Integrated multiomic predictors for ovarian cancer survival. Carcinogenesis. 2018;39:860–868. doi: 10.1093/carcin/bgy055. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Papp E, et al. Integrated genomic, epigenomic, and expression analyses of ovarian cancer cell lines. Cell Rep. 2018;25:2617–2633. doi: 10.1016/j.celrep.2018.10.096. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28.Wang, Z. & Wang, Y. Exploring dna methylation data of lung cancer samples with variational autoencoders. In 2018 IEEE International Conference on Bioinformatics and Biomedicine (BIBM) 1286–1289, 10.1109/BIBM.2018.8621365 (2018).
- 29.Zhang, X. et al. Integrated multi-omics analysis using variational autoencoders: Application to pan-cancer classification. arXiv preprintarXiv:1908.06278 (2019).
- 30.Fakoor, R., Ladhak, F., Nazi, A. & Huber, M. Using deep learning to enhance cancer diagnosis and classification. In Proceedings of the international conference on machine learning, vol. 28 (ACM New York, USA, 2013).
- 31.Kingma, D. P. & Welling, M. Auto-encoding variational bayes. arXiv preprintarXiv:1312.6114 (2013).
- 32.Titus, A. J., Wilkins, O. M., Bobak, C. A. & Christensen, B. C. An unsupervised deep learning framework with variational autoencoders for genome-wide dna methylation analysis and biologic feature extraction applied to breast cancer. bioRxiv 433763 (2018).
- 33.George, T. M. & Lio, P. Unsupervised machine learning for data encoding appliedto ovarian cancer transcriptomes. BioRxiv855593 (2019).
- 34.Way, G. P. & Greene, C. S. Evaluating deep variational autoencoders trained on pan-cancer gene expression. arXiv preprintarXiv:1711.04828 (2017).
- 35.Palazzo M, Beauseroy P, Yankilevich P. A pan-cancer somatic mutation embedding using autoencoders. BMC Bioinform. 2019;20:655. doi: 10.1186/s12859-019-3298-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 36.Zhao, S., Song, J. & Ermon, S. Infovae: Information maximizing variational autoencoders. arXiv preprintarXiv:1706.02262 (2017).
- 37.Basu, T. & Murthy, C. Effective text classification by a supervised feature selection approach. In 2012 ieee 12th international conference on data mining workshops 918–925 (IEEE, 2012).
- 38.Chao G, Luo Y, Ding W. Recent advances in supervised dimension reduction: A survey. Mach. Learn. Knowl. Extract. 2019;1:341–358. doi: 10.3390/make1010020. [DOI] [Google Scholar]
- 39.UCSC. xenabrowser. https://xenabrowser.net/datapages/ (2020). [accessed 8 Mar 2020].
- 40.UCSC. dataset: gene expression array - AffyU133a. https://tcga.xenahubs.net/download/TCGA.OV.sampleMap/HT_HG-U133A.gz (2020). [accessed 8 Mar 2020].
- 41.UCSC. dataset: copy number (gene-level)—gistic2. https://tcga.xenahubs.net/download/TCGA.OV.sampleMap/Gistic2_CopyNumber_Gistic2_all_data_by_genes.gz, year = 2020, note =.
- 42.UCSC. dataset: exon expression RNAseq - IlluminaHiSeq UNC. https://tcga.xenahubs.net/download/TCGA.OV.sampleMap/HiSeqV2_exon.gz (2020). [accessed 8 Mar 2020].
- 43.UCSC. dataset: DNA methylation - Methylation27k. https://tcga.xenahubs.net/download/TCGA.OV.sampleMap/HumanMethylation27.gz (2020). [accessed 8 Mar 2020].
- 44.UCSC. dataset: DNA methylation - Illumina Human Methylation 27. https://gdc.xenahubs.net/download/TCGA-OV.methylation27.tsv.gz (2020). [accessed 8 Dec 2020].
- 45.Han J, Pei J, Kamber M. Data Mining: Concepts and Techniques. Amsterdam: Elsevier; 2011. [Google Scholar]
- 46.Nguyen, H. M., Cooper, E. W. & Kamei, K. Borderline over-sampling for imbalanced data classification. Int. J. Knowl. Eng. Soft Data Paradigms3, 4–21 (2011).
- 47.Kullback S, Leibler RA. On information and sufficiency. Ann. Math. Stat. 1951;22:79–86. doi: 10.1214/aoms/1177729694. [DOI] [Google Scholar]
- 48.Chen, X. et al. Variational lossy autoencoder. arXiv preprintarXiv:1611.02731 (2016).
- 49.Sønderby, C. K., Raiko, T., Maaløe, L., Sønderby, S. K. & Winther, O. Ladder variational autoencoders. Adv. Neural Inf. Process. Syst.3738–3746 (2016).
- 50.Gretton, A., Borgwardt, K. M., Rasch, M., Schölkopf, B. & Smola, A. J. A kernel method for the two-sample-problem. In Proceedings of the 19th International Conference on Neural Information Processing Systems, NIPS’06, 513-520 (MIT Press, Cambridge, MA, USA, 2006).
- 51.Brownlee, J. How to Configure the Number of Layers and Nodes in a Neural Network. https://machinelearningmastery.com/how-to-configure-the-number-of-layers-and-nodes-in-a-neural-network/ (2020). [accessed 22 Nov 2020].
- 52.Ioffe, S. & Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift (2015). arXiv:1502.03167.
- 53.Tothill RW, et al. Novel molecular subtypes of serous and endometrioid ovarian cancer linked to clinical outcome. Clin. Cancer Res. 2008;14:5198–5208. doi: 10.1158/1078-0432.CCR-08-0196. [DOI] [PubMed] [Google Scholar]
- 54.Braicu EI, et al. Dynamic of molecular subtypes of high-grade serous ovarian cancer in paired primary and relapsed biopsies. J. Clin. Oncol. 2019;37:e17091–e17091. doi: 10.1200/JCO.2019.37.15_suppl.e17091. [DOI] [Google Scholar]
- 55.Shen H, Laird PW. Interplay between the cancer genome and epigenome. Cell. 2013;153:38–55. doi: 10.1016/j.cell.2013.03.008. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 56.Sun W, et al. The association between copy number aberration, DNA methylation and gene expression in tumor samples. Nucleic Acids Res. 2018;46:3009–3018. doi: 10.1093/nar/gky131. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 57.Damgacioglu H, Celik E, Celik N. Estimating gene expression from high-dimensional DNA methylation levels in cancer data: A bimodal unsupervised dimension reduction algorithm. Comput. Ind. Eng. 2019;130:348–357. doi: 10.1016/j.cie.2019.02.038. [DOI] [Google Scholar]
- 58.Seal DB, Das V, Goswami S, De RK. Estimating gene expression from DNA methylation and copy number variation: A deep learning regression model for multi-omics integration. Genomics. 2020;112:2833–2841. doi: 10.1016/j.ygeno.2020.03.021. [DOI] [PubMed] [Google Scholar]
- 59.Maaten LVD, Hinton G. Visualizing data using t-sne. J. Mach. Learn. Res. 2008;9:2579–2605. [Google Scholar]
- 60.Charrad M, Ghazzali N, Boiteau V, Niknafs A, Charrad MM. Package ‘nbclust’. J. Stat. Softw. 2014;61:1–36. doi: 10.18637/jss.v061.i06. [DOI] [Google Scholar]
- 61.Meyer, D. e1071: Misc Functions of the Department of Statistics, Probability Theory Group (Formerly: E1071), TU Wien. https://cran.r-project.org/web/packages/e1071/index.html (2019). [accessed 8-April-2020].
- 62.Harrell FE, Jr, Lee KL, Mark DB. Multivariable prognostic models: Issues in developing models, evaluating assumptions and adequacy, and measuring and reducing errors. Stat. Med. 1996;15:361–387. doi: 10.1002/(SICI)1097-0258(19960229)15:4<361::AID-SIM168>3.0.CO;2-4. [DOI] [PubMed] [Google Scholar]
- 63.Raykar, V. C., Steck, H., Krishnapuram, B., Dehing-Oberije, C. & Lambin, P. On ranking in survival analysis: Bounds on the concordance index. In Proceedings of the 20th International Conference on Neural Information Processing Systems, NIPS’07, 1209-1216 (Curran Associates Inc., Red Hook, NY, USA, 2007).
- 64.Therneau, T. M. survival: Survival Analysis. https://cran.r-project.org/web/packages/survival/index.html (2020). [accessed 9-April-2020].
- 65.Brier GW. Verification of forecasts expressed in terms of probability. Mon. Weather. Rev. 1950;78:1–3. doi: 10.1175/1520-0493(1950)078<0001:VOFEIT>2.0.CO;2. [DOI] [Google Scholar]
- 66.Gerds TA, Schumacher M. Consistent estimation of the expected brier score in general survival models with right-censored event times. Biom. J. 2006;48:1029–1040. doi: 10.1002/bimj.200610301. [DOI] [PubMed] [Google Scholar]
- 67.Potapov, S. survAUC: Estimators of prediction accuracy for time-to-event data. https://cran.r-project.org/web/packages/survAUC/index.html (2012). [accessed 9-April-2020].
- 68.Kingma, D. P. & Ba, J. Adam: A Method for Stochastic Optimization (2014). arXiv:1412.6980.
- 69.Kang F, et al. Ndrg2 gene expression pattern in ovarian cancer and its specific roles in inhibiting cancer cell proliferation and suppressing cancer cell apoptosis. J. Ovarian Res. 2020;13:1–10. doi: 10.1186/s13048-020-00649-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.