Abstract
Simple Summary
Cancer is the second leading cause of death worldwide. Predicting phenotype and understanding makers that define the phenotype are important tasks. We propose an interpretable deep learning model called T-GEM that can predict cancer-related phenotype prediction and reveal phenotype-related biological functions and marker genes. We demonstrated the capability of T-GEM on cancer type prediction using TGCA data and immune cell type identification using scRNA-seq data. The code and detailed documents are provided to facilitate easy implementation of the model in other studies.
Abstract
Deep learning has been applied in precision oncology to address a variety of gene expression-based phenotype predictions. However, gene expression data’s unique characteristics challenge the computer vision-inspired design of popular Deep Learning (DL) models such as Convolutional Neural Network (CNN) and ask for the need to develop interpretable DL models tailored for transcriptomics study. To address the current challenges in developing an interpretable DL model for modeling gene expression data, we propose a novel interpretable deep learning architecture called T-GEM, or Transformer for Gene Expression Modeling. We provided the detailed T-GEM model for modeling gene–gene interactions and demonstrated its utility for gene expression-based predictions of cancer-related phenotypes, including cancer type prediction and immune cell type classification. We carefully analyzed the learning mechanism of T-GEM and showed that the first layer has broader attention while higher layers focus more on phenotype-related genes. We also showed that T-GEM’s self-attention could capture important biological functions associated with the predicted phenotypes. We further devised a method to extract the regulatory network that T-GEM learns by exploiting the attributions of self-attention weights for classifications and showed that the network hub genes were likely markers for the predicted phenotypes.
Keywords: phenotypes prediction, interpretable deep learning, Transformer, cancer type prediction, immune cell type prediction
1. Introduction
The ever-growing genomics data from omics profiling and single-cell technologies have spurred rapid developments of deep learning (DL) methods for genomics, especially gene expression-based phenotype predictions [1,2]. In precision oncology, deep learning modeling of gene expression and other genomics data have offered exciting new solutions and perspectives to a variety of important questions cancer diagnosis using convolutional neural networks (CNNs) [3,4,5], patient survival prediction using graph convolution networks (GCNs) [6], drug response prediction using deep neural networks (DNN) [7,8], and cancer vulnerability prediction using DNN [9]. While producing accurate predictions is critical, developing deep interpretable DL models capable of informing the functions and mechanisms underlying the phenotype predictions becomes increasingly desirable [10].
However, developing interpretable DL models for gene expression-based phenotype prediction faces the following challenges. First, how to model unordered genes: A gene expression sample is commonly represented as a vector of unordered expression values. However, popular CNN was originally developed for images/texts, where features (pixels/words) have a fixed order. Therefore, different gene orders for the same sample could lead to distinct DL architectures with varying performances. Finding the best order is computationally inhibitive. Although DNN is order-agnostic, it is highly inefficient and prone to overfitting. Second, how to model gene–gene interactions: Functional gene interactions impose correlations between their expressions. However, interacting genes might be separated in the expression vector. Because CNNs exploit pixel/text correlations within a local region, they cannot fully capture functional interactions from the expression data. Existing attempts to map gene expression data into 2- or 3-D images are inadequate to model functionally interacting genes [11,12,13]. The recent development of graph CNN (GCNN) points to a promising direction for addressing this issue [6]. However, GCNN is restricted to working with a predefined network and cannot model context-specific interactions. Third, inferring functions and markers from (G)CNN/DNN is not trivial. As the convolution/dense layer bears little biological considerations, its computation muddles input genes’ identities, generating feature maps highly difficult to deconvolute. The disagreement between functional interactions and image local features hinders the existing DL interpretation tools from extracting much-desired insights into regulatory pathways and functions. Moreover, phenotypes can be defined by either a few markers or some complex signaling cascades. Because neighboring dependencies in image/text data encourage “smoothness,” the objective of revealing isolated markers or networks is at odds with the smooth constraint of many DL interpretation algorithms. Therefore, genomics data’s unique characteristics challenge the computer vision-inspired design of popular DL models such as CNN and ask for the need to develop interpretable DL models tailored for transcriptomics study.
One notable class of genomics-centric DL models is visible neural networks (VNNs) [14,15], which include models such as DCells [16], DrugCells [17], and P-NET [15]. They are DNN models whose connections and hierarchical layers are defined based on Gene Ontology (GO), pathways, or networks. They have been trained to predict gene dependencies in yeast and drug responses in cancer cells, where the model architectures could inform the associated functions [16,17]. Although VNNs facilitate interpretation, they are inflexible to model and uncover context-specific interactions as the connections and hierarchy are fixed. Moreover, they are oversized models for my genomics applications where the sample size is small. For instance, DCells and DrugCells have 12 layers but only model binary perturbations or mutations.
To address the above limitations, we propose a new interpretable DL modeling framework based on Transformer. Since their invention in 2017 [18], Transformer has taken natural language processing (NLP) by storm. The Transformer and the Transformer-based models have quickly achieved state-of-the-art performance for most NLP tasks [18,19]. The Transformer’s core is the self-attention mechanism, which learns global features’ dependencies. We recognize that self-attention addresses the existing DL models’ limitations to model genomics data. Specifically, unlike CNN, self-attention can model unordered input, such as genes in gene expression data. Furthermore, self-attention models’ interactions between all the features instead of local features, as in CNN, thus capable of learning important gene–gene interactions. Above all, self-attention facilitates interpretation. Unlike CNN or DNN, outputs of self-attention take the same input word identity but learn new representations of input words. This property has been exploited to explain that BERT learns the classical NLP pipeline in an explainable manner, encoding different syntactic and semantic forms [20,21,22]. These appealing features and benefits of self-attention inspired us to propose Transformer for Gene Expression Modeling or T-GEM.
In this paper, we hypothesize that:
T-GEM can accurately predict phenotypes based on gene expression;
T-GEM captures the biological function and potential marker genes.
To assess these hypotheses, we design the T-GEM model for modeling gene–gene interactions and evaluate its ability for gene expression-based predictions and functional interpretation of cancer-related phenotypes, including cancer type prediction and immune cell type classification. The remaining paper is organized as follows. In the Materials and Methods, we describe the datasets and preprocessing steps. We also detail the proposed T-GEM model and discuss the developed methods to interpret T-GEM for its learned functions, key regulatory networks, and marker genes. In the Results, we carefully assess T-GEM’s prediction performance and compare it with existing popular approaches. We further analyze the learning mechanism of T-GEM and show that T-GEM’s self-attention could capture important biological functions associated with the predicted phenotypes. We also extract the regulatory network that T-GEM learns by exploiting the attributions of self-attention weights for classifications and show that the network hub genes were likely markers for the predicted phenotypes. Concluding remarks are discussed in the Conclusion.
2. Materials and Methods
In this study, we proposed a novel interpretable deep learning model called T-GEM, or the Transformer for Gene Expression Modeling, to predict gene expression-based cancer-related phenotypes, including cancer type prediction and immune cell type classification. The interpretability of T-GEM is also presented.
2.1. Data Collection and Preprocessing
The T-GEM is evaluated on two gene expression-based datasets from RNA-seq and scRNA-seq, respectively.
2.1.1. TCGA RNA-Seq Data
The Cancer Genome Atlas (TCGA) [23] pan-cancer dataset was collected via R/Bioconductor package TCGAbiolinks [24] in December 2018. The dataset contains 10,340 and 713 samples for 33 cancer types and 23 normal tissues, respectively. All gene expression was processed by log-transferred and normalized by min-max. To facilitate the investigation of T-GEM’s learning mechanism and functional interpretation, we selected 1708 genes from 24 genesets containing 14 cancer functional genesets from CancerSEA [25] and 10 KEGG biological pathways that are less relevant to cancer.
2.1.2. PBMC scRNA-Seq Data
To test T-GEM’s generalizability and interpretability, we applied T-GEM for immune call type classification using the single-cell RNA-seq (scRNA-seq) gene expression profile of Human peripheral blood mononuclear cells (PBMC; Zhengsorted) [26]. The PBMC dataset includes 10 cell populations extracted through antibody-based bead enrichment and FACS sorting. Each cell population has 2000 cells and, therefore, there is a total of 20,000 cells from 10 cell classes. We followed the method in [27] to preprocess the data and selected 1000 highly variable genes as input features.
2.2. T-GEM Model
Let be the vector of normalized log-transformed gene expression of genes and be the classification label. The goal is to design a deep learning model parameterized by such that . The overall architecture of T-GEM follows that of a Transformer (Figure 1A), which includes multiple attention layers, each of which is composed of multiple self-attention heads. Similar to the classical Transformer, the attention layer generates new attention values, which are representations of the input gene expression . One important feature of the attention layer is that there is a one-to-one correspondence between the input gene and the output representation, i.e., is the representation of (Figure 1B). This input–out consistency is a key modeling feature of T-GEM that results in a biological-aware design of the self-attention units and an essential ingredient that enables an interpretable deep learning model. The core computing element in T-GEM is the self-attention module (Figure 1B), which takes the representation of each gene from the previous layer and computes a set of new representations. We use the self-attention module of the first attention layer to explain its inner operations (Figure 1B). To help understand the connections and differences between the self-attention modules of T-GEM and Transformer, we draw the parallel between an input gene for T-GEM and an input word for Transformer. However, a notable difference lies in the fact that a gene denotes its scalar expression value, whereas a word is embedded by a word vector. Similar to the Transformer, the self-attention module of T-GEM performs two sets of operations. First, three entities, i.e., Query (), Key (, and Value ( are computed for each gene , where , , and , with , , and being the weights to be learned. Compared to Transformer, they are similar in that all the computations are linear; they are nevertheless different because for Transformer, the linear weights are shared for all the words, whereas for T-GEM, the weights , and are gene-dependent. Introducing gene-dependent weights makes T-GEM more flexible and powerful to learn gene-specific representations. It is also computationally feasible because the input gene is a scalar as opposed to a word vector. In the second operation, the representations for each Query gene are computed. Specifically, for Query gene , an attention weight vector or a probability distribution of the similarity between and all other Key genes except is computed as
(1) |
where is the th attention weight in and is a similarity between and defined as
(2) |
Equation (2) makes Query attend to the Key gene that has a similar up/down behavior and consistently large magnitudes (either highly or lowly expressed) across different samples. Since the Key genes with such behavior are likely to be marker genes, Equation (2) would result in Query gene paying attention to marker genes.
Once the attention weights are computed, a representation of the Query gene is obtained as
(3) |
where is the vector of the values excluding gene . The computation in Equation (3) suggests that the genes that are deemed more “similar” by Equation (2) (i.e., correlated marker genes) will have bigger contributions to gene ’s representation. This progress is repeated for each gene until the representations for all the genes are computed. This concludes the operations for the self-attention module. When there are self-attention heads, the representations from each head will be linearly combined to obtain a single set of representations for layer 1
(4) |
where the summation represents the skip connection, is the weight vector, has its th element as the representation of the query gene in head , and represents the layer normalization function [18]. The entire computation from Equaions (1)–(4) will repeat to calculate the Query gene representations of additional layers. In the end, a classification layer is added
(5) |
where is an activation function and is the representations of last layer .
2.3. T-GEM Interpretation Methods
2.3.1. Entropy of the Attention Weights of a Head
The attention weight vector measures the attentions that the Query gene has on the Key genes, and it is in the form of a probability distribution. Therefore, to assess the characteristics of the attentions of a Query gene, we compute the entropy of
(6) |
Intuitively, a lower entropy demonstrates that the distribution of becomes more skewed, and therefore Query gene pays higher attention to specific genes or focused attention. In contrast, higher entropy indicates that the distribution of is flatter, and therefore Query gene pays similar attention to its Key genes or broad attention. We averaged the entropies across all samples to obtain the entropy of a Query gene.
2.3.2. Attribution Scores of Attention Weights of a Head and a Layer
To assess the contributions of the attention weights to the classification, we adopted the method proposed in [28] to compute the attribution scores of the attention weights using an Integrated gradient (IG). Given the T-GEM model with attention weights and some baseline weights , an attribution score is computed as
(7) |
where the integral is taken along a straight line from the to parameterized by the parameter α. To simply the computation, we set baseline as 0. We used the PyTorch interpretability library Captum [29] to compute the attribution scores in Equation (7).
2.3.3. Visualization of the Regulatory Networks of the Top Layer
A regulatory network representing the regulations of Query genes by Key genes was derived from the last layer using attention weight attribution scores computed from the testing samples of a specific label. We used Cytoscape [30] to visualize this network.
2.4. Training Procedure
The model was trained with optimizer Adam [31], with a batch size of 16 and an epoch of 50. The negative log-likelihood function was used as the loss function. The model was trained on multiple sets of hyper-parameters (Table 1), and the best model parameters were selected based on their validation performance. All of the baseline models were trained according to the same procedure. For CNN, AutoKeras [32] was applied to determine the best model architecture. All of the model training was conducted by using Pytorch 1.9.0 on NVIDIA A100 GPUs.
Table 1.
Hyperparameter | Value of Hyperparameter |
---|---|
The number of layers: | 1, 2, 3, 4 |
the number of heads | 1, 2, 3, 4, 5 |
the activation function of the classification layer | No activation, ReLu, GeLu |
2.5. Performance Evaluation Metrics
The proposed model was evaluated on two multi-labeled datasets while TCGA has the unbalance samples for each class and PBMC has the equal size for each class. We adopted two metrics to assess the models’ performance, including ACC (accuracy), MCC (Matthews correlation coefficient), and AUC (the area under the ROC curve). MCC is a reliable metric that produces a high score only if the model performs well in and . ACC and MCC is defined as
(8) |
(9) |
where and are the numbers of true positives, true negatives, false positives, and false negatives, respectively. The ROC curve is plotted by using FPR (false positive rate) against TPR (true positive rate)
(10) |
(11) |
and AUC was computed as the area under the ROC curve.
3. Results and Discussion
In this section, we first compare T-GEM’s performance with CNN and several classical classifiers on the TCGA and PBMC datasets. Then, we analyze the learning mechanism of T-GEM and examine the biological functions and marker genes captured by T-GEM.
3.1. T-GEM’s Performance for Cancer Type Classification Using TCGA Data
We evaluate the performance of the T-GEM model on cancer type prediction using TCGA pan-cancer gene expression data. The data contain 10,340 patients with 33 different cancer types and 713 normal samples. The classification includes 34 labels (33 cancer types and normal). For each label, we extracted 70% of the samples as the training set, 10% as the validation set, and 20% as the testing set. To facilitate the evaluation of interpretation, we selected genes from 14 cancer functional gene sets from CancerSEA, including apoptosis, angiogenesis, invasion, etc., and 10 KEGG biological pathways gene sets that are not related to cancer. A total of 1708 genes are used as input features (Detailed information about 14 functional gene sets and 10 KEGG pathways can be found in the Supplementary Materials file). Because we know which genes belong to cancer-related genesets, we could examine the genes that each T-GEM’s attention layers focus on and understand if T-GEM’s predictions are based on cancer-related genes and functions and what they are.
To determine the architecture and hyperparameter, T-GEM with multiple sets of hyper-parameters (Table 1) were trained, and the obtained model with the best validation performance included three layers, five heads, and no activation in the classification layer of Equation (5).
We evaluated the ACC, MCC, and AUC of T-GEM, CNN, and several classical classifiers (Table 2), where CNN was trained using AutoKeras so that its architecture and hyperparameters were optimized. As we can see, deep learning-based methods, including T-GEM and CNN, have overall better performance, whereas T-GEM achieves the overall best performance, although the improvement over CNN is small, especially on AUCs. To gain a better understanding of the differences in improvements by T-GEM shown by different performance metrics, we examined the confusion matrices of their predictions. We observed that most of the misclassifications were associated with two cancer types, namely colon adenocarcinoma (COAD) and rectum adenocarcinoma (READ). For SVM, it misclassified all 34 READ samples to COAD, while CNN misclassified about half of the 96 COAD samples to READ. In contrast, T-GEM correctly predicted 84/96 COAD samples and 15/34 READ samples. The lower misclassification by T-GEM was clearly revealed by ACC, but this could only register a slight AUC improvement because of the small sample sizes of these two cancer types. Collectively, T-GEM outperformed CNN and classical machine learning algorithms and could achieve lower misclassifications for especially cancer types with small samples.
Table 2.
ACC | MCC | AUC | |
---|---|---|---|
CNN(AUTOKERAS) | 94.34% | 0.9411 | 0.9985 |
SVM | 93.21% | 0.9292 | 0.9972 |
RANDOM FOREST | 91.60% | 0.9123 | 0.9970 |
DECISION TREE | 81.80% | 0.8097 | 0.9062 |
T-GEM | 94.92% | 0.9469 | 0.9987 |
3.1.1. Investigation of T-GEM’s Learning Mechanism
Next, we investigated how T-GEM learns to produce a classification prediction. Specially, we sought to understand the importance of each layer/head for classification, what T-GEM focuses on in each layer, and the biological functions of each layer. We designed three following analyses to address these questions.
To assess the importance of a layer/head, we first pruned that layer/head and checked the accuracies after pruning. Note that we consider the head output before the skip connection in this experiment. The larger the accuracy reduction that the pruning causes, the more important the head/layer is. To prune a head, we set this head’s corresponding weight in in Equation (4) as 0 while keeping the rest of elements in unchanged. Similarly, to prune a layer, we zeroed the weights of all the heads in that layer. The original test samples were used to compute this pruned model’s accuracy.
As can be seen in Table 3, the accuracy after pruning layer 1 drops the most from the original 94.92% to 57.48%, indicating that layer 1 is the essential layer. Examining different heads in layer 1 reveals that head 1 could be the most important head. Intriguingly, we observe much smaller accuracy drops for the heads than the layer. This could be due to heads in the same layer capturing similar features. Notice that there are slight drops associated with layers 2 and 3. Recall that each layer includes a skip connection, and thus, this result suggests that being able to pass the learned information from layer 1 to layers 2 and 3 could be an important mechanism to achieve the overall performance.
Table 3.
LAYER | HEAD 1 | HEAD 2 | HEAD 3 | HEAD 4 | HEAD 5 | |
---|---|---|---|---|---|---|
LAYER 1 | 57.48 | 87.19 | 94.92 | 95.10 | 94.92 | 94.83 |
LAYER 2 | 94.83 | 95.01 | 94.92 | 94.61 | 94.92 | 94.92 |
LAYER 3 | 94.92 | 94.97 | 94.92 | 94.92 | 94.92 | 94.97 |
To further support this observation, we directly assessed the classification performance of each head/layer. To that end, we took the output from a layer/head and trained an SVM. The ACCs are shown in (Table 4). Now, the larger the accuracy, the more important a head/layer is. Overall, all accuracies of all layers/heads are close to the original, indicating that each layer/head contains a large amount of information needed for accurate classification. Note that the accuracies decrease from layer 1 to layer 3, which supports the conclusion from the pruning experiment that passing outputs of layer 1 to other layers through skip connection ensures achieving the reported performance of the T-GEM model. We also notice that the accuracies of different heads, especially in layer 3 vary. This suggests that features extracted from these heads could be different and different heads might focus on learning information from different genes.
Table 4.
LAYER | HEAD 1 | HEAD 2 | HEAD 3 | HEAD 4 | HEAD 5 | |
---|---|---|---|---|---|---|
LAYER 1 | 93.57% | 93.48% | 91.69% | 92.36% | 91.91% | 93.12% |
LAYER 2 | 93.12% | 90.65% | 90.88% | 93.80% | 90.43% | 93.57% |
LAYER 3 | 91.24% | 93.35% | 90.29% | 90.61% | 89.57% | 88.18% |
Motivated by this observation, we then investigated what T-GEM focuses on in each layer. Recall from Equation (3), that the head output for each Query is computed as a linear combination of the Values weighted by attention weight vector . As an important part of T-GEM, reflects the attention in the form of the discrete probability distribution that Query gene pays to all Key genes. When this distribution is flat and close to the uniform distribution, the Query gene pays broad attention to Key genes. Otherwise, the Query gene put more attention on a subset of genes. This distribution varies in each layer and each head depending on what and how Query genes attend to Key genes. To understand if a head pays more attention on a subset of genes (i.e., smaller entropy) or has broad attention (i.e., large entropy), we computed the entropy of for every head in each layer averaged across all of the samples (Figure 2). We notice that the entropies of all the heads in layer 1 are similar and among the largest, suggesting that Query genes in layer 1 tend to pay similar and broad attention to key genes. In contrast, layer 2 and layer 3 contain heads with Query genes having lower entropies and layer 3 has some of the genes with the lowest average entropy. This suggests that layers 2 and 3 start to pay more attention to specific genes. To further understand what the specific genes the heads focus on, we examined five Query genes with the lowest entropy in layer 3 (Figure 2). We found that all of them are related to cancer. EMX2 is the candidate tumor suppressor that regulates tumor progression [33]. EPCAM is a well-known marker for tumor-initiating cancer stem cells, and it is a penitential target for cancer therapy [34]. POU5F1 is a penitential prognostic and diagnostic biomarker for various epithelial cancers [35]. A lack of BHMT impacts susceptibility to fatty liver and hepatocellular carcinoma [36]. Finally, GFAP is a classical marker of astrocytoma, and it is marker of less malignant and more differentiated astrocytoma [37].
In summary, T-GEM pays broad attention in layer 1 but has more concentrated attention in higher layers. The genes with more focused attention or lowest entropies in layer 3 are cancer-related genes.
3.1.2. T-GEM Makes Decisions by Focusing on Important Cancer Pathways
Next, we investigated the biological functions that T-GEM learns in each layer. To this end, we assessed the attribution of attention weights to the classification by computing their attribution scores using IG, as in Equation (7). Then, we averaged the weight attribution scores of the same Query and Key gene pairs from all heads in the same layer to obtain the attribution scores of the layer function. Then, according to [28], we thresholded the attribution scores to retain the Query and Key genes associated with the high weight attribution scores. A snapshot of layer 3 is shown in Figure 3A. We see that Query genes VCAN, MMP11, and FBN1 are connected with Key genes, indicating that the weight attribution scores between these Query and Key genes are large and, therefore, the information passed from these Key genes to the Query genes contributes to the classification. Consequently, we called these Query genes informative genes. On the contrary, we called Query genes with no linked Key genes (e.g., PGM1, PSMB4, LOX) non-informative genes.
To determine the biological function of each layer, we performed the functional enrichment of the associated informative genes. We investigated the predicted functions for BRAC classification as BRCA has the last number of samples (Figure 3B). Since our genes were selected from 14 cancer functional state genesets and 10 KEGG pathways less related to cancer, we used these 24 gene sets for enrichment.
As we can see from Figure 3B, the 4, 3, and 4 pathways were enriched for layers 12, and 3, respectively, (FDR < 0.05). Except one pathway (VALINE_LEUCINE_AND_ISOLEUCINE_DEGRADATION) in the first layer, all other ways are cancer functional states, indicating that T-GEM is able to quickly focus on genes related to cancer even starting from the first layer. Layer 2 and layer 3 focus on specific cancer pathways including metastasis, inflammation, and stemness, which are all highly correlated to breast cancer [38,39,40].
To investigate how the information related to these enriched cancer states is transferred from the input to generate the classification decision, we identified the input Key genes associated with the Query genes in all the significantly enriched pathways for each layer and performed the functional enrichment on these Key genes. Then, for each layer, we linked each enriched pathway with the pathways enriched in Key genes connected to Query genes in that pathway by virtue of Query–Key gene associations (Figure 3B). As shown in Figure 3B, the Key genes from layer 1 are mainly enriched in Invasion, Cell cycle, and DNA Repair, suggesting that only genes in these cancer pathways impact the informative Query genes and thus enriched pathways of layer 1. Interestingly, among these significantly enriched pathways of layer 1, only genes in invasion influence the significantly enriched pathways in layer 2, which are metastasis, inflammation, and stemness. Among these pathways, only invasion and metastasis contribute to the significantly enriched pathways of layer 3, which are invasion, metastasis, inflammation, and stemness, and further contribute to the classification of BRCA. Overall, this investigation reveals that T-GEM attends to more pathways in layers but manages to focus on genes in key cancer pathways in layers 2 and 3 to make decisions.
3.1.3. T-GEM Defines a Regulatory Network That Reveals Marker Genes for Different Cancer Types
Note that the Query–Key gene relationship obtained in the last section, as in Figure 3A, essentially defines a regulatory network that a layer learns with Key genes as the regulators. We extracted the network from the last T-GEM layer for BRCA and LUAD classification using their respective samples and examined the networks (Figure 4).
For BRAC, we first examined the nodes in this regulatory and tested their differential expression between breast cancer patients and normal samples by using the Mann–Whitney U test. Overall, 102/107 genes showed significant differential expression in breast cancer patients compared to normal samples (FDR < 0.05). We further examined the five not-differentially-expressed genes (ABAT, NFKBIA, GUCY1A1, CDH2, TPM2) and found that they do play an important role in breast cancer. For example, ABAT can suppress breast cancer metastasis [41], and NFKBIA’s deletion and low expression are highly related to the unfavorable survival of breast cancer patients [42]. CDH2 is strongly associated with several stem cell-related transcription factors, and it could be the targeted therapy for breast cancer [43]. TPM2 is a potential novel tumor suppressor gene for breast cancer [44]. This result suggests that T-GEM can focus on functional relevant genes even if they show small differential expression.
Close examination of the network (Figure 4) revealed several hub genes, including GATA3, FOXA1, COMP, IGFBP2, COX7A1, CDH2, VEGFD, AKAP12, and STC2. All of them are highly related to breast cancer, and some of them are known BRCA markers. For example, GATA3 is a well-known associated with favorable breast cancer pathologic features and a promising prognostic marker for breast cancer [45]; FOXA1 is a transcription factor that is predominantly expressed in breast cancer and is considered as breast cancer biomarker [46,47]; COMP is an emerging independent prognostic marker for breast cancer patients [48]. IGFBP2 is directly related to estrogen receptor status, and it is a potential biomarker for the therapeutic response to breast cancer [49]. Finally, STC2 has been shown to suppress breast cancer cell migration and invasion [50].
For LUAD, 183/199 nodes show significant differential expression in LUAD patients compared to normal samples (U-test, FDR < 0.05). Similarly, many of the remaining genes that are not differentially expressed (FBN1, COX7A2L, IGFBP2, WNT5B, CDKN1B, KIT, etc.) have been shown to be related to lung cancer. For example, FBN1 plays a crucial role in EMT, and its gene expression is highly correlated with the recurrence-free survival rate for LUAD patients [51]. COX7A2L is important for the assembly of mitochondrial respiratory supercomplexes, and it is related to the regulation of respiratory chain supercomplex assembly and function in lung cancer [52]. A high circulating level of IGFBP2 is highly associated with a poor survival rate, and it is a penitential prognostic biomarker for lung cancer [53].
Similar to BRCA, we observed hub gens NKX2-1, ST6GALNAC1, FOXA1, HSD17B6, AKAP12, DDIT4, ASPN, CEACAM5, MALAT1, GALNT6, and PIPOX (Figure 5), and they also highly associated with LUAD. NKX2-1 controls lung cancer development, and it could be a novel biomarker for lung cancer [54]. AKAP12 acts as a tumor promoter in LUAD and is negatively correlated with patients’ prognosis [55]. A high expression level of DDIT4 associates with poor survival in LUAD patients, and it is a predictor of overall survival [56]. ASPN is found to correlate with LUAD patients’ survival time [57]. HSD17B6 is shown as a potential tumor suppressor in LUAD and a promising prognostic indicator for LUAD patients [58]. CEACAM5 is indicated as a valid clinical biomarker and promising therapeutic target in lung cancer [59]. MALAT1 is a highly conserved nuclear noncoding RNA (ncRNA) and a predictive marker for metastasis development in lung cancer [60]. ST6GALNAC1 can promote lung cancer metastasis, and it is also a novel biomarker for LUAD [61,62]. FOXA1 plays an oncogenic role in lung cancer, and its high expression level is associated with a poor overall survival rate [63]. GALNT6 can promote invasion and metastasis of LUAD, and its expression level is associated with LUAD lymph node metastasis and poor prognosis [64,65]. PIPOX is a major enzyme in the sarcosine metabolism pathway, and it is positively correlated with shorter overall survival [66].
In summary, T-GEM’s attention mechanism models a regulatory network of Query genes by Key genes, and the hub genes of the network of the top layer reveal important marker genes of different cancer types.
3.2. T-GEM’ Performance for PBMC Single Cell Cell-Type Prediction and Interpretation Result
To further examine the generalizability of T-GEM, we applied it to the PBMC single-cell data set for cell type classification. The dataset contains 10 different cell types and 2000 samples for each cell type. Overall, 1000 highly variable genes were selected as the input features. We followed the same model training and selection process and obtained a T-GEM model with one layer and five heads with GeLU as the activation function of the classification layer.
T-GEM achieved among the best performances, outperforming random forest and decision trees by large margins but was competitive with CNN and SVM (Table 5). We further examined the confusion matrices of SVM and T-GEM separately and observed that both models have high misclassification rates for three T cell subtypes, CD4+ T helper, CD4+/CD25 T Reg, CD4+ /CD45RA+/CD25- Naïve T. However, T-GEM has lower misclassification rates (150/400 and 73/400) for T helper cell and T Reg cell than SVM (157/400, 94/400), although SVM has fewer misclassifications (50/400) for Naïve T cell than T-GEM (82/400). In conclusion, T-GEM achieved competitive performances with SVM or CNN for PBMC cell type classification using scRNA-seq data. However, CNN and SVM are not easy to interpret. We investigate the interpretability of T-GEM next.
Table 5.
ACC | MCC | AUC | |
---|---|---|---|
CNN(AUTOKERAS) | 89.00% | 0.8779 | 0.9945 |
SVM | 90.70% | 0.8970 | 0.9913 |
RANDOM FOREST | 82.53% | 0.8062 | 0.9870 |
DECISION TREE | 74.00% | 0.7112 | 0.8556 |
T-GEM | 90.73% | 0.8971 | 0.9964 |
3.2.1. T-GEM Learns the Biological Functions That Define Each Cell Types in PBMC
As shown for TCGA cancer type classification, T-GEM makes decisions by focusing on the genes in cancer-related pathways. Therefore, we used the same approach to examine the functions that T-GEM leans. We enriched the informative genes from the first layer in GO Biology Processes. The result shows that most of the informative genes are enriched on the ribosome- and translation-related pathways, including, for example, cytoplasmic translation, peptide biosynthetic process, ribosome biogenesis, ribosome assembly, etc.(the detailed enrichment result is in the Supplementary Materials file) Because these cell types are FACS-sorted based on the protein activities of surface markers, but T-GEM’s inputs are gene expression, T-GEM seems to focus more on translation-related processes that translate gene expression to protein activities. We subsequently compared the informative genes with the marker genes for these PBMC cell types defined in the CellMatch database [67] and identified the NK cell markers NKG7 and KLRB1 and the B cell markers CD79B and CD37. Note that many of the marker genes (such as CD4, CD19, etc.) were excluded from the 1000 HVGs selected for our prediction, suggesting that their expression levels are either low or not differentially expressed in the cell types. This result highlights T-GEM’s ability to learn the underlying processes that relate predicted phenotypes with inputs.
3.2.2. The T-GEM’s Regulatory Networks for NK Cell and B Cell
We next derived the regulatory network of Key-Query genes for the T-GEM layer for the NK cells and B cells (Figure 6 and Figure 7). Out of 123 nodes in this network, 114 showed differential expression in NK cells (FDR < 0.05). For NK cells, there are four hub genes, including MALAT1, CD52, RPLP1, and KLRB1 (Figure 6). KLRB1 is a C-type lectin-like receptor of NK cells and a known NK cell maker. It plays an important role as an inhibitory receptor of NK cells and also works as a co-stimulatory receptor to promote the secretion of IFNγ [68]. MALAT1 is a druggable long non-coding RNA and plays an important role in NK cell-mediated immunity [69].
We next examined the regulatory network for CD19+ B cells (Figure 7). Note that CD19 is not an HVG. Out of 128 nodes in this network, 127 showed differential expression in B cells (FDR < 0.05). CD53 is the only gene that is not differentially expressed. However, CD53 is found to promote B cell development by maintaining IL-&R signaling [70]. We examined the hub genes of the B cell regulatory network, which include HLA-DPB1, RPL35, RPLP0, RPLP1, RPL13, and MALAT1. Interestingly, RPL35, RPLP0, RPLP1, and RPL13 are ribosomal proteins, which could contribute to the enriched translation-related functions of this layer. MALAT1 is also a hub gene for the NK cell network. It has large expression variations across different PBMC cell types and is down-regulated in B cells [71]. HLA-DPB1 is shown to be highly expressed in B cell lines [72] and could serve as an expression marker for B cells. Taken together, these results demonstrate T-GEM’s ability to learn functions and marker genes associated with predicted phenotypes.
4. Conclusions
We proposed T-GEM, an interpretable deep-learning architecture inspired by Transformer for gene expression-based phenotype predictions using RNA-seq and scRNA-seq data. In the model architecture, the input gene expression is treated as the token, and a self-attention function attending to markers was proposed to model gene–gene interactions. We comprehensively investigated T-GEM’s performance and learned functions for cancer type classification using TCGA data and immune cell type classification using PBMC scRNA-seq. We show that
T-GEM can provide accurate gene expression-based prediction. We investigated T-GEM for cancer type classification using TCGA data and immune cell type classification using PBMC scRNA-seq. T-GEM has the best performance on both datasets compared with CNN and several classical classifiers.
T-GEM learns associated cancer functions. We showed that T-GEM had broad attention in the first layer and paid attention to specific cancer-related genes in layers 2 and 3 for cancer type classification. We also revealed that T-GEM learned cancer-associated pathways at every layer and could concentrate on specific pathways important for predicted phenotypes in layer 3.
We extracted the regulatory network of layer 3 and showed that the network hub genes were likely cancer marker genes. We also demonstrated the generalization of these results for immune cell type classification using PBMC scRNA-seq data.
As far as we know, this is the first time such a model has been proposed for modeling gene expression data. Inevitably, there are still many challenges to be addressed. High memory overhead associated with the training of the attention weights limits the input gene dimension. Right now, we have to carefully preselect a subset of genes, which could overlook important genes for T-GEM to model. Furthermore, the self-attention function in Equation (2) is general and allows the integration of prior biological networks such as protein–protein interaction networks to introduce biological meaningful inductive bias into the learning. Research into solutions to these challenges will be our future focus.
Supplementary Materials
The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/cancers14194763/s1.
Author Contributions
Conceptualization, Y.H. and T.-H.Z.; methodology, Y.H. and T.-H.Z.; software, T.-H.Z.; validation, T.-H.Z., M.M.H. and Y.-C.C.; formal analysis, T.-H.Z.; investigation, T.-H.Z.; resources, Z.-F.H. and M.F.; data curation, M.M.H. and Y.-C.C.; writing—original draft preparation, T.-H.Z.; writing—review and editing, Y.H., Y.C., M.F. and Y.-F.J.; visualization, T.-H.Z.; supervision, Y.H.; project administration, Y.H.; funding acquisition, Y.C. and Y.H. All authors have read and agreed to the published version of the manuscript.
Institutional Review Board Statement
Not applicable.
Informed Consent Statement
Not applicable.
Data Availability Statement
Code and data are available at: https://github.com/TingheZhang/T-GEM (accessed on 1 April 2022).
Conflicts of Interest
The authors declare no conflict of interest.
Funding Statement
This research was supported by Cancer Prevention and Research Institute of Texas (RP160732 and RP190346), and National Institutes of Health (K99CA248944 and R00CA248944).
Footnotes
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.
References
- 1.Chiu Y.C., Chen H.H., Gorthi A., Mostavi M., Zheng S., Huang Y., Chen Y. Deep learning of pharmacogenomics resources: Moving towards precision oncology. Brief. Bioinform. 2020;21:2066–2083. doi: 10.1093/bib/bbz144. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Huang K., Xiao C., Glass L.M., Critchlow C.W., Gibson G., Sun J. Machine Learning Applications for Therapeutic Tasks with Genomics Data. arXiv. 2021 doi: 10.1016/j.patter.2021.100328.2105.01171 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Mostavi M., Chiu Y.-C., Huang Y., Chen Y. Convolutional neural network models for cancer type prediction based on gene expression. BMC Med. Genom. 2020;13:44. doi: 10.1186/s12920-020-0677-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.Yoo S., Gujrathi I., Haider M.A., Khalvati F. Prostate cancer detection using deep convolutional neural networks. Sci. Rep. 2019;9:19518. doi: 10.1038/s41598-019-55972-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.Ragab D.A., Sharkas M., Marshall S., Ren J. Breast cancer detection using deep convolutional neural networks and support vector machines. PeerJ. 2019;7:e6201. doi: 10.7717/peerj.6201. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.Ramirez R., Chiu Y.C., Zhang S.Y., Ramirez J., Chen Y.D., Huang Y.F., Jin Y.F. Prediction and interpretation of cancer survival using graph convolution neural networks. Methods. 2021;192:120–130. doi: 10.1016/j.ymeth.2021.01.004. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Chiu Y.-C., Chen H.-I.H., Zhang T., Zhang S., Gorthi A., Wang L.-J., Huang Y., Chen Y. Predicting drug response of tumors from integrated genomic profiles by deep neural networks. BMC Med. Genom. 2019;12:18. doi: 10.1186/s12920-018-0460-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Sharifi-Noghabi H., Zolotareva O., Collins C.C., Ester M. MOLI: Multi-omics late integration with deep neural networks for drug response prediction. Bioinformatics. 2019;35:i501–i509. doi: 10.1093/bioinformatics/btz318. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.Chiu Y.-C., Zheng S., Wang L.-J., Iskra B.S., Rao M.K., Houghton P.J., Huang Y., Chen Y. DeepDEP: Deep learning of a cancer dependency map using cancer genomics. Nat. Commun. 2020. under review . [DOI] [PMC free article] [PubMed]
- 10.Banegas-Luna A.J., Peña-García J., Iftene A., Guadagni F., Ferroni P., Scarpato N., Zanzotto F.M., Bueno-Crespo A., Pérez-Sánchez H. Towards the Interpretability of Machine Learning Predictions for Medical Applications Targeting Personalised Therapies: A Cancer Case Survey. Int. J. Mol. Sci. 2021;22:4394. doi: 10.3390/ijms22094394. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Bazgir O., Zhang R., Dhruba S.R., Rahman R., Ghosh S., Pal R. Representation of features as images with neighborhood dependencies for compatibility with convolutional neural networks. Nat. Commun. 2020;11:4391. doi: 10.1038/s41467-020-18197-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Fatima N., Rueda L. iSOM-GSN: An integrative approach for transforming multi-omic data into gene similarity networks via self-organizing maps. Bioinformatics. 2020;36:4248–4254. doi: 10.1093/bioinformatics/btaa500. [DOI] [PubMed] [Google Scholar]
- 13.Sharma A., Vans E., Shigemizu D., Boroevich K.A., Tsunoda T. DeepInsight: A methodology to transform a non-image data to an image for convolution neural network architecture. Sci. Rep. 2019;9:11399. doi: 10.1038/s41598-019-47765-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Michael K.Y., Ma J., Fisher J., Kreisberg J.F., Raphael B.J., Ideker T. Visible machine learning for biomedicine. Cell. 2018;173:1562–1565. doi: 10.1016/j.cell.2018.05.056. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Elmarakeby H.A., Hwang J., Arafeh R., Crowdis J., Gang S., Liu D., AlDubayan S.H., Salari K., Kregel S., Richter C., et al. Biologically informed deep neural network for prostate cancer discovery. Nature. 2021;598:348–352. doi: 10.1038/s41586-021-03922-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Ma J.Z., Yu M.K., Fong S., Ono K., Sage E., Demchak B., Sharan R., Ideker T. Using deep learning to model the hierarchical structure and function of a cell. Nat. Methods. 2018;15:290–298. doi: 10.1038/nmeth.4627. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Kuenzi B.M., Park J., Fong S.H., Sanchez K.S., Lee J., Kreisberg J.F., Ma J.Z., Ideker T. Predicting Drug Response and Synergy Using a Deep Learning Model of Human Cancer Cells. Cancer Cell. 2020;38:672–684.e6. doi: 10.1016/j.ccell.2020.09.014. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Vaswani A., Shazeer N., Parmar N., Uszkoreit J., Jones L., Gomez A.N., Kaiser Ł., Polosukhin I. Proceedings of the Advances in Neural Information Processing Systems. Neural Information Processing Systems Foundation, Inc. (NeurIPS); La Jolla, CA, USA: 2017. Attention is all you need; pp. 5998–6008. [Google Scholar]
- 19.Devlin J., Chang M.-W., Lee K., Toutanova K. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv. 20181810.04805 [Google Scholar]
- 20.Tenney I., Das D., Pavlick E. BERT rediscovers the classical NLP pipeline. arXiv. 20191905.05950 [Google Scholar]
- 21.Jawahar G., Sagot B., Seddah D. Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics; Florence, Italy: 2019. What does BERT learn about the structure of language? [Google Scholar]
- 22.Jo J.-Y., Myaeng S.-H. Roles and Utilization of Attention Heads in Transformer-based Neural Language Models; Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics; Online. 5–10 July 2020; pp. 3404–3417. [Google Scholar]
- 23.The Cancer Genome Atlas Research Network. Weinstein J.N., Collisson E.A., Mills G.B., Shaw K.R.M., Ozenberger B.A., Ellrott K., Shmulevich I., Sander C., Stuart J.M. The Cancer Genome Atlas Pan-Cancer analysis project. Nat. Genet. 2013;45:1113–1120. doi: 10.1038/ng.2764. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24.Colaprico A., Silva T.C., Olsen C., Garofano L., Cava C., Garolini D., Sabedot T.S., Malta T.M., Pagnotta S.M., Castiglioni I., et al. TCGAbiolinks: An R/Bioconductor package for integrative analysis of TCGA data. Nucleic Acids Res. 2016;44:e71. doi: 10.1093/nar/gkv1507. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.Yuan H., Yan M., Zhang G., Liu W., Deng C., Liao G., Xu L., Luo T., Yan H., Long Z., et al. CancerSEA: A cancer single-cell state atlas. Nucleic Acids Res. 2019;47:D900–D908. doi: 10.1093/nar/gky939. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Zheng G.X.Y., Terry J.M., Belgrader P., Ryvkin P., Bent Z.W., Wilson R., Ziraldo S.B., Wheeler T.D., McDermott G.P., Zhu J., et al. Massively parallel digital transcriptional profiling of single cells. Nat. Commun. 2017;8:14049. doi: 10.1038/ncomms14049. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Wang T., Bai J., Nabavi S. Single-cell classification using graph convolutional networks. BMC Bioinform. 2021;22:364. doi: 10.1186/s12859-021-04278-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28.Hao Y., Dong L., Wei F., Xu K. Self-attention attribution: Interpreting information interactions inside transformer. arXiv. 2020 doi: 10.1609/aaai.v35i14.17533.2004.11207 [DOI] [Google Scholar]
- 29.Kokhlikyan N., Miglani V., Martin M., Wang E., Alsallakh B., Reynolds J., Melnikov A., Kliushkina N., Araya C., Yan S. Captum: A unified and generic model interpretability library for pytorch. arXiv. 20202009.07896 [Google Scholar]
- 30.Shannon P., Markiel A., Ozier O., Baliga N.S., Wang J.T., Ramage D., Amin N., Schwikowski B., Ideker T. Cytoscape: A software environment for integrated models of biomolecular interaction networks. Genome Res. 2003;13:2498–2504. doi: 10.1101/gr.1239303. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 31.Kingma D.P., Ba J. Adam: A method for stochastic optimization. arXiv. 20141412.6980 [Google Scholar]
- 32.Jin H., Song Q., Hu X. Auto-keras: An efficient neural architecture search system; Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining; Anchorage, AK, USA. 4–8 August 2019; pp. 1946–1956. [Google Scholar]
- 33.Qiu H., Yan Q., Luo X., Zhang H., Bao W., Wan X. EMX2 is downregulated in endometrial cancer and correlated with tumor progression. Int. J. Gynecol. Pathol. 2013;32:193–198. doi: 10.1097/PGP.0b013e31825d8049. [DOI] [PubMed] [Google Scholar]
- 34.Imrich S., Hachmeister M., Gires O. EpCAM and its potential role in tumor-initiating cells. Cell Adhes. Migr. 2012;6:30–38. doi: 10.4161/cam.18953. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35.He D., Zhang X., Tu J. Diagnostic significance and carcinogenic mechanism of pan-cancer gene POU5F1 in liver hepatocellular carcinoma. Cancer Med. 2020;9:8782–8800. doi: 10.1002/cam4.3486. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 36.Teng Y.-W., Mehedint M.G., Garrow T.A., Zeisel S.H. Deletion of betaine-homocysteine S-methyltransferase in mice perturbs choline and 1-carbon metabolism, resulting in fatty liver and hepatocellular carcinomas. J. Biol. Chem. 2011;286:36258–36267. doi: 10.1074/jbc.M111.265348. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37.van Bodegraven E.J., van Asperen J.V., Robe P.A.J., Hol E.M. Importance of GFAP isoform-specific analyses in astrocytoma. Glia. 2019;67:1417–1433. doi: 10.1002/glia.23594. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 38.Samaeekia R., Adorno-Cruz V., Bockhorn J., Chang Y.F., Huang S., Prat A., Ha N., Kibria G., Huo D., Zheng H., et al. miR-206 Inhibits Stemness and Metastasis of Breast Cancer by Targeting MKL1/IL11 Pathway. Clin. Cancer Res. 2017;23:1091–1103. doi: 10.1158/1078-0432.CCR-16-0943. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39.Aponte P.M., Caicedo A. Stemness in Cancer: Stem Cells, Cancer Stem Cells, and Their Microenvironment. Stem Cells Int. 2017;2017:5619472. doi: 10.1155/2017/5619472. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40.Wang X., Semba T., Phi L.T.H., Chainitikun S., Iwase T., Lim B., Ueno N.T. Targeting Signaling Pathways in Inflammatory Breast Cancer. Cancers. 2020;12:2479. doi: 10.3390/cancers12092479. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 41.Chen X., Cao Q., Liao R., Wu X., Xun S., Huang J., Dong C. Loss of ABAT-Mediated GABAergic System Promotes Basal-Like Breast Cancer Progression by Activating Ca(2+)-NFAT1 Axis. Theranostics. 2019;9:34–47. doi: 10.7150/thno.29407. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 42.Bredel M., Kim H., Thudi N.K., Scholtens D.M., Bonner J.A., Sikic B.I. NFKBIA deletion in triple-negative breast cancer. J. Clin. Oncol. 2013;31:1012. doi: 10.1200/jco.2013.31.15_suppl.1012. [DOI] [Google Scholar]
- 43.Yang C., Zhao X., Cui N., Liang Y. Cadherins Associate with Distinct Stem Cell-Related Transcription Factors to Coordinate the Maintenance of Stemness in Triple-Negative Breast Cancer. Stem Cells Int. 2017;2017:5091541. doi: 10.1155/2017/5091541. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 44.Zhang J., Zhang J., Xu S., Zhang X., Wang P., Wu H., Xia B., Zhang G., Lei B., Wan L., et al. Hypoxia-Induced TPM2 Methylation is Associated with Chemoresistance and Poor Prognosis in Breast Cancer. Cell. Physiol. Biochem. 2018;45:692–705. doi: 10.1159/000487162. [DOI] [PubMed] [Google Scholar]
- 45.Mehra R., Varambally S., Ding L., Shen R., Sabel M.S., Ghosh D., Chinnaiyan A.M., Kleer C.G. Identification of GATA3 as a breast cancer prognostic marker by global gene expression meta-analysis. Cancer Res. 2005;65:11259–11264. doi: 10.1158/0008-5472.CAN-05-2495. [DOI] [PubMed] [Google Scholar]
- 46.Nakshatri H., Badve S. FOXA1 as a therapeutic target for breast cancer. Expert Opin. Ther. Targets. 2007;11:507–514. doi: 10.1517/14728222.11.4.507. [DOI] [PubMed] [Google Scholar]
- 47.Dai X., Cheng H., Chen X., Li T., Zhang J., Jin G., Cai D., Huang Z. FOXA1 is Prognostic of Triple Negative Breast Cancers by Transcriptionally Suppressing SOD2 and IL6. Int. J. Biol. Sci. 2019;15:1030–1041. doi: 10.7150/ijbs.31009. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 48.Papadakos K.S., Hagerling C., Rydén L., Larsson A.M., Blom A.M. High Levels of Expression of Cartilage Oligomeric Matrix Protein in Lymph Node Metastases in Breast Cancer Are Associated with Reduced Survival. Cancers. 2021;13:5876. doi: 10.3390/cancers13235876. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 49.Pickard A., McCance D.J. IGF-Binding Protein 2—Oncogene or Tumor Suppressor? Front. Endocrinol. 2015;6:25. doi: 10.3389/fendo.2015.00025. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 50.Hou J., Wang Z., Xu H., Yang L., Yu X., Yang Z., Deng Y., Meng J., Feng Y., Guo X. Stanniocalicin 2 suppresses breast cancer cell migration and invasion via the PKC/claudin-1-mediated signaling. PLoS ONE. 2015;10:e0122179. doi: 10.1371/journal.pone.0122179. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 51.Haam S., Han J.-H., Lee H.W., Koh Y.W. Tumor Nonimmune-Microenvironment-Related Gene Expression Signature Predicts Brain Metastasis in Lung Adenocarcinoma Patients after Surgery: A Machine Learning Approach Using Gene Expression Profiling. Cancers. 2021;13:4468. doi: 10.3390/cancers13174468. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 52.Amoedo N.D., Sarlak S., Obre E., Esteves P., Bégueret H., Kieffer Y., Rousseau B., Dupis A., Izotte J., Bellance N., et al. Targeting the mitochondrial trifunctional protein restrains tumor growth in oxidative lung carcinomas. J. Clin. Investig. 2021;131 doi: 10.1172/JCI133081. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 53.Guo C., Lu H., Gao W., Wang L., Lu K., Wu S., Pataer A., Huang M., El-Zein R., Lin T., et al. Insulin-like growth factor binding protein-2 level is increased in blood of lung cancer patients and associated with poor survival. PLoS ONE. 2013;8:e74973. doi: 10.1371/journal.pone.0074973. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 54.Yang L., Lin M., Ruan W.J., Dong L.L., Chen E.G., Wu X.H., Ying K.J. Nkx2-1: A novel tumor biomarker of lung cancer. J. Zhejiang Univ. Sci. B. 2012;13:855–866. doi: 10.1631/jzus.B1100382. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 55.Chang J., Liu S., Li B., Huo Z., Wang X., Zhang H. MiR-338-3p improved lung adenocarcinoma by AKAP12 suppression. Arch. Med. Sci. AMS. 2021;17:462. doi: 10.5114/aoms.2019.90913. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 56.Song L., Chen Z., Zhang M., Zhang M., Lu X., Li C., Miao L. DDIT4 overexpression associates with poor prognosis in lung adenocarcinoma. J. Cancer. 2021;12:6422–6428. doi: 10.7150/jca.60118. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 57.Yu D.H., Ruan X.L., Huang J.Y., Liu X.P., Ma H.L., Chen C., Hu W.D., Li S. Analysis of the Interaction Network of Hub miRNAs-Hub Genes, Being Involved in Idiopathic Pulmonary Fibers and Its Emerging Role in Non-small Cell Lung Cancer. Front. Genet. 2020;11:302. doi: 10.3389/fgene.2020.00302. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 58.Tian T., Hong F., Wang Z., Hu J., Chen N., Lv L., Yi Q. HSD17B6 downregulation predicts poor prognosis and drives tumor progression via activating Akt signaling pathway in lung adenocarcinoma. Cell Death Discov. 2021;7:341. doi: 10.1038/s41420-021-00737-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 59.Beauchemin N., Arabzadeh A. Carcinoembryonic antigen-related cell adhesion molecules (CEACAMs) in cancer progression and metastasis. Cancer Metastasis Rev. 2013;32:643–671. doi: 10.1007/s10555-013-9444-6. [DOI] [PubMed] [Google Scholar]
- 60.Gutschner T., Hämmerle M., Eissmann M., Hsu J., Kim Y., Hung G., Revenko A., Arun G., Stentrup M., Gross M., et al. The noncoding RNA MALAT1 is a critical regulator of the metastasis phenotype of lung cancer cells. Cancer Res. 2013;73:1180–1189. doi: 10.1158/0008-5472.CAN-12-2850. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 61.Takamochi K., Ohmiya H., Itoh M., Mogushi K., Saito T., Hara K., Mitani K., Kogo Y., Yamanaka Y., Kawai J., et al. Novel biomarkers that assist in accurate discrimination of squamous cell carcinoma from adenocarcinoma of the lung. BMC Cancer. 2016;16:760. doi: 10.1186/s12885-016-2792-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 62.Lakshmanan I., Chaudhary S., Vengoji R., Seshacharyulu P., Rachagani S., Carmicheal J., Jahan R., Atri P., Chirravuri-Venkata R., Gupta R., et al. ST6GalNAc-I promotes lung cancer metastasis by altering MUC5AC sialylation. Mol. Oncol. 2021;15:1866–1881. doi: 10.1002/1878-0261.12956. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 63.Huang C., Liu J., Xiong B., Yonemura Y., Yang X. Expression and prognosis analyses of forkhead box A (FOXA) family in human lung cancer. Gene. 2019;685:202–210. doi: 10.1016/j.gene.2018.11.022. [DOI] [PubMed] [Google Scholar]
- 64.Song J., Liu W., Wang J., Hao J., Wang Y., You X., Du X., Zhou Y., Ben J., Zhang X., et al. GALNT6 promotes invasion and metastasis of human lung adenocarcinoma cells through O-glycosylating chaperone protein GRP78. Cell Death Dis. 2020;11:352. doi: 10.1038/s41419-020-2537-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 65.Chen C., Guo Q., Tang Y., Qu W., Zuo J., Ke X., Song Y. Screening and evaluation of the role of immune genes of brain metastasis in lung adenocarcinoma progression based on the TCGA and GEO databases. J. Thorac. Dis. 2021;13:5016–5034. doi: 10.21037/jtd-21-935. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 66.Cha Y.J., Kim D.H., Jung W.H., Koo J.S. Expression of sarcosine metabolism-related proteins according to metastatic site in breast cancer. Int. J. Clin. Exp. Pathol. 2014;7:7824–7833. [PMC free article] [PubMed] [Google Scholar]
- 67.Shao X., Liao J., Lu X., Xue R., Ai N., Fan X. scCATCH: Automatic Annotation on Cell Types of Clusters from Single-Cell RNA Sequencing Data. iScience. 2020;23:100882. doi: 10.1016/j.isci.2020.100882. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 68.Bláha J., Kalousková B., Skořepa O., Pažický S., Novák P., Vaněk O. High-level expression and purification of soluble form of human natural killer cell receptor NKR-P1 in HEK293S GnTI−cells. Protein Expr. Purif. 2017;140:36–43. doi: 10.1016/j.pep.2017.07.016. [DOI] [PubMed] [Google Scholar]
- 69.Guo F., Yuan Y., Chen Z., Gao F., Li X., Wang H., Wang X., Bai G. Downregulation of the long non-coding RNA MALAT1 in tenofovir-treated pregnant women with hepatitis B virus infection promotes immune recovery of natural killer cells via the has-miR-155-5p/HIF-1α axis. Int. Immunopharmacol. 2022;107:108701. doi: 10.1016/j.intimp.2022.108701. [DOI] [PubMed] [Google Scholar]
- 70.Greenberg Z.J., Monlish D.A., Bartnett R.L., Yang Y., Shen G., Li W., Bednarski J.J., Schuettpelz L.G. The Tetraspanin CD53 Regulates Early B Cell Development by Promoting IL-7R Signaling. J. Immunol. 2020;204:58–67. doi: 10.4049/jimmunol.1900539. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 71.Yang H., Liang N., Wang M., Fei Y., Sun J., Li Z., Xu Y., Guo C., Cao Z., Li S., et al. Long noncoding RNA MALAT-1 is a novel inflammatory regulator in human systemic lupus erythematosus. Oncotarget. 2017;8:77400–77406. doi: 10.18632/oncotarget.20490. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 72.Meurer T., Arrieta-Bolaños E., Metzing M., Langer M.-M., van Balen P., Falkenburg J.H.F., Beelen D.W., Horn P.A., Fleischhauer K., Crivello P. Dissecting Genetic Control of HLA-DPB1 Expression and Its Relation to Structural Mismatch Models in Hematopoietic Stem Cell Transplantation. Front. Immunol. 2018;9:2236. doi: 10.3389/fimmu.2018.02236. [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
Code and data are available at: https://github.com/TingheZhang/T-GEM (accessed on 1 April 2022).