Skip to main content
Scientific Reports logoLink to Scientific Reports
. 2022 Oct 20;12:17519. doi: 10.1038/s41598-022-22447-y

Heterogeneous graph neural network for lncRNA-disease association prediction

Hong Shi 1, Xiaomeng Zhang 1, Lin Tang 2, Lin Liu 1,
PMCID: PMC9585029  PMID: 36266433

Abstract

Identifying lncRNA-disease associations is conducive to the diagnosis, treatment and prevention of diseases. Due to the expensive and time-consuming methods verified by biological experiments, prediction methods based on computational models have gradually become an important means of lncRNA-disease associations discovery. However, existing methods still have challenges to make full use of network topology information to identify potential associations between lncRNA and disease in multi-source data. In this study, we propose a novel method called HGNNLDA for lncRNA-disease association prediction. First, HGNNLDA constructs a heterogeneous network composed of lncRNA similarity network, lncRNA-disease association network and lncRNA-miRNA association network; Then, on this heterogeneous network, various types of strong correlation neighbors with fixed size are sampled for each node by restart random walk; Next, the embedding information of lncRNA and disease in each lncRNA-disease association pair is obtained by the method of type-based neighbor aggregation and all types combination though heterogeneous graph neural network, in which attention mechanism is introduced considering that different types of neighbors will make different contributions to the prediction of lncRNA-disease association. As a result, the area under the receiver operating characteristic curve (AUC) and the area under the precision-recall curve (AUPR) under fivefold cross-validation (5FCV) are 0.9786 and 0.8891, respectively. Compared with five state-of-art prediction models, HGNNLDA has better prediction performance. In addition, in two types of case studies, it is further verified that our method can effectively predict the potential lncRNA-disease associations, and have ability to predict new diseases without any known lncRNAs.

Subject terms: Computational biology and bioinformatics, Biomarkers, Diseases

Introduction

Long non-coding RNAs (lncRNAs) are non-coding RNAs with more than 200 nt (nucleotides) in length1. More and more studies have shown that lncRNAs participates in many important biological processes, including gene transcription, cell differentiation and genetic regulation2. Moreover, Complex diseases that seriously endanger human health are also inseparable from the abnormal expression of lncRNAs, including diabetes3, cardiovascular diseases4, HIV5, mental disorders6 and some cancers such as lung cancer7, breast cancer8 and prostate cancer9. Therefore, identifying the associations between lncRNAs and diseases contributes to understanding the pathogenesis and principles of the diseases, and also provides help for the diagnosis, treatment and prevention of human disease. However, the traditional biological experiments take up a long time, cost much, and have some blindness, all of which will hinder the research process. In recent years, established lncRNA databases such as LncRNADisease2.010, Lnc2Cancer v2.011, NRED12, MNDR13, and GeneRIF14 have made it possible to develop computational methods for predicting potential lncRNA-disease associations. According to the different ideas of algorithms, the existing methods for predicting lncRNA-disease associations can be broadly classified into two categories. They are the method based on biological networks and machine learning, respectively.

Computational methods based on biological networks often rely on the known associations information between lncRNA and disease to build heterogeneous networks. Then lncRNA-disease association prediction is carried out based on this heterogeneous networks. For example, Sun et al.15 proposed a network-based computational model RWRlncD, that known lncRNA-disease association network is used to calculate the lncRNA similarity to predict the disease relevance of lncRNAs. Gu et al.16 proposed a model for a random walk on a global network (GrwLDA) that uses random walk in the lncRNA similarity network and the disease similarity network to predict potential lncRNA-disease associations. However, GrwLDA had difficulties in optimizing the model parameters. Wen et al.17 proposed the Lap-BiRWRHLDA model, which Laplace normalized the similarity matrix before constructing the lncRNA-disease networks, which integrated the two similarity networks through known lncRNA-disease associations, and then predicted lncRNA-disease associations using a double random walk on this heterogeneous networks. Zhang et al.18 propose a model LncRDNetFlow based on a global network framework that integrated multisource networks, including lncRNAs similarity network, proteins interaction network, diseases similarity network, and associations information among heterogeneous nodes. The model was able to predict potential associations information for an isolated disease. Zhao et al.19 developed a new random walk method MHRWR based on multisource networks. This method introduced disease-gene network and lncRNA-gene network to build a multi-layer network, so as to extract more potential information. Finally, a multi-layer random walk method was used to predict the associations of lncRNA-disease.

Computational methods based machine learning predict potential associations between lncRNAs and diseases by building lncRNA-disease association models, and train the model to improve accuracy using known lncRNA-disease associations data. Chen et al.20 assumed that similar diseases were often associated with functionally similar lncRNAs, and developed a model LRLSLDA based on a semi-supervised learning framework, where LRLSLDA effectively predicted potential lncRNA-disease associations by integrating known lncRNA-disease associations and lncRNA expression profiles. Nonetheless, LRLSLDA had the problem of optimize the model parameters. Subsequently, Chen et al.21 proposed a new lncRNA-disease prediction model named LNCSIM. LNCSIM further improved LRLSLDA model by introducing lncRNA-disease prediction similarity score. However, this method still could not solve the problem of parameter selection of semantic contribution factors. Zhao et al.22 developed a naive Bayesian-based computational approach that integrated various information of disease-related lncRNAs, including genomic, regulome, transcriptome, which resulted in successfully predicting 707 potential cancer-associated lncRNAs. Lan et al.23 proposed a novel computational method that used Katcher means to fuse the lncRNA and disease similarity matrixs of multiple data sources and predicted potential lncRNA-disease associations by the SVM classifier. Sheng et al.24 used random walk and convolution autoencoders to obtain new feature distributions and then input them into the model to reveal the potential associations between lncRNAs and diseases.

These two types of approaches still have methodological weaknesses. The methods based on biological network rely heavily on the constructed lncRNA-disease heterogeneous network. When network structure changes, this kind of method can’t effectively deal with it. The problem of the method based on machine learning is how to select the optimal features. Most existing machine learning methods do not take full advantage of the rich topological information contained in heterogeneous networks. To make full use of the lncRNAs and diseases feature information and the local and global information on the lncRNA-disease association data, the graph neural network approach appears in some new studies recently. For example, Xuan et al.25 used graph convolution network and convolutional neural network to learn the network structure information and the local network features of lncRNA-disease association pair. Wu et al.26 used graph convolutional network (GCN) as encoder to obtain the features of lncRNAs and diseases on the heterogeneous network, and then calculated the interaction score between lncRNA and disease by using the inner product of two potential factor vector. Zhang et al.27 utilized meta-paths to represent complex semantic information between entities in the network and introduced attention mechanisms to learn the weights of each neighborhood under the metapath and finally aggregate the potential features they obtained from the GCN model. A graph auto-encoder was leveraged to acquire low-dimensional features, finally used a random forest classifer for lncRNA-disease prediction28. Zhao et al.29 proposed a deep learning algorithm HGATLDA based on heterogeneous graph attention network. HGATLDA uses graph attention network to learn node embedding from isomorphic and heterogeneous subgraphs. In addition, a computational model based on graph attention network and multilayer perceptron (MLP) was proposed for association prediction30. However, these methods directly use graph attention network to extract features, and do not take into account the different number of neighbor nodes and the heterogeneity of node types. PANDA applied a graph convolutional auto-encoders for feature extraction and utilized a neural network to predict LDAs31.These graph neural network methods realize the capture and utilization of topological information in heterogeneous networks, but ignore the heterogeneity of nodes and edges in heterogeneous graphs.

Inspired by Zhang et al.32, the heterogeneity of structure and content in the heterogeneous graph is considered. We propose a novel method for lncRNA-disease association prediction called HGNNLDA. First, a heterogeneous network is constructed, which is composed of the similar network of lncRNAs, the known lncRNA-disease association network and the known lncRNA-miRNA association network. Then, a fixed-size sampling of strongly correlated neighbors is performed by restart random walk for each lncRNA and disease, and the sampled neighbors are grouped according to the types of nodes. Then, the feature vectors of sampled lncRNA, disease and miRNA are obtained by word2vec. The final embedding information of each lncRNA and disease is extracted by aggregating the sampling neighbors according to types and fusing different types, in which attention mechanism is introduced to indicate the importance of different types of neighbors. Finally, the embedding obtained from above steps of each lncRNA-disease association pair are used as the input of classifier, and the prediction score of association pair is calculated. The experimental results show that the AUC and AUPR values of HGNNLDA under fivefold cross validation (5FCV) are 0.9786 and 0.8891, respectively, which is superior to other state-of-art methods. In addition, two case studies show that HGNNLDA has the ability to predict disease-related lncRNA without any known association.

Results

Performance evaluation

We considered 2697 known lncRNA-disease associations as positive samples, but the number of positive samples only account for 2.7% of the total number of samples, so some previous studies3336 selected negative samples with the same number of positive samples from all unknown association pairs. We followed the same strategy and randomly selected 2697 lncRNA-disease associations from all the unknown lncRNA-disease associations to be the negative samples. After constructing the training set of the model, fivefold cross validation (5FCV) was used to evaluate the prediction performance of HGNNLDA. For 5FCV, the sample set was divided into 5 disjoint subsets on average, among which 4 subsets were utilized to train the model and the remaining subset was utilized for testing in each round. Then, HGNNLDA model trained was used to obtain the score of each test sample. The higher the score, the more likely it is that this lncRNA is related to the disease. Next, all test samples were sorted in descending order according to their scores. On this basis, we calculated the true positive rate (TPR) and false positive rate (FPR), Precision and Recall under different thresholds. The specific calculation is as follows:

TPR=TPTP+TNFPR=FPTN+FPPrecision=TPTP+FPRecall=TPTN+FN 1

Where TP (true positive) means that positive samples are correctly predicted as positive samples; FN (false negative) indicates that the positive sample is erroneously predicted as a negative sample; FP (false positive) means that the negative sample is erroneously predicted as a positive sample; TN (true negative) means that the negative sample is correctly predicted as a negative sample. Then, the ROC curve was drawn with TPR as the vertical axis and FPR as the horizontal axis, and the area under the ROC curve (AUC value) was used as the performance index to evaluate the prediction performance of the model. If the AUC value is larger, the prediction performance of this model is better. To improve the evaluation of the model performance when the positive and negative samples were seriously unbalanced, we also calculated AUPR value to evaluate the overall performance of the model.

Comparison with other models

In order to further evaluate the prediction performance of HGNNLDA method, we compared it with five state-of-art lncRNA-disease association prediction models, such as SIMCLDA37, MFLDA38, LDAP23, CNNLDA39 and GCNLDA25. Under the 5FCV, the average AUCs and AUPRs of all lncRNA-disease association prediction models as shown in Table 1. All experimental results of compared models come from Yao et al.’s previous study on lncRNA-disease association prediction, and the results of these models were tested on the same datasets40. ROC curve of each cross-validation of HGNNLDA is shown in Fig. 1. Other models didn’t take into account that the neighbor nodes of some nodes may not contain all types of nodes, and assumed that each type of node contributed the same to the prediction of lncRNA-disease association. HGNNLDA solves the problems of other models by restarting random walk and introducing the attention mechanism.

Table 1.

The mean AUCs and AUPRs of different methods.

Method AUC AUPR
SIMCLDA 0.746 0.095
MFLDA 0.626 0.066
LDAP 0.863 0.166
CNNLDA 0.952 0.251
GCNLDA 0.959 0.223
HGNNLDA 0.9786 0.8891

Figure 1.

Figure 1

ROC curves of HGNNLDA based on fivefold cross-validation.

Ablation study

To analyze the necessity of each component of our model, we adopt two variants of HGNNLDA (HGNNLDA-noNeigh and HGNNLDA-noAttention) as the comparison method. Specifically, HGNNLDA-noNeigh means that the embedded information of each node is only obtained by word2vec, and the information of any neighboring nodes is not aggregated. HGNNLDA-noAttention uses fully connected neural network instead of attention mechanism to aggregate the embedding of different types of neighbors, which means that different types of neighbor nodes are equally important for the final embedding of lncRNA and disease. Figure 2 shows the average AUC obtained using HGNNLDA and two variant models. HGNNLDA has better performance than HGNNLDA-noNeigh, which indicates that aggregating the information of neighboring nodes can better generate the embedded information of nodes. HGNNLDA gets better results than HGNNLDA-noAttention, which shows that attention mechanism can capture the influence of different types of nodes.

Figure 2.

Figure 2

Performance of HGNNLDA and its variants.

The effects of embedding size

Embedding size plays an important role in HGNNLDA, which is able to directly affect the performance of the model. In the experiment, we set different embedding dimension d (i.e. 8, 16, 32, 64, 128, 256), and evaluated the prediction performance under different setting. As can be seen from Fig. 3, within a certain range, the larger the embedding dimension, the better the node representation can be learned, and the higher the AUC value. However, when the embedding dimension increase continuously, the AUC value will become stable or slightly worse, which may be caused by over-fitting. Accounting for this factor, the embedding size is set to 128 in this paper.

Figure 3.

Figure 3

Impact of embedding size.

Case studies

To further verify the accuracy and effectiveness of HGNNLDA, we conducted two types of case studies.

For the first type of case study, we applied our proposed method to predict the potential lncRNA-disease associations of three common diseases (lung cancer, colon cancer and osteosarcoma). First, for a specific disease, we regarded all known associations between lncRNAs and diseases as training samples and unknown associations with this disease as candidate samples. Then, we scored all unknown candidate samples of lncRNA-/lung cancer/colon cancer/osteosarcoma, then sorted the scores in descending order and select the top 10 candidate associations related to this disease. The prediction results were verified using two databases (LncRNADisease10 database and LncRNA2Cancer11 database). Table 2 showed the top 10 results of predicting the potential associations with lung cancer, colon cancer and osteosarcoma, the accuracy reached 100%, 80% and 80% respectively. The results showed that our method can effectively predict the potential lncRNA-disease associations.

Table 2.

The top 10 predicted lncRNAs associated with lung cancer, colon cancer, osteosarcoma.

Rank LncRNA Disease Evidence
1 GAS5 Lung cancer LncRNAdisease
2 NEG8 Lung cancer Lnc2Cancer
3 HOOTTIP Lung cancer Lnc2Cancer
4 LINC00472 Lung cancer Lnc2Cancer
5 ZFAS1 Lung cancer Lnc2Cancer
6 HULC Lung cancer Lnc2Cancer
7 BCAR4 Lung cancer Lnc2Cancer
8 CASC15 Lung cancer Lnc2Cancer
9 BCYRN1 Lung cancer Lnc2Cancer
10 GHET1 Lung cancer Lnc2Cancer
1 UCA1 Colon cancer LncRNAdisease
2 OIP5-AS1 Colon cancer Lnc2Cancer
3 HOTTIP Colon cancer LncRNAdisease
4 HOTAIR Colon cancer LncRNAdisease
5 LINC00319 Colon cancer Unconfirmed
6 PVT1 Colon cancer Lnc2Cancer
7 GAS5 Colon cancer Lnc2Cancer
8 KCNQ10T1 Colon cancer Lnc2Cancer
9 DANCR Colon cancer Lnc2Cancer
10 BANCR Colon cancer Unconfirmed
1 NEAT1 Osteosarcoma Lnc2Cancer
2 XIST Osteosarcoma Lnc2Cancer
3 CCAT1 Osteosarcoma LncRNAdisease
4 EWSAT1 Osteosarcoma LncRNAdisease
5 AFAP1-AS1 Osteosarcoma Unconfirmed
6 KCNQ10T1 Osteosarcoma Lnc2Cancer
7 MIR155HG Osteosarcoma Unconfirmed
8 GAS5 Osteosarcoma Lnc2Cancer
9 PVT1 Osteosarcoma Lnc2Cancer
10 OIP5-AS1 Osteosarcoma Lnc2Cancer

For the second type of case study, We evaluated the ability of our proposed method to predict the new associations of diseases without any known related lncRNA. We took breast as an example in this case study. First, we set the known associations of breast cancer as unknown associations, and all lncRNAs were considered as candidate lncRNAs. The HGNNLDA was used to score these candidate lncRNAs associated with breast cancer. We found that 27 of the top 30 lncRNA were confirmed by LncRNAdisease database or LncRNA2Cancer database, as shown in Table 3. This result shows that HGNNLDA can effectively predict the potential associations of diseases without any known related lncRNAs.

Table 3.

The top 30 predicted lncRNAs associated with breast cancer.

LncRNA (1-15) Evidence LncRNA (16-30) Evidence
H19 Lnc2Cancer LSINCT5 Lnc2Cancer
HOTTIP Lnc2Cancer PVT1 Lnc2Cancer
CDKN2B-AS1 LncRNAdisease ZFAS1 Lnc2Cancer
AFAP1-AS1 Lnc2Cancer NCRUPAR Unconfirmed
KCNQ1OT1 Lnc2Cancer SOX2-OT LncRNAdisease
LINC00472 Lnc2Cancer TP53COR1 Unconfirmed
CASC16 LncRNAdisease BCAR4 Lnc2Cancer
MALAT1 Lnc2Cancer NPSR1-AS1 Unconfirmed
NEAT1 Lnc2Cancer GHET1 Lnc2Cancer
LINC00583 LncRNAdisease MIR17HG LncRNAdisease
XIST Lnc2Cancer LINC-ROR Lnc2Cancer
HOTAIR Lnc2Cancer NBAT1 Lnc2Cancer
CCAT2 Lnc2Cancer BANCR Lnc2Cancer
BCYRN1 LncRNAdisease HOTAIRM1 Lnc2Cancer
SPRY4-IT1 Lnc2Cancer DANCR Lnc2Cancer

Discussion

Identifying associations between lncRNAs and diseases will have a huge impact on our treatment and prevention of some diseases. Therefore, we propose a novel method HGNNLDA to predict the potential associations between lncRNAs and diseases. From the comparison of experimental results, it can be seen that HGNNLDA has superior performance for predicting lncRNA-disease associations. In addition, two types of cases also verify that HGNNLDA has the ability to identify potential lncRNA-disease associations, and can effectively predict new diseases without any known lncRNA.

The reliable performance of HGNNLDA is related to the following factors. First, the model integrates multiple sources of heterogeneous data to build a heterogeneous networks. Second, HGNNLDA gets all types of strong related neighbors of fixed size for each node by restarting random walk, which solves the defect that the direct related neighbors of some nodes are not representative enough. In addition, HGNNLDA is able to capture the strong correlation neighbor features of each node in this heterogeneous network, and fully exploiting the topology information of the heterogeneous network. Finally, HGNNLDA employs the attention mechanism to account for the differential impact of different types of nodes on lncRNA-disease association prediction. To sum up, HGNNLDA makes full use of the complex structural and semantic information of heterogeneous network, so as to achieves good prediction of lncRNA-disease associations.

However, our method still has some limitations. First, the data we use to build heterogeneous networks may contain noise and some outliers. Second, we randomly select the unknown lncRNA-disease association pairs as negative samples for training, which can’t guarantee that the lncRNA and disease in the unknown association pairs are completely unrelated, so it will have some influence on the prediction performance. Therefore, our future research will focus on how to overcome these problems.

Methods

The general overview of our proposed HGNNLDA framework is shown in Fig. 4, which consists of five key parts: (1) Construction of heterogeneous networks. First we downloaded lncRNA-disease associations, lncRNA-miRNA associations, and calculated the similarity between lncRNAs, and then constructed a heterogeneous network containing the three types of nodes of lncRNA, disease, and miRNA. (2) Sampling strong correlation neighbors and the feature representation of each neighbor. We sampled various types of fixed-sized neighbors for each lncRNA and disease by restart the random walk, and then extract the features of each neighbor node by word2vec. (3) Embedding learning. We used Bi-LSTM to obtain embedding for the three types of neighbors, lncRNA, disease, and miRNA. (4) Updating the node embedding. We introduced the attention mechanism, and aggregated the embedding of three types of neighbors and the embedding of nodes themselves based on the weights obtained. (5) LncRNA-disease association prediction. The embedding of lncRNA and disease were concatenated to get the embedding of lncRNA-disease association pair, then the prediction scores between lncRNA and disease were obtained by using fully connected and softmax layers, eventually optimizde the model by cross-entropy.

Figure 4.

Figure 4

The framework of HGNNLDA.

Datasets for lncRNA-disease associations prediction

Studies have shown that lncRNA can interact with the corresponding miRNA and perform biological functions together with miRNAs41. Therefore, all useful biological information can be assembled to construct a heterogeneous network including the lncRNA-lncRNA similarity network, the experimentally validated lncRNA-disease association network, and the lncRNA-miRNA association network. The data used in this paper were obtained from the previous study of lncRNA-disease association prediction by Fu et al.38. This dataset included 240 lncRNAs, 412 diseases, and 495 miRNAs. Among them, 2,697 verified lncRNA-disease associations are derived from LncRNADisease10, Lnc2Cancer11 and GeneRIF14 databases. In addition, 1002 lncRNA-miRNA associations came from starBase database42.

LncRNA functional similarity network

In this paper, the functional similarity of lncRNA is calculated by the method of Chen et al.21. LncRNA similarity is expressed by the similarity of lncRNA related diseases. Suppose that lncRNA l1 is associated with a group of diseases D1=d11,d12,,d(1m), lncRNA l2 is associated with a group of diseases D2=d21,d22,,d2n. Then the functional similarity between lncRNA l1 and l2 is represented by Sl1,l2 as follows:

Sl1,l2=1immax1jnDSSd(1i,d2j+1jnmax1imDSSd(2j,d1im+n 2
LFS=Sl1,l1Sl1,l240Sl240,l1Sl240,l240 3

where DSSd1i,d2j represents the semantic similarity between disease d1i and disease d2j, which adopts the method calculated by Wang et al.43; m and n represent the number of diseases in disease group D1 and D2, respectively; LFS is a functional similarity matrix of 240 × 240, and 240 represents the number of lncRNAs.

LncRNA-disease associations and lncRNA-miRNA associations

The datasets includes 2697 experimentally verified lncRNA-disease associations and 1002 experimentally verified lncRNA-miRNA associations38. The associations between lncRNAs and diseases are expressed by a 240 × 412 adjacency matrix LD, LDli,lj=1, if lncRNA li is related to disease dj, otherwise it is 0. Similarly, the associations between lncRNAs and miRNAs are represented by an adjacency matrix LM of 240 × 495, LMli,mj=1, if lncRNA li is related to miRNA mj, otherwise it is 0.

Heterogeneous network construction

As shown in Fig. 4a, we construct a heterogeneous network based on lncRNA functional similarity LFS, lncRNA-disease association network LD and lncRNA-miRNA association network LM. Heterogeneous networks can be expressed as:

G=N,E,NT,ET 4

where N represents the node set, which contains three types of nodes, namely NT=lncRNA,disease,miRNA, E represents the edge set, which contains three types of edges, namely ET=lncRNA-disease,lncRNA-lncRNA,lncRNA-miRNA.

Sampling heterogeneous neighbors with restart random walk

In heterogeneous networks, the neighbors of many nodes cannot include all types of nodes, and the number of neighbor nodes will vary32. For example, in Fig. 4a, no disease node is directly connected to the miRNA node, and d1 has two neighbor nodes, while l2 has seven neighbor nodes. Therefore, to make full use of the information of heterogeneous networks, we introduced restart random walk (RRW) to sample three types of strongly correlated neighbors for each node. The sampling operation of RRW in lncRNA-disease heterogeneous network includes two steps:

  • Selecting fixed size sampling length for RRW. Starting random walk from node vN, return to the starting node with probability p or iteratively move to the neighbor of the current node, where the probability q controls whether the walk is depth first select or breadth first select. When q>1, random walk tends to give priority to breadth; when q<1, random walk tends to give priority to depth. RRW runs until a fixed number of nodes are successfully collected, and the sampled nodes are denoted as RRWv. Moreover, the number of different types of nodes in RRWv is constrained, which ensure that all types of nodes are sampled.

  • Grouping neighbor nodes of lncRNA, disease and miRNA-type. For each node type t, the top kt nodes are selected based on the frequency of occurrence, and take them as the set of t-type correlated neighbors of node v.

In this way, three types of neighbors can be collected for each node, and classification by type is conducive to subsequently learn embedding of type.

Embedding learning

Word2vec is a word embedding technology proposed by Mikolov et al.44, it can provide a vectorized representation for each word or sentence while preserving semantic and contextual integrity. In the last step, each node obtained a fixed size sampling sequence by using the strategy based on RRW. Therefore, Word2vec can be used to get the embedding of each node. Next, to obtain the embedding of type, we can aggregate all the same type neighbors after sampling by using Bi-LSTM45. For example, we can express disease-type neighbors of node vN in the heterogeneous network as Ndv. Next, the embedding of each disease-type neighbor node is obtained by Word2vec44, as shown in Fig. 4b. Then we utilize Bi-LSTM to aggregate the embeddings of all disease-type neighbors, as shown in Fig. 4c. In the process of aggregating all lncRNA-type nodes, disease-type nodes and miRNA-type nodes, different Bi-LSTM are used to distinguish them. Bi-LSTM consists of a forward LSTM layer and a backward LSTM layer. The main structure of LSTM layer can be expressed as follows:

is=σws·fs+hs-1·ws+bffs=σwf·fs+hs-1·wf+bfos=σwo·fs+hs-1·wo+boc~s=σwc·fs+hs-1·wc+bccs=isc~s+fscs-1hs=ostanhcs 5

Where σ is sigmoid activation function; i, f, o and c represent input gate vector, forget gate vector, output gate vector and memory unit respectively; hs represents the output hidden vector by s-th node; w and b represent learnable parameters; represents dot product operation. Two different middle layer representations can be obtained through calculation. Then, after splicing the two middle layers, the general embedding of all disease-type neighbor nodes of node v can be obtained through the average pool layer, as shown follow:

hs=LSTMdhs-1,fshs=LSTMdhs,fsfdv=sNdvhshsNdv 6

Where fdvRd×1 is the general embedding of all disease-type neighbors of node v; hs and hs represent the forward and backward LSTM representations of s node respectively; the symbol indicates the connection operation.

Updating the node embedding with attention mechanism

In the previous step, the general embedding of lncRNA-type, disease-type and miRNA-type will be generated. Different types of neighbors will have different influences on the final embedding of node v32, for example, nodes of lncRNA, disease-type usually play a more important role in the prediction of lncRNA-disease association. So as to combine lncRNA-type, disease-type and miRNA-type general embeddings with node v embedding, we introduce the attention mechanism46. First, the importance of each type is learned, and then all heterogeneous types of nodes(including node v itself ) are aggregated to form the final embedding of node v. For any tNv, Nv=vNT, the importance βvt of t-type relative to node v is expressed as:

βvt=expσqTfvftvkNvexp(σqTfvfkv 7

Where σ is ReLU activation function; qTR2d×1 represents the attention vector; f(v) is that embedding of v obtained by word2vec; ftv is a general embedding based on t-type aggregating; indicates the connection operation; fkv=fv when k equals v. Then, the final embedding of node v can be aggregated by various types of embedding based on the corresponding importance coefficient. The details are as follows:

zv=σkNvβvtfkv 8

Where zvRd×1 represents the final embedding. To better understand the aggregation process of various types of nodes, explanation is shown in Fig. 4d.

LncRNA-disease association prediction

The final embedding of lncRNA li and the final embedding of disease dj are spliced to constitute the vector representation xi,jR2d×1 of the association pair li-dj:

xi,j=zlizdj 9

Where represents splicing operation. Then, each positive sample (there is an association between lncRNA and disease) is marked as 1, and each negative sample (there is no association between lncRNA and disease) is marked as 0. Then, we provide the embedding of the association pair li-dj to the fully connected layer and the softmax layer, and the score of association si,j0,1 between lncRNA li and disease dj is obtained. The specific si,j is expressed as follows:

si,j=softmaxWxi,j+b 10

Where WR2×2d is the parameter of the full connection layer and b is the bias; the larger the score of si,j, the greater the possibility of association between lncRNA li and disease dj. In our model, the cross-entropy loss between prediction and real association is defined as follows:

Loss=-i=1Tyilogsi 11

Where T is the number of training samples; si is the score of the association between lncRNA and disease of training sample; yi is the label of real association between lncRNA and disease.

Acknowledgements

This work was supported by the National Natural Science Foundation of China (No. 61862067), the Applied Basic Research Project in Yunnan Province (No. 202201AT070042) and the NSFC-Yunnan Union Key Grant (No. U1902201).

Author contributions

L.T., L.L. and H.S. conceived the presented idea. H.S. and X.M.Z. carried out the experiment and wrote the draft. L.T. and L.L. helped shape the research, analysis and manuscipt. All authors discussed the results and contributed to the final manuscript.

Data availability

The original datasets of our study was download from another lncRNA-disease association prediction study, the orginal datasets were available at https://github.com/ydkvictory/RFLDA. The processed data along with codes are available at https://github.com/hongshi940/HGNNLDA.

Competing interests

The authors declare no competing interests.

Footnotes

Publisher's note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

References

  • 1.Kapranov P, et al. RNA maps reveal new RNA classes and a possible function for pervasive transcription. Science. 2007;316:1484–1488. doi: 10.1126/science.1138341. [DOI] [PubMed] [Google Scholar]
  • 2.Mercer TR, Dinger ME, Mattick JS. Long non-coding RNAs: Insights into functions. Nat. Rev. Genet. 2009;10:155–159. doi: 10.1038/nrg2521. [DOI] [PubMed] [Google Scholar]
  • 3.Pasmant E, Sabbagh A, Vidaud M, Bièche I. ANRIL, a long, noncoding RNA, is an unexpected major hotspot in GWAS. FASEB J. 2011;25:444–448. doi: 10.1096/fj.10-172452. [DOI] [PubMed] [Google Scholar]
  • 4.Congrains A, et al. Genetic variants at the 9p21 locus contribute to atherosclerosis through modulation of ANRIL and CDKN2A/B. Atherosclerosis. 2012;220:449–455. doi: 10.1016/j.atherosclerosis.2011.11.017. [DOI] [PubMed] [Google Scholar]
  • 5.Zhang Q, Chen C-Y, Yedavalli VS, Jeang K-T. Neat1 long noncoding RNA and paraspeckle bodies modulate HIV-1 posttranscriptional expression. MBio. 2013;4:e00596-12. doi: 10.1128/mBio.00596-12. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Johnson R. Long non-coding RNAs in Huntington’s disease neurodegeneration. Neurobiol. Dis. 2012;46:245–254. doi: 10.1016/j.nbd.2011.12.006. [DOI] [PubMed] [Google Scholar]
  • 7.Ji P, et al. MALAT-1, a novel noncoding RNA, and thymosin β4 predict metastasis and survival in early-stage non-small cell lung cancer. Oncogene. 2003;22:8031–8041. doi: 10.1038/sj.onc.1206928. [DOI] [PubMed] [Google Scholar]
  • 8.Barsyte-Lovejoy D, et al. The c-Myc oncogene directly induces the H19 noncoding RNA by allele-specific binding to potentiate tumorigenesis. Cancer Res. 2006;66:5330–5337. doi: 10.1158/0008-5472.CAN-06-0037. [DOI] [PubMed] [Google Scholar]
  • 9.De Kok JB, et al. DD3PCA3, a very sensitive and specific marker to detect prostate tumors. Cancer Res. 2002;62:2695–2698. [PubMed] [Google Scholar]
  • 10.Bao Z, et al. LncRNADisease 20: An updated database of long non-coding RNA-associated diseases. Nucleic Acids Res. 2019;47:D1034–D1037. doi: 10.1093/nar/gky905. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Ning S, et al. Lnc2Cancer: A manually curated database of experimentally supported lncRNAs associated with various human cancers. Nucleic Acids Res. 2016;44:D980–D985. doi: 10.1093/nar/gkv1094. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Dinger ME, et al. NRED: A database of long noncoding RNA expression. Nucleic Acids Res. 2009;37:D122–D126. doi: 10.1093/nar/gkn617. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Wang Y, et al. Mammalian ncRNA-disease repository: A global view of ncRNA-mediated disease network. Cell Death Dis. 2013;4:e765–e765. doi: 10.1038/cddis.2013.292. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Lu, Z., Bretonnel Cohen, K. & Hunter, L. GeneRIF quality assurance as summary revision. In Biocomputing 2007, 269–280 (World Scientific, 2007). [DOI] [PMC free article] [PubMed]
  • 15.Sun J, et al. Inferring novel lncRNA-disease associations based on a random walk model of a lncRNA functional similarity network. Mol. BioSyst. 2014;10:2074–2081. doi: 10.1039/c3mb70608g. [DOI] [PubMed] [Google Scholar]
  • 16.Gu C, et al. Global network random walk for predicting potential human lncRNA-disease associations. Sci. Rep. 2017;7:1–11. doi: 10.1038/s41598-017-12763-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Wen Y, Han G, Anh VV. Laplacian normalization and bi-random walks on heterogeneous networks for predicting lncRNA-disease associations. BMC Syst. Biol. 2018;12:11–19. doi: 10.1186/s12918-018-0660-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Zhang J, Zhang Z, Chen Z, Deng L. Integrating multiple heterogeneous networks for novel lncRNA-disease association inference. IEEE/ACM Trans. Comput. Biol. Bioinform. 2017;16:396–406. doi: 10.1109/TCBB.2017.2701379. [DOI] [PubMed] [Google Scholar]
  • 19.Zhao, X., Yang, Y. & Yin, M. Mhrwr: Prediction of lncRNA-disease associations based on multiple heterogeneous networks. In IEEE/ACM Transactions on Computational Biology and Bioinformatics (2020). [DOI] [PubMed]
  • 20.Chen X, Yan G-Y. Novel human lncRNA-disease association inference based on lncRNA expression profiles. Bioinformatics. 2013;29:2617–2624. doi: 10.1093/bioinformatics/btt426. [DOI] [PubMed] [Google Scholar]
  • 21.Chen X, et al. Constructing lncRNA functional similarity network based on lncRNA-disease associations and disease semantic similarity. Sci. Rep. 2015;5:1–12. doi: 10.1038/srep11338. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Zhao T, et al. Identification of cancer-related lncRNAs through integrating genome, regulome and transcriptome features. Mol. BioSyst. 2015;11:126–136. doi: 10.1039/c4mb00478g. [DOI] [PubMed] [Google Scholar]
  • 23.Lan W, et al. LDAP: A web server for lncRNA-disease association prediction. Bioinformatics. 2017;33:458–460. doi: 10.1093/bioinformatics/btw639. [DOI] [PubMed] [Google Scholar]
  • 24.Sheng N, Cui H, Zhang T, Xuan P. Attentional multi-level representation encoding based on convolutional and variance autoencoders for lncRNA-disease association prediction. Brief. Bioinform. 2021;22:bbaa067. doi: 10.1093/bib/bbaa067. [DOI] [PubMed] [Google Scholar]
  • 25.Xuan P, Pan S, Zhang T, Liu Y, Sun H. Graph convolutional network and convolutional neural network based method for predicting lncRNA-disease associations. Cells. 2019;8:1012. doi: 10.3390/cells8091012. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Wu X, et al. Inferring lncRNA-disease associations based on graph autoencoder matrix completion. Comput. Biol. Chem. 2020;87:107282. doi: 10.1016/j.compbiolchem.2020.107282. [DOI] [PubMed] [Google Scholar]
  • 27.Zhang J, Jiang Z, Hu X, Song B. A novel graph attention adversarial network for predicting disease-related associations. Methods. 2020;179:81–88. doi: 10.1016/j.ymeth.2020.05.010. [DOI] [PubMed] [Google Scholar]
  • 28.Wu Q-W, Xia J-F, Ni J-C, Zheng C-H. GAERF: Predicting lncRNA-disease associations by graph auto-encoder and random forest. Brief. Bioinform. 2021;22:bbaa391. doi: 10.1093/bib/bbaa391. [DOI] [PubMed] [Google Scholar]
  • 29.Zhao X, Zhao X, Yin M. Heterogeneous graph attention network based on meta-paths for lncRNA-disease association prediction. Brief. Bioinform. 2022;23:bbab407. doi: 10.1093/bib/bbab407. [DOI] [PubMed] [Google Scholar]
  • 30.Lan W, et al. GANLDA: Graph attention network for lncRNA-disease associations prediction. Neurocomputing. 2022;469:384–393. [Google Scholar]
  • 31.Silva, A. B. O. V. & Spinosa, E. J. Graph convolutional auto-encoders for predicting novel lncRNA-disease associations. In IEEE/ACM Transactions on Computational Biology and Bioinformatics (2021). [DOI] [PubMed]
  • 32.Zhang, C., Song, D., Huang, C., Swami, A. & Chawla, N. V. Heterogeneous graph neural network. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, 793–803 (2019).
  • 33.Guo Z-H, You Z-H, Wang Y-B, Yi H-C, Chen Z-H. A learning-based method for lncRNA-disease association identification combing similarity information and rotation forest. iScience. 2019;19:786–795. doi: 10.1016/j.isci.2019.08.030. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Zhang Y, Ye F, Xiong D, Gao X. LDNFSGB: Prediction of long non-coding RNA and disease association using network feature similarity and gradient boosting. BMC Bioinform. 2020;21:1–27. doi: 10.1186/s12859-020-03721-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35.Madhavan, M. et al. Deep belief network based representation learning for lncRNA-disease association prediction. arXiv preprint arXiv:2006.12534 (2020).
  • 36.Zhu R, Wang Y, Liu J-X, Dai L-Y. IPCARF: Improving lncRNA-disease association prediction using incremental principal component analysis feature selection and a random forest classifier. BMC Bioinform. 2021;22:1–17. doi: 10.1186/s12859-021-04104-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37.Lu C, et al. Prediction of lncRNA-disease associations based on inductive matrix completion. Bioinformatics. 2018;34:3357–3364. doi: 10.1093/bioinformatics/bty327. [DOI] [PubMed] [Google Scholar]
  • 38.Fu G, Wang J, Domeniconi C, Yu G. Matrix factorization-based data fusion for the prediction of lncRNA-disease associations. Bioinformatics. 2018;34:1529–1537. doi: 10.1093/bioinformatics/btx794. [DOI] [PubMed] [Google Scholar]
  • 39.Xuan P, Cao Y, Zhang T, Kong R, Zhang Z. Dual convolutional neural networks with attention mechanisms based method for predicting disease-related lncRNA genes. Front. Genet. 2019;10:416. doi: 10.3389/fgene.2019.00416. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40.Yao D, et al. A random forest based computational model for predicting novel lncRNA-disease associations. BMC Bioinform. 2020;21:1–18. doi: 10.1186/s12859-020-3458-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41.Yang G, Lu X, Yuan L. LncRNA: A link between RNA and cancer. Biochim. Biophys. Acta Gene Regul. Mech. 2014;1839:1097–1109. doi: 10.1016/j.bbagrm.2014.08.012. [DOI] [PubMed] [Google Scholar]
  • 42.Li J-H, Liu S, Zhou H, Qu L-H, Yang J-H. starBase v20: Decoding miRNA-ceRNA, miRNA-ncRNA and proteinRNA interaction networks from large-scale CLIP-Seq data. Nucleic Acids Res. 2014;42:D92–D97. doi: 10.1093/nar/gkt1248. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43.Wang JZ, Du Z, Payattakool R, Yu PS, Chen C-F. A new method to measure the semantic similarity of go terms. Bioinformatics. 2007;23:1274–1281. doi: 10.1093/bioinformatics/btm087. [DOI] [PubMed] [Google Scholar]
  • 44.Mikolov, T., Sutskever, I., Chen, K., Corrado, G. S. & Dean, J. Distributed representations of words and phrases and their compositionality. Adv. Neural Inf. Process. Syst.26 (2013).
  • 45.Hochreiter S, Schmidhuber J. Long short-term memory. Neural Comput. 1997;9:1735–1780. doi: 10.1162/neco.1997.9.8.1735. [DOI] [PubMed] [Google Scholar]
  • 46.Veličković, P. et al. Graph attention networks. arXiv preprint arXiv:1710.10903 (2017).

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

The original datasets of our study was download from another lncRNA-disease association prediction study, the orginal datasets were available at https://github.com/ydkvictory/RFLDA. The processed data along with codes are available at https://github.com/hongshi940/HGNNLDA.


Articles from Scientific Reports are provided here courtesy of Nature Publishing Group

RESOURCES