Skip to main content
PLOS Computational Biology logoLink to PLOS Computational Biology
. 2022 Jan 20;18(1):e1009798. doi: 10.1371/journal.pcbi.1009798

CRBPDL: Identification of circRNA-RBP interaction sites using an ensemble neural network approach

Mengting Niu 1,2, Quan Zou 1,2, Chen Lin 3,*
Editor: Serdar Bozdag4
PMCID: PMC8806072  PMID: 35051187

Abstract

Circular RNAs (circRNAs) are non-coding RNAs with a special circular structure produced formed by the reverse splicing mechanism. Increasing evidence shows that circular RNAs can directly bind to RNA-binding proteins (RBP) and play an important role in a variety of biological activities. The interactions between circRNAs and RBPs are key to comprehending the mechanism of posttranscriptional regulation. Accurately identifying binding sites is very useful for analyzing interactions. In past research, some predictors on the basis of machine learning (ML) have been presented, but prediction accuracy still needs to be ameliorated. Therefore, we present a novel calculation model, CRBPDL, which uses an Adaboost integrated deep hierarchical network to identify the binding sites of circular RNA-RBP. CRBPDL combines five different feature encoding schemes to encode the original RNA sequence, uses deep multiscale residual networks (MSRN) and bidirectional gating recurrent units (BiGRUs) to effectively learn high-level feature representations, it is sufficient to extract local and global context information at the same time. Additionally, a self-attention mechanism is employed to train the robustness of the CRBPDL. Ultimately, the Adaboost algorithm is applied to integrate deep learning (DL) model to improve prediction performance and reliability of the model. To verify the usefulness of CRBPDL, we compared the efficiency with state-of-the-art methods on 37 circular RNA data sets and 31 linear RNA data sets. Moreover, results display that CRBPDL is capable of performing universal, reliable, and robust. The code and data sets are obtainable at https://github.com/nmt315320/CRBPDL.git.

Author summary

More and more evidences show that circular RNA can directly bind to proteins and participate in countless different biological processes. The calculation method can quickly and accurately predict the binding site of circular RNA and RBP. In order to identify the interaction of circRNA with 37 different types of circRNA binding proteins, we developed an integrated deep learning network based on hierarchical network, called CRBPDL. It can effectively learn high-level feature representations. The performance of the model was verified through comparative experiments of different feature extraction algorithms, different deep learning models and classifier models. Moreover, the CRBPDL model was applied to 31 linear RNAs, and the effectiveness of our method was proved by comparison with the results of current excellent algorithms. It is expected that the CRBPDL model can effectively predict the binding site of circular RNA-RBP and provide reliable candidates for further biological experiments.


This is a PLOS Computational Biology Methods paper.

Introduction

Circular RNA (circRNA) is a special circular endogenous noncoding RNA produced by selective shearing [1,2]. It has been proven to be widely present in Drosophila, mice, the hippocampus and human cells and tissues [3,4]. Although the RNA-binding proteins (RBP) binding sites on circular RNAs are less numerous than those on linear mRNAs, there is still strong evidence to support the interaction of RBPs with circular RNAs [5,6]. On the one hand, circRNAs can regulate RBPs in a variety of ways. CircRNAs can competitively bind to RBPs, regulate the function of RBPs, and act as sponges of RBPs, platforms for RBP assembly, and supertransporters that concentrate certain specific components [7,8]. RBP-adsorbed circRNA can be used as a regulatory factor for target gene transcription and splicing [9]. circRNA can also be used as a bait to retain RBPs in a specific intercellular space and as a scaffold to promote contact between two or more RBPs [10]. On the other hand, the influence of RBPs on circRNAs is becoming increasingly prominent. As a protein that binds to double-stranded or single-stranded RNA, RBPs are present throughout the life of RNA and mediate the maturation [11], transport [12], positioning and translation of RNA [13]. RBPs affect the entire process of the circRNA life cycle, and some RBPs are also involved in the generation of circRNAs, such as Quking (QKI), FUS, and HNRNPL. Moreover, they are involved in almost every aspect of the cyclic RNA life cycle, including generation [14], posttranscriptional regulation [15], and functional execution [16]. Some specific RBPs are tissue-specific or produced under pathological conditions, and their expression defects can cause a variety of diseases and other effects. Multiple studies have shown that the interaction between circular RNA and RBP has an important impact on cancer and other diseases and may be a disease of biomarkers [10,1721]. Therefore, predicting the binding site of RNA and RBP can provide insight into the mechanisms underlying diseases involving RBPs and help to further explore the role of circRNA in disease pathophysiology.

As a promising method, machine learning has been used to solve various biological problems, its superiority has been proven many times, and it has gradually been used to identify the binding sites of circular RNA-RBP [22]. Matizka et al. proposed the GraphProt method, which can learn secondary structure characteristics, and used support vector machine (SVM) to predict binding sites and affinity of RBPs in all tissues [23]. Corrado et al. applied recommendation algorithm to recommend RNA targets for RNA-binding proteins based on protein domain composition and RNA predicted secondary structure features [24]. Yu et al. employed the random forest algorithm (RF) to predict specific and general RBP sites based on motif information [25]. The above machine learning models are mainly based on the structural characteristics of RNA sequences to identify the binding sites of circular RNA-RBP [26,27].

Deep learning has fulfilled remarkable accomplishment in the field of bioinformatics recently [2830], which also includes the prediction of RNA-protein interactions The DeepBind method utilized convolutional neural network (CNN) to learn binding preference of individual RBPs and obtains better performance [31]. Pan et al. proposed iDeepE method, which uses the global CNN model to predict the binding site by studying RNA sequences [32]. In addition, they further used two separate CNNs and a long-term short-term memory network to learn the sites [33]. Pan et al. further used multilabel classification and deep learning to identify multiple RBPs that can interact with RNA [34]. Jia et al. constructed an hybrid deep neural network [35]. Zhang advanced a new stacked codon coding scheme and combined it with hybrid deep learning to complete the prediction [36]. Yang et al. constructed a multiscale neural network and predicted the binding site of circRBA-RBP based on contextual sequence information [37]. However, the feature learning network is relatively simple, and there is still potential for improvement in prediction performance.

In our work, we establish a novel computational predictor CRBPDL, which based on an ensemble deep network to identify circRNA-RBP interaction sites. First, we adopted 5 coding schemes to provide comprehensive feature information for model training, including k-nucleotide frequency (KNF), Doc2vec, electron-ion interaction pseudopotential (EIIP), nucleotide chemical properties (CCN) And cumulative nucleotide frequency (ANF). Due to the different distributions of feature descriptors, we first applied convolution filters to the features respectively, and then concatenated them into a feature matrix. Subsequently, to automatically extract high-order local and global context information from feature descriptors, we constructed a deep neural network architecture, which consists of a deep multi-scale residual network (ResNet) and a bidirectional gated recurrent unit with a self-attention mechanism (BiGRUs) network composition. We used deep multi-scale residual networks (MRSN) and BiGRUs to learn local and global contextual information, and can effectively represent high-level features. Then, used the self-attention mechanism to train the robustness of the model. After model training and selection, we can get the optimized deep learning model (for convenience, the deep learning model before integration is named "sig-CRBPDL"). Finally, the AdaBoost algorithm was used to integrate the deep learning model. We benchmarked CRBPDL and existing predictors on the unified circRNA dataset. The benchmark test results clearly showed the superiority of our proposed CRBPDL. In addition, CRBPDL has the potential to recognize linear RNA-RBP interaction sites. The benchmark results showed that CRBPDL also has stable performance in predicting linear RNA-RBP binding sites. The structure of the CRBPDL model is shown in Fig 1.

Fig 1. The overall framework of CRBPDL.

Fig 1

(A) The workflow of the development and assessment process of CRBPDL. (B) The structure of the sig-CRBPDL framework, including the input layer, convolutional layers, merger layers, inception module, attention layers, fully connected layers and output layer.

Results

Model performance under different network layers

Network depth has great effects on the performance of deep learning models. Different network depths will lead to diverse results. A relatively shallow network will make the model perform poorly, and an overly complex network will increase the calculation of the model. This section analyzed the model performance changes under different network layers. We compared the increase and decrease: reducing one MSRB block, that is, a 3-layer convolutional network layer, and adding an MSRB block, which means adding a 3-layer convolutional neural network, respectively named CRBP-3 and CRBP+3 for convenience of description. We calculated the prediction performance of CRBPDL, CRBPDL-3 and CRBPDL+3 (AUC as an evaluation index) and running time under 37 data sets (Fig 2A).

Fig 2.

Fig 2

(A) Comparison of model performance between different network depths visualized by box and fiddle charts. (B) Model performance analysis under different EPOCH. (C) Comparison of model performance under different learning rate schemes. (D) Comparison of model performance under different feature coding schemes.

First, it can be seen from the scatter plot that the AUC value distribution of CRBPDL is 0.9174, which is higher than that of CRBPDL-3 (0.8995), and the running time is the opposite. The AUCs of the CRBPDL and CRBPDL+3 (AUC is 0.9011) distributions are not very different, but the running times are quite different. On the other hand, by observing the distribution of the maximum, minimum, and average values in the box chart, it can be found that the midline positions of the three are similar, but the bottom positions of the CRBPDL-3 and CRBPDL+3AUC box charts are lower. The top position of the box chart for CRBPDL+3 time is higher. The prediction performance distributions of CRBPDL-3 and CRBPDL+3 were quite different, and the performance is not stable enough. In contrast, the distribution difference of CRBPDL is smaller, and the stability is better. In terms of time consumption, the performance of CRBPDL-3 is better than that of CRBPDL, but the difference is small. This proves the complexity of the network layer may impact the behavior of the network. It also shows in practical applications, when faced with the needs of different time consumption and prediction effects, both the progressive neural network and the deep neural network have research significance and value.

Model performance under different epoch times

This section statistically analyzes the changes in the loss and ACC of the training set and the validation set during the training phase and accordingly analyzes the impact of epoch on the model performance and the convergence of the model. If the graphs of all 37 data sets are displayed, there are too many pictures. Therefore, we only randomly selected 9 results for display, and can illustrate the effects of different data sets (the loss results of the remaining 28 data sets can be seen in S1 Text). This section mainly analyzes 9 out of 37 data sets which are AGO1, AGO2, U2AF65, DGCR8, FOX2, WTAP, EIF4A3, FMRP, and ZC3H7B. The results of the 9 data sets are shown in Fig 2B.

We can see that as the epoch time increases, the train-acc and validation-acc of CRBPDL both show an upward trend, and the overall train-loss and validation-loss show a downward trend and gradually stabilize; the model gradually converges, and training results are gradually optimized. The obvious performance of overfitting is that the performance of the training set is particularly fine, but the performance of the verification set is exceptionally poor. It can be found that in the 5th and 6th data sets of the 9 data sets, the trend of the acc curve is quite different. The performance effect of train-acc is obviously better than that of validation-acc. There is an obvious overfitting phenomenon, but not in the other 7 data sets. The reason for this difference may be that the data volumes of AGO3 and WTAP (that is, the 5th and 6th data sets) are small (1,210 and 892 data points, respectively), and the learning and training process of the CRBPDL model is not sufficient. In contrast, the data volume of the other data sets is on average one hundred times greater, achieving better training results. It can also be seen that the size of the data set is very important for the performance of deep learning neural networks.

Model performance under different learning rate

As a hyperparameter of the neural network, the learning rate can be used to improve the performance of the model. The lower the learning rate is, the slower the gradient rate. When determining the learning rate, it’s generally essential to rely on the comparison of old experience and multiple experiments. The section analyzes the effect of the learning rate. To compare the effect, we analyzed three learning rate attenuation schemes (step-based attenuation learning rate scheme, linear learning rate attenuation scheme, polynomial learning rate scheme) and two fixed learning rates (0.002, 0.0005). Fig 2C shows the comparison of the AUC of the network optimization process when uses different learning rates.

On all circRNA data sets, the Adam method achieved an average AUC value of 0.9284, which was significantly better than the 0.8926 of the linear scheme and the effect of two fixed learning rates (average AUC of 0.8167 and 0.8747, respectively). Although it is not much different from the average AUC value of 0.9273 of the step scheme, there are two abnormalities in the step. In contrast, Adam’s performance is relatively stable. By performing an experimental comparison of five case, it shows that the Adam linear learning rate plan is always better than other types of plans and has better performance. Therefore, we choose Adam as the learning rate learning plan.

Model Performance under different feature encoding schemes

To evaluate the contribution of the feature encoding schemes (named Fea-A) in this article, under the same CRBPDL architecture, we combined the proposed feature with the coding schemes of PASSION (named Fea-B) [35] and CRIP’s [36] stacked codon coding (named Fea-C). The AUC values of the 37 data sets are shown in the line graph of Fig 2D.

First of all, by observing the trend of the line chart, we can find that the AUC value of Fea-A is higher than that of Fea-B and Fea-C on multiple data. In addition, our method Fea-A obtains an average AUC value of 0.9201, which is not only upper than the 0.8928 of the Fea-B, but also superior than the 0.8792 of the Fea-C. For one thing, Fea-B uses 6 hand-designed features, and it is possible that a single hand-designed function is not suitable for advanced network architectures. For another thing, Fea-C is an improvement of one-hot encoding, which only uses feature type references in the pseudo-translation process. This may cause CRBPDL to fail to fully understand enough information in the circRNA-RBP interaction. Furthermore, the experimental results prove the validity of our feature encoding schemes.

In addition, we analyzed the different performance of the five feature encodes, and the results were shown in Fig 3A. It can be found that on 37 data sets, the Doc2vec coding scheme is relatively better than the other four. It shows that the global text characteristics of RBP binding sites are relatively obvious. Our word vector model seems to have learned the subtle sequence context from semantics, thereby improving the recognition performance. For circRNA data sets, the experimental results show that the self-learning word vector encoding scheme proposed in this paper has a good application prospect. Based on the word vectors obtained, whether the conservative motifs of the binding sites can be analyzed.

Fig 3.

Fig 3

(A) Performance comparison of five feature codes.(B) Heat maps of different network model performance under 37 data sets.(C) T -SNE scatter plot with original feature coding. (D) T-SNE Scatter Diagram of Deep Feature after Deep Convolutional Network.

Performance of neural network structures with different depths

To prove the effectiveness of our proposed CRBPDL, we input our features into different CNN to compare the prediction performance of different neural networks.We compared its performance with 5 structures: CNN-LSTM, iDeepE [32], ResNet [38], CRIP-RNN [36], and CNN-BiLSTM. CNN-LSTM includes two bidirectional LSTM layers and two fully connected layers; iDeepE combines output features of the global network and local neural network, and two layers of local multichannel neural networks (convolution, ReLU and max pooling) express high-level features and then input the feature map into two fully connected layers; ResNet uses a 21-layer local multichannel network, inserts a shortcut connection between the two networks, and makes the network into a corresponding residual network; CRIP uses two layers, a CNN that extracts high-level features and a RNN that acquires the long-term dependence of sequence; and CNN-BiLSTM uses bi-directional long-short term Memory to integrate data, including two bidirectional LSTM and two fully connected layers. These network structures can be built with reference to the literature or built by themselves, and the parameters of each model have the same parameters as CRBPDL. The experimental result is displayed in Fig 3B.

As shown in Fig 3B, we find that the average AUCs of all circular RNA data sets are 0.9174, 0.8778, 0.8854, 0.8877, 0.8760, 0.8733, 0.9148, and 0.9201. When the feature codes of this article are input into different neural networks, the results obtained are different. The difference is based on whether the network structure expresses high-level features accurately. Obviously, the CRBPDL model can learn more valuable sequence information for the identification of circRNA-RBP interaction sites.

To demonstrate how CRBPDL learns efficient feature representation, we take the “WATP” data set as an example, and use t-SNE graphs to visualize feature representation. Both dimensions automatically learn CRBPDL. The original features are shown in Fig 3C. We can find that it is challenging to visually distinguish two categories with primitive characteristics. In addition, the second level of full connectivity after feature representation (Fig 3D) can be used to better identify and separate positive and negative examples. Graphical display shows that CRBPDL can effectively learn excellent feature representations. Moreover, we further analyzed the different performance of MSRN and BiGRU, as shown in Fig 4A. It can be found that although the difference between the two is relatively small, the effect of MSRN is significantly better than that of BiGRU, indicating that in the CRBPDL model, the contribution of MSRN is greater.

Fig 4.

Fig 4

(A) Performance comparison of MSRN and BiGRU. (B) The performance comparison between CRBPDL integrated model and various classification algorithms (C) ROC curves of 37 datasets under the integration model (D) Radar chart of ACC indicators of CRBPDL model under 31 lncRNA datasets.

Performance evaluation of the integration model

In this section, the CRBPDL model and the sig-CRBPDL, RF, SVM, and logistics models and four other machine learning methods are compared. Among them, the evaluation indicators include ACC, SE, SP, and MCC. The parameters of the RF [39,40], SVM, and logistics algorithms are experimented with the default parameters, and the results (ACC, SE, SP, MCC) are shown in Fig 4B. We find that the value of our proposed integrated deep network model CRBPDL on ACC, SE, SP and MCC is significantly higher than the experimental results of RF, SVM and logistics, and the average values of SE, SP, MCC, ACC are 0.8548, 0.7796, 0.6897, and 0.8739. Compared with the single deep network model sig-CRBPDL, there is also a certain level of improvement, indicating that the integrated deep learning model CRBPDL improves the prediction performance of circRNA-RBP interaction sites.

Prediction performance of CRBPDL on 37 circRNA data sets

In this part, the prediction performance of CRBPDL, iCircRBP-DHN [37] and CRIP [36], PASSION [35], CSCRSites [41] and CircSLNN [42] and five other existing calculation methods are measured by AUC. CSCRSites was based on multiple convolutional thermal coding layers to identify cancer-specific RBP binding sites on circular RNAs. In contrast, CircSLNN used a sequence tagging network to recognize the interaction site. In terms of fairness, the six methods were tested on a unified benchmark data set, using the same sequence similarity threshold. In addition, the same setting environment is used as the model iCircRBP-DHN. The AUC results and average values of other comparative experiments are shown in Table 1, obtained directly from the literature (the maximum value was shown in bold) [37].

Table 1. Comparison of prediction performance under different classification models on 37 circRNA datasets.

dataset CRBPDL iCircRBP-DHN PASSION CRIP CSCRites CircSLNN
AGO1 0.9232 0.898±0.003 0.909±0.003 0.905±0.002 0851±0.002 0.844±0.002
AGO2 0.8233 0.797±0.004 0.822±0.003 0.811±0.001 0.755±0.127 0.715±0.013
AGO3 0.9472 0.920±0.016 0.909±0.008 0.895±0.002 0.820±0.004 0.860±0.006
ALKBH5 0.9952 0.979±0.004 0.752±0.030 0.721±0.008 0.798±0.009 0.583±0.014
AUF1 0.9810 0.985±0.002 0.979±0.003 0.980±0.000 0.940±0.001 0.971±0.001
C17ORF85 0.9881 0.987±0.002 0.860±0.021 0.813±0.007 0.815±0.012 0.721±0.010
C22ORF28 0.9088 0.913±0.004 0.894±0.008 0.876±0.002 0.878±0.002 0.797±0.001
CAPRIN1 0.8765 0.858±0.012 0.860±0.009 0.843±0.003 0.827±0.002 0.745±0.008
DGCR8 0.9236 0.906±0.002 0.917±0.002 0.914±0.001 0.870±0.002 0.846±0.002
EIF4A3 0.853 0.799±0.003 0.823±0.004 0.812±0.001 0.820±0.001 0.717±0.005
EWSR1 0.9443 0.942±0.004 0.938±0.006 0.936±0.001 0.882±0.002 0.906±0.002
FMRP 0.8966 0.892±0.002 0.900±0.002 0.898±0.001 0.890±0.001 0.826±0.003
FOX2 0.9618 0.958±0.005 0.830±0.034 0.815±0.006 0.755±0.010 0.602±0.033
FUS 0.8618 0.855±0.004 0.859±0.002 0.858±0.002 0.799±0.003 0.770±0.003
FXR1 0.9948 0.994±0.001 0.959±0.009 0.952±0.003 0.871±0.003 0.942±0.004
FXR2 0.9518 0.939±0.009 0.941±0.003 0.938±0.002 0.868±0.002 0.896±0.004
HNRNPC 0.9771 0.977±0.001 0.976±0.001 0.972±0.000 0.973±0.001 0.970±0.001
HUR 0.8758 0.867±0.005 0.879±0.006 0.874±0.001 0.850±0.001 0.796±0.009
IGF2BP1 0.8554 0.843±0.002 0.845±0.003 0.843±0.001 0.835±0.003 0.760±0.009
IGF2BP2 0.8426 0.831±0.004 0.827±0.009 0.821±0.002 0.752±0.126 0.740±0.004
IGF2BP3 0.8229 0.816±0.004 0.831±0.003 0.822±0.002 0.754±0.122 0.706±0.003
LIN28A 0.8751 0.857±0.007 0.875±0.005 0.865±0.001 0.840±0.002 0.777±0.003
LIN28B 0.9014 0.892±0.004 0.889±0.005 0.882±0.001 0.758±0.129 0.822±0.003
METTL3 0.8649 0.852±0.009 0.878±0.010 0.854±0.003 0.808±0.003 0.772±0.007
MOV10 0.8674 0.838±0.006 0.845±0.005 0.849±0.001 0.778±0.004 0.777±0.008
PTB 0.8347 0.822±0.006 0.829±0.004 0.826±0.001 0.692±0.157 0.738±0.007
PUM2 0.9758 0.970±0.004 0.952±0.004 0.953±0.001 0.936±0.001 0.932±0.002
QKI 0.9879 0.971±0.006 0.927±0.005 0.921±0.003 0.866±0.004 0.866±0.007
SFRS1 0.9684 0.964±0.000 0.965±0.003 0.964±0.001 0.963±0.001 0.926±0.003
TAF15 0.9945 0.992±0.002 0.967±0.002 0.965±0.001 0.941±0.002 0.968±0.002
TDP43 0.9336 0.926±0.002 0.934±0.002 0.930±0.001 0.923±0.001 0.896±0.003
TIA1 0.9666 0.961±0.004 0.935±0.006 0.932±0.003 0.915±0.009 0.901±003
TIAL1 0.9249 0.917±0.003 0.906±0.003 0.902±0.001 0.898±0.002 0.871±0.005
TNRC6 0.9797 0.967±0.002 0.785±0.010 0.741±0.007 0.729±0.010 0.662±0.015
U2AF65 0.9306 0.926±0.002 0.930±0.002 0.928±0.001 0.911±0.001 0.899±0.004
WTAP 0.9713 0.967±0.002 0.794±0.069 0.793±0.011 0.808±0.022 0.732±0.009
ZC3HB 0.8151 0.804±0.003 0.804±0.005 0.792±0.002 0.794±0.004 0.697±0.008
AVG 0.9188 0.908±0.06 0.884±0.06 0.876±0.07 0.842±0.07 0.809±0.010

As shown in Table 1, the average values of CRBPDL, iCircRBP-DHN, PASSION, CRIP, CSCRites and CircSLNN are 0.9188, 0.908±0.06, 0.884±0.06, 0.876±0.07, 0.842±0.07 and 0.809±0.010, respectively. Obviously, our model improves the state-of-the-art performance in 28 of the 37 and accomplishes the supreme average AUCs of 0.9174, specifically in AGO1, AGO2, ALKBH5 and MOV10. And we provide the ROC curve of CRBPDL, and the average ROC curve (Fig 4C). The results fully indicate the enhancement of CRBPDL. At the same time, we also noticed that on the 6 data sets, the performance of CRBPDL has a very small gap with iCircRBP-DHN and PASSION, especially 4 of them are slightly worse than PASSION. The underlying reason may be that PASSION has extracted 6 types. This shows that manual features including richer sequence information can be used, and integrated optimization algorithms can also be used. In addition, CRBPDL is better than CircSLNN, but CircSLNN is a sequence tagging method that can predict the location of the binding site. Therefore, as a new research direction, we can consider whether we can improve the accuracy of predicting the position of the binding site, not just as a binary classification problem.

Prediction performance of CRBPDL on 31 linear data sets

Similar to CRIP and PASSION, our CRBPDL also has the ability to identify linear RNA-protein interactions. To demonstrate the performance of our model CRBPDL, we compare it with ICIRCRBP-DHN, CRIP, iDeepS, DEEPBbind, CSCrites, and CIRCSLNN. To make a fair comparison, we used the same experimental data as the iCircRBP-DHN, and the results of the other comparative experiments were obtained directly from the literature [37]. The experimental results are shown in Table 2. From Table 2, CRBPDL obtained an average AUC of 0.9163, which is significantly superior than 0.895, 0.860, 0.842, 0.839, 0.833 and 0.803 of other methods. And on the 31 data sets, only the AUC of hnRNPC-1 is slightly lower than PASSION. In the remaining 30 data sets, our performance is still better than other methods. In addition, we have given the ACC on 31 data sets (Fig 4D), and we can find that the accuracy on the 31 data sets can meet the identification requirements of linear RNA-RBP binding sites.

Table 2. Comparison of prediction performance under different classification models on 31 linear RNA datasets.

dataset CRBPDL iCircRBP-DHN CRIP iDeepS DeepBind CSCRites CircSLNN
AGO1234 0.8178 0.788±0.041 0.737±0.005 0.708 0.687 0.708±0.004 0.662±0.015
AGO2MNAS 0.7995 0.736±0.069 0.598±0.009 0.564 0.535 0.583±0.005 0.557±0.007
2-bingding_1 0.9353 0.925±0.011 0.862±0.004 0.788 0.811 0.842±0.003 0.795±0.006
2-bingding_2 0.9439 0.929±0.014 0.852±0.007 0.831 0.814 0.828±0.003 0.754±0.004
AGO2 0.9088 0.800±0.010 0.638±0.006 0.639 0.585 0.636±0.005 0.562±0.017
eIF4AIII_1 0.9753 0.963±0.004 0.952±0.001 0.942 0.925 0.937±0.004 0.894±0.005
eIF4AIII_2 0.9851 0.963±0.006 0.954±0.001 0.945 0.933 0.944±0.002 0.897±0.006
ELVAL1-1 0.9482 0.939±0.006 0.918±0.002 0.914 0.903 0.910±0.001 0.882±0.005
ELVAL1-MNASE 0.7325 0.695±0.050 0.604±0.007 0.567 0.546 0.581±0.010 0.520±0.13
ELVAL1-A 0.9377 0.922±0.006 0.898±0.002 0.888 0.869 0.876±0.002 0.845±0.006
ELVAL1-2 0.9482 0.943±0.002 0.926±0.001 0.937 0.914 0.925±0.001 0.898±0.002
EWSR1 0.9217 0.918±0.005 0.912±0.003 0.919 0.882 0.884±0.001 0.851±.004
FUS 0.9596 0.947±0.007 0.941±0.001 0.934 0.926 0.907±0.002 0.905±0.007
mut-FUS 0.9621 0.946±0.006 0.939±0.001 0.938 0.915 0.907±0.002 0.9070.013
IGF2BP1-3 0.8156 0.781±0.031 0.693±0.005 0.691 0.685 0.703±0.005 0.597±0.013
hnRNPC-1 0.9514 0.952±0.009 0.963±0.001 0.966 0.954 0.936±0.004 0.935±0.004
hnRNPC-2 0.9891 0.974±0.002 0.985±0.000 0.982 0.976 0.967±0.002 0.962±0.001
hnRNPL-1 0.8583 0.829±0.032 0.748±0.005 0.659 0.761 0.650±0.007 0.670±0.011
hnRNPL-2 0.7995 0.761±0.027 0.740±0.007 0.671 0.74 0.636±0.004 0.654±0.014
HnRNPL-like 0.8008 0.779±0.021 0.685±0.010 0.644 0.708 0.632±0.010 0.636±0.014
MOV10 0.8971 0.885±0.010 0.814±0.002 0.807 0.803 0.803±0.003 0.764±0.010
NSUN2 0.8681 0.832±0.008 0.865±0.003 0.789 0.847 0.798±0.004 0.776±0.015
PUM2 0.9782 0.969±0.003 0.963±0.003 0.966 0.937 0.959±0.002 0.920±0.004
QKI 0.9786 0.962±0.002 0.967±0.001 0.972 0.955 0.956±0.001 0.929±0.004
SFRS1 0.9236 0.912±0.007 0.886±0.004 0.888 0.86 0.885±0.003 0.794±0.008
TAF1S 0.9336 0.971±0.002 0.963±0.001 0.961 0.956 0.922±0.004 0.925±0.002
TDP-43 0.946 0.928±0.013 0.911±0.002 0.914 0.902 0.913±0.003 0.841±0.009
TIA1 0.9636 0.945±0.011 0.930±0.001 0.916 0.908 0.891±0.001 0.894±0.005
TIAL1 0.9799 0.915±0.012 0.898±0.002 0.885 0.881 0.864±0.004 0.847±0.009
U2AF65 0.9782 0.971±0.007 0.968±0.001 0.965 0.959 0.918±0.002 0.932±0.003
Y2AF65 0.969 0.951±0.005 0.935±0.002 0.927 0.916 0.906±0.004 0.893±0.002
Ave 0.9163 0.895±0.08 0.860±0.012 0.842 0.839 0.833±0.012 0.803±0.013

Conclusion

In this paper, we design a new deep learning method, called CRBPDL, for circular RNA-RBP interaction site identification. Based on the MSRN framework, CRBPDL first connects the five codes into a single feature vector. Then MSRB is used to automatically explore higher-level local or global context dependencies and obtain high-level sequence features. Subsequently, the output of each MSRB is combined for global hierarchical feature fusion. And add self-Attention to grasp more critical and relevant features and improve prediction performance. Finally, an integrated deep learning network is constructed based on the Adaboost algorithm. Through the visualization of feature representation, this unique architecture has proven to be effective. To verify CRBPDL, we performed predictions of the binding sites of circRNA and linear RNA and evaluated the performance of different methods. The comparison of 37 circular RNA data sets and 31 linear RNAs not only proves the effectiveness of our method but also shows the potential of the model in the identification of circular RNA-RBP interaction sites. Currently, there are few data on known RBP binding sites. The positive and negative samples are unbalanced. Therefore, the most important thing is that future research is to expand the data set, collect RBP binding sites that bind to circRNA, lncRNA or other RNAs, explore their binding characteristics, and develop universal prediction software.

Materials and methods

Data sets

To prove the effectiveness of our proposed CRBPDL and make a fair comparison with other tools at the same time, we used the benchmark data set (named as ‘circRNA_RBP-37’) used in [35,37,42]. The data set consists of 37 RBPs downloaded from the circinteractome database (https://circinteractome.nia.nih.gov/) [19]. The database collects RBP bound to mature circular RNA and RBP bound to the upstream and downstream sequences of mature circular RNA. Since RBP binding may play a role in regulating splicing events near the splicing site, we considered all RBP binding sites in this study. In the end, we obtained a total of 32,216 circular RNAs related to 37 circular RNA data sets. Among them, the positive sample came from the interaction site on the circular RNA verified by the laboratory. In each CLIP-seq peak, the sequence fragment with a length of 101 nucleotides (nt) was centered and extends 50 nucleotides (nt) in both directions. At the same time, negative sequences were randomly selected from the left-over circular RNA fragments. Subsequently, we applied the same postprocessing method to extract the 101 nucleotide length (nt) binding sites/residual intermediate readings in the previous work [17,18]. Since sequence similarity will influence the consistency of the ML, we used CD-HIT to eliminate the sequence with a similarity threshold of 0.8, which is the same as in CRIP and PASSION. After removing sequence redundancy, we got the final data set, namely the positive and negative samples are 335,976 and 335,976 respectively. 80% of the data sets were selected as training set, and 20% were used as test set.

Additionally, refer to other studies [17,18], we compared the efficiency of CRBPDL to identify the linear RNA -RBP interaction sites. We downloaded the linear RNA data set from PASSION [35] and iDeepS [22], which includes the linear RNA dataset of CLIP-Seq data combined with 31 RBP. Each data set has 5,000 training sets and 1,000 test sets.

Feature encoding

k-nucleotide frequencies

To characterize the local context features of circular RNA sequences, we used KNF coding sequences. KNF describes the frequency of all possible polynucleotides of k nucleotides in the sequence. In this study, we took k = 1, 2, 3, namely single-nucleotide composition frequency, dinucleotide composition frequency and trinucleotide composition frequency. KNF retains a large number of original sequence patterns and integrates a variety of sequence information [43,44]. Compared with traditional single hot spot representation [45], KNF effectively compensates for the lack of information.

Doc2vec

In recent DL model research, to learn more sequence context and semantic information, an increasing number of sequence studies have adopted continuous, high-dimensional word embedding-based coding to substitute one-hot coding, and have achieved good results. Therefore, based on the circRNA corpus of circBase [46], we used the Distributed Memory Model of Paragraph Vectors (PV-DM) model of the Doc2Vec algorithm to vectorize the sequence [47] and train the vectorized model Doc2Vec.model. After that, sequence data were input into the model, 10-mer sequence fragments were taken as circular RNA words, and word embedding training was used to obtain feature vectors. In this way, learning as a continuous distribution representation of global context features expands the vocabulary and can capture the semantics and grammar in these subsequences for long-term dependency modeling.

Electron–ion interaction pseudopotential

The EIIP [48] describes the characteristics of free electron energy on the circRNA sequences. EIIP was widely used to predict the binding sites of RBPs. The EIIP values of the four characters that may appear in the sequence (ie, “A”, “T”, “C”, “G”) are 0.1260, 0.1335, 0.1340 and 0.0806. Hence the EIIP coding method can be used to encode DNA sequence as a digital vector. For example, AATCCGA encoding is a numeric vector consisting of (0.1260, 0.1260,0.1335, 0.1340, 0.1340, 0.0806,0.1260).

Chemical characteristic of nucleotide

Each nucleotide has three types of chemical characteristics (CCN): chemical functions (including amino and keto groups), ring structure (including bicyclic purines and monocyclic pyrimidines), and hydrogen bonds (including weak hydrogen bonds and strong hydrogen bonds) [49]. For the ring structure, A and G belong to purines, coded as 1, and C and T belong to pyrimidines, coded as 0. For chemical functions, A and C belong to amino groups, coded as 1; G and T belong to ketone groups, coded as 0. For hydrogen bonds, A and T belong to a weak hydrogen bond, coded as 1, while C and G belong to a strong hydrogen bond, coded as 0. For example, AATCCGA can be encoded as (1,1,1,1,1,1,0,0,1,0,1,0,0,1,0,1,0,0,1,1,1).

Accumulated nucleotide frequency

ANF presents the density characteristics of nucleotide sequence [49]. Suppose a circRNA sequence S = s1s2…si, where i is the length of S. Sj = s1s2…sj, j is the length of Sj. Sj is the j-th prefix sequence of S. Then the ANF calculation formula is:

ANFsj=f(sj)j (1)
f(sj)=n=1jT(st),T(st)={1,st=sj0,stsj (2)

Multiscale residual network

To obtain rich feature information, a multi-scale CNN layer is constructed to capture high-level features. Unlike traditional Convolutional Neural Networks, different from traditional CNN, multiscale residual network can improve the information trend flow and gradient of the whole network, reduce the computational complexity and improve the model performance [50].

In the MSRN framework, due to the different distributions of the five feature descriptors, we employed convolution filters on five characteristics with a convolution kernel of 128 and then cascaded, which is a common method to balance the distribution of biological features. Afterwards, the MSRN framework contained a shallow CNN extraction layer, and the size of the convolution kernel was 3. Then, the inception module, including 6 cascaded multiscale residual block (MSRB) modules, was used, and the convolution kernel was 64. Each MSRB includes a 3 convolutional layer. Based on the hierarchical feature fusion structure (HFFS), the output of each MSRB was combined to perform global feature fusion. Subsequently, following input to a layer of convolution kernel, there were 192 filters, and a 1×1 convolution can increase and decrease the number of channels, organize information across channels, and increase feature transformation with a small amount of calculation and nonlinear transformation to improve the network expression ability. After that, there was a merge layer with a dropout value of 0.4.

bidirectional gating recurrent unit

For circRNA sequences, besides local background information, there are also long-chain dependencies [51]. Multiscale residual block network can capture only the dependencies between sequences. Therefore, the study employed a Bidirectional GRU to obtain context information from the front and back at the same time to improve the performance.

Bidirectional GRU has only two gates, namely, the update and the reset gate. The update gate controls the extent to which the state information at the previous moment is brought into the current state. The larger the value of the update gate, the more the state information at the previous moment is brought in. The reset gate is used to control the degree of ignoring the state information at the previous moment. The smaller the reset gate, the more information is ignored. The bidirectional GRU can adaptively change its state according to the input, thereby solving the problem of vanishing gradient in RNN.

Self-Attention

The self-attention mechanism was to adaptively pay attention to and learn an important part according to the needs, and ignore the insignificant part. It was widely used in various deep learning applications, including vision processing, phosphorylation site prediction, drug target prediction, etc. [52]. The intention of the attention mechanism is to neglect insignificant word in the bulk of information, selectively filter out a particle of important information, and and express the importance of the information by calculating the weight of the information.

In this research, in our model CRBPDL, the output matrix of the BiGRU layer and its transposed matrix were input into the attention layer, and different features were given different weights, and important features were selected from the dimensional features. Abandon some secondary features and used sigmoid as the activation function.

Implementation

CRBPDL was implemented using the Keras 1.1.2 library in Python. First, we used 80% of the benchmark data set for the training and 20% for testing. Then, on the training set, 80% for training and 20% for verification. Acc was used to evaluate each parameter setting. The verification data set was applicable to monitor the astringency of each stage in the training process, and the training process can be quitted in advance. The study adopted the update method of the learning rate of the Adam gradient descent algorithm, where the initial learning rate is set to 0.001, the max epochs is 200, the epochs is 30, and the batch size is 50. In addition, we have also adopted a variety of techniques to prevent or reduce overfitting, such as batch normalization [50], dropout [51] and early stopping. We used the selected optimal parameter settings, used all training data to train the model, chose the model with the greatest performance as the base model, employed AdaBoost for ensemble, and applied the integrated model as the computational model. AdaBoost is an iterative algorithm. Its core idea is to train different classifiers (weak classifiers) for the same training set, and then group these weak classifiers to form a stronger final classifier (strong classifier) [53,54].

Evaluation metrics

In this study, we employed five evaluation metrics: sensitivity (SE), specificity (SP), accuracy (ACC), Matthew’s correlation coefficient (MCC) and AUC [5563], defined as follows:

SE=TPTP+FN (3)
SP=TNTN+FP (4)
ACC=TN+TPTN+FP+TP+FN (5)
MCC=(TP×TN)(FP×FN)(TP+FP)×(TP+FN)×(TN+FP)×(TN+FN) (6)

where TP, TN, FP and FN denote the numbers of true positives, true negatives, false positives and false negatives, respectively. Furthermore, the area under the curve (AUC) is the area under the ROC curve.

Supporting information

S1 Text. Model performance analysis under different EPOCH.

(DOCX)

Data Availability

All relevant files are available from https://github.com/nmt315320/CRBPDL.git.

Funding Statement

The work (QZ) was supported by the National Natural Science Foundation of China (No. 62131004, No.61922020),the Sichuan Provincial Science Fund for Distinguished Young Scholars (2021JDJQ0025), and the Special Science Foundation of Quzhou (2021D004). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1.Hao S, Lv J, Yang Q, Wang A, Li Z, Guo Y, et al. Identification of key genes and circular RNAs in human gastric cancer. Medical science monitor: international medical journal of experimental and clinical research. 2019;25:2488. doi: 10.12659/MSM.915382 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Bogard B, Francastel C, Hubé F. A new method for the identification of thousands of circular RNAs. Non-coding RNA Investig. 2018;2:5. [Google Scholar]
  • 3.Li S, Han L. Circular RNAs as promising biomarkers in cancer: detection, function, and beyond. Genome medicine. 2019;11(1):1–3. doi: 10.1186/s13073-018-0611-9 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Dori M, Alieh LHA, Cavalli D, Massalini S, Lesche M, Dahl A, et al. Sequence and expression levels of circular RNAs in progenitor cell types during mouse corticogenesis. Life science alliance. 2019;2(2). doi: 10.26508/lsa.201900354 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Zang J, Lu D, Xu A. The interaction of circRNAs and RNA binding proteins: an important part of circRNA maintenance and function. Journal of neuroscience research. 2020;98(1):87–97. doi: 10.1002/jnr.24356 [DOI] [PubMed] [Google Scholar]
  • 6.You X, Vlatkovic I, Babic A, Will T, Epstein I, Tushev G, et al. Neural circular RNAs are derived from synaptic genes and regulated by development and plasticity. Nature neuroscience. 2015;18(4):603–10. doi: 10.1038/nn.3975 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Zhang M, Huang N, Yang X, Luo J, Yan S, Xiao F, et al. A novel protein encoded by the circular form of the SHPRH gene suppresses glioma tumorigenesis. Oncogene. 2018;37(13):1805–14. doi: 10.1038/s41388-017-0019-9 [DOI] [PubMed] [Google Scholar]
  • 8.Hentze MW, Preiss T. Circular RNAs: splicing’s enigma variations. The EMBO journal. 2013;32(7):923–5. doi: 10.1038/emboj.2013.53 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Niu M, Wu J, Zou Q, Liu Z, Xu L. rBPDL: Predicting RNA-binding proteins using deep learning. IEEE Journal of Biomedical and Health Informatics. 2021. doi: 10.1109/JBHI.2021.3069259 [DOI] [PubMed] [Google Scholar]
  • 10.Okholm TLH, Sathe S, Park SS, Kamstrup AB, Rasmussen AM, Shankar A, et al. Transcriptome-wide profiles of circular RNA and RNA-binding protein interactions reveal effects on circular RNA biogenesis and cancer pathway expression. Genome medicine. 2020;12(1):1–22. doi: 10.1186/s13073-020-00812-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Xiao R, Chen J-Y, Liang Z, Luo D, Chen G, Lu ZJ, et al. Pervasive chromatin-RNA binding protein interactions enable RNA-based regulation of transcription. Cell. 2019;178(1):107–21. e18. doi: 10.1016/j.cell.2019.06.001 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Du WW, Fang L, Yang W, Wu N, Awan FM, Yang Z, et al. Induction of tumor apoptosis through a circular RNA enhancing Foxo3 activity. Cell death & differentiation. 2017;24(2):357–70. doi: 10.1038/cdd.2016.133 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Janas T, Janas MM, Sapoń K, Janas T. Mechanisms of RNA loading into exosomes. FEBS letters. 2015;589(13):1391–8. doi: 10.1016/j.febslet.2015.04.036 [DOI] [PubMed] [Google Scholar]
  • 14.Conn SJ, Pillman KA, Toubia J, Conn VM, Salmanidis M, Phillips CA, et al. The RNA binding protein quaking regulates formation of circRNAs. Cell. 2015;160(6):1125–34. doi: 10.1016/j.cell.2015.02.014 [DOI] [PubMed] [Google Scholar]
  • 15.Abdelmohsen K, Panda AC, Munk R, Grammatikakis I, Dudekula DB, De S, et al. Identification of HuR target circular RNAs uncovers suppression of PABPN1 translation by CircPABPN1. RNA biology. 2017;14(3):361–9. doi: 10.1080/15476286.2017.1279788 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Dudekula DB, Panda AC, Grammatikakis I, De S, Abdelmohsen K, Gorospe M. CircInteractome: a web tool for exploring circular RNAs and their interacting proteins and microRNAs. RNA biology. 2016;13(1):34–42. doi: 10.1080/15476286.2015.1128065 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Zhou H, Mangelsdorf M, Liu J, Zhu L, Wu JY. RNA-binding proteins in neurological diseases. Science China Life Sciences. 2014;57(4):432–44. doi: 10.1007/s11427-014-4647-9 [DOI] [PubMed] [Google Scholar]
  • 18.Pereira B, Billaud M, Almeida R. RNA-binding proteins in cancer: old players and new actors. Trends in cancer. 2017;3(7):506–28. doi: 10.1016/j.trecan.2017.05.003 [DOI] [PubMed] [Google Scholar]
  • 19.Han Y, Yang J, Qian X, Cheng W-C, Liu S-H, Hua X, et al. DriverML: a machine learning algorithm for identifying driver genes in cancer sequencing studies. Nucleic acids research. 2019;47(8):e45–e. doi: 10.1093/nar/gkz096 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Wang HT, Jijun; Ding, Yijie; Guo, Fei. Exploring associations of non-coding RNAs in human diseases via three-matrix factorization with hypergraph-regular terms on center kernel alignment. Briefings in Bioinformatics. 2021:1-. [DOI] [PubMed] [Google Scholar]
  • 21.Yijie Wang HD; Jijun Tang; Quan Zou; Fei Guo. Identify RNA-associated subcellular localizations based on multi-label learning using Chou’s 5-steps rule. BMC Genomics. 2021;22(56):1–. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Ferre F, Colantoni A, Helmer-Citterich M. Revealing protein–lncRNA interaction. Briefings in bioinformatics. 2016;17(1):106–16. doi: 10.1093/bib/bbv031 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Maticzka D, Lange SJ, Costa F, Backofen R. GraphProt: modeling binding preferences of RNA-binding proteins. Genome biology. 2014;15(1):1–18. doi: 10.1186/gb-2014-15-1-r17 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.Corrado G, Tebaldi T, Costa F, Frasconi P, Passerini A. RNAcommender: genome-wide recommendation of RNA–protein interactions. Bioinformatics. 2016;32(23):3627–34. doi: 10.1093/bioinformatics/btw517 [DOI] [PubMed] [Google Scholar]
  • 25.Yu H, Wang J, Sheng Q, Liu Q, Shyr Y. beRBP: binding estimation for human RNA-binding proteins. Nucleic acids research. 2019;47(5):e26-e. doi: 10.1093/nar/gky1294 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Adjeroh D, Allaga M, Tan J, Lin J, Jiang Y, Abbasi A, et al. Feature-based and string-based models for predicting RNA-protein interaction. Molecules. 2018;23(3):697. doi: 10.3390/molecules23030697 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Huang D-S, Yu H-J. Normalized feature vectors: a novel alignment-free sequence comparison method based on the numbers of adjacent amino acids. IEEE/ACM Transactions on Computational Biology and Bioinformatics. 2013;10(2):457–67. doi: 10.1109/TCBB.2013.10 [DOI] [PubMed] [Google Scholar]
  • 28.Niu M, Lin Y, Zou Q. sgRNACNN: identifying sgRNA on-target activity in four crops using ensembles of convolutional neural networks. Plant Molecular Biology. 2021;105(4):483–95. doi: 10.1007/s11103-020-01102-y [DOI] [PubMed] [Google Scholar]
  • 29.Liu B, Gao X, Zhang H. BioSeq-Analysis2.0: an updated platform for analyzing DNA, RNA and protein sequences at sequence level and residue level based on machine learning approaches. Nucleic Acids Research. 2019;47(20):e127. doi: 10.1093/nar/gkz740 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.Tang Y-J, Pang Y-H, Liu B. IDP-Seq2Seq: Identification of Intrinsically Disordered Regions based on Sequence to Sequence Learning. Bioinformaitcs. 2020;36(21):5177–86. [DOI] [PubMed] [Google Scholar]
  • 31.Alipanahi B, Delong A, Weirauch MT, Frey BJ. Predicting the sequence specificities of DNA-and RNA-binding proteins by deep learning. Nature biotechnology. 2015;33(8):831–8. doi: 10.1038/nbt.3300 [DOI] [PubMed] [Google Scholar]
  • 32.Pan X, Shen H-B. Predicting RNA–protein binding sites and motifs through combining local and global deep convolutional neural networks. Bioinformatics. 2018;34(20):3427–36. doi: 10.1093/bioinformatics/bty364 [DOI] [PubMed] [Google Scholar]
  • 33.Pan X, Yang Y, Xia CQ, Mirza AH, Shen HB. Recent methodology progress of deep learning for RNA–protein interaction prediction. Wiley Interdisciplinary Reviews: RNA. 2019;10(6):e1544. doi: 10.1002/wrna.1544 [DOI] [PubMed] [Google Scholar]
  • 34.Pan X, Fan Y-X, Jia J, Shen H-B. Identifying RNA-binding proteins using multi-label deep learning. Science China Information Sciences. 2019;62(1):1–3. [Google Scholar]
  • 35.Jia C, Bi Y, Chen J, Leier A, Li F, Song J. PASSION: an ensemble neural network approach for identifying the binding sites of RBPs on circRNAs. Bioinformatics. 2020;36(15):4276–82. doi: 10.1093/bioinformatics/btaa522 [DOI] [PubMed] [Google Scholar]
  • 36.Zhang K, Pan X, Yang Y, Shen H-B. CRIP: predicting circRNA–RBP-binding sites using a codon-based encoding and hybrid deep neural networks. Rna. 2019;25(12):1604–15. doi: 10.1261/rna.070565.119 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37.Yang Y, Hou Z, Ma Z, Li X, Wong K-C. iCircRBP-DHN: identification of circRNA-RBP interaction sites using deep hierarchical network. Briefings in Bioinformatics. 2020. [DOI] [PubMed] [Google Scholar]
  • 38.Lu Z, Jiang X, Kot A. Deep coupled resnet for low-resolution face recognition. IEEE Signal Processing Letters. 2018;25(4):526–30. [Google Scholar]
  • 39.Niu M, Ju Y, Lin C, Zou Q. Characterizing viral circRNAs and their application in identifying circRNAs in viruses. Briefings in Bioinformatics. 2021. doi: 10.1093/bib/bbab404 [DOI] [PubMed] [Google Scholar]
  • 40.Niu M, Zou Q. SgRNA-RF: identification of SgRNA on-target activity with imbalanced datasets. IEEE/ACM Transactions on Computational Biology and Bioinformatics. 2021:1–. doi: 10.1109/TCBB.2021.3079116 [DOI] [PubMed] [Google Scholar]
  • 41.Wang Z, Lei X, Wu F-X. Identifying Cancer-Specific circRNA–RBP Binding Sites Based on Deep Learning. Molecules. 2019;24(22):4035. doi: 10.3390/molecules24224035 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42.Ju Y, Yuan L, Yang Y, Zhao H. CircSLNN: identifying RBP-binding sites on circRNAs via sequence labeling neural networks. Frontiers in genetics. 2019;10:1184. doi: 10.3389/fgene.2019.01184 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43.Niu M, Zhang J, Li Y, Wang C, Liu Z, Ding H, et al. CirRNAPL: a web server for the identification of circRNA based on extreme learning machine. Computational and structural biotechnology journal. 2020;18:834–42. doi: 10.1016/j.csbj.2020.03.028 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 44.Shao J, Yan K, Liu B. FoldRec-C2C: protein fold recognition by combining cluster-to-cluster model and protein similarity network. Briefings in Bioinformatics. doi: 10.1093/bib/bbaa144 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 45.Rathi P, Maurer S, Summerer D. Selective recognition of N 4-methylcytosine in DNA by engineered transcription-activator-like effectors. Philosophical Transactions of the Royal Society B: Biological Sciences. 2018;373(1748):20170078. doi: 10.1098/rstb.2017.0078 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 46.Glažar P, Papavasileiou P, Rajewsky N. circBase: a database for circular RNAs. Rna. 2014;20(11):1666–70. doi: 10.1261/rna.043687.113 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 47.Le Q, Mikolov T, editors. Distributed representations of sentences and documents. International conference on machine learning; 2014: PMLR. [Google Scholar]
  • 48.Nair AS, Sreenadhan SP. A coding measure scheme employing electron-ion interaction pseudopotential (EIIP). Bioinformation. 2006;1(6):197. [PMC free article] [PubMed] [Google Scholar]
  • 49.Bari A, Reaz MR, Jeong B-S. Effective DNA encoding for splice site prediction using SVM. MATCH Commun Math Comput Chem. 2014;71:241–58. [Google Scholar]
  • 50.Li J, Fang F, Mei K, Zhang G, editors. Multi-scale residual network for image super-resolution. Proceedings of the European Conference on Computer Vision (ECCV); 2018. [Google Scholar]
  • 51.Chaabane M, Williams RM, Stephens AT, Park JW. circDeep: deep learning approach for circular RNA classification from other long non-coding RNA. Bioinformatics. 2020;36(1):73–80. doi: 10.1093/bioinformatics/btz537 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 52.Wang D, Zeng S, Xu C, Qiu W, Liang Y, Joshi T, et al. MusiteDeep: a deep-learning framework for general and kinase-specific phosphorylation site prediction. Bioinformatics. 2017;33(24):3909–16. doi: 10.1093/bioinformatics/btx496 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 53.Ao C, Zou Q, Yu L. NmRF: identification of multispecies RNA 2’-O-methylation modification sites from RNA sequences. Briefings in bioinformatics. 2021. doi: 10.1093/bib/bbab480 [DOI] [PubMed] [Google Scholar]
  • 54.Jiao S, Zou Q, Guo H, Shi L. iTTCA-RF: a random forest predictor for tumor T cell antigens. Journal of Translational Medicine. 2021;19(1):449. doi: 10.1186/s12967-021-03084-x [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 55.Wei L, Zhou C, Chen H, Song J, Su R. ACPred-FL: a sequence-based predictor using effective feature representation to improve the prediction of anti-cancer peptides. Bioinformatics. 2018;34(23):4007–16. doi: 10.1093/bioinformatics/bty451 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 56.Shao J, Liu B. ProtFold-DFG: protein fold recognition by combining Directed Fusion Graph and PageRank algorithm. Briefings in Bioinformatics. doi: 10.1093/bib/bbaa192 [DOI] [PubMed] [Google Scholar]
  • 57.Wei L, Liao M, Gao Y, Ji R, He Z, Zou Q. Improved and Promising Identification of Human MicroRNAs by Incorporating a High-Quality Negative Set. IEEE/ACM Transactions on Computational Biology and Bioinformatics. 2014;11(1):192–201. doi: 10.1109/TCBB.2013.146 [DOI] [PubMed] [Google Scholar]
  • 58.Wei L, Wan S, Guo J, Wong KKL. A novel hierarchical selective ensemble classifier with bioinformatics application. Artificial Intelligence in Medicine. 2017;83:82–90. doi: 10.1016/j.artmed.2017.02.005 [DOI] [PubMed] [Google Scholar]
  • 59.Wei L, Xing P, Zeng J, Chen J, Su R, Guo F. Improved prediction of protein-protein interactions using novel negative samples, features, and an ensemble classifier. Artificial Intelligence in Medicine. 2017;83:67–74. doi: 10.1016/j.artmed.2017.03.001 [DOI] [PubMed] [Google Scholar]
  • 60.Wang H, Ding Y, Tang J, Guo F. Identification of membrane protein types via multivariate information fusion with Hilbert-Schmidt Independence Criterion. Neurocomputing. 2020;383:257–69. [Google Scholar]
  • 61.Jijun Ding YT; Guo Fei. Identification of Drug-Target Interactions via Dual Laplacian Regularized Least Squares with Multiple Kernel Fusion. Knowledge-Based Systems. 2020;204. [Google Scholar]
  • 62.Ding Y, Tang J, Guo F. Identification of drug-target interactions via fuzzy bipartite local model. Neural Computing & Applications. 2020;23:10303–19. [Google Scholar]
  • 63.Shen Y, Tang J, Guo F. Identification of protein subcellular localization via integrating evolutionary and physicochemical information into Chou’s general PseAAC. Journal of Theoretical Biology. 2019;462:230–9. doi: 10.1016/j.jtbi.2018.11.012 [DOI] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

S1 Text. Model performance analysis under different EPOCH.

(DOCX)

Data Availability Statement

All relevant files are available from https://github.com/nmt315320/CRBPDL.git.


Articles from PLoS Computational Biology are provided here courtesy of PLOS

RESOURCES