Skip to main content
Briefings in Bioinformatics logoLink to Briefings in Bioinformatics
. 2024 Feb 12;25(2):bbae040. doi: 10.1093/bib/bbae040

ULDNA: integrating unsupervised multi-source language models with LSTM-attention network for high-accuracy protein–DNA binding site prediction

Yi-Heng Zhu 1, Zi Liu 2, Yan Liu 3, Zhiwei Ji 4,, Dong-Jun Yu 5,
PMCID: PMC10939370  PMID: 38349057

Abstract

Efficient and accurate recognition of protein–DNA interactions is vital for understanding the molecular mechanisms of related biological processes and further guiding drug discovery. Although the current experimental protocols are the most precise way to determine protein–DNA binding sites, they tend to be labor-intensive and time-consuming. There is an immediate need to design efficient computational approaches for predicting DNA-binding sites. Here, we proposed ULDNA, a new deep-learning model, to deduce DNA-binding sites from protein sequences. This model leverages an LSTM-attention architecture, embedded with three unsupervised language models that are pre-trained on large-scale sequences from multiple database sources. To prove its effectiveness, ULDNA was tested on 229 protein chains with experimental annotation of DNA-binding sites. Results from computational experiments revealed that ULDNA significantly improves the accuracy of DNA-binding site prediction in comparison with 17 state-of-the-art methods. In-depth data analyses showed that the major strength of ULDNA stems from employing three transformer language models. Specifically, these language models capture complementary feature embeddings with evolution diversity, in which the complex DNA-binding patterns are buried. Meanwhile, the specially crafted LSTM-attention network effectively decodes evolution diversity-based embeddings as DNA-binding results at the residue level. Our findings demonstrated a new pipeline for predicting DNA-binding sites on a large scale with high accuracy from protein sequence alone.

Keywords: protein–DNA interaction, deep learning, unsupervised protein language model, evolution diversity, LSTM-attention network

INTRODUCTION

Protein–DNA interactions are essential for a wide range of biological processes, such as gene expression, DNA replication, chromatin remodeling and signal transduction [1, 2]. Accurate recognition of protein–DNA binding sites is crucial for understanding the molecular mechanisms underlying various biological processes and thus advancing drug discovery [3–6]. Although the current biochemical experiments, such as X-ray crystallography [7] and Cryo-EM [8], are the most precise way for determining DNA-binding sites, they tend to be labor-intensive and time-consuming. Consequently, a large number of sequenced proteins still lack DNA-binding annotations up to now. As of June 2023, the UniProt database [9] contained about 246 million protein sequences, yet less than 0.1% of these sequences had available experimental annotations of DNA-binding sites. There is an immediate need to design efficient computational approaches for predicting protein–DNA binding sites with high accuracy [10–12].

Current methods for DNA-binding site prediction can be categorized into two groups, which are driven by template detection and machine learning, respectively [13]. In the early stage, template detection-based methods were the predominant force in protein–DNA interaction prediction [14, 15]. Specifically, these methods locate DNA-binding sites by detecting the templates with similar sequences or structures to the query. For example, S-SITE [16] identifies sequence templates using PSI-BLAST alignment [17], while PreDNA [18] and DBD-Hunter [19] search templates through designing structure alignment algorithms. Other notable predictors in this field include PreDs [20], DBD-Threader [21], DR_bind [22] and Morozov’s method [23].

A common shortcoming of template detection-based methods is that their accuracy highly depends on the availability of templates with experimentally annotated DNA-binding sites. To overcome this dependency, many machine learning-based methods have been developed. These methods involve extracting manually crafted features from protein sequences and structures (e.g. position-specific scoring matrix [24] and peptide backbone torsion angles [10]), which are further fed to machine learning models (e.g. support vector machine [25] and random forest [26]) to carry out DNA-binding site prediction, including classical examples such as DNAPred [13], TargetDNA [27], MetaDBSite [28] and TargetS [29].

Although machine learning-based methods achieved some progress, their prediction performance was still unsatisfactory. The main reason for this is the shortage of comprehensive and informative feature representations. Specifically, most of these methods are driven by simple and straightforward feature representation methods, such as sequence composition coding and evolution conservation analysis, which fail to capture the complex patterns of protein–DNA interaction [30, 31]. To partially address this challenge, deep learning techniques have been employed in recently proposed DNA-binding site prediction methods, such as Guan’s method [32], PredDBR [33], iProDNA-CapsNet [34] and GraphBind [35]. The significant advantage of deep learning techniques over traditional machine learning methods is that they tend to derive more discriminative feature representations using complicated networks. However, the training efficiency of deep neural network models is frequently constrained by the limited experimental annotation data consisting of only thousands of protein–DNA complexes from the Protein Data Bank (PDB) [36]. As a result, most deep learning models cannot achieve optimal prediction performance.

To relieve the problem arising from the inadequacy of experimentally annotated data, a viable solution is to employ unsupervised protein language models, which are pre-trained on a huge amount of amino acid sequences without DNA-binding annotations via deep learning techniques. Owing to thorough training and learning from extensive sequences, language models could capture crucial inter-residue correlations associated with DNA-binding patterns and encode them as discriminative feature embeddings. Meanwhile, several pre-trained language models have emerged in recent literature, such as TAPE [37] and SeqVec [38]. These methods are frequently employed through supervised deep neural networks in various bioinformatics tasks, including protein design [39, 40], function annotation [41, 42], structure prediction [43, 44] and ligand-binding prediction [45, 46].

In this study, we develop a novelty deep learning model, ULDNA, to accurately predict protein–DNA binding sites through integrating unsupervised protein language models from multiple database sources with the designed LSTM-attention network. Specifically, we utilize three recently proposed language models (i.e. ESM2 [44], ProtTrans [47] and ESM-MSA [48]), separately pre-trained on different large-scale sequence databases, to extract the complementary feature embeddings with evolution diversity, in which the complicated DNA-binding patterns are hidden. Then, an LSTM-attention architecture is specially crafted to effectively decode the evolution diversity-based feature embeddings as the confidence scores of DNA-binding sites at the residue level. ULDNA has been systematically tested on five protein–DNA binding site datasets. Results from computational experiments demonstrated that ULDNA significantly enhances the accuracy of DNA-binding site prediction compared to existing state-of-the-art approaches. The ULDNA online server is freely accessible for academic use through the URL http://csbio.njust.edu.cn/bioinf/uldna/.

MATERIALS AND METHODS

Benchmark datasets

The proposed methods were evaluated by five protein–DNA binding site datasets, i.e. PDNA-543, PDNA-41, PDNA-335, PDNA-52 and PDNA-316. PDNA-543 and PDNA-41 were collected by Hu et al. [27]. The former is comprised of 543 protein chains with DNA-binding annotations deposited in the PDB database before 10 October 2014, while the latter includes 41 DNA-binding protein chains that were deposited in the PDB after 10 October 2014. Here, the CD-HIT software [49] has been used to eliminate redundant proteins both within and across datasets under a sequence identity cut-off of 30%. PDNA-335 and PDNA-52 were collected by Yu et al. [29]. These two datasets consist of 335 and 52 DNA-binding protein chains, respectively, which were released in the PDB before and after 10 March 2010. The sequence identity within each dataset and between different datasets is reduced to 40% through the PISCES software [50]. PDNA-316 was collected by Si et al. [28] and composed of 316 DNA-binding chains deposited in the PDB before 31 December 2011, where the sequence identity of any two chains was reduced to 30% using the CD-HIT [49].

Table 1 presents a detailed summary of five datasets, where the definition of DNA-binding sites is described in Text S1 of Supporting Information (SI). Meanwhile, Figure S1 illustrates the frequencies of 20 native amino acids at DNA-binding and non-DNA-binding sites in each dataset.

Table 1.

Statistical summary of five protein-DNA binding site datasets

Dataset (Max_L, Min_L, Avg_L)a (Num_DBS, Num_NDBS)b
PDNA-543 (1937, 18, 266) (9549, 134,995)
PDNA-41 (1517, 20, 360) (734, 14,021)
PDNA-335 (1609, 51, 232) (6461, 71,320)
PDNA-52 (1132, 54, 331) (973, 16,225)
PDNA-316 (994, 36, 230) (5609, 67,109)

aMax_L/Min_L/Avg_L: the maximal/minimal/average sequence length.

bNum_DBS/Num_NDBS: the number of DNA-binding sites/non-DNA-binding sites.

The architecture of ULDNA

As depicted in Figure 1, ULDNA is a deep learning model for predicting protein–DNA binding sites, where the input is a query sequence with amino acids and the output includes the confidence scores of DNA-binding sites at the residue level. ULDNA comprises two procedures, i.e. (i) feature embedding extraction using multi-source language models and (ii) DNA-binding site prediction using the LSTM-attention network.

Figure 1.

Figure 1

The workflow of ULDNA.

Procedure I: feature embedding extraction using multi-source language models

The input sequence is fed to ESM2 [44] and ProtTrans [47] transformers to generate two feature embedding matrices with the scales of Inline graphic and Inline graphic, respectively. Meanwhile, we search the multiple sequence alignment (MSA) of the input sequence from the UniClust30 database [51]. This MSA is subsequently fed to the ESM-MSA transformer [48] to generate another feature embedding matrix with the scale of Inline graphic. Here, Inline graphic is the length of the input sequence, 2560, 1024 and 768 are preset hyper-parameters in transformer models. ESM2, ProtTrans, and ESM-MSA are both unsupervised attention networks with 36, 24 and 12 layers, respectively, and trained on Uniref50 [52], UniClust30 & Uniref50, and BFD (Big Fantastic Database) [53] & Uniref50 databases, respectively, where ‘&’ means that two databases are both used to train a transformer. Each transformer has learned abundant evolution knowledge from millions of sequences and could encode the input sequence (or MSA) as a feature embedding matrix with evolution diversity. Considering that the evolution knowledge from multiple database sources could be complementary, we concatenate the above-mentioned three feature embedding matrices from different transformer models as a hybrid embedding matrix with the scale of Inline graphic.

Procedure II: DNA-binding site prediction using the LSTM-attention network

The hybrid feature embedding is fed to the designed LSTM-attention network to generate a score vector with Inline graphic dimensions, indicating the confidence scores of belonging to DNA-binding sites for all residues in the query sequence. In the LSTM-attention network, a BiLSTM layer and a self-attention layer are combined to further strengthen the relationship between evolution diversity-based feature embeddings and DNA-binding patterns at the residue level to improve prediction accuracy.

Unsupervised protein language models

ESM2 transformer [44] is an unsupervised deep attention neural network with 36 layers, as depicted in Figure S2. Here, the input of ESM2 is a query sequence with amino acids, while the output is an evolution diversity-based feature embedding matrix. Each layer consists of 20 attention heads and a feed-forward network (FFN), where each head performs the scale dot-product operation to learn the evolution correlation between amino acids in the query sequence from an individual view. Meanwhile, the FFN fuses the evolution knowledge from all attention heads to capture the evolution diversity for the entire sequence. The ESM2 model with 3 billion parameters was trained on over 60 million proteins from the UniRef50 database, as carefully described in Text S2 of SI.

ProtTrans transformer [47] shares a similar architecture to the ESM2. This model is composed of 24 attention layers, with each layer including 32 attention heads. The ProtTrans model with 3 billion parameters was trained on over 45 million proteins from BFD and UniRef50 databases.

ESM-MSA transformer [48] aims to capture the co-evolution knowledge for the input MSA and encode it as a feature embedding matrix, as shown in Figure S3. ESM-MSA consists of 12 attention blocks, with each block including one row-attention layer and one column-attention layer that separately learn the co-evolution correlation between amino acids at the sequence and position level. The ESM-MSA model with 100 million parameters was trained on over 26 million MSAs from Unclust30 and UniRef50 databases, with details in Text S3.

LSTM-attention network

The designed LSTM-attention network includes a BiLSTM layer, a self-attention layer, a fully connected layer, and an output layer, as shown in Figure 1. The BiLSTM includes a forward LSTM and a backward LSTM, which have the same architecture consisting of 256 cells with reverse propagation directions. Each LSTM cell is mainly composed of two states (i.e. cell state Inline graphic and hidden state Inline graphic) and three gates (i.e. forget gate Inline graphic, input gate Inline graphic and output gate Inline graphic). The cell and hidden states are separately used to store and output the signals at the current time-step. The forget, input and output gates are used to control the ratios of incorporating the history signal, inputting the current signal and outputting the updated signal, respectively. Specifically, at time-step Inline graphic (Inline graphic, Inline graphic is the length of the input sequence), the above-mentioned states and gates are computed as follows:

graphic file with name DmEquation1.gif (1)
graphic file with name DmEquation2.gif (2)
graphic file with name DmEquation3.gif (3)
graphic file with name DmEquation4.gif (4)
graphic file with name DmEquation5.gif (5)
graphic file with name DmEquation6.gif (6)

where Inline graphic and Inline graphic are cell state and hidden state, respectively, at the time-step Inline graphic, Inline graphic is the input at the time-step Inline graphic (i.e. the feature embedding vector with 4352 dimensions of the Inline graphicth residue in the query sequence for DNA-binding prediction), Inline graphic is the weight, Inline graphic is the bias, Inline graphic is concatenation operation between two vectors and Inline graphic is the Sigmoid function. The output of the BiLSTM layer is represented as a Inline graphic matrix through concatenating the hidden states in all LSTM cells at all time-steps.

The self-attention layer consists of 10 attention heads, each of which performs the scale dot-product attention as follows:

graphic file with name DmEquation7.gif (7)
graphic file with name DmEquation8.gif (8)

where Inline graphic is the output matrix by the BiLSTM; Inline graphic is an attention matrix in the Inline graphicth attention head; Inline graphic, Inline graphic and Inline graphic are Query, Key and Value matrices with the scale of Inline graphic, respectively; Inline graphic is an Inline graphicweight matrix measuring the position-correlation of amino acid pairs in the query; and Inline graphic is a scale factor.

The attention matrices in all 10 heads are concatenated and then inputted into the fully connected layer containing 1024 neurons, followed by an output layer with one neuron:

graphic file with name DmEquation9.gif (9)
graphic file with name DmEquation10.gif (10)
graphic file with name DmEquation11.gif (11)

where Inline graphic is the linear rectification function and Inline graphic is a score vector with Inline graphic dimensions, indicating the confidence scores of belonging to DNA-binding sites for all residues in the query sequence.

Loss function

We use the cross-entropy loss [54] as the training loss of ULDNA:

graphic file with name DmEquation12.gif (12)

where Inline graphic is the confidence score of belonging to the DNA-binding site at the Inline graphicth residue in the query sequence; Inline graphic, if the Inline graphicth residue is a DNA-binding site annotated by experimental protocols; otherwise, Inline graphic. The training loss is minimized to optimize the hyper-parameters of the ULDNA model via the Adam optimization algorithm [55], where the learning rate, dropout rate and batch size are set to be 0.001, 0.2 and 1, respectively.

Implementation details

The five benchmark datasets were constructed by three individual works [27–29], leading to different definitions of protein–DNA binding sites (see details in Text S1 of SI). Therefore, we should use the datasets with the same definition of DNA-binding sites for training and testing the proposed ULDNA model. Specifically, we separately used PDNA-543 and PDNA-335 datasets to train models, which were then tested on PDNA-41 and PDNA-52, respectively, under independent validation. Moreover, there is no overlap between the two test datasets, because the release dates of proteins in the PDNA-52 are both earlier than those of proteins in the PDNA-41 (see details in the section of ‘Benchmark datasets’). The hyper-parameters and thresholds of the ULDNA model were determined on the corresponding training dataset under 10-fold cross-validation. In addition, the PDNA-316 dataset was used to further evaluate the performance of ULDNA over 10-fold cross-validation.

In the 10-fold cross-validation, the dataset was randomly split into 10-folds at the sequence level. Then, 9-folds were used to train the model, which was tested on the remaining 1-fold. This process was repeated 10 times, ensuring that each protein residue in the dataset was assigned a confidence score belonging to the DNA-binding site. Finally, an appropriate threshold was selected to evaluate the overall prediction performance of the model on the entire dataset under 10-fold cross-validation, where a protein residue was predicted as the DNA-binding site if its confidence score was higher than the selected threshold. To maximize the prediction performance of models over cross-validation, we optimized the hyper-parameters of models, e.g. the number of attention heads and training epochs, using the grid search strategy.

To reduce the influence of randomness, we repeatedly train the model 10 times and then use the average of the confidence scores outputted by 10 models as the final score for each protein residue.

Evaluation indices

Four indices are used to evaluate the performance of the proposed methods, including Sensitivity (Sen), Specificity (Spe), Accuracy (Acc) and Mathew’s Correlation Coefficient (MCC):

graphic file with name DmEquation13.gif (13)
graphic file with name DmEquation14.gif (14)
graphic file with name DmEquation15.gif (15)
graphic file with name DmEquation16.gif (16)

where TP, TN, FP and FN separately stand for numbers of true positives, true negatives, false positives and false negatives.

Since the four indices mentioned above depend on the preset threshold, choosing a suitable threshold is crucial for making fair comparisons between different models. In this study, the reported evaluation indices of the ULDNA model are determined by the threshold that yields the maximum MCC value on the training dataset under 10-fold cross-validation, unless stated otherwise. In addition, to evaluate the overall prediction performance of models, a threshold-independent index is utilized, i.e. the area under the receiver operating characteristic curve (AUROC) [56].

RESULTS AND DISCUSSION

Comparison with existing protein–DNA binding site predictors

To demonstrate the strong performance of the proposed ULDNA, we made a comparison with 12 existing popular DNA-binding site predictors, including BindN [57], ProteDNA [58], BindN+ [59], MetaDBSite [28], DP-Bind [60], DNABind [61], TargetDNA [27], iProDNA-CapsNet [34], DNAPred [13], Guan’s method [32], COACH [16] and PredDBR [33], on the PDNA-41 test dataset under independent validation, as summarized in Table 2.

Table 2.

Performance comparisons between ULDNA and 12 competing predictors on the PDNA-41 test dataset under independent validation

Method Sen Spe Acc MCC AUROC
BindNa 0.456 0.809 0.792 0.143 -
ProteDNAa 0.048 0.998 0.951 0.160 -
BindN+ (Inline graphic)a 0.241 0.951 0.916 0.178 -
BindN+ (Inline graphic)a 0.508 0.854 0.837 0.213 -
MetaDBSitea 0.342 0.934 0.904 0.221 -
DP-Binda 0.617 0.824 0.814 0.241 -
DNABinda 0.702 0.803 0.798 0.264 -
TargetDNA (Inline graphic)a 0.602 0.858 0.845 0.269 -
TargetDNA (Inline graphic)a 0.455 0.933 0.909 0.300 -
iProDNA-CapsNet (Inline graphic)b 0.753 0.753 0.753 0.245 -
iProDNA-CapsNet (Inline graphic)b 0.422 0.949 0.924 0.315 -
DNAPred (Inline graphic)c 0.761 0.767 0.761 0.260 0.858
DNAPred (Inline graphic)c 0.447 0.949 0.924 0.337 0.858
Guan’s methodd 0.476 0.964 0.949 0.357 -
COACHe 0.462 0.951 0.927 0.352 -
PredDBR (Inline graphic)e 0.764 0.758 0.758 0.264 -
PredDBR (Inline graphic)e 0.431 0.958 0.931 0.351 -
PredDBR (threshold = 0.5)e 0.391 0.968 0.939 0.359 -
ULDNA (Inline graphic) 0.824 0.899 0.895 0.458 0.935
ULDNA (Inline graphic) 0.556 0.970 0.950 0.499 0.935
ULDNA (threshold = 0.5) 0.271 0.994 0.958 0.417 0.935

a, b, c, d, eResults excerpted from TargetDNA [29], iProDNA-CapsNet [36], DNAPred [14], Guan et al [34] and PredDBR [35], respectively; ‘Inline graphic’ and ‘Inline graphic’ mean that the thresholds make Inline graphic and Inline graphic, respectively, on the PDNA-543 training dataset over 10-fold cross-validation. ‘-’ means that the corresponding value is unavailable.

It is observed that ULDNA obtains the highest MCC values among all 13 competing methods. Compared to the second best performer PredDBR (a recently proposed deep learning model), ULDNA gains 13.3% improvement of MCC values on average under three different thresholds. More importantly, four evaluation indices of ULDNA are both higher than those of PredDBR under Inline graphic and Inline graphic. Meanwhile, a similar trend but with more significant distinctions can be observed in comparison with other predictors. Taking DNAPred as an example, ULDNA shares the improvements of 6.3, 13.2, 13.4, 19.8 and 7.7%, respectively, on Sen, Spe, Acc, MCC and AUROC values under Inline graphic. It cannot escape from our notice that ProteDNA gains the highest Spe (0.998) but with the lowest Sen (0.048). This is due to that ProteDNA predicts too many false negatives.

Table 3 illustrates the performance comparison among ULDNA, DNABR [31], MetaDBSite [28], TargetS [29], DNAPred [13], COACH [16] and PredDBR [33] on the PDNA-52 test dataset under independent validation, where ULDNA achieves the highest MCC value among all control methods. Specifically, the improvements in MCC values between ULDNA and the other 6 predictors range from 6.6 to 33.2%.

Table 3.

Performance comparisons between ULDNA and 6 competing predictors on the PDNA-52 test dataset under independent validation

Method Sen Spe Acc MCC AUROC
DNABRa 0.407 0.873 0.846 0.185 -
MetaDBSitea 0.580 0.764 0.752 0.192 -
TargetSa 0.413 0.965 0.933 0.377 0.836
DNAPredb 0.518 0.949 0.925 0.405 0.876
COACHc 0.599 0.935 0.916 0.420 -
PredDBRc 0.539 0.958 0.935 0.451 -
ULDNA 0.704 0.944 0.931 0.517 0.945

a, b, cResults excerpted from TargetS [31], DNAPred [14] and PredDBR [35]. ‘-’ means that the corresponding value is unavailable.

We further compare our method with all the above-mentioned methods as well as the other 4 competing methods, including EC-RUS [62], DBS-PRED [63], DISIS [64] and BindN-rf [30], on three training datasets (i.e. PDNA-543, PDNA-335 and PDNA-316) under 10-fold cross-validation, as listed in Tables S1, S2 and S3 of SI. Again, the proposed ULDNA outperforms all other methods.

Contribution analysis of different protein language models

The contributions of three employed protein language models, i.e. ESM2, ProtTrans and ESM-MSA, could be analyzed by further benchmarking the performance of the designed LSTM-attention network with seven different feature embeddings, respectively. These include three individual embeddings extracted from ESM2, ProtTrans and ESM-MSA, and four hybrid embeddings generated by ProtTrans + ESM-MSA (PE), ESM2 + ESM-MSA (EE), ESM2 + ProtTrans (EP) and ESM2 + ProtTrans + ESM-MSA (EPE = ULDNA). Here, ‘+’ indicates that we directly concatenate individual embeddings of different language models as a hybrid embedding. Figure 2 presents the performance comparison between seven feature embeddings across three training datasets (PDNA-543, PDNA-335 and PDNA-316) under 10-fold cross-validation and two test datasets (PDNA-41 and PDNA-52) under independent validation, where the P-values of MCC and AUROC values between EPE and other six feature embeddings under two-sided Student’s t-test [65] are listed in Tables S4 and S5 and discussed in Text S4 of SI.

Figure 2.

Figure 2

The MCC and AUROC values of seven feature embeddings on five benchmark datasets.

It could be found that EPE achieves the best performance among the seven feature embeddings. From the view of MCC values, EPE gains the average improvements of 2.9, 4.3, 6.0, 2.2, 1.3 and 1.0% on five datasets in comparison with ESM2, ProtTrans, ESM-MSA, PE, EE and EP, respectively, where the P-values are both below 0.05 for all the comparisons. With respect to AUROC values, EPE occupies the top-1 position on four out of five datasets. Moreover, ESM2 shows the highest MCC and AUROC values among three individual embeddings; Meanwhile, the largest increase is caused by adding ESM2 to PE on each dataset.

These data demonstrate the following two conclusions. First, three language models pre-trained on different sequence database sources are complementary for improving DNA-binding site prediction. Second, ESM2 makes the most important contribution among the three language models.

Ablation study

We designed an ablation study to investigate the impact of algorithmic advancements in the ULDNA on its enhanced performance. Specifically, we began with a baseline model M0 and progressively incorporated ULDNA’s algorithmic elements to implement two improved models M1 and M2, where M2 is equivalent to ULDNA. The architectures of the three ablation models are depicted in Figure S4, with the following procedures.

M0

This model is built on the BiLSTM architecture, which is serially composed of a BiLSTM layer with 256 cells, a fully connected layer with 1024 neurons, and an output layer with one neuron. Meanwhile, the activation functions in the last two layers are employed by the linear rectification function and Sigmoid function, respectively. Here, the input sequence is encoded as the one-hot coding matrix [66], which is then fed to the BiLSTM architecture to output the confidence scores of belonging to DNA-binding sites for all residues. In addition, the loss function is designed as the cross-entropy loss, as shown in Equation (12).

M1

The one-hot coding matrix used in M0 is replaced by the hybrid feature embedding matrix concatenated by three individual embeddings from the ESM2, ProtTrans and ESM-MSA transformers. This hybrid embedding is further fed to the BiLSTM architecture employed by M0 to output the confidence scores of DNA-binding sites.

M2 (M2 = ULDNA)

We add a self-attention layer consisting of 10 attention heads after the BiLSTM layer in M1.

Figure 3 summarizes the performances of three ablation models across three training datasets under 10-fold cross-validation and two test datasets under independent validation. In comparison with M0, M1 shows a great performance improvement, with the MCC and AUROC values averagely rising by 31.4 and 17.7%, respectively, on five benchmark datasets. This observation demonstrates the significant importance of protein language models for improving DNA-binding site prediction. The performance advantage of M1 over M0 is mainly attributed to that the employed transformers learn the abundant knowledge, highly associated with protein–DNA interaction patterns, from complementary sequence database sources. After adding the self-attention layer, M2 achieves an average increase of 0.7% in MCC values on five datasets in contrast to M1. Although the AUROC values of M2 are slightly lower than those of M1 across the PDNA-543 and PDNA-41, they consistently increase on the other three datasets. These findings suggest that the inclusion of the self-attention layer helps improve the overall accuracy of DNA-binding site prediction, albeit to a lesser extent compared to the enhancements provided by protein language models.

Figure 3.

Figure 3

The MCC and AUROC values of three ablation models on five benchmark datasets.

Testing on recently released PDB targets

The proposed ULDNA was further compared with nine existing DNA-binding site prediction methods on 136 recently released DNA-binding protein chains from the PDB database, including DP-Bind [60], TargetS [29], TargetDNA [27], DNAPred [13], GraphBind [35], NCBRPred [67], GraphSite [10], PredDBR [33] and iDRNA-ITF [68]. Specifically, we collected 1096 DNA-binding protein chains under a cut-off of 30% sequence identity, where the 960 chains and 136 chains separately released in the PDB before and after 1 January 2023 were used as the training dataset (i.e. PDNA-960) and test dataset (i.e. PDNA-136), respectively, for the ULDNA model (see details in Text S5 of SI). Meanwhile, for the nine existing predictors, we downloaded the standalone software (or accessed the computation platforms) and implemented them on the PDNA-136 dataset using the default settings. Moreover, considering the unbalanced distribution of DNA binding sites, we added a new evaluation index, i.e. average precision (AP, see details in Text S6), in all comparisons.

Table 4 summarizes the prediction performance of ULDNA and 9 competing predictors on the PDNA-136 test dataset. It could be found that the proposed ULDNA achieves the best performance among 10 predictors in terms of MCC, AUROC and AP values. Meanwhile, the Sen and Spe values of ULDNA are separately ranked 3 and 2. Compared to the second-best performer GraphSite learning DNA-binding patterns from feature embeddings of AlphaFold2 [69], our method achieves 6.1, 5.8 and 1.6% improvement for MCC, AP and AUROC values, respectively. Moreover, ULDNA is ranked number 1 for all seven evaluation indices in comparison with TargetS, TargetDNA, DNAPred, NCBRPred, GraphSite and PredDBR.

Table 4.

Performance comparisons between ULDNA and nine state-of-the-art predictors on the PDNA-136 test dataset under independent validation

Method Sen Spe Acc MCC AP AUROC
DP-Bind 0.622 0.787 0.779 0.199 0.144 -
TargetS 0.266 0.959 0.929 0.211 0.264 -
TargetDNA 0.455 0.907 0.886 0.238 0.209 0.802
DNAPred 0.432 0.934 0.912 0.275 0.260 0.820
GraphBind 0.628 0.925 0.911 0.379 0.303 0.898
NCBRPred 0.372 0.947 0.921 0.261 0.203 0.799
GraphSite 0.541 0.950 0.931 0.390 0.302 0.907
PredDBR 0.351 0.947 0.920 0.246 0.234 0.775
iDRNA-ITF 0.325 0.966 0.937 0.282 0.208 -
ULDNA 0.544 0.965 0.947 0.451 0.360 0.923

For each competing predictor, we used the default threshold in the corresponding program to calculate evaluation indices. ‘-’ means that the AUROC value is unavailable due to that the corresponding predictor can only output the binary prediction results (‘0’ and ‘1’) rather than confidence scores.

Case study

To delve deeper into the effects of different DNA-binding site prediction approaches, we chose two proteins with PDB IDs of 2MXF_A and 3ZQL_A from our test datasets as case examples. For each protein, we used four in-house methods (denoted as LA-ESM2, LA-ProtTrans, LA-ESM-MSA and ULDNA) and a competing method (PredDBR [33]) to predict the corresponding DNA-binding sites. Four in-house methods use the same LSTM-attention network with different feature embeddings from ESM2, ProtTrans, ESM-MSA and ESM2 + ProtTrans+ESM-MSA, respectively. Here, ‘+’ indicates that we directly concatenate individual embeddings of different language models as a hybrid embedding. Table 5 summarizes the modeling results of two proteins for five DNA-binding site prediction methods, where the corresponding visualization results are illustrated in Figure 4. In addition, the predicted and native DNA-binding sites of two proteins by five methods are listed in Table S6 of SI.

Table 5.

The modeling results of five DNA-binding site prediction methods on two representative examples

Method 2MXF_A 3ZQL_A
TP FP TN FN MCC TP FP TN FN MCC
LA-ESM2 12 2 29 4 0.710 12 9 213 2 0.678
LA-ProtTrans 12 4 27 4 0.621 11 8 214 3 0.651
LA-ESM-MSA 12 1 30 4 0.760 14 10 212 0 0.746
ULDNA 13 0 31 3 0.861 14 8 214 0 0.783
PredDBR 8 2 29 6 0.564 14 18 204 0 0.634

Figure 4.

Figure 4

Visualization of prediction results for two proteins (2MXF_A and 3ZQL_A) using five DNA-binding site prediction models: (A) LA-ESM2, (B) LA-ProtTrans, (C) LA-ESM-MSA, (D) ULDNA, (E) PredDBR. The atomic-level native structure of each protein is downloaded from the PDB database and then plotted as the cartoon picture using PyMOL software [70]. The color scheme is used as follows: DNA in orange, true positives in blue, false positives in red and false negatives in green.

From the experiment data, we observed several interesting phenomena. First, the protein language models are critical to improve DNA-binding site prediction. Specifically, four in-house methods with pre-trained protein language models both show higher MCC values than the competing PredDBR without language models on two proteins. Taking ULDNA as an example, it gains the MCC increases by 29.7 and 14.9%, respectively, on 2MXF_A and 3ZQL_A in comparison with PredDBR.

Second, the combination of complementary protein language models can further increase the accuracy of ULDNA. In 2MXF_A, three in-house methods (i.e. LA-ESM2, LA-ProtTrans and LA-ESM-MSA) with different language models detect a total of 14 true positives. This number surpasses the true positives predicted by each individual method, suggesting that three language models (i.e. ESM2, ProtTrans and ESM-MSA) extract complementary knowledge from multiple sequence databases. Meanwhile, the false positives predicted by one in-house method can be corrected by the other two methods. For example, LA-ESM2 generates two false positives (10P and 11H), which are correctly predicted as non-DNA-binding sites by LA-ProtTrans and LA-ESM-MSA. As a result, by taking the combination of three language models, ULDNA gains the most true positives without false positives among all methods. Occasionally, one in-house method could capture all the true positives identified by other methods. In the case of 3ZQL_A, LA-ESM-MSA encompasses all the true positives predicted by both LA-ESM2 and LA-ProtTrans. Despite this overlap, the overall accuracy of the final ULDNA is still improved by including all individual methods to reduce false positives.

CONCLUSIONS

In this work, a novelty deep learning model, ULDNA, is developed to predict DNA-binding sites from protein sequences through leveraging an LSTM-attention architecture embedded with protein language transformer models. The results from benchmark testing have shown that ULDNA significantly surpasses existing popular methods in the accuracy of predicting DNA-binding sites. The performance enhancement of ULDNA stems from two advancements. First, three transformer models pre-trained on multiple large-scare sequence databases could capture the complementary feature embeddings with evolution diversity, which are highly associated with protein–DNA interactions. Second, the specifically designed LSTM-attention network further strengthens the relationship between evolution diversity-based feature embeddings and DNA-binding patterns to improve prediction accuracy.

Although the prediction performance is promising, there remains substantial potential for further advancements. First, the serial feature concatenation strategy, currently utilized in the ULDNA, cannot perfectly deal with the redundant information among the feature embeddings from different transformers. Thus, designing a more advanced approach to feature fusion could help reduce the adverse effects arising from information redundancy in the future. Second, with the development of protein structure prediction models (e.g. AlphaFold2 [69] and ESMFold [44]), the predicted structures will have the huge potential for improving DNA-binding site prediction. Resarches in these directions are currently ongoing.

Key Points

  • Accurate recognition of protein–DNA binding sites is crucial for understanding the molecular mechanisms underlying various biological processes and thus advancing drug discovery. This study has designed a novelty deep learning model ULDNA to accurately predict DNA-binding sites from protein sequences through integrating three unsupervised protein language models from multiple database sources with the designed LSTM-attention network.

  • Results from computational experiments have revealed that ULDNA significantly surpasses existing popular methods in the accuracy of DNA-binding site prediction. The major strength of ULDNA stems from employing three transformer language models that can effectively capture complementary feature embeddings with evolution diversity that are highly associated with complicated DNA-binding patterns.

  • An online server for predicting protein–DNA binding sites is freely accessible through the URL http://csbio.njust.edu.cn/bioinf/uldna/.

Supplementary Material

SI_bbae040
si_bbae040.docx (7MB, docx)

Author Biographies

Yi-Heng Zhu received his PhD degree in control science and engineering from Nanjing University of Science and Technology in 2023. He is currently a lecturer at the College of Artificial Intelligence, Nanjing Agricultural University. His research interests include bioinformatics, machine learning and pattern recognition.

Zi Liu received his PhD degree from the School of Computer Science and Engineering, Nanjing University of Science and Technology. From 2017 to 2019, he participated in a joint PhD program at the prestigious University of Michigan, USA. He is a lecturer at the School of Information and Engineering, Jingdezhen Ceramic University. His research interests include bioinformatics, machine learning and pattern recognition.

Yan Liu received his PhD degree in computer science from Nanjing University of Science and Technology in 2023. He is currently a young hundred distinguished professor at the School of Information Engineering, Yangzhou University, China. His research interests include pattern recognition, machine learning and bioinformatics.

Zhiwei Ji joined in the School of Artificial Intelligence in Nanjing Agricultural University (NJAU) on 21 August 2020 as a full professor. Prior to this position, he was an assistant professor at the University of Texas Health Science Center at Houston (UTHealth). He has been working on Systems Biology, Bioinformatics, Pattern Recognition, Big Data Analysis and Modeling for over 10 years. Currently, he is the director of Center for Data Science and Intelligent Computing at NJAU.

Dong-Jun Yu is a full professor at the School of Computer Science and Engineering, Nanjing University of Science and Technology, Nanjing, China. His research interests include pattern recognition, machine learning and bioinformatics. He is a senior member of the China Computer Federation (CCF) and a senior member of the China Association of Artificial Intelligence (CAAI).

Contributor Information

Yi-Heng Zhu, College of Artificial Intelligence, Nanjing Agricultural University, Nanjing 210095, China.

Zi Liu, School of Computer Science and Engineering, Nanjing University of Science and Technology, Nanjing 210094, China.

Yan Liu, School of Information Engineering, Yangzhou University, Yangzhou 225000, China.

Zhiwei Ji, College of Artificial Intelligence, Nanjing Agricultural University, Nanjing 210095, China.

Dong-Jun Yu, School of Computer Science and Engineering, Nanjing University of Science and Technology, Nanjing 210094, China.

FUNDING

This work is supported by the National Natural Science Foundation of China (62372234, 62072243, 61772273 and 62306142), the Natural Science Foundation of Jiangsu (BK20201304 and BK20211210), the Foundation of National Defense Key Laboratory of Science and Technology (JZX7Y202001SY000901) and Jiangsu Funding Program for Excellent Postdoctoral Talent (No. 2023ZB224).

CODE AVAILABILITY

The datasets and source codes can be freely downloaded at https://github.com/yiheng-zhu/ULDNA.

References

  • 1. Stormo GD, Zhao Y. Determining the specificity of protein-DNA interactions. Nat Rev Genet  2010;11(11):751–60. [DOI] [PubMed] [Google Scholar]
  • 2. Gallagher LA, Velazquez E, Peterson SB, et al.  Genome-wide protein-DNA interaction site mapping in bacteria using a double-stranded DNA-specific cytosine deaminase. Nat Microbiol  2022;7(6):844–55. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3. Esmaeeli R, Bauzá A, Perez A. Structural predictions of protein-DNA binding: MELD-DNA. Nucleic Acids Res  2023;51(4):1625–36. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4. Hua Y, Li J, Feng Z, et al.  Protein drug interaction prediction based on attention feature fusion. J Comput Res Develop  2022;59(9):2051–65. [Google Scholar]
  • 5. Kim E, Kim Y-J, Ji Z, et al.  ROR activation by Nobiletin enhances antitumor efficacy via suppression of IκB/NF-κB signaling in triple-negative breast cancer. Cell Death Dis  2022;13(4):374. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6. Shao H, Peng T, Ji Z, et al.  Systematically studying kinase inhibitor induced signaling network signatures by integrating both therapeutic and side effects. PloS One  2013;8(12):e80832. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7. Orengo CA, Michie AD, Jones S, et al.  CATH: a hierarchic classification of protein domain structures. Structure  1997;5(8):1093–109. [DOI] [PubMed] [Google Scholar]
  • 8. Yu Y, Li S, Ser Z, et al.  Cryo-EM structure of DNA-bound Smc5/6 reveals DNA clamping enabled by multi-subunit conformational changes. Proc Natl Acad Sci  2022;119(23):e2202799119. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9. UniProt Consortium . UniProt: a worldwide hub of protein knowledge. Nucleic Acids Res  2019;47(D1):D506–15. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10. Yuan Q, Chen S, Rao J, et al.  AlphaFold2-aware protein-DNA binding site prediction using graph transformer. Brief Bioinform  2022;23(2):bbab564. [DOI] [PubMed] [Google Scholar]
  • 11. Qu K, Wei L, Zou Q. A review of DNA-binding proteins prediction methods. Curr Bioinform  2019;14(3):246–54. [Google Scholar]
  • 12. Zhu Y-H, Zhang C, Liu Y, et al.  TripletGO: integrating transcript expression profiles with protein homology inferences for gene function prediction. Genom Proteom Bioinform  2022;20(5):1013–27. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13. Zhu Y-H, Hu J, Song X-N, Yu D-J. DNAPred: accurate identification of DNA-binding sites from protein sequence by ensembled hyperplane-distance-based support vector machines. J Chem Inf Model  2019;59(6):3057–71. [DOI] [PubMed] [Google Scholar]
  • 14. Jones S, Shanahan HP, Berman HM, Thornton JM. Using electrostatic potentials to predict DNA-binding sites on DNA-binding proteins. Nucleic Acids Res  2003;31(24):7189–98. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15. Tsuchiya Y, Kinoshita K, Nakamura H. Structure-based prediction of DNA-binding sites on proteins using the empirical preference of electrostatic potential and the shape of molecular surfaces. Proteins  2004;55(4):885–94. [DOI] [PubMed] [Google Scholar]
  • 16. Yang J, Roy A, Zhang Y. Protein-ligand binding site recognition using complementary binding-specific substructure comparison and sequence profile alignment. Bioinformatics  2013;29(20):2588–95. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17. Altschul SF, Madden TL, Schäffer AA, et al.  Gapped BLAST and PSI-BLAST: a new generation of protein database search programs. Nucleic Acids Res  1997;25(17):3389–402. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18. Li T, Li Q-Z, Liu S, et al.  PreDNA: accurate prediction of DNA-binding sites in proteins by integrating sequence and geometric structure information. Bioinformatics  2013;29(6):678–85. [DOI] [PubMed] [Google Scholar]
  • 19. Gao M, Skolnick J. DBD-hunter: a knowledge-based method for the prediction of DNA-protein interactions. Nucleic Acids Res  2008;36(12):3978–92. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20. Tsuchiya Y, Kinoshita K, Nakamura H. PreDs: a server for predicting dsDNA-binding site on protein molecular surfaces. Bioinformatics  2005;21(8):1721–3. [DOI] [PubMed] [Google Scholar]
  • 21. Gao M, Skolnick J. A threading-based method for the prediction of DNA-binding proteins with application to the human genome. PLoS Comput Biol  2009;5(11):e1000567. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22. Chen YC, Wright JD, Lim C. DR_bind: a web server for predicting DNA-binding residues from the protein structure based on electrostatics, evolution and geometry. Nucleic Acids Res  2012;40(W1):W249–56. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23. Morozov AV, Havranek JJ, Baker D, Siggia ED. Protein-DNA binding specificity predictions with structural models. Nucleic Acids Res  2005;33(18):5781–98. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24. Jones DT. Protein secondary structure prediction based on position-specific scoring matrices. J Mol Biol  1999;292(2):195–202. [DOI] [PubMed] [Google Scholar]
  • 25. Noble WS. What is a support vector machine?  Nat Biotechnol  2006;24(12):1565–7. [DOI] [PubMed] [Google Scholar]
  • 26. Biau G, Scornet E. A random forest guided tour. Test  2016;25:197–227. [Google Scholar]
  • 27. Hu J, Li Y, Zhang M, et al.  Predicting protein-DNA binding residues by weightedly combining sequence-based features and boosting multiple SVMs. IEEE/ACM Trans Comput Biol Bioinform  2016;14(6):1389–98. [DOI] [PubMed] [Google Scholar]
  • 28. Si J, Zhang Z, Lin B, et al.  MetaDBSite: a meta approach to improve protein DNA-binding sites prediction. BMC Syst Biol  2011;5(1):1–7. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29. Yu D-J, Hu J, Yang J, et al.  Designing template-free predictor for targeting protein-ligand binding sites with classifier ensemble and spatial clustering. IEEE/ACM Trans Comput Biol Bioinform  2013;10(4):994–1008. [DOI] [PubMed] [Google Scholar]
  • 30. Wang L, Yang MQ, Yang JY. Prediction of DNA-binding residues from protein sequence information using random forests. BMC Genom  2009;10:1–9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31. Ma X, Guo J, Liu H-D, et al.  Sequence-based prediction of DNA-binding residues in proteins with conservation and correlation information. IEEE/ACM Trans Comput Biol Bioinform  2012;9(06):1766–75. [DOI] [PubMed] [Google Scholar]
  • 32. Guan S, Zou Q, Wu H, Ding Y. Protein-DNA binding residues prediction using a deep learning model with hierarchical feature extraction. IEEE/ACM Trans Comput Biol Bioinform  2022;20(5):2619–28. [DOI] [PubMed] [Google Scholar]
  • 33. Hu J, Bai Y-S, Zheng L-L, et al.  Protein-DNA binding residue prediction via bagging strategy and sequence-based cube-format feature. IEEE/ACM Trans Comput Biol Bioinform  2021;19(6):3635–45. [DOI] [PubMed] [Google Scholar]
  • 34. Nguyen BP, Nguyen QH, Doan-Ngoc G-N, et al.  iProDNA-CapsNet: identifying protein-DNA binding residues using capsule neural networks. BMC Bioinform  2019;20:1–12. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35. Xia Y, Xia C-Q, Pan X, Shen H-B. GraphBind: protein structural context embedded rules learned by hierarchical graph neural networks for recognizing nucleic-acid-binding residues. Nucleic Acids Res  2021;49(9):e51–1. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36. Berman HM, Westbrook J, Feng Z, et al.  The protein data bank. Nucleic Acids Res  2000;28(1):235–42. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37. Rao R, Bhattacharya N, Thomas N, et al.  Evaluating protein transfer learning with TAPE. Adv Neural Inf Process Syst  2019;32:9689–701. [PMC free article] [PubMed] [Google Scholar]
  • 38. Heinzinger M, Elnaggar A, Wang Y, et al.  Modeling aspects of the language of life through transfer-learning protein sequences. BMC Bioinform  2019;20(1):1–17. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39. Madani A, Krause B, Greene ER, et al.  Large language models generate functional protein sequences across diverse families. Nat Biotechnol  2023;41(8):1099–106. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40. Ferruz N, Schmidt S, Höcker B. ProtGPT2 is a deep unsupervised language model for protein design. Nat Commun  2022;13(1):4348. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41. Littmann M, Heinzinger M, Dallago C, et al.  Protein embeddings and deep learning predict binding residues for various ligand classes. Sci Rep  2021;11(1):23916. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42. Zhu Y-H, Zhang C, Yu D-J, Zhang Y. Integrating unsupervised language model with triplet neural networks for protein gene ontology prediction. PLoS Comput Biol  2022;18(12):e1010793. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43. Rives A, Meier J, Sercu T, et al.  Biological structure and function emerge from scaling unsupervised learning to 250 million protein sequences. Proc Natl Acad Sci  2021;118(15):e2016239118. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 44. Lin Z, Akin H, Rao R, et al.  Evolutionary-scale prediction of atomic-level protein structure with a language model. Science  2023;379(6637):1123–30. [DOI] [PubMed] [Google Scholar]
  • 45. Jiang Z, Shen Y-Y, Liu R. Structure-based prediction of nucleic acid binding residues by merging deep learning-and template-based approaches. PLoS Comput Biol  2023;19(9):e1011428. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 46. Song Y, Yuan Q, Zhao H, Yang Y. Accurately identifying nucleic-acid-binding sites through geometric graph learning on language model predicted structures. Brief Bioinform  2023;24(6):bbad360. [DOI] [PubMed] [Google Scholar]
  • 47. Elnaggar A, Heinzinger M, Dallago C, et al.  ProtTrans: toward understanding the language of life through self-supervised learning. IEEE Trans Pattern Anal Mach Intell  2021;44(10):7112–27. [DOI] [PubMed] [Google Scholar]
  • 48. Rao RM, Liu J, Verkuil R, et al.  MSA transformer. Int Conf Mach Learn  2021;139:8844–56. [Google Scholar]
  • 49. Li W, Godzik A. Cd-hit: a fast program for clustering and comparing large sets of protein or nucleotide sequences. Bioinformatics  2006;22(13):1658–9. [DOI] [PubMed] [Google Scholar]
  • 50. Wang G, DunbrackRL, Jr. PISCES: a protein sequence culling server. Bioinformatics  2003;19(12):1589–91. [DOI] [PubMed] [Google Scholar]
  • 51. Mirdita M, Von Den Driesch L, Galiez C, et al.  Uniclust databases of clustered and deeply annotated protein sequences and alignments. Nucleic Acids Res  2017;45(D1):D170–6. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 52. Suzek BE, Huang H, McGarvey P, et al.  UniRef: comprehensive and non-redundant UniProt reference clusters. Bioinformatics  2007;23(10):1282–8. [DOI] [PubMed] [Google Scholar]
  • 53. Steinegger M, Mirdita M, Söding J. Protein-level assembly increases protein sequence recovery from metagenomic samples manyfold. Nat Methods  2019;16(7):603–6. [DOI] [PubMed] [Google Scholar]
  • 54. Zhang Z, Sabuncu MR. Generalized cross entropy loss for training deep neural networks with noisy labels. Adv Neural Inf Process Syst  2018;31:1–11. [Google Scholar]
  • 55. Kingma DP, Ba J. Adam: a method for stochastic optimization. The 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, May 7-9, 2015, Conference Track Proceedings. 10.48550/arXiv.1412.6980. [DOI] [Google Scholar]
  • 56. Xie S, Xie X, Zhao X, et al.  HNSPPI: a hybrid computational model combing network and sequence information for predicting protein–protein interaction. Brief Bioinform  2023;24(5):bbad261. [DOI] [PubMed] [Google Scholar]
  • 57. Wang L, Brown SJ. BindN: a web-based tool for efficient prediction of DNA and RNA binding sites in amino acid sequences. Nucleic Acids Res  2006;34(suppl_2):W243–8. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 58. Chu W-Y, Huang Y-F, Huang C-C, et al.  ProteDNA: a sequence-based predictor of sequence-specific DNA-binding residues in transcription factors. Nucleic Acids Res  2009;37(suppl_2):W396–401. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 59. Wang L, Huang C, Yang MQ, Yang JY. BindN+ for accurate prediction of DNA and RNA-binding residues from protein sequence features. BMC Syst Biol  2010;4:1–9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 60. Hwang S, Gou Z, Kuznetsov IB. DP-Bind: a web server for sequence-based prediction of DNA-binding residues in DNA-binding proteins. Bioinformatics  2007;23(5):634–6. [DOI] [PubMed] [Google Scholar]
  • 61. Liu R, Hu J. DNABind: a hybrid algorithm for structure-based prediction of DNA-binding residues by combining machine learning-and template-based approaches. Proteins  2013;81(11):1885–99. [DOI] [PubMed] [Google Scholar]
  • 62. Ding Y, Tang J, Guo F. Identification of protein-ligand binding sites by sequence information and ensemble classifier. J Chem Inf Model  2017;57(12):3149–61. [DOI] [PubMed] [Google Scholar]
  • 63. Ahmad S, Gromiha MM, Sarai A. Analysis and prediction of DNA-binding proteins and their binding residues based on composition, sequence and structural information. Bioinformatics  2004;20(4):477–86.14990443 [Google Scholar]
  • 64. Ofran Y, Mysore V, Rost B. Prediction of DNA-binding residues from sequence. Bioinformatics  2007;23(13):i347–53. [DOI] [PubMed] [Google Scholar]
  • 65. De Winter JC. Using the Student's t-test with extremely small sample sizes. Pract Assess Res Eval  2019;18(1):10. [Google Scholar]
  • 66. Rodríguez P, Bautista MA, Gonzalez J, Escalera S. Beyond one-hot encoding: lower dimensional target embedding. Image Vision Comput  2018;75:21–31. [Google Scholar]
  • 67. Zhang J, Chen Q, Liu B. NCBRPred: predicting nucleic acid binding residues in proteins based on multilabel learning. Brief Bioinform  2021;22(5):bbaa397. [DOI] [PubMed] [Google Scholar]
  • 68. Wang N, Yan K, Zhang J, Liu B. iDRNA-ITF: identifying DNA-and RNA-binding residues in proteins based on induction and transfer framework. Brief Bioinform  2022;23(4):bbac236. [DOI] [PubMed] [Google Scholar]
  • 69. Jumper J, Evans R, Pritzel A, et al.  Highly accurate protein structure prediction with AlphaFold. Nature  2021;596(7873):583–9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 70. Yuan S, Chan HS, Hu Z. Using PyMOL as a platform for computational drug design. Wiley Interdiscip Rev: Comput Mol  2017;7(2):e1298. [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

SI_bbae040
si_bbae040.docx (7MB, docx)

Articles from Briefings in Bioinformatics are provided here courtesy of Oxford University Press

RESOURCES