Abstract
Background
Protein secondary structure prediction method based on probabilistic models such as hidden Markov model (HMM) appeals to many because it provides meaningful information relevant to sequence-structure relationship. However, at present, the prediction accuracy of pure HMM-type methods is much lower than that of machine learning-based methods such as neural networks (NN) or support vector machines (SVM).
Results
In this paper, we report a new method of probabilistic nature for protein secondary structure prediction, based on dynamic Bayesian networks (DBN). The new method models the PSI-BLAST profile of a protein sequence using a multivariate Gaussian distribution, and simultaneously takes into account the dependency between the profile and secondary structure and the dependency between profiles of neighboring residues. In addition, a segment length distribution is introduced for each secondary structure state. Tests show that the DBN method has made a significant improvement in the accuracy compared to other pure HMM-type methods. Further improvement is achieved by combining the DBN with an NN, a method called DBNN, which shows better Q3 accuracy than many popular methods and is competitive to the current state-of-the-arts. The most interesting feature of DBN/DBNN is that a significant improvement in the prediction accuracy is achieved when combined with other methods by a simple consensus.
Conclusion
The DBN method using a Gaussian distribution for the PSI-BLAST profile and a high-ordered dependency between profiles of neighboring residues produces significantly better prediction accuracy than other HMM-type probabilistic methods. Owing to their different nature, the DBN and NN combine to form a more accurate method DBNN. Future improvement may be achieved by combining DBNN with a method of SVM type.
Background
Over past decades, the prediction accuracy of protein secondary structure has gained some improvements, largely due to the successful application of machine learning tools such as neural network (NN) and support vector machine (SVM). Qian and Sejnowski designed one of the earliest NN methods [1]. Rost and Sander introduced the alignment profile with multiple sequence alignment into the prediction. Their method, named as PHD, performed much better than previous ones, because of the use of alignment profile as the network's input [2]. Jones made an important improvement by pioneering the use of position-specific scoring matrices (PSSM) to generate the so-called PSI-BLAST profile and developed the method called PSIPRED [3]. Recently, new advances have been made in developing NN-based prediction methods [4-7]. Similarly, SVM-based methods were developed for protein secondary structure prediction, first taking the alignment profile as inputs and then being improved to use the PSI-BLAST profile [8-12]. Generally speaking, the Q3 of a modern NN or SVM-based method can reach over 76%.
In contrast to NN and SVM, probabilistic methods for protein secondary structure prediction such as those based on hidden Markov model (HMM) have had very limited accuracy [13-18]. Most of them were designed for single sequence prediction with prediction accuracy generally less than 70%. Recently, two profile-based HMM methods were proposed, which take either the alignment profile or PSI-BLAST profile as inputs [16,18]. Both of the methods treat the profile as production from a multinomial distribution with 20 possible outcomes (20 amino acids), and thus lose the information about the correlation between entries of the profile. As a result, the prediction accuracy of the two methods, which is around 72%, is still much lower than the common level of NN or SVM-based methods. It is notable that there is a special HMM-type method, SAM-T04 [19], which has shown comparable accuracy to NN and SVM-based methods. However, with using a neural network for the sequence-to-structure prediction while building the HMM only at the secondary structure level [19,20], SAM-T04 should not be regarded as a pure HMM-type method.
It would be interesting to break this apparent asymmetry in accuracy between machine learning-based methods and probabilistic model-based methods. The probabilistic model is of somewhat different nature from machine learning tools, and provides a complement to the latter. Thus, combining the two kinds of model is likely to produce a consensus prediction that has better accuracy than the prediction of individual program [21]. In addition, the probabilistic model outputs a set of knowledge about the property of secondary structure in an explicit way, including specific correlation structure between neighboring residues, while such information is implicit in NN or SVM. Hence, the development of an appropriate probabilistic model is interesting for understanding the mechanism by which sequence determines structure.
In this paper we introduce a new probabilistic model, dynamic Bayesian network (DBN), for protein secondary structure prediction. DBN represents a directed graphical model of a stochastic process, often regarded as a generalized HMM capable of describing correlation structure in a more flexible way [22]. A novel feature of our method is the introduction of a multivariate Gaussian distribution for the profile of each residue, which takes into account the correlation between entries of the PSSM. In addition, our method considers a high-ordered dependency between profiles of neighboring residues and introduces a segment length distribution for each secondary structure state. Testing results show that the DBN method has made a significant improvement in accuracy over previous pure HMM-type methods. Further improvement is achieved by combining the DBN with an NN, a method named DBNN, which has achieved better Q3 accuracy than many other popular methods and is competitive to the current state-of-the-arts. The most interesting feature of DBN/DBNN is that a significant improvement in the prediction accuracy is achieved when combined with other methods by a simple consensus.
Results and Discussion
Training and testing datasets
Three public datasets are employed for training and testing, i.e. CB513 [21], EVA [23] common set, and a large dataset containing 3,223 chains (denoted by EVAtrain) constructed by G. Karypis [12]. The first dataset contains 513 protein sequences with guaranteed non-redundancy via a strict criterion (z-score ≥ 5) for the sequence similarity; this dataset is used independently from two other datasets. The second is obtained from EVA server, where several secondary structure prediction servers are evaluated with sequences deposited in PDB [24]. In particular, a set labeled as "common set 6" (denoted by EVAc6) is selected, which contains 212 protein chains and has been used to test several popular prediction methods [25]. The third dataset, EVAtrain, is used in conjunction with EVAc6, with the former for training and the latter for testing. EVAtrain has been guaranteed to have less than 25% sequence identities to EVAc6.
Furthermore, we have built a fourth dataset based on the known tertiary structural similarity from the SCOP [26] database (release 1.69), to evaluate the performance of our methods when dealing with proteins of remote evolutionary relation. One protein domain for each superfamily of the four classes (all α, all β, α and β, α/β) is selected. The domains of multi-segment, of NMR structure, and of low X-ray resolution (> 2.5Å) are removed. Also, too short (< 30 residues) or too long (> 500 residues) sequences are removed. The final dataset contains 576 protein sequences and is referred to as SD576.
For all the datasets described above, the secondary structure is assigned by DSSP program [27], and the eight-state secondary structure is converted to three, according to the rule: H, G, and I to H (helix); E and B to E (sheet); all others to C (coil).
Window sizes
The window sizes, denoted by LAA and LSS for profile and secondary structure respectively, describe the range of dependency of current site on its neighbors. The correlation between the Q3 accuracy of DBN and window sizes is studied via a set of seven-fold cross-validation tests of DBNsigmoid (see Methods) on SD576 using different window sizes. Due to the limitation in the computational resources, the upper bounds of LAA and LSS are set to be 5 and 4, respectively.
As shown in Fig. 1, Q3 is improved significantly when LSS > 0, and saturated when LSS > 1, which indicates that there is strong short-range dependency between the profile of a residue and the secondary structure states of its neighbors. A similar phenomenon occurs for profiles' dependency of neighboring sites. Note that the model with either LAA = 0 or LSS = 0 is a special case of DBN, in which the distribution of the profile of each residue is independent from neighboring profiles or neighboring secondary structure states, respectively. As a result, its topology is different from that of a full-DBN version (LAA > 0 and LSS > 0) due to the removal of Ri or di nodes (see Fig. 2(c)).
Our results are in partial agreement with the conclusions of Crooks and Brenner, who claimed that each amino acid was dependent on the neighboring secondary structure states but was essentially independent from neighboring amino acids [16]. We argue, however, that the PSI-BLAST profile has quite different correlation structure from a single amino acid sequence, from which Crooks et al. derived their conclusions. In fact, the dependency between neighboring profiles are significant and helpful for improving the prediction accuracy.
Fig. 1 also shows that the most accurate model occurs when using the set (LAA = 4, LSS = 4), for which Q3 reaches about 77.5%. However, test shows that this model is very time-consuming. We choose a more economical set (LAA = 4, LSS = 3) which offers a similar Q3 (see Fig. 1) with a big saving in computational cost, for all the DBN models used in current study.
The accuracy improvements through combinations
All the basic DBN- and NN-based models described in Methods are tested on the SD576 dataset, and the results shown in Table 1 report the performance of these models, as well as of their combinations. Specifically, both DBNlinear (combination of DBNlinear+NC and DBNlinear+CN) and DBNsigmoid (combination of DBNsigmoid+NC and DBNsigmoid+CN) have significantly improved the performance in all the measures, indicating that the two directions of the sequence (i.e. from N-terminus to C-terminus and reverse) contain complementary information. In addition, the combination of the two different PSSM-transformation strategies (i.e. the combination of DBNlinear and DBNsigmoid to produce DBNfinal) also contributes to the accuracy improvement, increasing Q3 and SOV by 0.8% and 0.9%, respectively, for DBN-based models. Note that for NN-based models, the accuracy improvement by combination is much less evident, indicating that NN is not sensitive to PSSM-transformation strategies.
Table 1.
Model | Q3 (%) | SOV (%) | CH | CE | CC |
DBNlinear+NC | 75.1 | 74.0 | 0.69 | 0.60 | 0.55 |
DBNlinear+CN | 74.6 | 73.3 | 0.68 | 0.61 | 0.53 |
DBNlinear | 77.0 | 75.8 | 0.72 | 0.64 | 0.58 |
DBNsigmoid+NC | 75.8 | 74.5 | 0.72 | 0.60 | 0.56 |
DBNsigmoid+CN | 74.6 | 73.3 | 0.69 | 0.61 | 0.54 |
DBNsigmoid | 77.4 | 75.9 | 0.74 | 0.64 | 0.59 |
DBNfinal | 78.2 | 76.8 | 0.74 | 0.65 | 0.60 |
NNlinear | 77.6 | 73.2 | 0.72 | 0.64 | 0.60 |
NNsigmoid | 77.1 | 71.0 | 0.72 | 0.63 | 0.59 |
NNfinal | 77.8 | 73.3 | 0.73 | 0.64 | 0.60 |
DBNN | 80.0 | 78.1 | 0.77 | 0.68 | 0.63 |
All the eleven models listed in the table are described in Methods. The average results of seven-fold cross-validation are shown.
Table 1 shows that DBNfinal has improved by 3.5% over NNfinal in SOV. It can be understood, because DBN-based models explicitly incorporate the segment length distributions while NN-based models miss such information.
Finally, the combination of all the basic DBN- and NN-based models, which produces the resultant DBNN, has achieved further improvement in the accuracy, increasing Q3 and SOV by 1.8% and 1.3%, respectively, compared to DBNfinal (see Table 1). This implies that the two types of models are indeed complementary.
Secondary structure segment length distributions
To study the significance of the secondary structure segment length distributions introduced in DBN models, we define a degenerate DBN (denoted by DBNgeo), which has the same structure to DBNfinal except Dmax = 1 [see Eq. (10)]. As described in Methods, Dmax = 1 implies a geometric distribution for the segment lengths. The segment length distributions of the predicted secondary structure by both DBNfinal and DBNgeo are calculated and compared to the true distributions observed in the SD576 dataset, as shown in Fig. 3(b)–(d). In particular, Fig. 3(b) shows that, for helices, the segments of one and two residues are over-predicted, while those of three residues are under-predicted, by both DBNfinal and DBNgeo. But longer segments are all predicted correctly by both models. Generally speaking, DBNfinal has better performance than DBNgeo: the prediction of DBNfinal for segments of 3 and 5–7 residues is much better than that of DBNgeo.
Fig. 3(c) and 3(d) show the segment length distributions for sheets and coils, respectively. Both DBNfinal and DBNgeo have missed a rich population of one residue, and over-predicted segments of 3–5 residues, for sheets. DBNgeo has predicted a spurious peak for segments of 3 and 4 residues, which is absent in the true distribution. On the contrary, DBNfinal gives a distribution closer to the observation, in which the peak is located at segments of about 5 residues. Fig. 3(d) shows that DBNfinal and DBNgeo have very similar performance for coils: both under-predict the segments of 1 and 2 residues and over-predict those of 3 and 4 residues. However, DBNfinal predicts a much better distribution for long coils (over 8 residues) than DBNgeo.
It is interesting to study whether we can modify the a priori segment length distribution, gα(n) in Eq. (10), to get a predicted (posterior) distribution closer to the observation shown in Fig. 3(a). A calculation is made by using a modified version of DBNfinal, denoted by DBNmod, which is constructed as following: take the a priori segment length distribution directly from the training set, then run the prediction and calculate the posterior distribution, and finally modify the a priori distribution according to the following equation:
(1) |
where gαold(n) is the a priori segment length distribution before the modification, gαpre(n) is the predicted distribution, gαobs(n) is the observed distribution, α = H, E, or C, and n = 1, 2, ... Dmax. The quantity gαnew(n) is then normalized to form the new a priori segment length distribution. The Eq. (1) enhances the population of deficient segments and reduces that of over-represented ones, in a linear fashion. All the three models, DBNfinal, DBNgeo, and DBNmod, are tested on SD576, and the performance on segment length distributions prediction is measured by "relative entropies", defined by
(2) |
where gαobs(n), gαpre(n), and Dmax have the same definitions as above, and α = H, E, or C.
The results presented in Table 2 show that DBNgeo has much higher relative entropies indicating a strong deviation of the predicted distributions from the observation, than other two models. Note that Q3 and SOV of DBNgeo are also much lower than that of DBNfinal (Table 2), implying that the segment length distributions do have an effect on the prediction accuracy. On the other hand, DBNmod shows the lowest relative entropies for all the three secondary structure states with almost the same Q3 and SOV to DBNfinal (see Table 2), which indicates that Eq. (1) has effectively improved the prediction of segment length distributions.
Table 2.
Model | Q3 (%) | SOV (%) | Relative entropy (bit) | |||
Helix | Sheet | Coil | Average | |||
DBNgeo | 76.7 | 74.3 | 0.247 | 0.170 | 0.290 | 0.236 |
DBNfinal | 78.2 | 76.8 | 0.236 | 0.096 | 0.210 | 0.181 |
DBNmod | 78.2 | 76.3 | 0.214 | 0.038 | 0.110 | 0.121 |
The seven-fold cross-validation test results on three models with different segment length distributions are explained in the text. The performance is measured by Q3, SOV, and the relative entropies between the observed segment length distributions from SD576 and the model's predictions [Eq. (2)]. Clearly, DBNfinal and DBNmod have visible improvement over DBNgeo.
Comparison between DBN and leading HMM-type methods
The DBN method (DBNfinal) developed in this work is also evaluated on the widely used CB513 dataset, and its performance is compared to two recently published HMM-type methods, denoted by HMMCrooks [16] and HMMChu [18], respectively, both of which have also been tested on the same or a similar dataset. In comparison, we have calculated the significant-difference margin (denoted by ErrSig) for each score, which is defined as the standard deviation divided by the square root of the number of proteins and was used by others [12]. The results presented in Table 3 show that DBNfinal has made improvements for all measures compared to the two methods mentioned above. Specifically, DBNfinal improves Q3 by 3.5% over HMMCrooks and 4.1% over HMMChu, and improves SOV by 4.4% over HMMChu. Since the ErrSig for Q3 and SOV are 0.41 and 0.63, respectively, the improvements are judged to be significant. Matthews' coefficients [28] shown in Table 3 indicate that DBNfinal is particularly good at the prediction of helices and sheets, compared to above two methods.
Table 3.
Method | Q3 (%) | SOV (%) | CH | CE | CC |
HMMCrooks | 72.8 | -- | -- | -- | -- |
HMMChu | 72.2 | 68.3 | 0.61 | 0.52 | 0.51 |
DBNdiag/ErrSig | 72.5/0.42 | 65.9/0.63 | 0.66/0.01 | 0.55/0.01 | 0.51/0.01 |
DBNfinal/ErrSig | 76.3/0.41 | 72.7/0.63 | 0.71/0.01 | 0.61/0.01 | 0.57/0.01 |
DBNfinal and DBNdiag are methods developed in this work and their descriptions can be found in the text. Entries marked with "--" mean that the data could not be obtained from the literature. HMMChu has been trained and tested on the CB480 dataset (a reduced version of CB513), while all other methods have been trained and tested on the CB513 dataset. The average results of seven-fold cross-validation are shown.
The improvements made by DBNfinal are believed mainly due to the use of a conditional linear Gaussian distribution to model the PSI-BLAST profile of each residue, in which the correlation between the 20 entries in the profile is considered (see Methods). In contrast, both HMMCrooks and HMMChu employ a multinomial distribution to model the profile, which lacks the above correlation information [16,18]. The supporting experiment of our conjecture consists in constructing a degenerate DBN model (denoted by DBNdiag) that has the similar architecture to DBNfinal but only has a diagonal covariance matrix for the distribution of AAi [Eq. (7)], so that the correlation between entries of the profile is ignored. We have tested this model on the CB513 dataset, and the results (Table 3) show that the Q3 of DBNdiag drops down to 72.5%, similar to those of HMMCrooks and HMMChu, which highlights the importance of the non-diagonal entries in the covariance matrix.
Comparison between DBNN and other popular methods
CB513 dataset
The best models developed in this work, DBNN, is then tested on the CB513 dataset and compared to other popular methods. Specifically, the methods SVM [8], PMSVM [11], SVMpsi [9], JNET [7], SPINE [6], and YASSPP [12] are selected for comparison, because they have been tested on the same (or a similar) dataset. Table 4 shows that DBNN has the best Q3 accuracy among all the methods mentioned above, with improvements ranging from 0.3% to 4.6%. Since the ErrSig is 0.41/0.40, this indicates that for all methods except YASSPP, the improvement made by DBNN is significant. In SOV measure, DBNN ranks second, below YASSPP but above SVMpsi. The comparison of the Matthews' coefficients between DBNN and YASSPP indicates that the two methods are complementary and may be combined to obtain further improvement in the prediction accuracy: DBNN has a better CH while YASSPP has a better CC.
Table 4.
Method | Q3 (%) | SOV (%) | CH | CE | CC |
SVM | 73.5 | -- | 0.65 | 0.53 | 0.54 |
PMSVM | 75.2 | -- | 0.71 | 0.61 | 0.61 |
SVMpsi | 76.6 | 73.5 | 0.68 | 0.60 | 0.56 |
JNET | 76.9 | -- | -- | -- | -- |
YASSPP | 77.8 | 75.1 | 0.58 | 0.64 | 0.71 |
†SPINE | 76.8 | -- | -- | -- | -- |
DBNN/ErrSig | 78.1/0.41 | 74.0/0.62 | 0.74/0.01 | 0.64/0.01 | 0.60/0.01 |
†DBNN/ErrSig | 78.0/0.40 | 74.0/0.62 | 0.74/0.01 | 0.64/0.01 | 0.60/0.01 |
The description of DBNN can be found in Methods. Entries marked with "--" mean that the data could not be obtained from literatures. JNET has been trained and tested on the CB480 dataset (a reduced version of CB513), while all other methods have been trained and tested on the CB513 dataset. Methods marked with "†" have been evaluated using ten-fold cross-validation, while others have been evaluated using seven-fold cross-validation.
EVA dataset
DBNN is also compared to some live prediction servers by using the EVAc6 dataset and EVA website. The methods selected to compare are: Prospect [29], PROF_king [30], SAM-T99 [31], PSIPRED [3], PROFsec (unpublished), and PHDpsi [32], and their evaluation results on EVAc6 are obtained directly from the EVA website [33]. Because not all sequences are tested against all methods, the EVAc6 dataset is rearranged into five subsets, and the comparison is made between methods that are tested on the same subset (see Table 5).
Table 5.
Method | Q3 (%) | SOV (%) | CH | CE | CC |
Subset 1 (80 chains) | |||||
Prospect | 71.1 | 68.7 | 0.59 | 0.69 | 0.49 |
DBNN/ErrSig | 78.8/1.34 | 74.8/1.74 | 0.72/0.03 | 0.64/0.04 | 0.62/0.02 |
Subset 2 (175 chains) | |||||
PROF_king | 71.7 | 66.9 | 0.62 | 0.68 | 0.49 |
DBNN/ErrSig | 77.3/0.86 | 71.9/1.27 | 0.71/0.02 | 0.64/0.03 | 0.57/0.02 |
Subset 3 (179 chains) | |||||
SAM-T99 | 77.1 | 74.4 | 0.66 | 0.68 | 0.53 |
DBNN/ErrSig | 77.3/0.86 | 71.9/1.28 | 0.71/0.02 | 0.64/0.02 | 0.57/0.02 |
Subset 4 (212 chains) | |||||
PSIPRED | 77.8 | 75.4 | 0.69 | 0.74 | 0.56 |
PROFsec | 76.7 | 74.8 | 0.68 | 0.72 | 0.56 |
PHDpsi | 75.0 | 70.9 | 0.66 | 0.69 | 0.53 |
DBNN/ErrSig | 77.8/0.79 | 72.4/1.16 | 0.71/0.02 | 0.65/0.02 | 0.58/0.01 |
Subset 5 (73 chains) | |||||
SAM-T99 | 76.3 | 72.9 | 0.71 | 0.64 | 0.56 |
PSIPRED | 75.8 | 72.1 | 0.70 | 0.64 | 0.57 |
PROFsec | 75.3 | 73.0 | 0.68 | 0.61 | 0.54 |
PHDpsi | 73.3 | 69.2 | 0.66 | 0.56 | 0.52 |
PROF_king | 70.7 | 64.9 | 0.63 | 0.57 | 0.50 |
DBNN/ErrSig | 76.4/1.48 | 72.4/2.06 | 0.73/0.04 | 0.67/0.04 | 0.59/0.03 |
CM1/ErrSig | 77.2/1.14 | 73.2/1.87 | 0.73/0.04 | 0.66/0.04 | 0.58/0.02 |
CM2/ErrSig | 77.7/1.17 | 73.4/1.78 | 0.74/0.04 | 0.67/0.04 | 0.60/0.02 |
CM3/ErrSig | 78.1/1.17 | 74.4/1.76 | 0.75/0.04 | 0.67/0.04 | 0.60/0.02 |
DBNN and the three consensus methods (CM1, CM2, and CM3) developed in this work are compared with other leading methods on five subsets of EVAc6; each comparison is carried out with maximum number of common sequences. The results of the six existing methods, Prospect, PROF_king, SAM-T99, PROFsec, PHDpsi, and PSIPRED, are obtained directly from the EVA website.
Table 5 shows that DBNN has generally a better Q3 than all other existing methods. In addition, the ErrSigs indicate that, for Prospect, PROF_king, and PHDpsi, the improvement made by DBNN is significant. In SOV, however, DBNN is modest: it is better than Prospect, PROF_king, and PHDpsi, but less well than SAM-T99, PROFsec, and PSIPRED, as shown in Table 5. Note that DBNN has the best CH among all the methods.
The t-tests are also performed for rigorous pairwise comparison between different methods. Specifically, we test the hypothesis that "method X" gives a significantly higher mean score than "method Y", by calculated t-values as , where d = (x-y); x is the accuracy score of "method X", and y is of "method Y"; , and n = the number of proteins. We have evaluated all the methods on the subset 5 of EVAc6 (containing 73 chains), of which the prediction data of existing methods can be obtained directly from EVA website (Prospect is removed from the comparison because of the too many missing data for this method). The results shown in Table 6 indicate that DBNN has significantly better prediction, in both Q3 and SOV, than PROF_king and PHDpsi, and has competitive performance to the three state-of-the-arts: PSIPRED, SAM-T99, and PROFsec.
Table 6.
Method Y | |||||||||
Method X | PROF_king | SAM-T99 | PSIPRED | PROFsec | PHDpsi | DBNN | CM1 | CM2 | CM3 |
Q3: | |||||||||
PROF_king | -- | -4.70 | -3.99 | -3.56 | -1.88 | -4.52 | -6.19 | -6.93 | -6.88 |
SAM_T99 | 4.70 | -- | 0.50 | 0.93 | 2.45 | -0.16 | -1.41 | -2.09 | -3.02 |
PSIPRED | 3.99 | -0.50 | -- | 0.53 | 2.18 | -0.63 | -2.01 | -2.62 | -3.38 |
PROFsec | 3.56 | -0.93 | -0.53 | -- | 2.31 | -0.94 | -2.87 | -3.22 | -3.72 |
PHDpsi | 1.88 | -2.45 | -2.18 | -2.31 | -- | -2.48 | -4.55 | -5.11 | -5.10 |
DBNN | 4.52 | 0.16 | 0.63 | 0.94 | 2.48 | -- | -0.91 | -1.61 | -2.50 |
CM1 | 6.19 | 1.41 | 2.01 | 2.87 | 4.55 | 0.91 | -- | -1.65 | -2.82 |
CM2 | 6.93 | 2.09 | 2.62 | 3.22 | 5.11 | 1.61 | 1.65 | -- | -1.48 |
CM3 | 6.88 | 3.02 | 3.38 | 3.72 | 5.10 | 2.50 | 2.82 | 1.48 | -- |
SOV: | |||||||||
PROF_king | -- | -4.05 | -3.89 | -3.80 | -1.99 | -3.69 | -5.30 | -5.66 | -5.86 |
SAM_T99 | 4.05 | -- | 0.54 | -0.06 | 2.43 | 0.36 | -0.20 | -0.35 | -1.21 |
PSIPRED | 3.89 | -0.54 | -- | -0.62 | 1.77 | -0.19 | -0.97 | -1.22 | -2.57 |
PROFsec | 3.80 | 0.06 | 0.62 | -- | 2.93 | 0.37 | -0.15 | -0.28 | -1.12 |
PHDpsi | 1.99 | -2.43 | -1.77 | -2.93 | -- | -1.67 | -3.30 | -3.30 | -3.82 |
DBNN | 3.69 | -0.36 | 0.19 | -0.37 | 1.67 | -- | -0.58 | -0.83 | -1.83 |
CM1 | 5.30 | 0.20 | 0.97 | 0.15 | 3.30 | 0.58 | -- | -0.27 | -2.03 |
CM2 | 5.66 | 0.35 | 1.22 | 0.28 | 3.30 | 0.83 | 0.27 | -- | -2.55 |
CM3 | 5.86 | 1.21 | 2.57 | 1.12 | 3.82 | 1.83 | 2.03 | 2.55 | -- |
The t-values are calculated for the differences in accuracy scores between "method X" and "method Y" (x-y) tested on EVAc6 subset 5. The descriptions of DBNN, CM1, CM2 and CM3 can be found in the text. Underlined are where calculated t > tabulated t (significant). The tabulated t = 1.67 for α = 0.05 and degree of freedom = 72.
All the above evaluation work shows that prediction accuracy of protein secondary structure by any individual program seems to reach a limit, no better Q3 than 78% (see Table 5). Previous studies [21,34] show that a simple way to achieve further improvement is to construct a consensus over several independent predictors. The consensus would be effective if the individual predictors are mutually complementary (more independent). So, the study of consensus performance is also a way to judge if a new method or program brings in new (complementary) information. This study is carried out with a design of three consensus methods (CM) using a simple "weighted vote" strategy to generate the final output: CM1 combines the five existing popular methods, PROF_king, SAM-T99, PSIPRED, PROFsec, and PHDpsi; CM2 repeatedly replaces one of the above five methods by DBNfinal, and CM3 is the same as CM2 except DBNN is in the place of DBNfinal. The weight for the vote of each method is set to be the success rate of the method for each type of secondary structure, which is derived from an individual evaluation of its own. The CM-series are evaluated on the subset 5 of EVAc6. The results shown in Table 5 indicate that CM3 has the top performance and that DBNN brings in complementary information to the family of existing methods. Note that CM2 ranks second (better than CM1 in both Q3 and SOV), indicating that the success of DBNN is derived from DBN.
The t-tests between the CM-series and the individual methods are also performed, and the results shown in Table 6 indicate that a simple combination of the five existing methods does not make significant improvement in accuracy: the individual method SAM-T99 has competitive Q3 to CM1. On the other hand, the inclusion of DBN or DBNN (both CM2 and CM3) has given rise to significantly better Q3 than all individual methods including SAM-T99. This is further enhanced by a direct comparison between CM3 and CM1; significant improvements in both Q3 and SOV are clearly evidenced. Finally, let us note that none of the consensus methods shows significant improvement in SOV over all individual methods, indicating that SOV is particularly hard to improve.
Conclusion
A new method for protein secondary structure prediction of probabilistic nature based on dynamic Bayesian networks is developed and evaluated by several measures, which has shown significantly better prediction accuracy than previous pure HMM-type methods such as HMMCrooks and HMMChu. The improvement is mainly due to the use of a multivariate Gaussian distribution for the PSI-BLAST profile of each residue and the consideration of dependency between profiles of neighboring residues. In addition, because of the introduction of secondary structure segment length distributions in the model, DBN shows much better SOV than a typical NN.
The essentially different nature of DBN and NN inspires a model that combines the two and forms the DBNN with significant further improvements in both Q3 and SOV. DBNN is shown to be better than most of popular methods and competitive compared to the three state-of-the-art programs. We are then encouraged to explore further with consensus methods that combine all the best existing methods together. This study has demonstrated again the uniqueness of DBNN: the best consensus method is achieved by the inclusion of DBNN. This provides the evidence that DBNN brings in complementary information to the family of existing methods.
An interesting feature of our work here, compared to NN or SVM, is that it provides a set of distributions which have specific meanings and which can be studied further to improve our understanding of the model's behavior behind the prediction. An example is provided regarding the secondary structure segment length distributions used by the DBN, which is set to be an a priori distribution but can further be adjusted and improved. This points to a way for further improving the performance of DBN, by including modifications on more distributions, such as the transition probabilities between secondary structure states or the distribution of the profile of each residue. These distributions are also interesting for advancing the understanding of such fundamental problems as protein dynamics and protein folding, for which the information in implicit form in NN or SVM is of little use.
It appears that the limits of secondary structure prediction are being reached as no new method over the past decade has shown any major improvement since PSIPRED. All of the top methods are between 77%–80% accurate, in terms of Q3, depending on data set used. This implies that the complexity of the sequence-structure relationship is such that any single tool, when it attempts to extract (during learning) and to extrapolate (during predicting) the knowledge, can only represent some facets of this relationship, but not the whole. Further hope lies in the possibility that more facets are covered by new models, and that new models are integrated with the existing ones. The consensus methods reported above are just a simple approach in that direction; more sophisticated strategy for combining multiple scores can be sought in the future.
Methods
Generation of the PSI-BLAST profile
Each protein sequence in the datasets described above is used as query to search against the NR database [35] by using PSI-BLAST program [36]. The number of iterations in running PSI-BLAST is set to be 3; all other options are set to be defaults. The PSSM produced by the program is a matrix of integers typically in the range of ± 7 (see Fig. 2(a)). Each row of the PSSM is a 20-dimension vector corresponding to 20 amino acids, which is used to derive the PSI-BLAST profile of the corresponding residue.
Transformation of the PSSM
Similar to other secondary structure prediction methods [3,6,11], we transform the PSSM into the range from 0 to 1 before using it as input of models. Two strategies are employed for the transformation: one follows the function
(3) |
and is referred to as "linear transformation"; the other follows the function
(4) |
and is referred to as "sigmoid transformation".
Assessment of the prediction accuracy
Several measures are adopted to assess the performance of our methods in a comprehensive way. The first is the overall three-state prediction accuracy, Q3, defined by
(5) |
where n is the number of correctly predicted residues and N is the total number of residues. The second, SOV, is a segment-level measure of the prediction accuracy, and its most recent definition can be found in [37]. At last, the Matthews' correlation coefficient [28] is used for each class of secondary structure, which is defined by
(6) |
where ni is the number of residues correctly predicted to be secondary structure of class i, mi is the number of residues correctly not predicted to be secondary structure of class i, ui is the number of residues observed but not predicted to be secondary structure of class i, and oi is the number of residues predicted but not observed to be secondary structure of class i (i = H, E, and C).
The dynamic Bayesian network
DBN is a directed graphical model in which nodes represent random variables and arcs represent dependency between nodes. The architecture of our DBN model is illustrated in Fig. 2(b). There are totally six nodes for each residue. Specifically, the node AAi (i = 1, 2, 3...) contains the PSI-BLAST profile of residue i, which is a 20-dimensional vector corresponding to 20 scores in the PSSM. The node Ri stores replica of the profiles of a series of residues before i, i.e. the profiles of residues i-1, i-2, i-3, ... i-LAA, as shown in Fig. 2(b), where LAA is a profile window size indicating the range of the dependency for the profiles. As shown in Fig. 2(b), all the dependency between AAi and its neighboring sites, AAi-1, AAi-2, ... AAi-LAA, can be summarized into one single connection to Ri, simplifying the topology of the graph. The state-space of Ri is 21·LAA-dimensional, with 20·LAA storing the profiles of the past residues and extra LAA dimensions representing the "over-terminus" state.
The node SSi is used to describe the secondary structure state of residue i, which has a discrete state-space of three elements: H, E, and C. The node di has a similar role as Ri, but describes here the joint distribution with the secondary structure states of residues i-1, i-2, ... i-LSS, where LSS is the secondary structure window size indicating the range of the dependency, as shown in Fig. 2(b). Again, the node di is introduced to simplify the topology of the graph, yet to keep a long-range dependency between profile (AAi) and secondary structure (SSi-1, SSi-2, ...). The dimension of di is 4·LSS, where 3·LSS are from the joint past secondary structure states and the extra LSS from the "over-terminus" situation.
The nodes Di and Fi are introduced to mimic a duration-HMM [22], with a specified parameter Dmax and two elements, respectively. Specifically, Di represents the distance (measured by the number of residues) from the position i to the end of the corresponding secondary structure segment. For example, in a segment with end residue at position j, the value of Di is set to be j-i+1. Note that the state-space of Di requires that the maximum length of segments should not exceed Dmax. In order to cope with longer segments, a modified definition of Di is introduced as following: when the length of a segment ≤ Dmax, the value of Di is set as described above; when the length of the segment > Dmax, for example Dmax+3, the Di is set to be Dmax for the first four residues of the segment and is set to be Dmax-1, Dmax-2, ... 1 for the rest. In this way, the lengths of segments longer than Dmax are modeled by a geometric distribution (see below). The value of the node Fi is deterministically dependent on Di: if Di > 1, Fi = 1; if Di = 1, Fi = 2.
Each node described above is assigned a specific conditional probability distribution (CPD) function according to the connections' pattern shown in Fig. 2(b), except for Ri, which is a "root" node [22] with no "parent node", and which is observable in both training and predicting. Specifically, the CPD of AAi (i = 1, 2, 3) is modeled using a conditional linear Gaussian function, which is defined by:
P(AAi = y | Ri = u, SSi = α, di = γ) = N(y;wα,γu + cα,γ, Σα,γ), | (7) |
where N(y;μ, Σ) represents a Gaussian distribution with mean μ and covariance Σ, u is a 21·LAA-dimensional vector, α is one of H, E, and C, and γ is one of the LSS-tuples formed by four elements: O, H, E, and C (O represents the "over-terminus" state). The distribution function is characterized by the mean μα,γ = wα,γu + cα,γ, where wα,γ is a 20 × 21 LAA matrix and cα,γ is a 20-dimensional vector, and the covariance Σα,γ. The subscripts α and γ indicate that the parameters wα,γ, cα,γ, and Σα,γ are dependent on the states of SSi and di. Second, the CPD of SSi (i = 2, 3, 4...) is defined by
(8) |
where Tα(β) is the transition probability from the secondary structure state α to the state β. Third, the CPD of di (i = 2, 3, 4...) is defined by
(9) |
where λj and γj (j = 1, 2, ... LSS) are the jth elements of the LSS-tuples λ and γ, respectively. Fourth, the CPD of Di (i = 2, 3, 4...) is defined by
(10) |
where gα(n) is the segment length distribution given the secondary structure state α and hα is the probability for Di to maintain the value Dmax given SSi = α and Di-1 = Dmax. Using this function, the probability of producing a segment with length n (n > = Dmax) is proportional to (1-hα)hαn-Dmax, i.e. a geometric distribution. The validity of using such a distribution to model segments of length longer than Dmax is supported by Fig. 3(a), in which all the helices, sheets, and coils show exponential tails in their segment length distributions. Fig. 3(a) also indicates that a proper Dmax should be 13, after which all the distributions can be fitted well to exponential functions (see the inset of Fig. 3(a)). At last, the CPD of Fi (i = 1, 2, 3...) is defined by
(11) |
Note that the CPDs of SS1, d1, and D1 have similar definition to CPDs of SSi, di, and Di (i = 2, 3, 4...) but with an independent set of parameters.
The parameters of the CPDs described above are derived by applying the maximum likelihood (ML) method to the training set. In prediction, the marginal probability distribution of SSi (i = 1, 2, 3...) is computed by using the forward-backward (FB) algorithm [22], and then the state of SSi with the maximum probability is the prediction of residue i. Both ML and FB algorithms are implemented by using the Bayes Net Toolbox [38].
The neural network
The typical three-layered feed-forward back-propagation architecture is used in our NN-based models. The sliding window-based training and testing strategy are employed with an optimal window size of 15 derived from an empirical evaluation of varying window sizes from 7 to 19. The momentum terms and learning rates of the network are set to be 0.9 and 0.005, respectively, and the number of hidden units is set to be 75.
Training and combinations
Training is done in two different ways, depending on datasets involved. For the dataset CB513 and SD576, the standard N-fold cross-validation testing strategy is adopted, where N is either 7 or 10. That is, the dataset is split into N subsets with approximately equal numbers of sequences in each, and then N-1 of them are used for training while the remaining one is used for testing; the process continues N times with a rotation of the testing subset, making sure that every protein sequence is tested once. The second way of training concerns the dataset EVAc6, for which there exists a separate large dataset EVAtrain with low sequence identity (< 25%) to EVAc6. So, it is customary to use EVAtrain as the training set and EVAc6 as the test set.
Note that the DBN and NN models are usually trained on the same training set, in order to make a comparison and to be combined later to form DBNN. However, the detailed training process of DBN is somewhat different from NN, owing to different architectures of the model. The DBN takes two sets of data as input, one for profile and the other for secondary structure; each set is a sliding window with the "current" residue located at the right end. The correlation information between "current" residue and its neighbors is stored in the data, but depends on the direction in which the window slides (from N-terminus to C-terminus or reverse). We actually run the DBN model in both directions and then average the results (see below). On the other hand, the NN takes only one sliding-window, with the "current" residue located at the center of the window. Finally, the training for DBNN is simple the training of DBN and NN on the same dataset.
When a sequence is selected for either training or testing, the original PSSM generated by PSI-BLAST can be transformed into [0 1] in two strategies: linear transformation [Eq. (3)] or sigmoid transformation [Eq. (4)]. In addition, as mentioned above, the direction from either N-terminus to C-terminus (NC) or the reverse (CN) gives rise to different correlation structure, so we treat them separately. As a result, four basic DBN models are generated corresponding to four above combinations: (i) DBNlinear+NC, (ii) DBNlinear+CN, (iii) DBNsigmoid+NC, and (iv) DBNsigmoid+CN, where the subscripts are self-explanatory. On the other hand, NN is split into two kinds according to the transformation for PSSM, and the corresponding models are denoted by NNlinear and NNsigmoid, respectively.
The six basic models described above are believed to contain complementary information and need to be combined to form three final models. Two strategies for forming the final models are used. The first is a simple averaging of the output scores and is used to form the two architecture-based final models, DBNfinal and NNfinal. It is done in two steps. One first averages the outputs of DBNlinear+NC and DBNlinear+CN to form DBNlinear, and of DBNsigmoid+NC and DBNsigmoid+CN to form DBNsigmoid. Then, DBNlinear and DBNsigmoid are further combined to form DBNfinal. Similarly, NNlinear and NNsigmoid are combined to form NNfinal.
The second strategy consists in using a new neural network, which has the same architecture to basic NN models except that it takes as inputs, the outputs of all the other scores (DBNlinear+NC, DBNlinear+CN, DBNsigmoid+NC, DBNsigmoid+CN, NNlinear, and NNsigmoid). This final model is named DBNN, and is the one that shows the best performance among the models mentioned above.
Availability
All the codes and datasets described above are available from our homepage [39].
Authors' contributions
ZSS and HQZ supervised the whole process of the work. XQY wrote the codes and did the tests. XQY, HQZ, and ZSS draft the manuscript.
Acknowledgments
Acknowledgements
We acknowledge the support by the National Natural Science Foundation of China (No. 10225210 and No. 30300071), and the National Basic Research Program of China (973 Program) under grant No. 2003CB715905.
Contributor Information
Xin-Qiu Yao, Email: yxq@ctb.pku.edu.cn.
Huaiqiu Zhu, Email: hqzhu@pku.edu.cn.
Zhen-Su She, Email: she@pku.edu.cn.
References
- Qian N, Sejnowski TJ. Predicting the secondary structure of globular proteins using neural network models. J Mol Biol. 1988;202:865–884. doi: 10.1016/0022-2836(88)90564-5. [DOI] [PubMed] [Google Scholar]
- Rost B, Sander C. Prediction of protein secondary structure at better than 70% accuracy. J Mol Biol. 1993;232:584–599. doi: 10.1006/jmbi.1993.1413. [DOI] [PubMed] [Google Scholar]
- Jones DT. Protein secondary structure prediction based on position-specific scoring matrices. J Mol Biol. 1999;292:195–202. doi: 10.1006/jmbi.1999.3091. [DOI] [PubMed] [Google Scholar]
- Pollastri G, McLysaght A. Porter: a new, accurate server for protein secondary structure prediciton. Bioinformatics. 2005;21:1719–1720. doi: 10.1093/bioinformatics/bti203. [DOI] [PubMed] [Google Scholar]
- Adamczak R, Porollo A, Meller J. Combining prediction of secondary structure and solvent accessiblility in proteins. Proteins. 2005;59:467–475. doi: 10.1002/prot.20441. [DOI] [PubMed] [Google Scholar]
- Dor O, Zhou Y. Achieving 80% ten-fold cross-validated accuracy for secondary structure prediction by large-scale training. Proteins. 2007;66:838–845. doi: 10.1002/prot.21298. [DOI] [PubMed] [Google Scholar]
- Cuff JA, Barton GJ. Application of multiple sequence alignment profiles to improve protein secondary structure prediction. Proteins. 2000;40:502–511. doi: 10.1002/1097-0134(20000815)40:3<502::AID-PROT170>3.0.CO;2-Q. [DOI] [PubMed] [Google Scholar]
- Hua S, Sun Z. A novel method of protein secondary structure prediction with high segment overlap measure: support vector machine approach. J Mol Biol. 2001;308:397–407. doi: 10.1006/jmbi.2001.4580. [DOI] [PubMed] [Google Scholar]
- Kim H, Park H. Protein secondary structure prediction based on an improved support vector machines approach. Protein Eng. 2003;16:553–560. doi: 10.1093/protein/gzg072. [DOI] [PubMed] [Google Scholar]
- Ward JJ, McGuffin LJ, Buxton BF, Jones DT. Secondary structure prediction with support vector machines. Bioinformatics. 2003;19:1650–1655. doi: 10.1093/bioinformatics/btg223. [DOI] [PubMed] [Google Scholar]
- Guo J, Chen H, Sun Z, Lin Y. A novel method for protein secondary structure prediction using dual-layer SVM and profiles. Proteins. 2004;54:738–743. doi: 10.1002/prot.10634. [DOI] [PubMed] [Google Scholar]
- Karypis G. YASSPP: better kernels and coding schemes lead to improvements in protein secondary structure prediction. Proteins. 2006;64:575–586. doi: 10.1002/prot.21036. [DOI] [PubMed] [Google Scholar]
- Stultz CM, White JV, Smith TF. Structural analysis based on state-space modeling. Protein Sci. 1993;2:305–314. doi: 10.1002/pro.5560020302. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Thompson MJ, Goldstein RA. Predicting protein secondary structure with probabilistic schemata of evolutionarily derived information. Protein Sci. 1997;6:1963–1975. doi: 10.1002/pro.5560060917. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Schmidler SC, Liu JS, Brutlag DL. Bayesian segmentation of protein secondary structure. J Comput Biol. 2000;7:233–248. doi: 10.1089/10665270050081496. [DOI] [PubMed] [Google Scholar]
- Crooks GE, Brenner SE. Protein secondary structure: entropy, correlations and prediction. Bioinformatics. 2004;20:1603–1611. doi: 10.1093/bioinformatics/bth132. [DOI] [PubMed] [Google Scholar]
- Aydin Z, Altunbasak Y, Borodovsky M. Protein secondary structure prediction for a single-sequence using hidden semi-Markov models. BMC Bioinformatics. 2006;7:178–192. doi: 10.1186/1471-2105-7-178. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Chu W, Ghahramani Z, Podtelezhnikov A, Wild DL. Bayesian segmental models with multiple sequence alignment profiles for protein secondary structure and contact map prediction. IEEE Trans Comput Biol Bioinfo. 2006;3:98–113. doi: 10.1109/TCBB.2006.17. [DOI] [PubMed] [Google Scholar]
- Karplus K, Katzman S, Shackleford G, Koeva M, Draper J, Barnes B, Soriano M, Hughey R. SAM-T04: what is new in protein-structure prediction for CASP6. Proteins. 2005;Suppl 7:135–142. doi: 10.1002/prot.20730. [DOI] [PubMed] [Google Scholar]
- Prediction scheme of SAM-T04 http://www.soe.ucsc.edu/research/compbio/SAM_T06/faq.html
- Cuff JA, Barton GJ. Evaluation and improvement of multiple sequence methods for protein secondary structure prediction. Proteins. 1999;34:508–519. doi: 10.1002/(SICI)1097-0134(19990301)34:4<508::AID-PROT10>3.0.CO;2-4. [DOI] [PubMed] [Google Scholar]
- Murphy KB. Computer Science. PHD. U. C. Berkeley; 2002. Dynamic Bayesian networks: representation, inference and learning; p. 225. [Google Scholar]
- Koh IY, Eyrich VA, Marti-Renom MA, Przybylski D, Madhusudhan MS, Eswar N, Grana O, Pazos F, Valencia A, Sali A, Rost B. EVA: evaluation of protein structure prediction servers. Nucleic Acids Res. 2003;31:3311–3315. doi: 10.1093/nar/gkg619. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Berman HM, Westbrook J, Feng Z, Gilliland G, Bhat TN, Weissig H, Shindyalov IN, Bourne PE. The protein data bank. Nucleic Acids Res. 2000;28:235–242. doi: 10.1093/nar/28.1.235. [DOI] [PMC free article] [PubMed] [Google Scholar]
- EVA common set 6 http://cubic.bioc.columbia.edu/eva/sec/set_com6.html
- Andreeva A, Howorth D, Brenner SE, Hubbard TJP, Chothia C, Murzin AG. SCOP database in 2004: refinements integrate structure and sequence family data. Nucleic Acids Res. 2004;32:D226–D229. doi: 10.1093/nar/gkh039. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kabsch W, Sander C. Dictionary of protein secondary structure: pattern recognition of hydrogen-bonded and geometrical features. Biopolymers. 1983;22:2577–2637. doi: 10.1002/bip.360221211. [DOI] [PubMed] [Google Scholar]
- Matthews BW. Comparison of the predicted and observed secondary structure of T4 phage lysozyme. Biochim Biophys Acta. 1975;405:442–451. doi: 10.1016/0005-2795(75)90109-9. [DOI] [PubMed] [Google Scholar]
- Xu Y, Xu D. Protein threading using PROSPECT: design and evaluation. Proteins. 2000;40:343–354. doi: 10.1002/1097-0134(20000815)40:3<343::AID-PROT10>3.0.CO;2-S. [DOI] [PubMed] [Google Scholar]
- Ouali M, King RD. Cascaded multiple classifiers for secondary structure prediction. Protein Sci. 2000;9:1162–1176. doi: 10.1110/ps.9.6.1162. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Karplus K, Barrett C, Cline M, Diekhans M, Grate L, Hughey R. Predicting protein structure using only sequence information. Proteins. 1999;Suppl. 3:121–125. doi: 10.1002/(SICI)1097-0134(1999)37:3+<121::AID-PROT16>3.0.CO;2-Q. [DOI] [PubMed] [Google Scholar]
- Przybylski D, Rost B. Alignments grow, secondary structure prediction improves. Proteins. 2002;46:197–205. doi: 10.1002/prot.10029. [DOI] [PubMed] [Google Scholar]
- EVA results http://cubic.bioc.columbia.edu/eva/sec/common3.html
- McGuffin LJ, Jones DT. Benchmarking secondary structure prediction for fold recognition. Proteins. 2003;52:166–175. doi: 10.1002/prot.10408. [DOI] [PubMed] [Google Scholar]
- NR database ftp://ftp.ncbi.nih.gov/blast/db
- Altschul SF, Madden TL, Schaffer AA, Zhang J, Zhang Z, Miller W, Lipman DJ. Gapped BLAST and PSI-BLAST: a new generation of protein database search programs. Nucleic Acids Res. 1997;25:3389–3402. doi: 10.1093/nar/25.17.3389. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Zemla A, Venclovas C, Fidelis K, Rost B. A modified definition of sov, a segment-based measure for protein secondary structure prediction assessment. Proteins. 1999;34:220–223. doi: 10.1002/(SICI)1097-0134(19990201)34:2<220::AID-PROT7>3.0.CO;2-K. [DOI] [PubMed] [Google Scholar]
- Bayes net toolbox http://bnt.sourceforge.net
- DBNN homepage http://ctb.pku.edu.cn/main/SheGroup/Software/DBNN