Skip to main content
Bioinformatics logoLink to Bioinformatics
. 2019 Jan 11;35(17):2957–2965. doi: 10.1093/bioinformatics/btz016

MULTiPly: a novel multi-layer predictor for discovering general and specific types of promoters

Meng Zhang 1,1, Fuyi Li 2,3,1, Tatiana T Marquez-Lago 4,5, André Leier 4,5, Cunshuo Fan 6, Chee Keong Kwoh 7, Kuo-Chen Chou 8,, Jiangning Song 2,3,9,, Cangzhi Jia 1,6,
Editor: John Hancock
PMCID: PMC6736106  PMID: 30649179

Abstract

Motivation

Promoters are short DNA consensus sequences that are localized proximal to the transcription start sites of genes, allowing transcription initiation of particular genes. However, the precise prediction of promoters remains a challenging task because individual promoters often differ from the consensus at one or more positions.

Results

In this study, we present a new multi-layer computational approach, called MULTiPly, for recognizing promoters and their specific types. MULTiPly took into account the sequences themselves, including both local information such as k-tuple nucleotide composition, dinucleotide-based auto covariance and global information of the entire samples based on bi-profile Bayes and k-nearest neighbour feature encodings. Specifically, the F-score feature selection method was applied to identify the best unique type of feature prediction results, in combination with other types of features that were subsequently added to further improve the prediction performance of MULTiPly. Benchmarking experiments on the benchmark dataset and comparisons with five state-of-the-art tools show that MULTiPly can achieve a better prediction performance on 5-fold cross-validation and jackknife tests. Moreover, the superiority of MULTiPly was also validated on a newly constructed independent test dataset. MULTiPly is expected to be used as a useful tool that will facilitate the discovery of both general and specific types of promoters in the post-genomic era.

Availability and implementation

The MULTiPly webserver and curated datasets are freely available at http://flagshipnt.erc.monash.edu/MULTiPly/.

Supplementary information

Supplementary data are available at Bioinformatics online.

1 Introduction

The first and most critical step of gene expression is the initiation of transcription, requiring a dynamic cooperation between the RNA polymerase (RNAP) and the promoter (Ramprakash and Schwarz, 2008). Promoters are chromosome regions that facilitate the transcription of particular genes, and they are located proximal to the transcription start sites of genes, towards the 5′ region of the sense strand. In bacteria, the promoter is recognized by the RNA polymerase and correlated function-specific sigma factors that are labelled on the basis of their molecular weights (σ24,σ28,σ32,σ38,σ54 and σ70), which in turn are often brought to the promoter by regulatory proteins that bind to specific sites nearby (Barrios et al., 1999; Helmann and Chamberlin, 1988; Towsey et al., 2008). The types of promoters are defined according to how the σ factors identify the promoter.

The precise recognition of promoters is crucial to regulation of the expression of each gene and each transcription unit in the genome. However, the precise prediction of promoters remains a challenging task, because individual promoters usually differ from the consensus at one or even more positions (Mrozek et al., 2014, 2016).

In recent years, a number of computational methods have been developed to rapidly differentiate DNA sequences as promoters or non-promoters, aimed at complementing with experimental efforts and overcoming certain experimental bottlenecks. For instance, position weight matrices (PWMs) were used to predict σ70 promoters in Escherichia coli, based on the conservation of the -10 and -35 hexamers (with the consensus sequences ‘TATAAT’ and ‘TTGACA’, respectively) and the distribution of promoters from the start of the gene (Hertz and Stormo, 1996; Huerta and Collado-Vides, 2003); however, the latter approach achieved a relatively lower accuracy. In 2009, Kemal, a new method that integrated feature selection and a fuzzy-AIRS classifier system to predict E.coli promoter gene sequences was proposed (Polat and Güneş, 2009). More recently, with machine learning techniques booming, many promoter prediction tools have been developed and made available, including 70ProPred, iPro54-PseKNC, iPromoter-2L and bTSSfinder (He et al., 2018; Liang et al., 2017; Lin et al., 2014, 2017; Liu et al., 2018; Shahmuradov et al., 2017). We note that, amongst previously developed tools, only iPromoter-2L is able to predict whether a query sequence sample is a promoter or not (Task 1), and identify which specific promoter type it would belong to if it is identified as a promoter (Task 2). iPromoter-2L reached an overall accuracy of 81.68% for identifying promoters and non-promoters on the 5-fold cross-validation test. However, with respect to the prediction of specific promoter types, except for the identification of the σ24 promoter, the performance results on other types of promoters were not entirely satisfactory. For σ28, σ32, σ38 and σ54 promoters, iPromoter-2L achieved a specificity (Sp) of higher than 99%, but achieved a much worse sensitivity (Sn) of lower than 54%. In addition, for σ70 promoter prediction, the Sn was 95.34%, while the Sp was only 59.35%. A major reason for the observed large discrepancy might be attributed to the different numbers of the six distinct types of promoters.

To address this complexity and improve the effectiveness of promoter prediction, in this work, we developed MULTiPly, a multi-layer two-task predictor designed to both recognize the promoters and identify their specific types in E.coli. Firstly, both the sequences themselves and the information measures including k-tuple nucleotide composition (KNC), dinucleotide-based auto covariance (DAC) and the global information of the whole samples including bi-profile Bayes (BPB) and k-nearest neighbour feature (KNN), were taken into consideration; subsequently, the F-score feature selection method was applied to identify the optimal feature combination. To overcome the complexity associated with the analysis of varying numbers of samples for six types of known promoters, the method learns to differentiate between one (positive) promoter subset and the joint set of all other promoter subsets with less samples than the positive dataset (negative). We established a total of five binary sub-classifiers in the second task according to the dataset size. In the first sub-classifier, the largest subset S+(σ70) was regarded as the positive class, while the union of the other five types of promoter samples were considered as negative samples to train the classifier for identifying the σ70 promoters. Then, we successively deemed S+(σ24), S+(σ32), S+(σ38) and S+(σ28) as the positive class, and the rest promoters that were not classified jointly as the negative class. Comprehensive benchmarking experiments using 5-fold cross-validation, jackknife test and independent test based on our newly constructed independent test dataset consistently showed the effectiveness of the proposed MULTiPly approach, especially for distinguishing specific types of promoters.

2 Materials and methods

As suggested in a series of recent publications (Chen et al., 2018a,b,c; Cheng et al., 2018a,b; Li et al., 2018a,b; Song et al., 2018a,b,c), we followed the guidelines of Chou’s 5-step rule (Chou, 2011), in an effort to make the presentation of this paper more clear and transparent, enable others to repeat analysis steps, and ensure that the proposed predictor can be easily and widely used by the majority of experimental scientists. The five detailed steps include: (i) construct a valid benchmark dataset and an independent test dataset; (ii) extract the features that can truly reflect their intrinsic correlations with the target to be predicted; (iii) introduce a powerful algorithm (or prediction engine) to operate the prediction; (iv) properly perform cross-validation tests to objectively evaluate the predictor’s accuracy; (v) establish a user-friendly web-server as an implementation of the predictor that is freely accessible to the wider research community. A graphical illustration of the five steps involved in the development of MULTiPly is shown in Figure 1.

Fig. 1.

Fig. 1.

The overall framework of MULTiPly. The five steps are formulated and illustrated according to Chou’s 5-step rule (Chou, 2011)

2.1 Datasets

The models of MULTiPly were trained using a most-recent dataset, constructed in (Liu et al., 2018a,b). All collected promoter samples were experimentally verified (each with 81 bp) and retreived from the RegulonDB database (Version 9.3). RegulonDB (available at http://regulondb.ccg.unam.mx/) is one of the most useful public resources on bacterial gene regulation in the model organism E.coli K-12. If a reported promoter belonged to two sigma types, we accordingly assigned it to the type that had a larger number of experimentally verified sequences. After the raw data processing, the final benchmark dataset S was defined as follows:

S=S+S-S+=S+σ24S+σ28S+σ32S+σ38S+σ54S+σ70  (1)

where S+ denotes the positive dataset containing 2860 promoter sequences, S- denotes the negative dataset containing 2860 non-promoter sequences, while the symbol denotes the ‘union’ in the Set Theory. S+ contains all six types of promoter sequences; specifically, there existed 484 promoter sequences of σ24, 134 of σ28, 291 of σ32, 163 of σ38, 94 of σ54 and 1694 of σ70, respectively. The length of each sequence in our used datasets is 81. As RegulonDB was updated in 18/06/2018, we collected the recently experimentally verified promoter samples from the current version of RegulonDB (Version 10.0) as the independent test dataset, which was denoted as Stest, to test the performance of MULTiPly. Lastly, a total of 54 promoter sequences were collected in Stest, including 46 sequences of σ70, 1 of σ24, 2 of σ32, 4 of σ38 and 1 of σ28.

2.2 Feature extraction strategy

In general, feature extraction refers to the formulation of an effective mathematical expression representing a nucleotide sequence. In this study, features were extracted incorporating both global features (i.e. BPB and KNN features) and local (i.e. KNC and DAC features) features, in order to derive more representative and useful information from promoter and non-promoter samples. BPB features reflect the nucleotide distribution within the whole samples, while KNN features describe whether each sample sequence is more similar to the positive or negative samples. KNC was used to encode the compositions of nucleotides and di-nucleotides in a single DNA sample. DAC measures the correlation between two di-nucleotides which have the similar physicochemical index. The feature extraction procedures are described in the following sections.

2.2.1 Bi-profile bayes (BPB)

BPB has proven useful for improving the prediction performance of machine learning-based models in a number of different bioinformatics studies, such as predicting protein methylation sites (Shao et al., 2009), caspase cleavage sites (Song et al., 2010, 2012a,b; Wang et al., 2014) and strong and weak enhancer (Jia and He, 2016). BPB considers the position-specific information from both positive and negative training samples simultaneously. The latter is the main reason why BPB outperforms other feature encoding schemes in many cases.

Each of the DNA samples S can be expressed as:

S=R1R2R3RiRL (i=1, 2, 3,,L) (2)

where Ri is one of the nucleotides A, G, C and T; i represents a nucleotide position, and L denotes the length of the nucleotide sequence. In this study, L =81, which is the same as that used in previous works (Liu et al., 2018a,b). The sequence S is encoded as a feature vector VBPB=p1,p2,,pL,pL+1,p2L, where pi(i=1, 2, , L) represents the posterior probability of each nucleotide at the ith position in all positive samples, and pi(i=L+1, L+2,, 2L) denotes the posterior probability of each nucleotide at the ith position in all negative samples. When the numbers of positive and negative samples were equal and sufficiently large, the frequency of each nucleotide at each position would be a close approximation to the true probability of the occurrence. Accordingly, the posterior probabilities of the positive and negative samples were calculated as the occurrence frequencies for each nucleotide to appear at each position in the positive and negative training datasets, respectively. The dimension of the BPB feature vector was 162, the 1st–81th features were derived from the overall probability distribution of the positive samples, while the 82th–162th features were derived from the overall probability distribution of the negative samples.

2.2.2 KNN features

In the fields of bioinformatics and computational biology, the KNN features have been successfully applied to the analysis and prediction of protein, DNA and RNA sequences (Chen et al., 2013; Jia et al., 2016, 2018; Li et al., 2018a,b; Wang et al., 2017). By extracting relevant features from similar sequences in both the positive and negative datasets using the KNN algorithm, the KNN scores could capture the local sequence similarity in the promoter and non-promoter samples (Gao et al., 2010).

For two local sequences P1 and P2, the distance DistP1,P2 can be defined as follows:

DistP1,P2=i=1LSim(P1i,P2(i))  (3)

where L represents the number of nucleotides in a DNA sequence (L =81 in this study), while P1i and P2(i) denote the nucleotides at the ith position of sequences P1 and P2, respectively. For two nucleotides a and b, their similarity score is defined as (Jia et al., 2018)

Sima, b=+2, if a=b;-1, otherwise. (4)

For a query DNA sequence (potential promoter or non-promoter sequence), the local sequence similarity would be first considered. Then, the KNN scores were calculated based on the proportion of the positive and negative samples in the set of k neighbours, respectively. The detailed procedures for calculating the KNN scores are described as follows: (i) form a comparison set that contains all the positive and negative samples; (ii) calculate the distances between a query sequence and the other samples in the comparison set; (iii) sort the distances in the ascending order and generate the top k nearest neighbours; (iv) calculate the KNN scores, which is the percentage of the positive neighbours in its k nearest neighbours. To obtain the best features, different values of k (k =10, 20, 30… 200) were assessed in this study. More specifically, if the dimension of KNN features was d (1 ≤ d 20), the numbers of 10, 20, …, 10d neighbours would be successively selected.

2.2.3 k-tuple nucleotide composition (KNC)

The type and position of nucleotides within a DNA sequence contained crucial information. Accordingly, strategies for extracting such information in an effective manner have been extensively researched in a number of previous studies. The KNC can characterize the occurrence frequency and the permutation order of nucleotides in each sequence, and this measure has been widely used in many previous studies (Chen et al., 2015; Ioshikhes et al., 1996; Jia et al., 2013; Kabir and Hayat, 2016; Li et al., 2015a,b). After various trials, the 1-tuple (mononucleotide) and 2-tuple (dinucleotide) compositions (referred to as MNC and DNC, respectively) were determined to construct the feature vector. The MNC feature vector can be formulated as follows:

D=fA, fC, fG, fT' (5)

where fi represents the frequency of occurrence of each nucleotide. The DNC feature vector can be defined as follows:

D=fAA,fAC,fAG,fAT,,fTT' (6)

where fi represents the frequency of occurrence of each dinucleotide i.

2.2.4 Dinucleotide-based auto-covariance (DAC)

DAC measures the correlation between two di-nucleotides separated by a distance (λ) along the sequence with the same physicochemical index (Dong et al., 2009; Guo et al., 2008; Liu et al., 2015, 2017a,b). It can be calculated as:

DACu,λ=i=1L-λ-1(PuRiRi+1-P¯u)(PuRi+λRi+λ+1-P¯u)/(L-λ-1) (7)

where u is a physicochemical index, L is the length of the promoter sequence S, PuRiRi+1 denotes the numerical value of the physicochemical index u for the dinucleotide RiRi+1 at the position i, and P¯u is the average value for the physicochemical index u along the whole sequence, which is defined as:

P¯u=j=1L-1PuRjRj+1/(L-1) (8)

In such a way, the length of the DAC feature vector can be defined as N*Λ, where N is the number of physicochemical indices while Λ is the maximum of λ (λ= 1, 2…Λ). In this study, we selected six physicochemical indices, including Base stacking, Dinucleotide GC content, A-philicity, Rise, Roll and Stability and set the parameter Λ as 2. The feature vector can then be generated using the very powerful, publicly available Pse-in-One web server, documented in the literature (Friedel et al., 2009; Liu et al., 2017a,b).

2.3 Feature optimization

When multiple features are incorporated to train a model, the dimension of the resulting hybrid feature vectors becomes very large. As the initial features might contain redundant and noisy information, we presumed that this could exert a negative effect on model training. Therefore, to filter out the noisy and irrelevant features and select a subset of optimal features, the most important features were identified by a feature selection method known as F-score (Bui, 2016; Lin and Ding, 2011; Zuo and Jia, 2017). The F-score of the jth feature is defined as:

F-scorej=(x-j+-x-j)2+(x-j--x-j)21m+-1k=1m+(x-k,j+-x-j+)2+1m--1k=1m-(x-k,j--x-j-)2 (9)

where x-j, x-j+ and x-j- denote the average values of the jth feature in the combined (i.e. positive and negative), the positive and the negative datasets, respectively. m+ denotes the number of positive samples, m- denotes the number of negative samples, x-k,j+ denotes the jth feature of the kth positive instance, and x-k,j- denotes the jth feature of the kth negative instance. A feature with a larger F-score value indicates that such feature can distinguish well between the positive and negative samples, and thus is regarded as being more useful for classification.

2.4 Model training

Support vector machine (SVM) is a powerful and popular supervised machine-learning method, and can be used to solve both linear and nonlinear data classification, regression and prediction tasks (Jia and Yun, 2017; Jia et al., 2013; Wee and Low, 2012; Ying and Keong, 2004; Zhang et al., 2007; Zou et al., 2016). In this study, SVM was trained with the LIBSVM package (Chang and Lin, 2011) to build the model to differentiate both promoter and non-promoter samples. We adopted the radial basis function (RBF) Kxi,xj=exp(-γxi-xj2) as the kernel function. Based on 5-fold cross-validation test, the penalty parameter C and kernel parameter γ were optimized for different types of input features using the SVMcg function of the LIBSVM package. This procedure was conducted for each task separately. In the first task, different types of feature sets (i.e. BPB, MNC, DNC, KNN and DAC) as well as their combined feature sets were evaluated by means of jackknife and cross-validation. Finally, the optimal parameters C=32 and γ=0.01056 were identified, and assigned for the prediction of promoters and non-promoters. In the second task, there were five binary sub-classifiers all of which had distinct parameters from each other. Among those five sub-classifiers, C=1.4142 and γ=0.01121 were the final parameters used for the first sub-classifier, C=2.8284 and γ=2 for the second sub-classifier, C=5.6569 and γ=1 for the third sub-classifier, C=32 and γ=0.25 for the fourth sub-classifier and C=1.4142 and γ=2 for the fifth sub-classifier.

2.5 Performance assessment

To examine the combination of the optimal features and evaluate the prediction performance of the trained models, 5-fold cross-validation, jackknife and independent dataset tests were performed in the present study, as suggested in a number of previous studies (Chen et al., 2017, 2018a,b,c; Chou and Zhang, 1995; Jia et al., 2015; Li et al., 2015a,b, 2016, 2018a,b; Song et al., 2018a,b,c). In addition, we also calculated four commonly used performance measurements, i.e. Sensitivity (Sn), Specificity (Sp), Accuracy (Acc) and the Matthew’s Correlation Coefficient (MCC), which are respectively defined as:

 Sn=1-N-+N+ 0 Sn1 Sp=1-N+-N- 0Sp1  Acc= Λ=1-N-++N+-N++N-  0Acc1   MCC= 1-N-+N++N+-N-1+N+--N-+N+ 1+N-+-N+-N- -1MCC1 (10)

where N+ represents the total number of positive samples, N-+ represents the total number of false negatives, N- represents the total number of negative samples, while N+- represents the total number of false positives, respectively.

2.6 Multiple classification process

MULTiPly is a two-task seamless predictor. The role of the first task is to distinguish a query DNA sequence as a promoter or non-promoter, which is a classic binary classification problem. The second task is to further predict which of the six types of promoters the identified promoter in the first task belongs to. Therefore, this second task is a multi-classification problem. As revealed in the process of constructing the benchmark dataset, the numbers of examples included in the six promoter subsets were quite unbalanced. For example, the largest promoter subset S+(σ70) contained 1694 samples while the smallest promoter subset S+(σ54) contained only 94 samples. To overcome the data imbalance problem, we developed five binary sub-classifiers. In the first sub-classifier, the subset S+(σ70) was regarded as the positive dataset, while the subset S+σ24 S+σ28S+σ32S+σ38S+σ54 was regarded as the negative dataset. In this way, a query DNA sequence sample can be classified as belonging to the σ70 promoter class or to the non-σ70 promoter class. If the query sequence was classified as the non-σ70 promoter class, the next sub-classifier was started. To train the second sub-classifier, the subset S+(σ24) was considered as positive samples and the subset S+σ28S+σ32S+σ38S+σ54 was considered as negative samples. Similar to our description above, the second sub-classifier can predict the query DNA sequence as belonging to the σ24 promoter or non-σ24 promoter class. This process was proceeded until the fifth sub-classifier, the subset S+σ28 was regarded as the positive dataset and S+σ54 regarded as the negative dataset, respectively. Through the subsequent evaluation, standard performance measures indicate the above approach based on the five binary sub-classifiers could not only address the data imbalance problem but, as a by-product, could also accurately predict which of the six types the identified promoter belonged to. The flowchart of this multi-layer classifier is shown in Figure 2.

Fig. 2.

Fig. 2.

The flowchart of the proposed multi-layer classifier

3 Results and discussion

3.1 Selection of the basic features

The combination of different heterogeneous features often leads to different prediction results; accordingly, how to effectively select the basic and essential features to incorporate into the model is a crucial but hard problem to solve. In this study, features that achieved the best prediction performance were chosen as the basic features. Since the dimension of BPB was large, we sorted the 162 components of the characteristic vector using the F-Score, and then chose a step size of 10 entries in the vector to increase the number of components. The other features types were selected using a step size of 2 according to the F-Score. Selection of the optimal feature combination was based on the jackknife test that had the only output result making it easy to compare to (Chou, 2011).

The detailed performance results for the selection of single feature types are given in the Supplementary Tables S1 and S2. For the sake of convenience and intuitive understanding, Tables 1 and 2 show the best performance results for all single types of features, and the corresponding feature dimension at which the best performance was achieved. For the first task, the KNN features with 15 dimensions [KNN (15) for short] were regarded as the basic features, and were then incorporated into the BPB with a step size of 10 entries to further improve the prediction performance. Supplementary Table S3 showed that for KNN(15) combined with BPB of 130 dimension [BPB(130)], the MCC value improved to some extent (for brevity, the encoding scheme was represented by KNN(15)+BPB(130), so on and so forth). Next, KNN(15)+BPB(130) were further incorporated with the component of DNC one by one, and as a result KNN(15)+BPB(130)+DNC(9) reached the best performance with an Acc of 86.80% and an MCC of 0.7360. This process was terminated at the feature combination KNN(15) + BPB(130) + DNC(9) + MNC(1) + DAC(10), which reached a Sn of 87.27%, a Sp of 86.57%, an Acc of 86.92% and an MCC of 0.7385.

Table 1.

The best performance achieved by single type of features for the first task

Features Dimension Sn(%) Sp(%) Acc(%) MCC
KNN 15 85.56 86.68 86.12 0.7224
BPB 120 82.03 81.40 81.71 0.6343
DNC 12 74.86 80.84 77.85 0.558
MNC 4 73.25 80.59 76.92 0.5399
DAC 12 74.48 76.15 75.31 0.5064

Table 2.

The best performance achieved by single types of features for the second task

Sub-classifier Features Dimension Sn (%) Sp (%) Acc(%) MCC
1st KNN 15 90.26 75.64 84.30 0.6723
BPB 162 88.55 76.76 83.74 0.6609
DNC 4 89.08 29.67 64.86 0.237
MNC 3 88.84 27.10 63.67 0.2055
DAC 12 90.2 24.01 63.22 0.1925
2nd KNN 3 86.36 91.06 89.11 0.7754
BPB 130 89.05 92.67 91.17 0.8179
DNC 8 21.9 90.32 61.92 0.1698
MNC 3 2.89 98.24 58.66 0.0378
DAC 12 33.06 84.31 63.04 0.2037
3rd KNN 11 80.07 86.7 83.87 0.6696
BPB 80 83.51 87.47 85.78 0.7094
DNC 10 26.80 85.93 60.7 0.159
MNC 2 1.72 99.49 57.77 0.0592
DAC 6 13.75 92.58 58.94 0.1038
4th KNN 5 82.82 89.04 86.45 0.7206
BPB 70 82.21 86.40 84.65 0.6850
DNC 14 42.33 78.51 63.43 0.2238
MNC 3 26.99 87.28 62.15 0.1806
DAC 12 49.08 75.00 64.19 0.2488
5th KNN 1 96.27 82.98 90.79 0.8107
BPB 140 94.78 91.49 93.42 0.8641
DNC 10 79.10 60.64 71.49 0.4046
MNC 3 91.04 7.45 56.58 −0.0269
DAC 10 76.12 58.51 68.86 0.3509

The purpose of the second task is to predict the specific subtype that a predicted promoter belonged to. To select an optimal combination of features for each of the sub-classifiers, we employed the same strategy and method as described for the first task. The detailed results on the jackknife test are shown in Supplementary Table S4.

For the first sub-classifier, to identify σ70 promoters, the feature combination of KNN(15) + BPB(130) + DAC(6) yielded an Acc of 85.24% and an MCC of 0.6923, respectively. For the second sub-classifier, to identify σ24 promoters, BPB(130) +KNN(17)+DAC(1)+ DNC(12) achieved an Acc of 91.68% and an MCC of 0.8286, respectively. The prediction performance for the third sub-classifier, to identify σ32 promoters, reached an Acc of 87.98% and an MCC of 0.7534, respectively, based on the feature combination of BPB(80)+ KNN(15)+DNC(2). The fourth sub-classifier, to identify σ38 promoters, achieved an Acc of 86.96% and an MCC of 0.7331, respectively, based on only two types of features, KNN(5)+BPB(80). For the last sub-classifier, to distinguish σ28 and σ54 promoters, it used the feature combination of BPB(140)+KNN(3)+DNC(1)+DAC (3) and yielded an Acc of 95.18% and an MCC of 0.9003, respectively.

For dimensionality reduction, we followed two rules: (i) if two kinds of feature combinations achieved the same Acc value, we selected the dimensional features that achieved the larger Sn; and (ii) if all performance indices were identical, we selected the features with the fewest dimensions. Supplementary Tables S5 and S6 provide the best performance results for each combination, for the purpose of easing the interpretation of performance trends.

3.2 Comparison with existing methods on the same training dataset

In general, if one uses different training datasets and validation methods to compare the performance of different prediction tools, the results will vary greatly among them (Li and Lin, 2006; Lin et al., 2014; Liu et al., 2018; Silva et al., 2014; Song, 2012a,b). Therefore, to avoid bias, we applied the same training dataset used in (Liu et al., 2018). The results are shown in Figure 3, which indicate that MULTiPly uniformly achieved a superior performance compared with all other methods. Specifically, the Sn was 7.79% higher than the second-best predictor, iPromoter-2L. Note that only two methods iPromoter-2L and MULTiPly were able to recognize the specific types of promoters. As such, we were more interested in comparing the performance of the two methods for the second task. As shown in Figure 4 and Supplementary Table S7, MULTiPly achieved better MCCs for all six types of promoters, implying that Sn and Sp values were not extremely different, as a higher Sn (or Sp) and a lower Sp (or Sn) at the same time would lead to a lower MCC value. However, the only exception for MULTiPly was in the case of differentiating σ70 promoters, for which the value of Sn was 90.43%, which was 13.5% higher than the value of Sp. In contrast, iPromoter-2L had a larger divergence between the Sn and Sp values: when either its Sn (or Sp) was over 95%, the other measurement was lower than 60% at the same time.

Fig. 3.

Fig. 3.

Performance comparison results between MULTiPly, PCSF, vwZ-curve, Stability, iPro54 and iPromoter-2L for identifying promoters for the first task on 5-fold cross-validation test

Fig. 4.

Fig. 4.

Performance comparison between MULTiPly and iPromoter-2L for the second task in terms of MCC on 5-fold cross-validation test

To further illustrate the effectiveness of the developed MULTiPly method, we assessed and compared its performance with a direct multi-class SVM classifier (Supplementary Table S8). It can be seen that for σ32, σ38 and σ28 types of promoters, none of the promoters were predicted correctly by the multi-class SVM classifier. The worse performance of the multi-class SVM classifier might be explained by the fact that it did not consider the effects brought upon by different numbers of different types of known promoters.

3.3 Performance comparison on the independent test dataset

We compared the proposed MULTiPly method with other existing methods (Li and Lin, 2006; Lin et al., 2014; Liu et al., 2018; Silva et al., 2014; Song, 2012a,b) on an independent test dataset containing 54 newly found promoters. Because no web servers were available for PSCF, vwZ-curve and Stability, we only compared the prediction performance of iPro54, iPromoter-2L and MULTiPly. Performance comparison results between the three methods are provided in Table 3. For the first task, iPro54 only correctly predicted 22 promoter sequences, while iPromoter-2L and MULTiPly achieved the best performance, with all promoter sequences being correctly predicted. Next, we further compared the performance of MULTiPly and iPromoter-2L for the second task of identifying the specific type of promoters. In this regard, iPromoter-2L and MULTiPly achieved a similar performance across all types of promoters (Table 3).

Table 3.

Performance comparison between MULTiPly, iPromoter-2L and iPro54 for identifying promoters and their types on the independent test dataset

Promoter Method TPa FNb
promoter iPro54 22 32
iPromoter-2L 54 0
MULTiPly 54 0
σ70-promoter iPromoter-2L 44 2
MULTiPly 43 3
σ24-promoter iPromoter-2L 1 0
MULTiPly 1 0
σ32-promoter iPromoter-2L 1 1
MULTiPly 1 1
σ38-promoter iPromoter-2L 1 3
MULTiPly 1 3
σ28-promoter iPromoter-2L 1 0
MULTiPly 1 0
a

TP represents the number of predicted (σi)-promoter sequences.

b

FN represents the number of predicted non-(σi)-promoter sequences, where i =70, 24, 32, 38 or 28.

3.4 Performance comparison with other machine learning classifiers

Based on the feature combination determined by SVM, we compared the prediction performance between six commonly used machine learning algorithms, including random forest (RF) (Breiman, 2001; Wei et al., 2018a,b,c), naive Bayes (NB) (Rish, 2001), Ensemble for Boosting (Maclin and Opitz, 1999), discriminant analysis (Cao and Sanders, 1996), gradient boosting decision tree (GBDT) (Friedman, 2001) and SVM (Feng et al., 2018; Wei et al., 2018a,b,c). We performed jackknife tests to examine if there was still room for performance improvement. By and large, the quantity of trees has a bearing on the performance of the RF algorithm. As a consequence, we set out to search for the optimal RF parameters in the two-task predictor. The results are shown in Supplementary Table S9. For GBDT, the learning rate for every tree was set to 0.1, the boosting number was set to 1000 and the depth for every tree was set to 3, respectively. Through a comprehensive performance comparison of these algorithms, we verified the correctness and effectiveness of the SVM classification model, reflected by its higher MCC values. The results are shown in Supplementary Table S10. However, it is worth noting that for the identification of promoters and non-promoters, as well as σ70-promoters and σ32-promoters, the other classifiers instead of the SVM also achieved similar prediction results. Overall, while the results are very promising, it seems that there could be further room for the performance improvement through continued tests and research.

3.5 Web server implementation

As pointed out in Chou and Shen (2009) and suggested in a number of recent publications (see, e.g. Chen et al., 2018a,b,c; Cheng et al., 2018a,b; Feng et al., 2017; Liu et al., 2017a,b; Qiu et al., 2018; Su et al., 2018; Wei et al., 2018a,b,c; Xiao et al., 2017; Xu et al., 2017), user-friendly and publicly accessible web servers represent the future direction for the development of practically useful prediction methods and bioinformatics tools. As a matter of fact, a great variety of practically useful web servers have significantly increased the impact of bioinformatics on medical science (Chou, 2015), driving medicinal chemistry into an unprecedented revolution (Chou, 2017). In view of this, we have implemented and made available the MULTiPly (http://flagshipnt.erc.monash.edu/MULTiPly/) web server via which users can readily obtain their desired prediction results of potential promoters.

The MULTiPly web server was implemented using MATLAB and Java Server Pages, managed by Tomcat 8 and configured on a 64-bit windows server equipped with an 8-core CPU, 1TB hard disk and 32 GB memory. The web server requires DNA sequences in the FASTA format as the input. Supplementary Figure S1 shows an example of the prediction webpages of the web server with the detailed prediction outputs.

4 Conclusion

In this study, we present MULTiPly, a novel bioinformatics tool for identifying bacterial promoters and the specific promoter types they belong to. MULTiPly is capable of recognizing the specific type of promoters in a layer-by-layer manner, which overcomes the complexity brought upon by different numbers of available types of promoters in the datasets. Extensive benchmarking experiments on 5-fold cross-validation and jackknife tests demonstrate the strategy used by MULTiPly is effective and can deal with the data imbalance problems. We expect that MULTiPly will be used as a useful tool for expediting the discovery of both general and specific types of promoters in the future.

Funding

This work was supported by Fundamental Research Funds for the Central Universities (No. 3132016306, 3132018227), the National Natural Science Foundation of Liaoning Province (20180550307) and the National Scholarship Fund of China for Studying Abroad. JS was supported by grants from the National Health and Medical Research Council of Australia (NHMRC) (APP490989, APP1127948 and APP1144652), the Australian Research Council (ARC) (LP110200333 and DP120104460), the National Institute of Allergy and Infectious Diseases of the National Institutes of Health (R01 AI111965), a Major Inter-Disciplinary Research (IDR) project awarded by Monash University and the Collaborative Research Program of Institute for Chemical Research, Kyoto University (2018-28). TML and AL were supported in part by the Informatics Institute of the School of Medicine at UAB.

Conflict of Interest: none declared.

Supplementary Material

btz016_Supplementary_Materials

References

  1. Barrios H. et al. (1999) Compilation and analysis of sigma(54)-dependent promoter sequences. Nucleic Acids Res., 27, 4305–4313. [DOI] [PMC free article] [PubMed] [Google Scholar]
  2. Breiman L. (2001) Random forests. Mach. Learn., 45, 5–32. [Google Scholar]
  3. Bui V.M. (2016) SOHSite: incorporating evolutionary information and physicochemical properties to identify protein S-sulfenylation sites. BMC Genomics, 17, 9.. [DOI] [PMC free article] [PubMed] [Google Scholar]
  4. Cao J., Sanders D.B. (1996) Multivariate discriminant analysis of the electromyographic interference pattern: statistical approach to discrimination among controls, myopathies and neuropathies. Med. Biol. Eng. Comput., 34, 369–374. [DOI] [PubMed] [Google Scholar]
  5. Chang C.C., Lin C.J. (2011) LIBSVM: a library for support vector machines. ACM Transactions on Intelligent Systems and Technology, 2, 1–27. [Google Scholar]
  6. Chen X. et al. (2013) Incorporating key position and amino acid residue features to identify general and species-specific Ubiquitin conjugation sites. Bioinformatics, 29, 1614–1622. [DOI] [PubMed] [Google Scholar]
  7. Chen W. et al. (2015) Pseudo nucleotide composition or PseKNC: an effective formulation for analyzing genomic sequences. Mol. Biosyst., 11, 2620–2634. [DOI] [PubMed] [Google Scholar]
  8. Chen W. et al. (2017) iDNA4mC: identifying DNA N4-methylcytosine sites based on nucleotide chemical properties. Bioinformatics, 33, 3518–3523. [DOI] [PubMed] [Google Scholar]
  9. Chen W. et al. (2018a) iRNA-3typeA: identifying three types of modification at RNA’s adenosine sites. Mol. Ther. Nucl. Acids, 11, 468–474. [DOI] [PMC free article] [PubMed] [Google Scholar]
  10. Chen Z. et al. (2018b) Large-scale comparative assessment of computational predictors for lysine post-translational modification sites. Brief. Bioinf., bby089. [DOI] [PMC free article] [PubMed] [Google Scholar]
  11. Chen Z. et al. (2018c) iFeature: a python package and web server for features extraction and selection from protein and peptide sequences. Bioinformatics, 34, 2499–2502. [DOI] [PMC free article] [PubMed] [Google Scholar]
  12. Cheng X. et al. (2018a) pLoc_bal-mGneg: predict subcellular localization of Gram-negative bacterial proteins by quasi-balancing training dataset and general PseAAC. J. Theor. Biol., 458, 92–102. [DOI] [PubMed] [Google Scholar]
  13. Cheng X. et al. (2018b) pLoc-mEuk: predict subcellular localization of multi-label eukaryotic proteins by extracting the key GO information into general PseAAC. Genomics, 110, 50–58. [DOI] [PubMed] [Google Scholar]
  14. Chou K.C., Zhang C.T. (1995) Prediction of protein structural classes. Crit. Rev. Biochem. Mol. Biol., 30, 275–349. [DOI] [PubMed] [Google Scholar]
  15. Chou K.C. (2011) Some remarks on protein attribute prediction and pseudo amino acid composition (50th Anniversary Year Review). J. Theor. Biol., 273, 236–247. [DOI] [PMC free article] [PubMed] [Google Scholar]
  16. Chou K.C. (2015) Impacts of bioinformatics to medicinal chemistry. Med. Chem., 11, 218–234. [DOI] [PubMed] [Google Scholar]
  17. Chou K.C. (2017) An unprecedented revolution in medicinal chemistry driven by the progress of biological science. Curr. Top. Med. Chem., 17, 2337–2358. [DOI] [PubMed] [Google Scholar]
  18. Chou K.C., Shen H.B. (2009) Recent advances in developing web-servers for predicting protein attributes. Nat. Sci., 01, 63–92. [Google Scholar]
  19. Dong Q.W. et al. (2009) A new taxonomy-based protein fold recognition approach based on autocross-covariance transformation. Bioinformatics, 25, 2655–2662. [DOI] [PubMed] [Google Scholar]
  20. Feng C.Q. et al. (2018) iTerm-PseKNC: a sequence-based tool for predicting bacterial transcriptional terminators. Bioinformatics, bty827–bty827. [DOI] [PubMed] [Google Scholar]
  21. Feng P. et al. (2017) iRNA-PseColl: identifying the occurrence sites of different RNA modifications by incorporating collective effects of nucleotides into PseKNC. Mol. Ther. Nucleic Acids, 7, 155–163. [DOI] [PMC free article] [PubMed] [Google Scholar]
  22. Friedel M. et al. (2009) DiProDB: a database for dinucleotide properties. Nucleic Acids Res., 37, D37.. [DOI] [PMC free article] [PubMed] [Google Scholar]
  23. Friedman J.H. (2001) Greedy function approximation: a gradient boosting machine. Ann. Stat., 29, 1189–1232. [Google Scholar]
  24. Gao J.J. et al. (2010) Musite, a tool for global prediction of general and kinase-specific phosphorylation sites. Mol. Cell Proteomics, 9, 2586–2600. [DOI] [PMC free article] [PubMed] [Google Scholar]
  25. Guo Y.Z. et al. (2008) Using support vector machine combined with auto covariance to predict proteinprotein interactions from protein sequences. Nucleic Acids Res., 36, 3025–3030. [DOI] [PMC free article] [PubMed] [Google Scholar]
  26. He W.Y. et al. (2018) 70ProPred: a predictor for discovering sigma70 promoters based on combining multiple features. BMC Syst. Biol., 12, [DOI] [PMC free article] [PubMed] [Google Scholar]
  27. Helmann J.D., Chamberlin M.J. (1988) Structure and function of bacterial sigma factors. Annu. Rev. Biochem., 57, 839–872. [DOI] [PubMed] [Google Scholar]
  28. Hertz G.Z., Stormo G.D. (1996) Escherichia coli promoter sequences: analysis and prediction. Method Enzymol., 273, 30–42. [DOI] [PubMed] [Google Scholar]
  29. Huerta A.M., Collado-Vides J. (2003) Sigma70 promoters in Escherichia coli: specific transcription in dense regions of overlapping promoter-like signals. J. Mol. Biol., 333, 261–278. [DOI] [PubMed] [Google Scholar]
  30. Ioshikhes I. et al. (1996) Nucleosome DNA sequence pattern revealed by multiple alignment of experimentally mapped sequences. J. Mol. Biol., 262, 129–139. [DOI] [PubMed] [Google Scholar]
  31. Jia C. et al. (2018) NucPosPred: predicting species-specific genomic nucleosome positioning via four different modes of general PseKNC. J. Theor. Biol., 450, 15–21. [DOI] [PubMed] [Google Scholar]
  32. Jia C., Yun Z. (2017) S-SulfPred: a sensitive predictor to capture S-sulfenylation sites based on a resampling one-sided selection undersampling-synthetic minority oversampling technique. J. Theor. Biol., 422, 84–89. [DOI] [PubMed] [Google Scholar]
  33. Jia C.Z., He W.Y. (2016) EnhancerPred: a predictor for discovering enhancers based on the combination and selection of multiple features. Sci. Rep. UK, 6. [DOI] [PMC free article] [PubMed] [Google Scholar]
  34. Jia C.Z. et al. (2013) O-GlcNAcPRED: a sensitive predictor to capture protein O-GlcNAcylation sites. Mol. Biosyst., 9, 2909–2913. [DOI] [PubMed] [Google Scholar]
  35. Jia C.Z. et al. (2016) RNA-MethylPred: a high-accuracy predictor to identify N6-methyladenosine in RNA. Anal. Biochem., 510, 72–75. [DOI] [PubMed] [Google Scholar]
  36. Jia J. et al. (2015) iPPI-Esml: an ensemble classifier for identifying the interactions of proteins by incorporating their physicochemical properties and wavelet transforms into PseAAC. J. Theor. Biol., 377, 47–56. [DOI] [PubMed] [Google Scholar]
  37. Kabir M., Hayat M. (2016) iRSpot-GAEnsC: identifing recombination spots via ensemble classifier and extending the concept of Chou's PseAAC to formulate DNA samples. Mol. Genet. Genomics, 291, 285–296. [DOI] [PubMed] [Google Scholar]
  38. Li F. et al. (2016) GlycoMine(struct): a new bioinformatics tool for highly accurate mapping of the human N-linked and O-linked glycoproteomes by incorporating structural features. Sci. Rep., 6, 34595.. [DOI] [PMC free article] [PubMed] [Google Scholar]
  39. Li F. et al. (2015a) GlycoMine: a machine learning-based approach for predicting N-, C- and O-linked glycosylation in the human proteome. Bioinformatics, 31, 1411–1419. [DOI] [PubMed] [Google Scholar]
  40. Li W.C. et al. (2015b) iORI-PseKNC: a predictor for identifying origin of replication with pseudo k-tuple nucleotide composition. Chemometr. Intell. Lab., 141, 100–106. [Google Scholar]
  41. Li F. et al. (2018a) Quokka: a comprehensive tool for rapid and accurate prediction of kinase family-specific phosphorylation sites in the human proteome. Bioinformatics, bty522–bty522. [DOI] [PMC free article] [PubMed] [Google Scholar]
  42. Li F. et al. (2018b) Twenty years of bioinformatics research for protease-specific substrate and cleavage site prediction: a comprehensive revisit and benchmarking of existing methods. Brief. Bioinf., bby077–bby077. [DOI] [PMC free article] [PubMed] [Google Scholar]
  43. Li Q.Z., Lin H. (2006) The recognition and prediction of sigma(70) promoters in Escherichia coli K-12. J. Theor. Biol., 242, 135–141. [DOI] [PubMed] [Google Scholar]
  44. Liang Z.Y. et al. (2017) Pro54DB: a database for experimentally verified sigma-54 promoters. Bioinformatics, 33, 467–469. [DOI] [PubMed] [Google Scholar]
  45. Lin H. et al. (2014) iPro54-PseKNC: a sequence-based predictor for identifying sigma-54 promoters in prokaryote with pseudo k-tuple nucleotide composition. Nucleic Acids Res., 42, 12961–12972. [DOI] [PMC free article] [PubMed] [Google Scholar]
  46. Lin H., Ding H. (2011) Predicting ion channels and their types by the dipeptide mode of pseudo amino acid composition. J. Theor. Biol., 269, 64–69. [DOI] [PubMed] [Google Scholar]
  47. Lin H. et al. (2017) Identifying sigma70 promoters with novel pseudo nucleotide composition. IEEE/ACM Trans. Comput. Biol. Bioinf., 99, 1–1. [DOI] [PubMed] [Google Scholar]
  48. Liu B. et al. (2015) repDNA: a Python package to generate various modes of feature vectors for DNA sequences by incorporating user-defined physicochemical properties and sequence-order effects. Bioinformatics, 31, 1307.. [DOI] [PubMed] [Google Scholar]
  49. Liu B. et al. (2017a) Pse-in-One 2.0: an improved package of web servers for generating various modes of pseudo components of DNA, RNA, and protein sequences. Nat. Sci., 09, 67–91. [DOI] [PMC free article] [PubMed] [Google Scholar]
  50. Liu L.M. et al. (2017b) iPGK-PseAAC: identify lysine phosphoglycerylation sites in proteins by incorporating four different tiers of amino acid pairwise coupling information into the general PseAAC. Med. Chem., 13, 552–559. [DOI] [PubMed] [Google Scholar]
  51. Liu B. et al. (2018) iPromoter-2L: a two-layer predictor for identifying promoters and their types by multi-window-based PseKNC. Bioinformatics, 34, 33–40. [DOI] [PubMed] [Google Scholar]
  52. Maclin R., Opitz D. (1999) Popular ensemble methods: an empirical study. J. Artif. Intell. Res., 11, 169–198. [Google Scholar]
  53. Mrozek D. et al. (2016) HDInsight4PSi: boosting performance of 3D protein structure similarity searching with HDInsight clusters in Microsoft Azure cloud. Inform. Sci., 349, 77–101. [Google Scholar]
  54. Mrozek D. et al. (2014) Cloud4Psi: cloud computing for 3D protein structure similarity searching. Bioinformatics, 30, 2822–2825. [DOI] [PMC free article] [PubMed] [Google Scholar]
  55. Polat K., Güneş S. (2009) A new method to forecast of Escherichia coli promoter gene sequences: integrating feature selection and Fuzzy-AIRS classifier system. Expert. Syst. Appl., 36, 57–64. [Google Scholar]
  56. Qiu W.-R. et al. (2018) iKcr-PseEns: identify lysine crotonylation sites in histone proteins with pseudo components and ensemble classifier. Genomics, 110, 239–246. [DOI] [PubMed] [Google Scholar]
  57. Ramprakash J., Schwarz F.P. (2008) Energetic contributions to the initiation of transcription in E. coli. Biophys. Chem., 138, 91–98. [DOI] [PubMed] [Google Scholar]
  58. Rish I. (2001) An empirical study of the naive Bayes classifier. J. Universal Comput. Sci., 1, 127. [Google Scholar]
  59. Shahmuradov I.A. et al. (2017) bTSSfinder: a novel tool for the prediction of promoters in cyanobacteria and Escherichia coli. Bioinformatics, 33, 334–340. [DOI] [PMC free article] [PubMed] [Google Scholar]
  60. Shao J.L. et al. (2009) Computational identification of protein methylation sites through bi-profile bayes feature extraction. PLoS One, 4. [DOI] [PMC free article] [PubMed] [Google Scholar]
  61. Silva S.D.E. et al. (2014) DNA duplex stability as discriminative characteristic for Escherichia coli sigma(54)- and sigma(28)- dependent promoter sequences. Biologicals, 42, 22–28. [DOI] [PubMed] [Google Scholar]
  62. Song J. et al. (2010) Cascleave: towards more accurate prediction of caspase substrate cleavage sites. Bioinformatics, 26, 752–760. [DOI] [PubMed] [Google Scholar]
  63. Song J. et al. (2012a) PROSPER: an integrated feature-based tool for predicting protease substrate cleavage sites. PLoS One, 7, e50300.. [DOI] [PMC free article] [PubMed] [Google Scholar]
  64. Song K. (2012b) Recognition of prokaryotic promoters based on a novel variable-window Z-curve method. Nucleic Acids Res., 40, 963–971. [DOI] [PMC free article] [PubMed] [Google Scholar]
  65. Song J. et al. (2018a) PROSPERous: high-throughput prediction of substrate cleavage sites for 90 proteases with improved accuracy. Bioinformatics, 34, 684–687. [DOI] [PMC free article] [PubMed] [Google Scholar]
  66. Song J. et al. (2018b) PREvaIL, an integrative approach for inferring catalytic residues using sequence, structural, and network features in a machine-learning framework. J. Theor. Biol., 443, 125–137. [DOI] [PubMed] [Google Scholar]
  67. Song J. et al. (2018c) iProt-Sub: a comprehensive package for accurately mapping and predicting protease-specific substrates and cleavage sites. Brief. Bioinf., bby028–bby028. [DOI] [PMC free article] [PubMed] [Google Scholar]
  68. Su R. et al. (2018) Developing a multi-dose computational model for drug-induced hepatotoxicity prediction based on toxicogenomics data. IEEE/ACM Trans. Comput. Biol. Bioinf, doi: 10.1109/TCBB.2018.2858756. [DOI] [PubMed]
  69. Towsey M. et al. (2008) The cross-species prediction of bacterial promoters using a support vector machine. Comput. Biol. Chem., 32, 359–366. [DOI] [PubMed] [Google Scholar]
  70. Wang L.N. et al. (2017) Computational prediction of species-specific malonylation sites via enhanced characteristic strategy. Bioinformatics, 33, 1457–1463. [DOI] [PubMed] [Google Scholar]
  71. Wang M. et al. (2014) Cascleave 2.0, a new approach for predicting caspase and granzyme cleavage targets. Bioinformatics, 30, 71–80. [DOI] [PubMed] [Google Scholar]
  72. Wee L.J.K., Low H.M. (2012) SVM-based prediction of the calpain degradome using Bayes Feature Extraction. 5534–5540. Eng. Med. Biol. Soc. [DOI] [PubMed] [Google Scholar]
  73. Wei L. et al. (2018a) Comparative analysis and prediction of quorum-sensing peptides using feature representation learning and machine learning algorithms. Brief. Bioinf., bby107–bby107. [DOI] [PubMed] [Google Scholar]
  74. Wei L. et al. (2018b) Exploring sequence-based features for the improved prediction of DNA N4-methylcytosine sites in multiple species. Bioinformatics, bty824–bty824. [DOI] [PubMed] [Google Scholar]
  75. Wei L. et al. (2018c) ACPred-FL: a sequence-based predictor using effective feature representation to improve the prediction of anti-cancer peptides. Bioinformatics, 34, 4007–4016. [DOI] [PMC free article] [PubMed] [Google Scholar]
  76. Xiao X. et al. (2017) pLoc-mGpos: incorporate key gene ontology information into general PseAAC for predicting subcellular localization of Gram-positive bacterial proteins. Nat. Sci., 9, 331–349. [Google Scholar]
  77. Xu Y. et al. (2017) iPreny-PseAAC: identify C-terminal cysteine prenylation sites in proteins by incorporating two tiers of sequence couplings into PseAAC. Med. Chem., 13, 544–551. [DOI] [PubMed] [Google Scholar]
  78. Ying Z., Keong K.C. (2004) Fast leave-one-out evaluation and improvement on inference for LS-SVMs. Int. C Patt. Recog., 494–497. [Google Scholar]
  79. Zhang G.L. et al. (2007) Prediction of supertype-specific HLA class I binding peptides using support vector machines. J. Immunol. Methods, 320, 143–154. [DOI] [PMC free article] [PubMed] [Google Scholar]
  80. Zou Q. et al. (2016) Protein folds prediction with hierarchical structured SVM. Curr. Proteomics, 13, 79–85. [Google Scholar]
  81. Zuo Y., Jia C.Z. (2017) CarSite: identifying carbonylated sites of human proteins based on a one-sided selection resampling method. Mol. Biosyst., 13, 2362–2369. [DOI] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

btz016_Supplementary_Materials

Articles from Bioinformatics are provided here courtesy of Oxford University Press

RESOURCES