Skip to main content
Computational and Structural Biotechnology Journal logoLink to Computational and Structural Biotechnology Journal
. 2018 Oct 24;16:412–420. doi: 10.1016/j.csbj.2018.10.007

iGHBP: Computational identification of growth hormone binding proteins from sequences using extremely randomised tree

Shaherin Basith a, Balachandran Manavalan a,, Tae Hwan Shin a,b, Gwang Lee a,b,
PMCID: PMC6222285  PMID: 30425802

Abstract

A soluble carrier growth hormone binding protein (GHBP) that can selectively and non-covalently interact with growth hormone, thereby acting as a modulator or inhibitor of growth hormone signalling. Accurate identification of the GHBP from a given protein sequence also provides important clues for understanding cell growth and cellular mechanisms. In the postgenomic era, there has been an abundance of protein sequence data garnered, hence it is crucial to develop an automated computational method which enables fast and accurate identification of putative GHBPs within a vast number of candidate proteins. In this study, we describe a novel machine-learning-based predictor called iGHBP for the identification of GHBP. In order to predict GHBP from a given protein sequence, we trained an extremely randomised tree with an optimal feature set that was obtained from a combination of dipeptide composition and amino acid index values by applying a two-step feature selection protocol. During cross-validation analysis, iGHBP achieved an accuracy of 84.9%, which was ~7% higher than the control extremely randomised tree predictor trained with all features, thus demonstrating the effectiveness of our feature selection protocol. Furthermore, when objectively evaluated on an independent data set, our proposed iGHBP method displayed superior performance compared to the existing method. Additionally, a user-friendly web server that implements the proposed iGHBP has been established and is available at http://thegleelab.org/iGHBP.

Keywords: Extremely randomised tree, Growth hormone binding protein, Machine learning, Random forest, Support vector machine

Highlights

  • We developed an identification method for GHBPs using extremely randomized tree.

  • iGHBP displayed superior performance compared to the existing method.

  • We constructed a user-friendly web server that implements the proposed iGHBP method.

1. Introduction

Circulating growth hormones (GH) exist in a partially complexed form with binding proteins. The high affinity growth hormone binding protein (GHBP) is one such predominant GH binding protein that represents the extracellular ligand-binding domain of the GH receptor (GHR) [[1], [2], [3], [4]]. In humans, GHBP is generated by the proteolytic cleavage of the GHR at the cell surface using the tumor necrosis enzyme factor-α-converting enzyme (TACE), thereby releasing the extracellular domain of GHR (i.e., GHBP) [[5], [6], [7]]. By contrast, GHBP is produced in rodents by the alternative processing of the GHR transcript [8]. Binding GH to the GHR triggers the physiological functions of the hormone. Previous studies suggested that the biological effects of GHBP is dependent on the serum level of GH [5], as low levels of GH lead to a dwarf phenotype but increases the life longevity [1,9], while high levels lead to acromegaly, kidney damage, and diabetic eye. Therefore, the study of GHBP is gaining momentum from functional proteomics, leading to its clinical identification.

Traditionally, GHBPs were identified and characterised using biochemical experiments including immunoprecipitation, ligand immunofunctional assays, chromatography, and cross-linking assays [[10], [11], [12], [13]]. To identify GHBP from a protein sequence using these methods seems to be highly expensive, time-consuming, and overly complex to be utilised in a high-throughput manner. Thus, the development of sequence-based computational methods is needed to identify potential GHBP candidates. Recently, Tang et al. developed an Support vector machine (SVM)-based prediction model called HBPred [14], where the authors have used an optimal feature set obtained from dipeptide composition (DPC) using an incremental feature selection strategy. HBPred is the only publicly available method, which was developed using the same data set as our method. Although the existing method has a specific advantage in GHBP prediction, the accuracy and transferability of the prediction model still require improvement.

In this study, we proposed a novel sequence-based predictor, called iGHBP, for the identification of GHBPs from given protein sequences (Fig. 1). Firstly, we collected GHBPs from UniProt and constructed nonredundant benchmarking and independent data sets. Secondly, we investigated five different machine learning (ML) algorithms [SVM, random forest (RF), extremely randomized tree (ERT), gradient boosting (GB), adaBoost (AB)], five compositions [amino acid composition (AAC), amino acid index (AAI), DPC, chain-transition-distribution (CTD), and physiochemical properties (PCP)], and 16 hybrid features (a linear combination of various compositions). In total, we generated 21 models for each ML method and selected the best model. Thirdly, we applied a two-step feature selection protocol on the above selected best model to improve the prediction performance. Finally, we evaluated these models against the state-of-the art method, HBPred, on the independent data set. Experimental results showed that the ERT-based prediction model produced consistent performance on both the benchmarking and independent data sets, hence, we named iGHBP as the superior model, demonstrated by outperforming the existing predictor as well as other predictors tested in this study. Therefore, it can be expected that iGHBP can be an effective tool for identifying GHBPs.

Fig. 1.

Fig. 1

Overview of the proposed methodology for predicting GHBPs that involved the following steps. (i) data set construction; (ii) feature extraction; (iii) feature ranking; (iv) exploration of various machine learning algorithms and an appropriate selection based on the performance produced using sequential forward search; (v) construction of the final prediction model that separates the input into putative GHBPs and non-GHBPs.

2. Methods

The iGHBP methodology development involved five major stages: Data set construction, feature extraction, feature ranking, model training and validation, and the construction of the final prediction model. Each of these major stages is described in the following section.

2.1. Data set construction

2.1.1. Benchmarking data set

We utilised the data set constructed by Tang et al., [14] which was specifically used for the classification of GHBPs or non-GHBPs. The reason for considering this data set is as follows: (i) they have applied several filtering schemes to construct such a reliable data set; (ii) it is an nonredundant data set, and none of the sequences possesses pairwise sequence identity (>60%) with any other sequence; (iii) furthermore, it enables a fair comparison between our method and the existing method, which was developed using the same benchmarking data set. Thus, the benchmarking data set can be formulated as:

S=S+S (1)

where the subsets S+ and S respectively contain 123 GHBPs and 123 non-HBPs, and the symbol ∪ denotes a union, in set theory.

2.1.2. Independent data set

To assess the performance of iGHBP with other related methods, we constructed an independent data set. Firstly, we considered 355 manually annotated and reviewed GHBP proteins from Universal Protein Resource (UniProt) using hormone-binding keywords in molecular function item of Gene Ontology. After this, we used CD-HIT [15], which is widely used to perform sequence clustering and to remove highly similar sequences, by setting a threshold of 0.6. The final data set contained 31 GHBPs and was supplemented with an equal number of non-GHBPs. Basically, these non-GHBPs are other functional proteins such as cancer lectins and phage virion proteins. Note that none of the protein sequences in the independent data set appeared in the benchmarking data set, ensuring a fair comparison of prediction model performance.

2.2. Feature representation of proteins

A protein sequence (P) can be represented as:

P=R1R2R3..RN (2)

where R1, R2 and R3 respectively denote the 1st, 2nd, and 3rd residues in the protein P and so forth. N denotes the protein length. It should be noted that the residue Ri is an element of the standard amino acid {A, C, D, E, F, G, H, I, K, L, M, N, P, Q, R, S, T, V, W, Y}. To develop a ML model, we formulated proteins with diverse-length as fixed-length feature vectors. We exploited five different compositions that cover different aspects of sequence information as described below:

2.2.1. AAC

AAC is the percentage of standard amino acids; it has a fixed length of 20 features. AAC can be formulated as follows:

AACP=f1f2f3f20 (3)

where fi=RiNi=12320 is the percentage of the composition with amino acid type i, Ri is the quantity of type i appearing in the protein, and N is the protein length.

2.2.2. DPC

DPC is the rate of dipeptides normalised by all possible dipeptide combinations; it has a fixed length of 400 features. DPC can be formulated as follows:

DPCP=f1f2f3f400 (4)

where fi=Rii=1400Rii=123400 is the percentage of the composition with dipeptide type i and Ri is the quantity of type i appearing in the protein.

2.2.3. CTD

CTD was introduced by Dubchak, et al. [16] for predicting protein-folding classes. A detailed description of computing CTD features was presented in our previous study [37]. Briefly, the twenty standard amino acids are classified into three different groups, namely: polar, neutral, and hydrophobic. Composition (C) consists of percentage composition values from these three groups for a target protein. Transition (T) consists of the percentage frequency of a polar followed by a neutral residue, or that of a neutral followed by a polar residue. This group may also contain a polar followed by a hydrophobic residue or a hydrophobic followed by a polar residue. Distribution (D) consists of five values for each of the three groups, and measures the percentage of a target sequence length within which 25, 50, 75, and 100% of the amino acids of a specific property are located. CTD generates 21 features for each PCP; hence, seven different PCPs (hydrophobicity, polarisability, normalised van der Waals volume, secondary structure, polarity, charge, and solvent accessibility) yield a total of 147 features.

2.2.4. AAI

The AAIndex database contains a variety of physiochemical and biochemical properties of amino acids [17]. However, utilising all the information present in the AAIndex database as input features to the ML algorithm may affect the model’s performance, due to redundancy. To this end, Saha et al., [18] applied a fuzzy clustering method on the AAIndex database and classified it into eight clusters, where the central indices of each cluster were considered as high-quality amino acid indices. The accession numbers of the eight amino acid indices in the AAIndex database are BLAM930101, BIOV880101, MAXF760101, TSAJ990101, NAKH920108, CEDJ970104, LIFS790101, and MIYS990104. These high-quality indices encode the target protein sequences as 160-dimensional vectors. However, the average of these eight high-quality amino acid indices (a 20-dimensional vector) was used as an additional input feature to save the computational time.

2.2.5. PCP

PCP computed from the target protein sequence includes: (i) hydrophobic residues (i.e., F, I, W, L, V, M, Y, C, A); (ii) hydrophilic residues (i.e., S, Q, T, R, K, N, D, E); (iii) neutral residues (i.e., H,G, P); (iv) positively charged residues (i.e., K, H, R); (v) negatively charged residues (i.e., D, E); (vi) n (sequence length); (vii) fraction of turn-forming residues (i.e., [N+G+P+S]/n); (viii) absolute charge per residue (R+KDEn0.03); (ix) molecular weight; and (x) aliphatic index (i.e., [A+2.9V+3.9I+3.9L]/n).

Briefly, we extracted five feature encoding schemes based on composition and physicochemical properties, which includes AAC, DPC, CTD, AAI, and PCP respectively generates 20-, 400-, 147-, 20-, and 10-dimensional vectors.

2.3. Machine learning algorithms

In this study, we explored five different ML algorithms, including RF, ERT, SVM, GB, and AB for binary classification (GHBP or non-GHBP). All these ML algorithms were implemented using the Scikit-Learn package (v0.18) [19]. A brief description of these methods and how they were used given in the following sections:

2.3.1. Random forest

RF is one of the most successful ML methods, and utilises hundreds or thousands of independent decision trees to perform classification and regression [20]. RF combines the concepts of bagging and random feature selection. For a given training data set (D), generate a new training data set (Di) by uniformly drawing N bootstrapped samples from D. Grow a tree using Di and repeat the following steps at each node of the tree until its fully grown: (i) select mtry random features from the original features and select the best variable by optimising the impurity criteria, and (ii) split the node into two child nodes. The tree grows until the amount of data in the node is below the given threshold (nsplit). Repeat the above-mentioned steps to build a large quantity (ntree) of classification trees. To classify a test data, input features are passed through from the root to the end node of each tree based on predetermined splits. The majority of the class from the forest is considered as the final classification.

2.3.2. Extremely randomised tree

Geurts et al. [21] proposed the ERT algorithm, which utilises hundreds or thousands of independent decision trees to perform classification and regression problems, and has been applied in a large number of biological problems [22,23]. ERT aims to further decrease the variance of the prediction model by including stronger randomisation techniques. The ERT algorithm is similar to RF, but with the following differences: (i) ERT does not apply a bagging procedure for the construction of each tree. Instead, it uses the whole input training set for the construction of each tree. (ii) ERT selects a node split very randomly (both a variable index and variable splitting values are chosen randomly), whereas RF finds the best split (optimised by a variable index and a variable splitting value) among a random subset of variables. Furthermore, Grid search was performed for optimising the regularisation parameters ntree, mtry, and nsplit. The search space for ntree, mtry, and nsplit are:

40ntree1000with step ∆ntree=201mtry15with step ∆mtry=11nsplit10with step ∆nsplit=1 (5)

2.3.3. Support vector machine

SVM is a well-known supervised ML algorithm [24], which has been widely used in various biological problems [25,26]. It maps the original feature vectors into a higher Hilbert space using different kernel functions and then searches an optimal hyperplane in Hilbert space. In this study, radial basis kernel function was utilized to construct a SVM model. Grid search was performed for optimizing regularisation parameters C and the kernel width parameter γ with the search space as mentioned in [27].

2.3.4. Adaptive boosting

Fruend [28] proposed AB algorithm that combines a several weak classifiers to build a strong classifier. In this study, we treated decision tree as a base classifier with the default parameters as implemented in Scikit package. However, the number of estimators at which boosting terminated is optimized in the range of 50–500 with an interval of 50.

2.3.5. Gradient boosting

Friedman proposed the GB algorithm [29], which is a forward learning ensemble method that produces a final strong prediction model based on the ensemble of weak models (decision trees), which has been widely used in bioinformatics and computational biology [27,30]. In GB, the two most influential parameters are ntree, and nsplit, we optimized with the search space as mentioned in [27].

In addition to the above ML algorithms, we note that there are other ML algorithms such as deep belief network, recurrent neural network, deep learning, and two-layer neural network have been successfully applied in various biological problems [[31], [32], [33], [34], [35], [36]]. However, these methods will be considered in our future studies.

2.4. Cross-validation

Generally, three cross-validation methods, namely an independent data set test, a sub-sampling (or k-fold cross-validation) test, and a leave-one-out cross-validation (LOOCV) test, are often used to evaluate the anticipated success rate of a predictor. Among the three methods, however, the LOOCV test is deemed the least arbitrary and most objective as demonstrated by Eqs. 28-32 of [37], and hence it has been widely recognised and increasingly adopted by investigators to examine the quality of various predictors [[38], [39], [40], [41], [42], [43], [44], [45], [46], [47], [48]]. Accordingly, the LOOCV test was also used to examine the performance of the model proposed in the current study. In the LOOCV test, each sequence in the training data set is in turn singled out as an independent test sample and all the rule-parameters are calculated without including the one being identified.

2.5. Performance evaluation

To evaluate the performance of the constructed models, we used four measurements that were commonly used in binary classification tasks, including sensitivity, specificity, accuracy, and Matthews correlation coefficient (MCC). They are calculated as follows:

Sensitivity=TPTP+FN
Specificity=TNTN+FP
Accuracy=TP+TNTP+TN+FP+FN
MCC=TP×TNFP×FNTP+FPTP+FNTN+FPTN+FN (6)

where TP is the number of true positives (i.e., GHBPs classified correctly as GHBPs) and TN is the number of true negatives (i.e., non-GHBPs classified correctly as non-GHBPs). FP is the number of false positives (i.e., GHBPs classified incorrectly as non-GHBPs) and FN is the number of false negatives (i.e., non-GHBPs classified incorrectly as GHBPs).

Additionally, the receiver operating characteristic (ROC) curve, which is a plot of the true positive rate against the false positive rate under different classification thresholds, is depicted to visually measure the comprehensive performance of different classifiers.

2.6. Feature selection

To improve the feature representation capability and identify the subset of optimal features that contribute for correctly classifying GHBPs and non-GHBPs, we employed a novel two-step feature selection strategy. Notably, the two-step feature selection protocol employed here is similar to the one used in our recent studies [26,[49], [50], [51]], where the features were ranked according to feature importance scores (FISs) using the RF algorithm in the first step, and feature subsets were selected manually based on the FISs in the second step. In this study, the first step is identical to our previous protocol. However, in the second step, a sequential forward search (SFS) was employed to select the optimal feature subset, rather than using manual feature subset selection.

In the first step, we inputted a given set of features for the RF algorithm and carried out a 10-fold cross-validation (CV). For each round of CV, we built 1000 trees using a mtry range from 1–50. The average FISs from all the trees were used to rank the features.

D=F1F2F3FNT (7)

where F1 is the first feature with the maximum FIS; F2 is the second feature with the second maximum FIS; F3 is the third feature with the third maximum FIS and so on; N and T are the total number of features and the transpose operator, respectively.

In the second step, we utilised SFS to identify and select the optimal features from a ranked feature set based on the following steps. (i) The first feature subset only contained the first feature in the ranked set D. The second feature subset contains the first and the second feature in D, and so on. Finally, we obtained N feature subsets. (ii) All the N feature subsets were inputted to ERT to develop their corresponding prediction model using a LOOCV test. Finally, the best performance produced by the feature subset was considered as the optimal feature set.

3. Results and discussion

3.1. Performance comparison of various models using different feature encodings

In this study, we considered 21 feature encodings that include individual composition-based features and hybrid features (a linear combination of different individual compositions), which were inputted to five different ML algorithms, developing their corresponding models using a LOOCV procedure. In total, 105 prediction models were developed and the performance of each model in terms of accuracy with respect to the different feature encodings and ML algorithms is shown in Fig. 2. Among these methods, ERT and RF perform consistently better than other three algorithms (SVM, GB, and AB). Here, the model that achieved the highest accuracy was regarded as the best model. Accordingly, five models were selected from each ML method. Surprisingly, these five ML models produced their best performances using hybrid features (ERT: H8 (DPC+AAI); RF and AB: H5 (DPC+AAC); SVM: H4 (AAC+DPC+AAI+CTD+PCP); and GB: H10 (DPC+AAI+CTD+PCP)), indicating that various aspects of sequence information may be needed for a better prediction. Table 1 shows the performance comparison of five different ML methods, where the methods are ranked according to MCC and it can be considered as one of the best measures in binary classification [22,52]. Among these methods, RF, ERT, and GB produced a similar performance with an MCC and accuracy of 0.546 and 0.772, respectively, which is slightly better than AB and significantly better than SVM. Therefore, we selected only four ML-based models (RF, ERT, AB, and GB) and applied feature selection protocol on these models.

Fig. 2.

Fig. 2

Performance of different ML-based models using the benchmarking data set. AAC: amino acid composition; DPC: dipeptide composition; CTD: chain-transition-distribution; AAI: amino acid index; PCP: physicochemical properties; H1: AAC + AAI; H2: AAC + DPC + AAI; H3: AAC + DPC + AAI + CTD; H4: AAC + DPC + AAI + CTD + PCP; H5: AAC + DPC; H6: AAC + CTD; H7: AAC + PCP; H8: AAI + DPC; H9: AAI + DPC + CTD; H10: AAI + DPC + CTD + PCP; H11: AAI + CTD; H12: AAI + PCP; H13: DPC + CTD; H14: DPC + CTD + PCP; H15: DPC + PCP; and H16: CTD + DPC.

Table 1.

The performance of the best model for each ML method obtained from different feature encodings.

Methods Features MCC Accuracy Sensitivity Specificity AUC
ERT H8 (420) 0.546 0.772 0.740 0.805 0.813
RF H5 (420) 0.546 0.776 0.829 0.724 0.805
GB H10 (577) 0.545 0.772 0.789 0.756 0.806
AB H5 (420) 0.531 0.764 0.715 0.813 0.767
SVM H4 (597) 0.457 0.728 0.772 0.683 0.746

The first column represents the method name developed in this study. The second column represents the hybrid model and its corresponding number of features. The third, fourth, fifith, sixth, and seventh columns, respectively, represent the MCC, accuracy, sensitivity, specificity, and AUC. RF: random forest; ERT: extra tree classifier; SVM: support vector machine; GB: gradient boosting; and AB: adaBoost.

3.2. Construction of iGHBP

To identify the most informative features that improves a prediction performance, a feature selection protocol was employed to remove noisy and redundant features [[53], [54], [55], [56]]. In an effort to construct the optimal or best predictive model, we applied a two-step feature selection protocol to identify an optimal feature set from the hybrid features that improves the prediction performance. In the first step, we applied the RF algorithm to rank the features, according to FIS, with hybrid features H5 (Fig. 3A), H8 (Fig. 3B) and H10 (Fig. 3C).

Fig. 3.

Fig. 3

Feature importance score computed for the hybrid features H5 (A), H8 (B) and H10 (C) using the RF algorithm.

SFS approach was used in the second step to select the optimal feature set from the ranked feature list. Fig. 3A shows the feature importance scores of 420-dimensional vectors. These features were ranked according to FIS and generated 420 feature sets (see methods). Each feature set was inputted to the ERT algorithm, and their corresponding models were developed using an LOOCV test. We plotted the SFS curve in Fig. 4A by using accuracy as Y-axis and feature number as X-axis. The maximum accuracy of 84.96% was observed with an optimal feature set of 190 features, while the other metrics such as MCC, sensitivity, specificity, and AUC are 0.701, 88.62, 81.30, and 0.896, respectively. Surprisingly, the obtained performance is identical to HBPred, where both methods use identical cross-validation methods and benchmarking data sets, however the number of features and the choice of ML algorithms are different. We also dramatically reduced the considered features from 420 to 190, indicating that our proposed feature selection technique could pick out the optimal dipeptides and AAI so as to further improve the prediction quality.

Fig. 4.

Fig. 4

SFS curve for discriminating GHBPs and non-GHBPs. (A) -. The maximum accuracy (i.e., SFS peak) obtained in leave-one-out cross-validation is shown in the red circle.

The above procedure was followed for other three methods (RF, GB, and AB). The best performance in terms of accuracy for RF, GB, and AB peaked at 80.5% (Fig. 4B), 81.7% (Fig. 4C), and 83.1% (Fig. 4D), respectively, with corresponding X-axis of 241, 161, and 167. These results show that a two-step feature selection protocol significantly improves the performances of the respective models. Next, we compared the performances of four different ML-based methods. To be specific, the accuracy of the ERT-based prediction model is ~1.9–4.4 higher than the other three methods, indicating the superiority of the ERT-based method in GHBP prediction. Hence, we named ERT-based prediction model as iGHBP.

3.3. Performance comparison between the optimal model and the control

To show the efficiency of our feature selection protocol, we compared the performance of the optimal model and the control without feature selection or using all features. Fig. 5 shows that our two-step feature selection protocol significantly improved the prediction performances of all four ML-based methods. Specifically, ERT, RF, GB and AB, whose accuracy values were respectively 7.7%, 2.9%, 4.5%, and 6.6% higher than the control, indicating an effectiveness of feature selection protocol. A similar protocol has been used in previous studies and has shown that the corresponding optimal models improved in performance [53,54,56,57].

Fig. 5.

Fig. 5

Performance comparison between the control (without feature selection) and optimal feature set-based models of four different ML algorithms. In the x-axis, normal and bold font respectively represent the control and the final model using the optimal feature set.

3.4. Analysis of feature selection

Although feature selection protocol significantly improved the performances of the respective ML-based methods, we specifically investigated the effectiveness of our feature selection protocol on ERT-based method (iGHBP). Here, we computed each feature average of GHBPs and non-GHBPs separately and compared their distribution for the hybrid features (Fig. 6A) and the optimal features (Fig. 6B). Results show that GHBPs and non-GHBPs were distributed more differentially in the feature space using optimal feature set when compared to the hybrid features, demonstrating why our feature descriptor led to the most informative prediction of GHBPs.

Fig. 6.

Fig. 6

Distribution of the GHBPs and non-GHBPs in the benchmarking data set using our hybrid features (A) and the optimal feature set (B).

3.5. Performance assessment for GHBP prediction based on the independent data set

Generally, it is essential to evaluate the proposed model using an independent data set to check whether the prediction model has generalisation capability or robustness [26]. In order to check the robustness of iGHBP, we further compared against three other ML methods developed in this study and against the state-of-the-art predictor (HBPred) on the independent data set. To make a fair comparison, we ensure lower sequence identities between the benchmarking and independent data sets, as it would otherwise lead to an overestimation of performance if the sequences in the independent data set had higher identities that those in the benchmarking data set. The results are summarised in Table 2, where the methods are ranked according to MCC. It can be observed that the proposed predictor iGHBP achieved the best performance with the following metrics with MCC, accuracy, specificity, and AUC, values of 0.646, 82.3%, 83.9, and 0.813, respectively. Specifically, the MCC and accuracy of iGHBP were 17.4–45% and 9.7–22.6% higher when compared to the other methods, thus demonstrating the superiority of iGHBP. Furthermore, we computed a pairwise comparison of AUCs between iGHBP and HBPred using two-tailedt test [58] and obtained the P-value of 0.009, demonstrating iGHBP significantly outperformed the HBPpred.

Table 2.

Performances of various methods on the independent data set.

Methods Features MCC Accuracy Sensitivity Specificity AUC
ERT 190 0.646 0.823 0.807 0.839 0.813
RF 241 0.472 0.726 0.871 0.581 0.777
GB 161 0.331 0.661 0.774 0.548 0.700
AB 167 0.324 0.661 0.613 0.710 0.675
HBPred 73 0.196 0.597 0.677 0.516 0.600

The first column represents the method name as used in this study. The second column represents the number of features present in the optimal feature set. The third, fourth, fifth, sixth and seventh columns, respectively, represent the MCC, accuracy, sensitivity, specificity, and AUC.

It is worth mentioning that both iGHBP and HBPred produced identical performance with the benchmarking data set, although there was variation in the input feature dimension and ML algorithm. However, only iGHBP produced a similar and consistent performance in both the benchmarking and independent data sets (Fig. 7, A and B), indicating that the current predictor is more stable and reliable. Notably, the optimal feature set contains 190 features, which is ~ 3-fold higher than the features used in the previous study. It is understandable that a larger and optimal feature set plays an important role in capturing the key components between the actual GHBPs and non-GHBPs and improve the performance. This is remarkable progress in biological research because a more reliable tool for the identification of biological macromolecules can vastly reduce the experimental cost. Hence, the iGHBP can be expected to be a tool with a high availability for the identification of GHBPs.

Fig. 7.

Fig. 7

Receiver operating characteristic curves of the various prediction models. (A) Leave-one-out cross-validation on the benchmarking data set and (B) independent data set. Higher AUC value indicates better performance of a particular method.

3.6. Web server implementation

As pointed out in [37] and shown in many follow-up publications [25,49,[59], [60], [61], [62], [63], [64], [65], [66], [67], [68], [69], [70], [71], [72]], user-friendly and publicly accessible web servers are the future of direction for developing more useful predictors. To this end, an online prediction server for iGHBP was developed, and it is available at www.thegleelab.org/iGHBP. All data sets utilized in the current study can be downloaded from our web server. Below, we give researchers a step-by-step guideline on how to use the webserver to get their desired results. In the first step, users need to submit the query sequences into the input box. Note that the input sequences should be in FASTA format. Examples of FASTA-formatted sequences can be seen by clicking on the button FASTA format above the input box. Finally, clicking on the button Submit, you will get the predicted results on the screen of your computer.

4. Conclusions

The biological significance of GHBPs has motivated the development of computational tools that facilitate accurate prediction. In this work, we developed a novel GHBP predictor called iGHBP. Here, we systematically assessed the use and performance of various composition-based features and their combinations along with various ML approaches in GHBP prediction. Our main findings are as follows: (i) Among five classifiers, ERT performed the best according to our performance measures (MCC, accuracy, and AUC), based on LOOCV. (ii) Of those five different compositions, an optimal feature set using a combination of DPC and AAI achieved the highest performance, emphasising the arrangement of particular local ordering dipeptides and biochemical properties. (iii) Experiment results from independent tests show that the proposed predictor iGHBP is more promising and effective for the GHBPs identification. As an application of this method, we have also made available an iGHBP webserver for the wider research community to use. It is expected that iGBHP will be a useful tool for discovering hypothetical GHBPs in a high-throughput and cost-effective manner, facilitating characterisation of their functional mechanisms. Furthermore, our proposed methods, along with the increasing availability of experimentally verified data and novel features, will greatly improve the prediction of GHBP.

Conflict of interest

The authors declare that there is no conflict of interest.

Authors Contributions

BM and GL conceived and designed the experiments. SB and BM performed the experiments. BM, SB, and TS analyzed the data. GL contributed reagents/materials/software tools. BM, SB, and GL wrote the manuscript.

Acknowledgements

This work was supported by the Basic Science Research Program through the National Research Foundation (NRF) of Korea funded by the Ministry of Education, Science, and Technology [2018R1D1A1B07049572 and 2018R1D1A1B07049494] and ICT & Future Planning [2016M3C7A1904392].

Contributor Information

Balachandran Manavalan, Email: bala@ajou.ac.kr.

Gwang Lee, Email: glee@ajou.ac.kr.

References

  • 1.Baumann G. Growth hormone binding protein. The soluble growth hormone receptor. Minerva Endocrinol. 2002;27:265–276. [PubMed] [Google Scholar]
  • 2.Fisker S. Physiology and pathophysiology of growth hormone-binding protein: methodological and clinical aspects. Growth Horm IGF Res. 2006;16:1–28. doi: 10.1016/j.ghir.2005.11.001. [DOI] [PubMed] [Google Scholar]
  • 3.Herington A.C., Ymer S., Stevenson J. Identification and characterization of specific binding proteins for growth hormone in normal human sera. J Clin Invest. 1986;77:1817–1823. doi: 10.1172/JCI112507. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Leung D.W., Spencer S.A., Cachianes G., Hammonds R.G., Collins C., Henzel W.J. Growth hormone receptor and serum binding protein: purification, cloning and expression. Nature. 1987;330:537–543. doi: 10.1038/330537a0. [DOI] [PubMed] [Google Scholar]
  • 5.Schilbach K., Bidlingmaier M. Growth hormone binding protein - physiological and analytical aspects. Best Pract Res Clin Endocrinol Metab. 2015;29:671–683. doi: 10.1016/j.beem.2015.06.004. [DOI] [PubMed] [Google Scholar]
  • 6.Wang X., He K., Gerhart M., Huang Y., Jiang J., Paxton R.J. Metalloprotease-mediatedGH receptor proteolysis and GHBP shedding. Determination of extracellular domain stem region cleavage site. J Biol Chem. 2002;277:50510–50519. doi: 10.1074/jbc.M208738200. [DOI] [PubMed] [Google Scholar]
  • 7.Saito Y., Teshima R., Takagi K., Ikebuchi H., Yamazaki T., Sawada J. Activation of protein kinase C alpha enhances human growth hormone-binding protein release. Mol Cell Endocrinol. 1998;146:197–205. doi: 10.1016/s0303-7207(98)00151-8. [DOI] [PubMed] [Google Scholar]
  • 8.Amit T., Youdim M.B., Hochberg Z. Clinical review 112: Does serum growth hormone (GH) binding protein reflect human GH receptor function? J Clin Endocrinol Metab. 2000;85:927–932. doi: 10.1210/jcem.85.3.6461. [DOI] [PubMed] [Google Scholar]
  • 9.Edens A., Talamantes F. Alternative processing of growth hormone receptor transcripts. Endocr Rev. 1998;19:559–582. doi: 10.1210/edrv.19.5.0347. [DOI] [PubMed] [Google Scholar]
  • 10.Sohm F., Manfroid I., Pezet A., Rentier-Delrue F., Rand-Weaver M., Kelly P.A. Identification and modulation of a growth hormone-binding protein in rainbow trout (Oncorhynchus mykiss) plasma during seawater adaptation. Gen Comp Endocrinol. 1998;111:216–224. doi: 10.1006/gcen.1998.7106. [DOI] [PubMed] [Google Scholar]
  • 11.Zhang Y., Marchant T.A. Identification of serum GH-binding proteins in the goldfish (Carassius auratus) and comparison with mammalian GH-binding proteins. J Endocrinol. 1999;161:255–262. doi: 10.1677/joe.0.1610255. [DOI] [PubMed] [Google Scholar]
  • 12.Einarsdottir I.E., Gong N., Jonsson E., Sundh H., Hasselberg-Frank L., Nilsen T.O. Plasma growth hormone-binding protein levels in Atlantic salmon Salmo salar during smoltification and seawater transfer. J Fish Biol. 2014;85:1279–1296. doi: 10.1111/jfb.12473. [DOI] [PubMed] [Google Scholar]
  • 13.Fisker S., Frystyk J., Skriver L., Vestbo E., Ho K.K., Orskov H. A simple, rapid immunometric assay for determination of functional and growth hormone-occupied growth hormone-binding protein in human serum. Eur J Clin Invest. 1996;26:779–785. doi: 10.1046/j.1365-2362.1996.2010558.x. [DOI] [PubMed] [Google Scholar]
  • 14.Tang H., Zhao Y.W., Zou P., Zhang C.M., Chen R., Huang P. HBPred: a tool to identify growth hormone-binding proteins. Int J Biol Sci. 2018;14:957–964. doi: 10.7150/ijbs.24174. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Huang Y., Niu B., Gao Y., Fu L., Li W. CD-HIT Suite: a web server for clustering and comparing biological sequences. Bioinformatics. 2010;26:680–682. doi: 10.1093/bioinformatics/btq003. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Dubchak I., Muchnik I., Holbrook S.R., Kim S.-H. Prediction of protein folding class using global description of amino acid sequence. Proc Natl Acad Sci. 1995;92:8700–8704. doi: 10.1073/pnas.92.19.8700. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Kawashima S., Pokarowski P., Pokarowska M., Kolinski A., Katayama T., Kanehisa M. AAindex: amino acid index database progress report 2008. Nucl Acids Res. 2007;36:D202–D205. doi: 10.1093/nar/gkm998. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Saha I., Maulik U., Bandyopadhyay S., Plewczynski D. Fuzzy clustering of physicochemical and biochemical properties of amino acids. Amino Acids. 2012;43:583–594. doi: 10.1007/s00726-011-1106-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Abraham A., Pedregosa F., Eickenberg M., Gervais P., Mueller A., Kossaifi J. Machine learning for neuroimaging with scikit-learn. Front Neuroinform. 2014;8:14. doi: 10.3389/fninf.2014.00014. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Breiman L. Random forests. Mach Learning. 2001;45:5–32. [Google Scholar]
  • 21.Geurts P., Ernst D., Wehenkel L. Extremely randomized trees. Mach Learning. 2006;63:3–42. [Google Scholar]
  • 22.Manavalan B., Basith S., Shin T.H., Choi S., Kim M.O., Lee G. MLACP: machine-learning-based prediction of anticancer peptides. Oncotarget. 2017;8:77121–77136. doi: 10.18632/oncotarget.20365. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Manavalan B., Subramaniyam S., Shin T.H., Kim M.O., Lee G. Machine-learning-based prediction of cell-penetrating peptides and their uptake efficiency with improved accuracy. J Proteome Res. 2018;17:2715–2726. doi: 10.1021/acs.jproteome.8b00148. [DOI] [PubMed] [Google Scholar]
  • 24.Cortes C., Vapnik V. Support-vector networks. Machine Learning. 1995;20:273–297. [Google Scholar]
  • 25.Feng C.Q., Zhang Z.Y., Zhu X.J., Lin Y., Chen W., Tang H. iTerm-PseKNC: a sequence-based tool for predicting bacterial transcriptional terminators. Bioinformatics. 2018 doi: 10.1093/bioinformatics/bty827. [DOI] [PubMed] [Google Scholar]
  • 26.Manavalan B., Lee J. SVMQA: support-vector-machine-based protein single-model quality assessment. Bioinformatics. 2017;33:2496–2503. doi: 10.1093/bioinformatics/btx222. [DOI] [PubMed] [Google Scholar]
  • 27.Manavalan B., Govindaraj R.G., Shin T.H., Kim M.O., Lee G. iBCE-EL: A new ensemble learning framework for improved linear B-cell epitope prediction. Front Immunol. 2018;9:1695. doi: 10.3389/fimmu.2018.01695. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Freund Y. Vol. 121. 1995. Boosting a weak learning algorithm by majority, Information and computation; pp. 256–285. [Google Scholar]
  • 29.Friedman J.H. Greedy function approximation: a gradient boosting machine. Annals of statistics. 2001:1189–1232. [Google Scholar]
  • 30.Rawi R., Mall R., Kunji K., Shen C.H., Kwong P.D., Chuang G.Y. PaRSnIP: sequence-based protein solubility prediction using gradient boosting machine. Bioinformatics. 2018;34:1092–1098. doi: 10.1093/bioinformatics/btx662. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31.Yang J., Cao R., Si D. Proceedings of the 2018 ACM International Conference on Bioinformatics, Computational Biology, and Health Informatics. ACM; 2018. EMNets: A convolutional autoencoder for protein surface retrieval based on cryo-electron microscopy imaging; pp. 639–644. [Google Scholar]
  • 32.Cao R., Adhikari B., Bhattacharya D., Sun M., Hou J., Cheng J. QAcon: single model quality assessment using protein structural and contact information with machine learning techniques. Bioinformatics. 2017;33:586–588. doi: 10.1093/bioinformatics/btw694. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Cao R., Bhattacharya D., Hou J., Cheng J. DeepQA: improving the estimation of single protein model quality with deep belief networks. BMC Bioinform. 2016;17:495. doi: 10.1186/s12859-016-1405-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Cao R., Freitas C., Chan L., Sun M., Jiang H., Chen Z. ProLanGO: Protein function prediction using neural machine translation based on a recurrent neural network. Molecules. 2017;22:1732. doi: 10.3390/molecules22101732. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35.Tang H., Cao R.-Z., Wang W., Liu T.-S., Wang L.-M., He C.-M. A two-step discriminated method to identify thermophilic proteins. Int J Biomath. 2017;10:1750050. [Google Scholar]
  • 36.Stephenson N., Shane E., Chase J., Rowland J., Ries D., Justice N. Survey of machine learning techniques in drug discovery. Curr Drug Metabol. 2018;273:236–247. doi: 10.2174/1389200219666180820112457. [DOI] [PubMed] [Google Scholar]
  • 37.Chou K.-C. Some remarks on protein attribute prediction and pseudo amino acid composition. J Theor Biol. 2011;273:236–247. doi: 10.1016/j.jtbi.2010.12.024. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 38.Chen W., Feng P., Yang H., Ding H., Lin H., Chou K.C. iRNA-3typeA: Identifying three types of modification at RNA's adenosine sites. Mol Ther Nucleic Acids. 2018;11:468–474. doi: 10.1016/j.omtn.2018.03.012. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39.Chen W., Feng P.M., Lin H., Chou K.C. iRSpot-PseDNC: identify recombination spots with pseudo dinucleotide composition. Nucleic Acids Res. 2013;41:e68. doi: 10.1093/nar/gks1450. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40.Chen W., Tang H., Ye J., Lin H., Chou K.C. iRNA-PseU: Identifying RNA pseudouridine sites. Mol Ther Nucleic Acids. 2016;5:e332. doi: 10.1038/mtna.2016.37. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41.Feng P.M., Chen W., Lin H., Chou K.C. iHSP-PseRAAAC: Identifying the heat shock protein families using pseudo reduced amino acid alphabet composition. Anal Biochem. 2013;442:118–125. doi: 10.1016/j.ab.2013.05.024. [DOI] [PubMed] [Google Scholar]
  • 42.Chen W., Yang H., Feng P., Ding H., Lin H. iDNA4mC: identifying DNA N4-methylcytosine sites based on nucleotide chemical properties. Bioinformatics. 2017;33:3518–3523. doi: 10.1093/bioinformatics/btx479. [DOI] [PubMed] [Google Scholar]
  • 43.Lai H.Y., Chen X.X., Chen W., Tang H., Lin H. Sequence-based predictive modeling to identify cancerlectins. Oncotarget. 2017;8:28169–28175. doi: 10.18632/oncotarget.15963. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 44.Lin H., Ding C., Song Q., Yang P., Ding H., Deng K.J. The prediction of protein structural class using averaged chemical shifts. J Biomol Struct Dyn. 2012;29:643–649. doi: 10.1080/07391102.2011.672628. [DOI] [PubMed] [Google Scholar]
  • 45.Lin H., Liang Z.Y., Tang H., Chen W. Identifying sigma70 promoters with novel pseudo nucleotide composition. IEEE/ACM Trans Comput Biol Bioinform. 2017;2016:1–7. doi: 10.1109/TCBB.2017.2666141. [DOI] [PubMed] [Google Scholar]
  • 46.Yang H., Tang H., Chen X.X., Zhang C.J., Zhu P.P., Ding H. Identification of secretory proteins in Mycobacterium tuberculosis using pseudo amino acid composition. Biomed Res Int. 2016;(2016):5413903. doi: 10.1155/2016/5413903. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 47.Zhao Y.W., Su Z.D., Yang W., Lin H., Chen W., Tang H. IonchanPred 2.0: A tool to predict ion channels and their types. Int J Mol Sci. 2017;18:1838. doi: 10.3390/ijms18091838. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 48.Cao R., Cheng J. Protein single-model quality assessment by feature-based probability density functions. Sci Rep. 2016;6:23990. doi: 10.1038/srep23990. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 49.Manavalan B., Shin T.H., Kim M.O., Lee G. AIPpred: Sequence-based prediction of anti-inflammatory peptides using random forest. Front Pharmacol. 2018;9:276. doi: 10.3389/fphar.2018.00276. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 50.Manavalan B., Shin T.H., Lee G. PVP-SVM: Sequence-based prediction of phage virion proteins using a support vector machine. Front Microbiol. 2018;9:476. doi: 10.3389/fmicb.2018.00476. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 51.Manavalan B., Shin T.H., Lee G. DHSpred: support-vector-machine-based human DNase I hypersensitive sites prediction using the optimal features selected by random forest. Oncotarget. 2018;9:1944–1956. doi: 10.18632/oncotarget.23099. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 52.Manavalan B., Shin T.H., Kim M.O., Lee G. PIP-EL: A new ensemble learning method for improved proinflammatory peptide predictions. Front Immunol. 2018;9:1783. doi: 10.3389/fimmu.2018.01783. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 53.Wei L., Zhou C., Chen H., Song J., Su R. ACPred-FL: a sequence-based predictor based on effective feature representation to improve the prediction of anti-cancer peptides. Bioinformatics. 2018 doi: 10.1093/bioinformatics/bty451. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 54.Wei L., Xing P., Su R., Shi G., Ma Z.S., Zou Q. CPPred-RF: a sequence-based predictor for identifying cell-penetrating peptides and their uptake efficiency. J Proteome Res. 2017;16:2044–2053. doi: 10.1021/acs.jproteome.7b00019. [DOI] [PubMed] [Google Scholar]
  • 55.Wei L., Tang J., Zou Q. SkipCPP-Pred: an improved and promising sequence-based predictor for predicting cell-penetrating peptides. BMC Genom. 2017;18:1. doi: 10.1186/s12864-017-4128-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 56.Wei L., Chen H., Su R. M6APred-EL: A sequence-based predictor for identifying N6-methyladenosine sites using ensemble learning. Mol Ther Nucl Acids. 2018:635–644. doi: 10.1016/j.omtn.2018.07.004. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 57.Yang H., Qiu W.-R., Liu G., Guo F.-B., Chen W., Chou K.-C. iRSpot-Pse6NC: Identifying recombination spots in Saccharomyces cerevisiae by incorporating hexamer composition into general PseKNC. Int J Biol Sci. 2018;14:883. doi: 10.7150/ijbs.24616. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 58.Hanley J.A., McNeil B.J. The meaning and use of the area under a receiver operating characteristic (ROC) curve. Radiology. 1982;143:29–36. doi: 10.1148/radiology.143.1.7063747. [DOI] [PubMed] [Google Scholar]
  • 59.Manavalan B., Lee J., Lee J. Random forest-based protein model quality assessment (RFMQA) using structural features and potential energy terms. PLoS One. 2014;9:e106542. doi: 10.1371/journal.pone.0106542. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 60.Manavalan B., Govindaraj R., Lee G., Choi S. Molecular modeling-based evaluation of dual function of IkappaBzeta ankyrin repeat domain in toll-like receptor signaling. J Mol Recognit. 2011;24:597–607. doi: 10.1002/jmr.1085. [DOI] [PubMed] [Google Scholar]
  • 61.Manavalan B., Basith S., Choi Y.M., Lee G., Choi S. Structure-function relationship of cytoplasmic and nuclear IkappaB proteins: an in silico analysis. PLoS One. 2010;5:e15782. doi: 10.1371/journal.pone.0015782. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 62.Basith S., Manavalan B., Govindaraj R.G., Choi S. In silico approach to inhibition of signaling pathways of Toll-like receptors 2 and 4 by ST2L. PLoS One. 2011;6:e23989. doi: 10.1371/journal.pone.0023989. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 63.Basith S., Manavalan B., Gosu V., Choi S. Evolutionary, structural and functional interplay of the IkappaB family members. PLoS One. 2013;8:e54178. doi: 10.1371/journal.pone.0054178. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 64.Wei L., Xing P., Tang J., Zou Q. PhosPred-RF: a novel sequence-based predictor for phosphorylation sites using sequential information only. IEEE Trans Nanobiosci. 2017;16:240–247. doi: 10.1109/TNB.2017.2661756. [DOI] [PubMed] [Google Scholar]
  • 65.Wei L., Xing P., Shi G., Ji Z.-L., Zou Q. Fast prediction of protein methylation sites using a sequence-based feature selection technique. IEEE/ACM Trans Comput Biol Bioinform. 2017 doi: 10.1109/TCBB.2017.2670558. [DOI] [PubMed] [Google Scholar]
  • 66.Wei L., Tang J., Zou Q. Local-DPP: An improved DNA-binding protein prediction method by exploring local evolutionary information. Information Sci. 2017;384:135–144. [Google Scholar]
  • 67.Wei L., Su R., Wang B., Li X., Zou Q. Integration of deep feature representations and handcrafted features to improve the prediction of N 6-methyladenosine sites. Neurocomputing. 2018 [Google Scholar]
  • 68.Su R., Wu H., Xu B., Liu X., Wei L. Developing a multi-dose computational model for drug-induced hepatotoxicity prediction based on toxicogenomics data. IEEE/ACM Trans Comput Biol Bioinform. 2018 doi: 10.1109/TCBB.2018.2858756. [DOI] [PubMed] [Google Scholar]
  • 69.Bhattacharya D., Nowotny J., Cao R., Cheng J. 3Drefine: an interactive web server for efficient protein structure refinement. Nucleic Acids Res. 2016;44:W406–W409. doi: 10.1093/nar/gkw336. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 70.Wei L., Luan S., Nagai L.A.E., Su R., Zou Q. Exploring sequence-based features for the improved prediction of DNA N4-methylcytosine sites in multiple species. Bioinformatics. 2018 doi: 10.1093/bioinformatics/bty824. [DOI] [PubMed] [Google Scholar]
  • 71.Qiang X., Zhou C., Ye X., Du P.F., Su R., Wei L. CPPred-FL: a sequence-based predictor for large-scale identification of cell-penetrating peptides by feature representation learning. Brief Bioinform. 2018 doi: 10.1093/bib/bby091. [DOI] [PubMed] [Google Scholar]
  • 72.Bhattacharya D., Cao R., Cheng J. UniCon3D: de novo protein structure prediction using united-residue conformational search via stepwise, probabilistic sampling. Bioinformatics. 2016;32:2791–2799. doi: 10.1093/bioinformatics/btw316. [DOI] [PMC free article] [PubMed] [Google Scholar]

Articles from Computational and Structural Biotechnology Journal are provided here courtesy of Research Network of Computational and Structural Biotechnology

RESOURCES