Abstract
Background
Several computational tools for predicting protein Ubiquitylation and SUMOylation sites have been proposed to study their regulatory roles in gene location, gene expression, and genome replication. However, existing methods generally rely on feature engineering, and ignore the natural similarity between the two types of protein translational modification. This study is the first all-in-one deep network to predict protein Ubiquitylation and SUMOylation sites from protein sequences as well as their crosstalk sites simultaneously. Our deep learning architecture integrates several meta classifiers that apply deep neural networks to protein sequence information and physico-chemical properties, which were trained on multi-label classification mode for simultaneously identifying protein Ubiquitylation and SUMOylation as well as their crosstalk sites.
Results
The promising AUCs of our method on Ubiquitylation, SUMOylation and crosstalk sites achieved 0.838, 0.888, and 0.862 respectively on tenfold cross-validation. The corresponding APs reached 0.683, 0.804 and 0.552, which also validated our effectiveness.
Conclusions
The proposed architecture managed to classify ubiquitylated and SUMOylated lysine residues along with their crosstalk sites, and outperformed other well-known Ubiquitylation and SUMOylation site prediction tools.
Keywords: Protein ubiquitylation site, Protein SUMOylation site, Convolution neural network, Deep learning, Ensemble learning
Background
Ubiquitin [1, 2] is a small protein composed of 76 amino acids in eukaryotes. Through the catalytic action of activating enzyme (E1), binding enzyme (E2), and ligase (E3) [3, 4], ubiquitins can covalently connect to the lysine residues of the target proteins [5, 6]. As a major member of the family, small ubiquitin-related modifier (SUMO) proteins have similar 3D structures and biological modification processes to ubiquitins [7, 8]. They are both highly conserved in evolution and related to diverse cellular activities including gene location, gene expression, and genome replication [9]. However, numerous potential Ubiquitylation and SUMOylation sites remain to be discovered from protein sequences.
Since most ubiquitinated and SUMOylated proteins are short-lived proteins with poor stability, the experimental approaches to identify protein Ubiquitylation and SUMOlytion sites might be costly and time-consuming [10]. Therefore, it is worthwhile to study the computational approaches.
At present, several sequence-based approaches have been proposed to carry out the prediction of protein Ubiquitylation and SUMOylation sites respectively. Huang et al. [11] developed a method called UbiSite, using an efficient radial basis function (RBF) network to identify protein Ubiquitylation sites. Next, Chen et al. [12] established UbiProber, which extracted a set of features including physico-chemical property (PCP) and amino acid composition(AAC) to make Ubiquitylation site prediction. Subsequently, Radivojac et al. [13] proposed a random-forest based predictor UbPred, in which 586 sequence attributes were detected from the input features. GPS-sumo [14] employed a group-based prediction system (GPS) by a similarity clustering strategy to identify SUMOlytion sites. JASSA by Guillaume et al. [15] uses a scoring system based on a position frequency matrix. Then, pSumo-cd [16] applied a covariance discriminant algorithm in combination with a pseudo amino acid composition model. A recent work HseSUMO [17] only employed four half-sphere exposure-based features to predict SUMOylation sites. In addition to the individual prediction of Ubiquitylation or SUMOylation sites, mUSP was proposed to predict their crosstalk. They treated these three types as three binary problems independently. However, these traditional machine learning methods employed feature engineering, which may lead to incomplete representations and biased results.
Deep learning as a cutting-edge representation learning technique enables the production of high-level semantic features without handcrafted design [18], it has been widely applied to several PTM problems with large datasets [19, 20]. Wang et al. [21, 22] proposed a deep learning predictor MusiteDeep, based on convolutional neural networks, to predict and visualize protein post translational modification sites. Chen et al. [23] built a computation model, MUscADEL, based on the long short term memory (LSTM) recurrent neural network. Fu et al. [24] used Matlab to implement deepUbi, a protein Ubiquitylation site prediction tool. Due to that its backend Matlab is a closed commercial software, its availability is limited. Although deep learning has been applied to PTM problems, the similarities between the two PTMs have not been recognized or fully exploited. To our best knowledge, there is no site prediction tool based on deep learning to predict protein Ubiquitylation and SUMOylation sites simultaneously.
In this paper, we proposed an ensemble deep learning based predictor for identifying protein Ubiquitylation and SUMOylation sites as well as their crosstalk sites simultaneously. The ensemble learning layer integrated different types of physico-chemical properties of amino acids. The network can learn the high-level representation from the raw protein sequence and its corresponding physico-chemical properties. Owing to the similarity of biochemical processes of Ubiquitylation and SUMOylation, the Ubiquitylation and SUMOylation datasets were used for training simultaneously, which not only circumvents the scarcity of training data but also endows the model with more discerning power.
Result
Overview
Figure 1 provides an overview of our workflow. We formulated the process of protein Ubiquitylation and SUMOylation site prediction as a multi-label classification problem. First, we collected the protein sequences of Ubiquitylation and SUMOylation from UniProt/Swiss-Prot, used CD-HIT to remove the redundant sequences that have more than 40 sequence identity, and split the remaining data into the training set, validation set, and testing set. Next, all the fragments of sequences were encoded and inputted into seven respective deep networks. Then, we proposed an ensemble learning layer to integrate multiple protein representations as shown in Fig. 2. It integrates seven supervised learning subnets, each of which utilized convolution layers or fully connected layers, to extract deep representations from protein sequence features. At last, since we targeted two categories for multi-label classification, we defined the output of our tool using dummy code, in which Ubiquitylation and SUMOylation sites independently associated with different labels. The 2-dimensional code 10 was set to represent Ubiquitylation sites and 01 was assigned to SUMOylation sites, while code 11 denoted the crosstalk (both Ubiquitylation and SUMOylation) sites and code 00 was encoded for negatives. The output layer of our deep model was set to 2 neurons to generate multi-label results by using the sigmoid [25] activation function, which independently produced a probability for each category.
Comparative results among the ensemble model and seven meta classifiers
We compare between the ensemble model and seven meta classifiers to provide deeper insight into the advantage of our ensemble learning strategy. As we have pretrained each meta subnet before integrating to the ensemble architecture, the performance of the meta classifiers can be easily assessed on the same test data by loading pretrained weights. The comparative results are shown in Fig. 3.
From this figure, we can observe the meta classifiers showed varying degrees of effectiveness, and offered deep representations from different perspectives. Such meta classifiers with sufficient precision and diversity provided a good ensemble foundation. In addition, these meta classifiers performed differently between protein Ubiquitylation and SUMOylation sites. For instance, the physico-chemical subnet performed top-3 rank out of the 7 meta classifiers on Ubiquitylation sites while it ranked as last of the 7 meta-classifiers on SUMOylation sites. This demonstrated how an adaptive ensemble was required to properly combine all meta classifiers for different categories.
Results of protein Ubiquitylation and SUMOylation sites prediction
We compared our method with several popular and accessible protein ubiquitination and SUMOylation site prediction tools (Ubisite [11], Ubiprober [12], Ubpred [13], psumo-cd [16], JASSA [15], sumoplot [26], GPSsumo [14], and MUscADEL) [23] by submitting our testing dataset to their websites. Their performance was plotted as ROC and PR curves in Fig. 4. Our AUC values were 0.838 on Ubiquitylation site prediction and 0.888 on SUMOylation site prediction respectively. A similar situation appeared on the PR curves, where the AP value of Ubiquitylation site prediction was 0.683 and the AP value of SUMOylation site prediction was 0.804. As shown in Fig. 4, the performance of the proposed deep learning architecture was superior to other protein Ubiquitylation and SUMOylation site prediction tools for each measure.
The graphical sequence logo was generated by the WebLogo tool to visualize amino acid residue conservation at a given position as Fig. 5. Amino acid residues Glu (E), Glu(G), Lys(K), Leu (L) appeared more frequently in positive samples of Ubiquitylation fragments, while Glu (E), Leu (L), Pro(P), Arg(R) were more enriched in positive samples of SUMOylation fragments. The results indicated the dependencies of upstream and downstream amino acid sites, which is consistent with the article of Chen et al. [27].
t-SNE [28] plot was employed to visualize the discriminating ability of the raw inputs and merged deep representations from three classes as Fig. 6 shown. Different colors represented different classes. It clearly showed that the distributions of original features were disordered and messy. After mappings of multiple hidden layers, the sample distribution tended to separate, which implied that our multi-label classification model may detect distinguishing representations and fuse seven subnets to further enhance the discriminative ability of our model. But in the meanwhile, some overlaps resulting in not complete distinctive boundaries also can be observed from the t-SNE plot. We reasoned such heterogeneous samples located closely at feature space implied crosstalk sites and some potential unlabeled positive samples. We investigated the crosstalk sites from our experimental data and found they accounted for nearly 2% out of total Ubiquitylation and SUMOylation sites. Such crosstalk samples reflected characteristics of Ubiquitylation and SUMOylation sites and were marked either Ubiquitylation sites (in red) or SUMOylation sites (in blue) in Fig. 6. Therefore, the overlaps between red and blue dots likely represented the crosstalk sites. Almost negatives (in green) in Fig. 6 were concentrated in a region except a small part scattered at the zones enriched in red dots and blue dots.
Compared with dimensional reduction using PCA
Principal component analysis (PCA) is a popular feature selection method that conducts a linear transformation to convert the original variables to a set of new orthogonal variables [29], which enables to avoid manual feature selection. Venkatarajan et al. reduced the multidimensional scaling of 237 physicochemical properties to 5-dimensional representations by using PCA [30]. We used this method to reduce the dimension from 402 to 6 as the original input, to include the principle componenets of physical-chemical properties for comparisions (Fig. 7).
Independent dataset results
For fair evaluations, we built another independent set whose protein sequences were collected from UniProt/Swiss-Prot database updated after November 2020. At that point we had completed our data collection and all other tools had been published. Thereby, this dataset never appeared in our tool nor others, and served as an independent comparison. The details of this newly collected dataset are summarized in Additional file 1: Table S1. The performance of our method and others on the independent set is shown in Fig. 8. We also obtained an AUC and AP of 0.765 and 0.441 on crosstalk site prediction respectively.
Discussion
In our pipeline, ensemble learning is used to simultaneously identify protein Ubiquitylation sites and SUMOylation sites as well as their crosstalk sites. Different from common ensemble approaches of using a simple average or predefined weights, our ensemble subnet enabled learning combined weights in a data-driven fashion.The led us to outperform other meta classifiers on both Ubiquitylation site prediction and SUMOylation site prediction.
Since we used the ensemble layer to synthesize two types of input features, the model can adaptively learn effective features. In addition, two PTMs mutually supported and boosted the multi-label prediction performance. Because crosstalk cannot be positive in both categories for ROC and PR curves, we calculated its AUC and AP of 0.862 and 0.552 respectively.
We also explored overlapping samples with similar deep features in t-SME to Ubiquitylation sites and/or SUMOylation sites that were potentially unrevealed positive sites, since only a small fraction of protein post-translational modification (PTM) sites were experimentally annotated in the records from Swiss-Prot [31].
The comparative results of PCA showed our deep architecture enables to generate more informative representations. As shown in Fig. 7, our model can obtain more potential associations between features and better classification results.
Our tool showed stronger discerning power than all other listed tools on the unseen samples in terms of AUC and AP. This can be attributed in part to adaptively integrating richer input modalities with seven descriptors and applying bootstrapping strategy to balance positive and negative samples in modeling.
Methods
Benchmark dataset
We built a benchmark dataset by collecting annotations from UniProt/Swiss-Prot (Nov 2020 release) [32]. This database provides high-quality protein sequences and manual annotations, including the descriptions of amino acid residue modification. To avoid the overestimation of model performance caused by homogenous sequences, we used the Cluster Database at High Identity with Tolerance (CD-HIT) [33] to remove proteins that have more than 40 sequence identity. Afterwards, 1983 proteins of Ubiquitylation and 4728 proteins of SUMOylation remained. Experimentally validated lysine (K) residues, based on the annotation from UniProt/Swiss-Prot, were taken as positive samples. The rest of the lysine (K) residues in the proteins were regarded as negative samples. In total, we obtained 4222 ubiquitylated sites, 56544 non-ubiquitylated sites, 16432 SUMOylated sites, 203533 non-SUMOylated sites. Then, we retrieved the crosstalk sites from Swiss-Prot database by the keyword ‘cross-link’. Through this search, we collected 401 crosstalk sites for our basic datasets. We organized all details of the datasets including the number of Ubiquitylation and SUMOylation sites, and the ratio between positive and negative samples into the Additional file 2: Table S2. For the further details of the involved proteins, we saved all sequence names and fragments of the retrieved proteins in Additional file 3: Tables S3 and Additional file 4: Table S4 respectively.
In this study, we employed tenfold cross-validation to evaluate the performance of the model. In this process, all the proteins of Ubiquitylation and SUMOylation were partitioned into 10 equal parts. The ratio of the training, validation and testing sets was 8:1:1. The details of the tenfold cross-validation dataset and the independent testing set are listed in the Additional file 1: Table S1 and Additional file 2: Table S2.
According to our previous grid search, a sliding window with a length of 242+1=49 to intercept the protein sequence containing lysine residues (K) in the middle was optimal to deliver robust Ubiquitylation site prediction. The same settings were applied in this study consider the similarity between protein Ubiquitylation and SUMOylation. The details of optimized window size is explored in our previous work [34]. If the number of upstream and downstream positions was less than 24, then the placeholder was used to supplement. Moreover, the identifier “X” (unknown) was used to represent amino acids in the sequence that was not recognized by current sequencing techniques. “X” will be assigned an average value of 20 amino acids.
Encoding of protein fragments
The following two types of encoding were adopted to encode the amino acid composition of the original protein fragment [35].
The first encoding converted 20 amino acids and one placeholder to a binary feature matrix. The corresponding state of the amino acids on each vector was 1, and the remaining indeices were 0. All of the 49-length amino acid fragments were then organized as a matrix of size 49*21.
We also utilize physical-chemical properties (PCPs) encoding, which can be found and downloaded from the AAindex database [36]. In this study, all physical-chemical properties were divided into six highly correlated clusters. Then, each sequence fragment was coded into 6 two-dimensional (2D) matrices. The details of six physical-chemical properties are shown in Table 1.
Table 1.
Subnets | Layer category | Hyper-parameters | |||
---|---|---|---|---|---|
Activation function | Size | Filters | Dropout | ||
Sequence | 1D Convolution | Relu | 2 | 201 | 0.4 |
Relu | 3 | 151 | 0.4 | ||
Relu | 5 | 101 | 0.4 | ||
Dense | Relu | 256 | – | 0.3 | |
Relu | 128 | – | 0 | ||
Sigmoid | 2 | – | – | ||
Physico-O | Dense | Relu | 256 | – | 0.2 |
Relu | 128 | – | 0.1 | ||
Sigmoid | 2 | – | – | ||
Physico-P | Dense | Relu | 512 | – | 0.3 |
Relu | 256 | – | 0.2 | ||
Relu | 128 | – | 0.1 | ||
Sigmoid | 2 | – | – | ||
Physico-H | Dense | Relu | 1024 | – | 0.4 |
Relu | 512 | – | 0.3 | ||
Relu | 256 | – | 0.2 | ||
Relu | 128 | – | 0.1 | ||
Sigmoid | 2 | – | – | ||
Physico-C | 1D Convolution | Relu | 2 | 201 | 0.2 |
Relu | 3 | 151 | 0.1 | ||
Dense | Sigmoid | 2 | – | – | |
Physico-B | 1D Convolution | Relu | 2 | 201 | 0.3 |
Relu | 3 | 151 | 0.2 | ||
Relu | 5 | 101 | 0.1 | ||
Dense | Sigmoid | 2 | – | – | |
Physico-A | 1D Convolution | Relu | 2 | 201 | 0.4 |
Relu | 3 | 151 | 0.3 | ||
Relu | 5 | 101 | 0.2 | ||
Relu | 7 | 51 | 0.1 | ||
Dense | Sigmoid | 2 | – | – | |
Ensemble | Dense | Relu | 7 | – | – |
Sigmoid | 2 | – | – |
Deep learning architecture
Our deep learning architecture consists of seven subnets to handle seven input modalities (one-hot and six physical-chemical properties encoding matrices). The structure and detailed hyper-parameters of these subnets can be found in Table 1.
The first subnet was designed to extract the internal correlation between adjacent amino acids, and highlight the meaningful part of the feature maps. Thereafter, we merged all the newly generated feature maps with three dense layers, to produce a low dimensional feature representation. The rest of the subnets would each take one of the six groups of the physical-chemical properties defined by Tomii et al. in the AAindex database [36]. These included alpha and turn propensities, beta propensity, composition, hydrophobicity, physicochemical properties, and other properties. According to the dimension of the input feature matrices, the layers and hyperparameters of the network structure were adjusted accordingly.
These subnets separately detected intra-correlations and generated deep representations for each group of physical-chemical properties. Baseline experiments were conducted by training convolution layers and fully-connected layer with sigmoid activation function and nesterov adaptive moment estimation (Nadam) optimization algorithm, and a categorical cross-entropy loss function [37]. Since the number of samples between each class and negatives were imbalanced, a class-weight was set inversely proportional to the number of samples in the class to equalize the contribution of each class and the negatives. The class weights were calculated as follows:
1 |
The number of samples for each class is represented by n and C represents the weight of the class i. category. And binary cross-entropy [37] was used as the loss function to drive model fitting. The output layer independently maps the embedding from previous layers to generate two probabilities for Ubiquitylation and SUMOylation via the sigmoid function. All models were respectively trained using a maximum of 100 epochs and stopped early if there was no further improvement in loss for any 10 continuous epochs.
Ensemble learning
Our deep learning predictor incorporated an ensemble learning strategy, to predict protein Ubiquitylation sites and SUMOylation sites precisely and conveniently as well as their crosstalk sites. The seven well-trained subnets can be considered as seven meta classifiers for the parallel completion for the prediction task [38]. We included an additional fully connected layer to integrate the outputs from the seven subnets. Such stacking-based ensemble learning enables us to adaptively coordinate inter-class diverse meta-learners and generate better predictions [39, 40]. When training the whole ensemble network, we loaded the pretrained weights of all layers before the logits of each meta subnet respectively, and carried out the training procedure with the same training settings.
Since only a small number of lysine post-translational modification sites occur in protein sequences, the distribution of positive and negative samples is extremely imbalanced. Therefore, we employed bootstrapping for resampling the training data.This can help generate a more stable and unbiased model. Assuming that pos and neg represented the number of positive samples and negative samples, bootstrapping randomly selected pos positive samples and neg negative samples during the sampling process to form a balanced training set. Therefore, the network can be trained N (N= the number of negatives / the number of positives) times to learn the weight. In this study, by randomly resampling the negative samples in equal proportionto the positive number, we were able to balance positive samples (including Ubiquitylation sites and SUMOylation sites) and negative samples involved in each training iteration. According to the distribution between positives and negatives, as seen in Additional file 2: Table S2, such boostrapping procedure would go through 13 iterations in a training epoch to take as many negatives as possible in training. The specific data points from each class after balanced are described in Additional file 5: Table S5.
Conclusion
In this paper, we proposed a novel ensemble deep learning based predictor for simultaneously identifying protein Ubiquitylation sites and SUMOylation sites as well as their crosstalk sites. Overall, the highlight of our method is mainly due to its data-driven feature, multi-label formulation for Ubiquitylation and SUMOylation sites, and ensemble learning. Because of the natural structural and functional similarity between Ubiquitylation and SUMOylation, the data regarding two PTMs supported each other and boosted the multi-label prediction performance. The designed ensemble learning layer synthesized the results of multiple meta classifiers and avoided the possibility of poor generalization performance due to a single classifier.
By comparing with the results of similar tools, our ROC curves and PR curves were stable at a higher level. This demonstrated the effectiveness of our method and robustness of the ensemble models, and reflected the potential of the deep learning algorithm in the field of Ubiquitylation and SUMOylation protein sites prediction. Because the input of our architecture is not particularly designed for Ubiquitylation and SUMOylation, our architecture can extend to other PTMs easily without any adjustment. Further research will explore incorporating newly updated Ubiquitylation and SUMOlytion data to incremenetally upgrade our model, and extending our architecture to other types of PTMs.
Supplementary information
Acknowledgements
We thank Joshua Thompson for his language edits and all the staff of the journal for their efforts and those who have helped us in the course of our work.
Abbreviations
- ROC
Receiver operation characteristic
- AUC
Area under the receiver operating characteristic curve
- PR
Precision-Recall
- AP
Average precision
- SUMO
Small ubiquitin-like modifier
- t-SNE
t-distributed stochastic neighbor embedding
- PCA
Principal component analysis
Authors' contributions
FH designed the method ,conceived the algorithm and guided the work ideas. JYL implemented the experiment, wrote the manuscript. RW prepared the datasets. XWZ solved the work technical problems. YH revised the manuscript. Everyone finally reviewed the manuscript. All authors read and approved the final manuscript.
Funding
This research was funded by National Natural Science Funds of China (61802057), Jilin Provincial Natural Science Foundation (20210101174JC), and the Fundamental Research Funds for the Central Universities, JLU (93K172020K22) to F.H., Jilin Provincial Natural Science Foundation (20200201288JC) and the science and technology research project of “13th Five-Year” of the Education Department of Jilin province (JJKH20200330KJ) to Y.H. All the funding bodies were not involved in the design of the study and collection, analysis, and interpretation of data and in writing the manuscript.
Availability of data and materials
The source code and the datasets used in this work are available at https://github.com/lijingyimm/MultiUbiSUMO.
Declarations
Ethics approval and consent to participate
Not applicable.
Consent for publication
Not applicable.
Competing interests
The authors declare that they have no competing interests.
Footnotes
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Contributor Information
Fei He, Email: hef740@nenu.edu.cn.
Jingyi Li, Email: lijy801@nenu.edu.cn.
Rui Wang, Email: wangr921@nenu.edu.cn.
Xiaowei Zhao, Email: zhaoxw303@nenu.edu.cn.
Ye Han, Email: yeh@jlau.edu.cn.
Supplementary Information
The online version contains supplementary material available at 10.1186/s12859-021-04445-5.
References
- 1.Goldstein G, Scheid M, Hammerling U, Boyse EA, Niall SHD. Isolation of a polypeptide that has lymphocyte-differentiating properties and is probably represented universally in living cells. Proc Natl Acad Sci USA. 1975;72(1):11–15. doi: 10.1073/pnas.72.1.11. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Wilkinson KD. The discovery of ubiquitin-dependent proteolysis. Proc Natl Acad Sci USA. 2005;102(43):15280–15282. doi: 10.1073/pnas.0504842102. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Pickart CM, Eddins MJ. Ubiquitin: structures, functions, mechanisms. Biochim Et Biophys Acta Mol Cell Res. 2004. [DOI] [PubMed]
- 4.Welchman RL, Gordon C, Mayer RJ. Ubiquitin and ubiquitin-like proteins as multifunctional signals. Nat Rev Mol Cell Biol; 2005. [DOI] [PubMed]
- 5.Bienko M, Green CM, Crosetto N, Rudolf F, Zapart G, Coull B, Kannouche P, Wider G, Peter M, Lehmann AR. Ubiquitin-binding domains in y-family polymerases regulate translesion synthesis. Science. 2005;310(5755):1821–4. doi: 10.1126/science.1120615. [DOI] [PubMed] [Google Scholar]
- 6.Peng J, Schwartz D, Elias JE, Thoreen CC, Cheng D, Marsischky G, Roelofs J, Finley D, Gygi SP. A proteomics approach to understanding protein ubiquitination. Nat Biotechnol. 2003;21(8):921–6. doi: 10.1038/nbt849. [DOI] [PubMed] [Google Scholar]
- 7.Geiss-Friedlander R, Melchior F. Concepts in sumoylation: a decade on. Nat Rev Mol Cell Biol. 2007;8(12):947–56. doi: 10.1038/nrm2293. [DOI] [PubMed] [Google Scholar]
- 8.Gill G. Sumo and ubiquitin in the nucleus: different functions, similar mechanisms? Genes Dev. 2004;18(17):2046–59. doi: 10.1101/gad.1214604. [DOI] [PubMed] [Google Scholar]
- 9.Seeler JS, Dejean A. Nuclear and unclear functions of sumo. Nat Rev Mol Cell Biol. 2003;4(9):690. doi: 10.1038/nrm1200. [DOI] [PubMed] [Google Scholar]
- 10.Tung CW, Ho SY. Computational identification of ubiquitylation sites from protein sequences. BMC Bioinform. 2009; 9. [DOI] [PMC free article] [PubMed]
- 11.Ubisite: incorporating two-layered machine learning method with substrate motifs to predict ubiquitin-conjugation site on lysines. BMC Syst Biol. 2016; 10(1), 6. [DOI] [PMC free article] [PubMed]
- 12.Xiang C, Jian-Ding Q, Shao-Ping S, Sheng-Bao S, Shu-Yun H, Ru-Ping L. Incorporating key position and amino acid residue features to identify general and species-specific ubiquitin conjugation sites. Bioinformatics. 2013;13:1614–1622. doi: 10.1093/bioinformatics/btt196. [DOI] [PubMed] [Google Scholar]
- 13.Radivojac P, Vacic V, Haynes C, Cocklin RR, Mohan A, Heyen JW, Goebl MG, Iakoucheva LM. Identification, analysis, and prediction of protein ubiquitination sites. Proteins Struct Funct Bioinform. 2010;78(2):365–380. doi: 10.1002/prot.22555. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Qi Z, Yubin X, Yueyuan Z, Shuai J, Wenzhong L, Weiping M, Zexian L, Yong Z, Yu X, Jian R. Gps-sumo: a tool for the prediction of sumoylation sites and sumo-interaction motifs. Nucleic Acids Res. (W1), 2014;325. [DOI] [PMC free article] [PubMed]
- 15.Guillaume B, Antoine BN, Jean-François Z, Ali S, Alessia Z. Jassa: a comprehensive tool for prediction of sumoylation sites and sims. Bioinformatics. 2015;21:3483–91. doi: 10.1093/bioinformatics/btv403. [DOI] [PubMed] [Google Scholar]
- 16.Jianhua J, Liuxia Z, Zi L, Xuan X, Kuo-Chen C. psumo-cd: predicting sumoylation sites in proteins with covariance discriminant algorithm by incorporating sequence-coupled effects into general pseaac. Bioinformatics. 2016;(20):387. [DOI] [PubMed]
- 17.Sharma A, Lysenko A, López Y, Dehzangi A, Sharma R, Reddy H, Sattar A, Tsunoda T. Hsesumo: sumoylation site prediction using half-sphere exposures of amino acids residues. BMC Genomics 2019;19(S9). [DOI] [PMC free article] [PubMed]
- 18.Schmidhuber J. Deep learning in neural networks: an overview. Neural Netw. 2015;61:85–117. doi: 10.1016/j.neunet.2014.09.003. [DOI] [PubMed] [Google Scholar]
- 19.Thapa N, Chaudhari M, McManus S, Roy K, Newman RH, Saigo H. Deepsuccinylsite: a deep learning based approach for protein succinylation site prediction. BMC Bioinform. 2020;21(S3). [DOI] [PMC free article] [PubMed]
- 20.Duolin W, Yanchun L, Dong X. Capsule network for protein post-translational modification site prediction. Bioinformatics (14), 2019. [DOI] [PMC free article] [PubMed]
- 21.Duolin W, Dongpeng L, Jiakang Y, Fei H, Yuexu J, Siteng C, Jingyi L, Dong X. Musitedeep: a deep-learning based webserver for protein post-translational modification site prediction and visualization. Nucleic Acids Research, 2020. [DOI] [PMC free article] [PubMed]
- 22.Duolin Wang, Shuai Zeng, Chunhui Xu, Wangren Qiu, Yanchun Liang. Musitedeep: a deep-learning framework for general and kinase-specific phosphorylation site prediction. Bioinformatics, 2017. [DOI] [PMC free article] [PubMed]
- 23.Zhen C, Liu X, Li F, Chen L, Song J. Large-scale comparative assessment of computational predictors for lysine post-translational modification sites. Briefings in Bioinformatics, 2018. [DOI] [PMC free article] [PubMed]
- 24.Fu H, Yang Y, Wang X, Wang H, Xu Y. Deepubi: a deep learning framework for prediction of ubiquitination sites in proteins. Bmc Bioinformatics 2019;20(1). [DOI] [PMC free article] [PubMed]
- 25.Ito Y. Representation of functions by superpositions of a step or sigmoid function and their applications to neural network theory. Neural Netw. 1991;4(3):385–394. doi: 10.1016/0893-6080(91)90075-G. [DOI] [Google Scholar]
- 26.Tripurani SK, Lei W, Lee KB, Wee G, Yao J. The sumo-conjugating enzyme ubc9 is developmentally regulated and interacts with activating transcription factor 7 interacting protein 2 (atf7ip2) in cattle. Biol Reprod. 2011;85(1):184–196. [Google Scholar]
- 27.Chen YZ, Chen Z, Gong YA, Ying G. Sumohydro: A novel method for the prediction of sumoylation sites based on hydrophobic properties. PLoS ONE, 2012. [DOI] [PMC free article] [PubMed]
- 28.Pérezgonzález A, Vergara M, Sanchobru JL, Van D, Hinton GE, Shanmugapriya D, Padmavathi G, Kubo J, Gantz P, Science I. Visualizing data using t-sne, 2015.
- 29.Zhu S, Zhou L, Gao P, Bao Y, He Y, Feng L. Near-infrared hyperspectral imaging combined with deep learning to identify cotton seed varieties. Molecules. 2019;24(18). [DOI] [PMC free article] [PubMed]
- 30.Venkatarajan MS, Braun W. New quantitative descriptors of amino acids based on multidimensional scaling of a large number of physical-chemical properties. Molecular modeling annual. 2001;7(12):445–453. doi: 10.1007/s00894-001-0058-5. [DOI] [Google Scholar]
- 31.Lee TY, Huang HD, Hung JH, Huang HY, Wang TH. dbptm: An information repository of protein post-translational modification. Nucleic Acids Research 34(Database issue), 2006;622–7. [DOI] [PMC free article] [PubMed]
- 32.Boutet, E., O’Donovan, C.: The universal protein resource (uniprot) in 2010 (2009) [DOI] [PMC free article] [PubMed]
- 33.Li Y.H.N.G.F. Cd-hit suite: a web server for clustering and comparing biological sequences. Bioinformatics. 2010;26(5):680–2. doi: 10.1093/bioinformatics/btp698. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34.He F, Wang R, Li J, Bao L, Zhao X. Large-scale prediction of protein ubiquitination sites using a multimodal deep architecture. BMC Systems Biology 2018;12(S6). [DOI] [PMC free article] [PubMed]
- 35.Farabet C, Couprie C, Najman L, Lecun Y. Learning hierarchical features for scene labeling. IEEE Transactions on Pattern Analysis Machine Intelligence, 2012;35. [DOI] [PubMed]
- 36.Shuichi K, Piotr P, Maria P, Andrzej K, Toshiaki K, Minoru K. Aaindex: amino acid index database, progress report 2008. Nucleic Acids Research 36(Database issue), 2008;202–205. [DOI] [PMC free article] [PubMed]
- 37.Boer PTD, Kroese DP, Mannor S, Rubinstein RY. A tutorial on the cross-entropy method. Ann Oper Res. 2005;134(1):19–67. doi: 10.1007/s10479-005-5724-z. [DOI] [Google Scholar]
- 38.Rczkowski U, Moejko M, Zambonelli J, Szczurek E. Ara: accurate, reliable and active histopathological image classification framework with bayesian deep learning. Scientific Reports 2019;9(1). [DOI] [PMC free article] [PubMed]
- 39.Naimi Ashley I, Balzer Laura B. Stacked generalization: an introduction to super learning. European Journal of Epidemiology, 2018. [DOI] [PMC free article] [PubMed]
- 40.Wolpert DH. Stacked generalization. Neural Netw. 2017;5(2):241–259. doi: 10.1016/S0893-6080(05)80023-1. [DOI] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
The source code and the datasets used in this work are available at https://github.com/lijingyimm/MultiUbiSUMO.