Skip to main content
Bioinformatics logoLink to Bioinformatics
. 2025 Sep 4;41(9):btaf489. doi: 10.1093/bioinformatics/btaf489

Accurate prediction of toxicity peptide and its function using multi-view tensor learning and latent semantic learning framework

Ke Yan 1,2, Shutao Chen 3, Bin Liu 4,5,6,, Hao Wu 7,
Editor: Jianlin Cheng
PMCID: PMC12457739  PMID: 40905623

Abstract

Motivation

Therapeutic peptide is an important ingredient in the treatment of various diseases and drug discovery. The toxicity of peptides is one of the major challenges in peptide drug therapy. With the abundance of therapeutic peptides generated in the post-genomics era, it is a challenge to promptly identify toxicity peptides using computational methods. Although several efforts have been made, few algorithms are designed to identify whether a query peptide exhibits toxicity. Considering the varied levels of biological activities, the toxicity peptides should be further classified into multi-functional peptides.

Results

This study introduces a two-level predictor, ToxPre-2L, developed using the multi-view tensor learning and latent semantic learning framework. The proposed method utilized multi-label learning with feature induced labels to avoid the redundancy of information from each view. Then the multi-view tensor learning was employed to establish the latent semantic information among different views, while low-rank constraint learning was leveraged to exploit the correlation information among multi-labels. Finally, we constructed an updated toxicity peptide benchmark dataset to assess the effectiveness of the proposed method. Experimental results demonstrated that ToxPre-2L achieves a better performance than alternative computational methods in the prediction of toxicity peptides and their multi-functional types.

Availability and implementation

The source code and data of ToxPre-2L can be accessed at http://bliulab.net/ToxPre-2L.

1 Introduction

Therapeutic peptide-based drugs have achieved an unprecedented surge of interest over the last decades, resulting in a robust development process. Therapeutic peptides are short chains with 50 or fewer amino acids and are pivotal in addressing a spectrum of pathological conditions within clinical treatment paradigms (Wei et al. 2021a, Gohil and Thirugnanasambandan 2021). Compared to small molecules, therapeutic peptides exhibit heightened biological activity and selectivity, resulting in a relatively lower incidence of side effects (Jin et al. 2024). Therefore, therapeutic peptides have unique characteristics to attribute to the therapeutic agents and drug development (Su et al. 2024).

Therapeutic peptides, considered as the category of biodrugs, have several inherent drawbacks, such as the issues of toxicity, immunogenicity, and stability (Wei et al. 2021b). Several strategies have been suggested to enhance the stability of the peptides, e.g. peptide cyclization. Similarly, several in silico tools are proposed to predict the immunogenicity properties of the peptides (Wei et al. 2018). However, the current state of peptide toxicity research reveals a paucity of proposed methodologies that adequately address the limitation, a critical factor essential for optimizing the therapeutic application. While peptides show considerable promise as therapeutic agents, their clinical translation faces significant challenges due to potential cytotoxicity, particularly their demonstrated hemolytic activity against red blood cells and other eukaryotic cell membranes.

The conventional machine learning methods for the prediction of toxicity peptides (TXP) contain two stages, including extracting the discriminative features to represent the peptides and the taxonomy-based predictors to distinguish the toxic peptides from the nontoxic ones(Fang et al. 2023). In the first stage, several hand-crafted features are designed to represent the inherent characteristics of peptide sequences, such as the kmer (Yan et al. 2022), Pse-AAC (Shen and Chou 2008), etc. In the second stage, several discriminative basic classifiers are utilized to predict the toxicity of peptides (Hearst et al. 1998, Qi 2012), such as linear regression (LR) (Weisberg 2005), etc. ToxinPred (Gupta et al. 2013) utilized the support vector machine (SVM) in silico method to predict the TXP, employing several sequence features, including the amino acid component (AAC) (Li et al. 2021), dipeptide composition (DC) (Li et al. 2021), etc. ClanTox (Naamati et al. 2009) extracted 545-D features to train a boosted stump classifier for animal toxins. ToxTeller (Wang and Sung 2024) employs physicochemical properties and sequence composition features, combined with four machine learning algorithms (LR, SVM, random forest (RF) (Lv et al. 2019), and XGBoost (Yang et al. 2021)), for TXP prediction. These methods utilized the conventional machine to make contribution to the TXP prediction. However, these two methods do not incorporate sequence-order or positional dependency information—a significant limitation for amino acid and nucleic acid sequence analysis (Wei et al. 2022).

Deep learning methods have found widespread application in TXP prediction, demonstrating a preference for generating high-latent features from sequences over conventional machine learning-based approaches, leveraging the limited peptide knowledge available (Wei et al. 2021b, Sreeraman et al. 2023, Wang et al. 2023a, Chen et al. 2025). Numerous predictors utilizing different deep learning frameworks have been proposed for therapeutic peptide prediction (Veltri et al. 2018, Wei et al. 2018, Zhang et al. 2021, Tang et al. 2022, Yan et al. 2022, Ispano et al. 2023, Qi and Zou 2023, Yan et al. 2023a,b, Ren et al. 2024), such as the attention mechanism (Vaswani et al. 2017, Su et al. 2024). ATSE (Wei et al. 2021b) and ToxIBTL (Wei et al. 2022) utilized the peptide evolutionary, structural and physicochemical information and deep learning framework to predict the TXP. However, ATSE relies on the position-specific scoring matrix (PSSM) for evolutionary information. This approach is time-consuming because PSSM construction requires PSI-BLAST (Altschul et al. 1997) searches across a large dataset. Pan et al. (2020) developed a deep learning-based approach named ToxDL, which employs convolutional neural networks to predict protein toxicity by leveraging both protein structural information and domain features. But this method is limited by its reliance on querying the UniProt database to obtain protein domain information for each target protein and its specificity to protein toxicity prediction rather than broader TXP assessment (Wei et al. 2022). ToxGIN (Yu et al. 2024) utilized the graph isomorphism networks to predict TXP. Although several previous methods have been proposed, those methods still have some drawbacks. (i) TXP represents a protein functional collection characterized by a diverse range of subgroups, such as anti-bacterial peptide (ABP) (Yu et al. 2021), anti-microbial peptide (AMP) (Javadpour et al. 1996), anti-cancer peptide (ACP) (Zhao et al. 2021), anti-fungal peptide (AFP) (de Ullivarri et al. 2020). This diversity suggests that the peptide toxicity within each subgroup may exhibit varying biological activities. (ii) The presented methods fail to address the crucial label correlations between different subgroups (Liu et al. 2024), limiting their performance in TXP prediction.

Multi-view learning frameworks integrate peptide sequence information from diverse data sources to extract more discriminative features than single-view approaches (Yan et al. 2019). Recent advances have expanded these frameworks to better model latent semantics through subspace learning, capturing shared and complementary information across views (Xiao et al. 2024). Notably, TPpred-ATMV (Yan et al. 2022) employed a multi-view tensor learning (MVTL) framework to predict therapeutic peptides. Building on these foundations, newer studies (Kang et al. 2025) have further optimized multi-view integration using deep representation learning, demonstrating improved generalizability in peptide function prediction. AmpHGT (He et al. 2025) leverages multi-view structural representations to construct heterogeneous graph networks and introduces an end-to-end deep learning architecture for AMP prediction.

Despite significant advancements in peptide toxicity prediction, current methods still exhibit limitations in classification accuracy, leading to a non-negligible rate of misclassified samples. Developing more precise predictive models is essential to enhance the reliability of toxicity assessments and strengthen confidence in the identification of TXP and their subtype classifications. In this work, a novel computational predictor, ToxPre-2L, was developed within a two-level framework to predict the TXP and their multi-functional types, such as anti-bacterial, anti-fungal, etc. To our knowledge, ToxPre-2L is the first computation predictor that adopts MVTL to predict TXPs and their multi-functional types. The first-level classifier utilizes the MVTL model to predict whether the query peptide is TXP. The second-level classifier utilizes the MVTL and binary relevance (BR) with the low-rank model to predict which multi-functional types the peptide belongs to if it has been identified as a TXP in the first level. We make several contributions as follows:

  1. Our model employs a hybrid peptide encoding strategy that synergistically combines sequence-order information and physicochemical properties to capture both structural and functional determinants (Lemanov et al. 2003, Wei et al. 2022). The sequence-order features, particularly implemented through distance-based residue (DR) and distance pair (DP) representations, explicitly account for residue interactions at varying sequence intervals, thereby preserving critical spatial patterns. The pseudo amino acid composition (Pse-AAC) integrates intrinsic physicochemical attributes (e.g. hydrophobicity, hydrophilicity) with sequence composition, and k-mer measures the peptide composition information. Therefore, this multi-view encoding framework enables a comprehensive representation of peptides.

  2. We employed a multi-view tensor learning framework to supervise feature learning, enabling the construction of an optimized latent subspace that effectively integrates complementary information from multiple feature views while maximizing the retention of discriminative information for TXP prediction.

  3. The proposed framework employs a low-rank constraint to explicitly exploit higher-order label correlations across functional subgroups, effectively capturing latent label relationships and consequently boosting multi-label classification performance.

  4. The proposed method supports interpretable insights into the latent semantic subspace. Furthermore, we have developed ToxPre-2L, a publicly accessible webserver available at http://bliulab.net/ToxPre-2L.

2 Materials and methods

2.1 Benchmark dataset

We construct a new toxicity dataset for TXP prediction. For the first-level binary classification task, we construct a benchmark dataset, which contains the positive sequences and negative sequences. The positive sequences sourced from our preceding work (Lv et al. 2023) are used to predict the TXP and the other therapeutic functional peptides. The positive sequences contain 2345 TXP sequences with a range of 10–50 residues, which are experimentally validated. For the negative nontoxic peptide (Non-TXP) sequences collection, the following steps were taken: (i) The peptides were sourced from the Swiss-prot dataset (Bairoch and Apweiler 1996) and the entries containing the related keywords were removed, including “Toxic,” “Toxin,” “KW-0020,” “KW-0800.” (ii) The Non-TXP with a length of 10–50 sequences were selected to keep the same distribution with the TXP sequences, and the Non-TXP with nonstandard residues were eliminated. (iii) The duplicated peptides that appeared from the previous dataset (Wei et al. 2021b) were removed. (iv) In order to reduce the homology bias and redundancy, the sequences with similarity exceeding 90% (Khosravian et al. 2013, Veltri et al. 2018, Burdukiewicz et al. 2020, Kavousi et al. 2020) within the nontoxic subset were filtered out using CD-HIT (Huang et al. 2010), resulting in 4125 Non-TXP sequences as the negative dataset. In order to construct the balanced dataset, we randomly selected 2345 Non-TXP sequences from the remaining negative dataset. Finally, we constructed a balanced dataset, where the number of TXP and Non-TXP sequences is the same. The first-level benchmark dataset is described by:

S=STXPSNon-TXP (1)

where STXP is the TXP positive dataset. SNon-TXP is the Non-TXP negative dataset.

For the second-level multi-functional benchmark dataset, the positive TXP dataset can be further divided into nine functional categories, i.e.

STXP=S1TXP with mono-labelS2ABPS9AVP (2)

where subscripts 1,,9 is the “TXP with mono-label”, “ABP”, “ACP”, “AMP”, “AFP”, “cell-penetrating peptide (CPP)”, “anti-parasitic peptide (APP)”, “drug delivery vehicle peptide (DDV)”, “anti-viral peptide (AVP)”, respectively. The statistical information of the second-level multi-functional dataset is illustrated in Supplementary Material 1, Fig. 1, available as supplementary data at Bioinformatics online. For parameters optimization and model evaluation, we employed 10-fold cross-validation strategy to roughly split the benchmark datasets S and STXP into the training, validation and test datasets with the ratio of 8:1:1.

2.2 Method overview

ToxPre-2L consists of two-level framework to identify the TXP and their subfunctions. Initially, ToxPre-2L uses various biological tools to represent peptide sequences, extracting multi-view feature representations. Subsequently, ToxPre-2L proposes a two-level stages architecture based on the multi-view tensor learning and latent semantic learning framework. The first-level stage of ToxPre-2L detects whether a query peptide is a TXP or Non-TXP, while the second-level stage further determines TXP among nine functional attributes. The model of the first-level stage was trained with the benchmark dataset1 S [cf. Equation (1)] and the model of the second-level stage was trained with the benchmark dataset2 STXP [cf. Equation (2)]. Finally, the optimization algorithms of the ToxPre-2L are presented in Supplementary Material 2, available as supplementary data at Bioinformatics online, in detail. The model architecture is illustrated in Fig. 1.

Figure 1.

Figure 1.

The framework of ToxPre-2L contains four stages. (A) The toxicity peptide dataset construction phase. (B) The feature extraction phase. The peptides are embedded by four feature-encoded methods that rely on sequential and physicochemical information. (C) The first-level stage. The input peptide sequences are first predicted by ToxPre-2L as TXP or Non-TXP. (D) The second-level stage. The predicted TXP sequences are further identified by ToxPre-2L as multi-label functional types.

2.3 Feature extraction methods

We utilized four feature extraction methods to encode the sequences, including the k-mer, DR, DP, and Pse-AAC. The K-mer (Khatun et al. 2020, Zulfiqar et al. 2024) feature considers the local sequence information by calculating the composition of subsequences with a fixed length K. The DR (Liu et al. 2014) feature calculates the composition of amino acid pairs between the LG spaces in the peptide. The Pse-AAC feature (Shen and Chou 2006) incorporates both amino acid composition and physicochemical properties, while the DP feature (Liu et al. 2014, Zhu et al. 2023, Zou et al. 2023) captures pairwise residue interactions by combining spatial distance constraints with sequence-order information providing complementary representations of peptide sequences. The detailed information of four features is described in Supplementary Material 1, available as supplementary data at Bioinformatics online. The above four features were extracted by Bioseq-Analysis 2.0 (Liu et al. 2019, Li et al. 2021) with default parameters.

2.4 Architecture of ToxPre-2L

The training dataset contains n peptides {xi,yi}i=1n, where xiRm is the feature of the ith peptide sequence and yi{-1,1}l represents the label vector corresponding to xi. l denotes the total number of potential labels. yis=1 (or −1) indicates the ith peptide sequence’s sth label (s[1,L]) is relevant (or irrelevant). The goals of the first-level stage and the second-level stage of the ToxPre-2L are to learn the binary labels classifier (L=2) and the multi-label classifier (L=9), respectively.

Suppose that n training peptide sequences and r test peptide sequences are encoded in D views. Let Xtr=[xtr1,(d);;xtrn,(d)]Rn×md be the training peptides from the dth(d[1,,D]) view, where xtri,(d)Rmd is a md-dimensional real-valued training peptide sequence vector from the dth view. Y=[y1,,yn]Rn×l is the label matrix of the Xtr peptides. The r test peptide sequences from the dth view are represented as Xtt=[xts1,(d);;xtsr,(d)]Rr×md.

To extract the effective information of each view feature, we utilized the multi-label information latent semantic index (MLSI) (Kai et al. 2005) strategy to reduce the dimensionality of the feature. MLSI selected the best feature subset to represent the intricate information structure inherent in multi-label features. The MLSI utilized the kernel function to map the input features into high-dimensional non-linear space. kx(xi,xj) (radial basis function) represents the inner product as

kx(xi,xj)=exp(-θxi-xj2) (3)

where θ is the bandwidth. Subsequently, MLSI selected the subspace features that collectively represent the intricate information structure inherent in each view feature representation. Therefore, MLSI focused on the relevant features and discovered the latent semantic space information with each view feature. We selected each view latent semantic feature subset XtrRn×s and XttRr×s from source multi-view information Xtr and Xtt by using the MLSI.

2.4.1 First-level stage: the peptide toxicity sub-predictor by adaptive weighted MVTL (AdaptMVTL) model

Motivated by the MVTL framework, we embedded the peptides from multi-view different features into the common latent space and captured the correlation between multi-view features in the following model:

minP(d),zd=1DXtr(d)P(d)-YF2+λ1d=1DP(d)22+λ2z (4)

where z={Xtr(d)P(d)}d=1D, z denotes the tensor nuclear norm. λ1>0 and λ2>0 are the tradeoff factors. To capture crucial information across different views, the proposed method utilized the t-SVD (Wu et al. 2019) to capture the largest singular values of the slice of the f-diagonal tensor.

Furthermore, we utilized non-negative automatic weights, denoted as a(d)(a(d)0), corresponding to the dth view. A greater weight a(d) indicates a more significant role of the corresponding view feature in multi-view learning. Accordingly, the final sub-predictor function AdaptMVTL turns out to be

mina(d), P(d),zd=1Da(d)Xtr(d)P(d)-YF2+λ1d=1DP(d)22+λ2z+γa22
s.t. a(d)0,d=1Da(d)=1 (5)

where γ>0 is a penalty parameter and a=[a(1),a(2),,a(D)] is the weight vector, where a(d) is the relative weight of the dth view.

When in prediction for a test peptide sequence xtt(d) from D views, we first achieve the f=d=1Da(d)xtt(d)P(d)(fRl), then get the predicted binary label result

ytt=argmax(f) (6)

Then the model obtains l scores corresponding to the TXP and Non-TXP and each score is accumulated from D views. A large score denotes the test peptide sequence xtt is more likely to belong to the corresponding label. The optimization solution of the AdaptMVTL model is presented in Supplementary Material 2, available as supplementary data at Bioinformatics online.

In summary, the first stage sub-predictor AdaptMVTL of ToxPre-2L has the following merits: (i) our model constructs the common representation among multi-view features by using latent semantic learning; and (ii) our model utilizes the MVTL to capture the strong correlation between different views.

2.4.2 Second-level stage: a unified TXP’s functional types sub-predictor for multi-label multi-view tensor learning with low-rank and BR (MLMVTLowRankBin) model

To capture the diverse characteristics of multiple functional types, the proposed method constructs L binary predictors using BR. This BR step is then integrated into the AdaptMVTL framework, which can be formulated as follows:

mina(d), P(d),z,td=1Da(d)Xtr(d)P(d)-YF2+λ1d=1DP(d)22+λ2z+γa22+d=1Di=1nj=1Ljyi+pj,(d),xtri,(d)t(xtri,(d))+jyi-pj,(d),xtri,(d)t(xtri,(d))s.t. a(d)0,d=1Da(d)=1 (7)

where yi+ (yi-) represents the index set of relevant (irrelevant) labels linked to the peptide sequence xtri. t(xtri,(d)) denotes the ideal thresholding value. g equals 1 when the g holds, and 0 otherwise.

Due to the threshold t(xtri,(d)) being associated with the parameter matrix P(d) and the corresponding sequence xtri,(d), the t(xtri,(d)) is difficult to optimization. Inspired by the RBRL (Wu et al. 2020), we fixed the threshold values t(xtri,(d))=0(i[1,n]) for n peptide sequences for simplicity. Moreover, we approximated the thresholding 0–1 loss function with the surrogate least squared hinge loss loss(y,f(x))=max(0,1-yf(x))2=(|1-yf(x)|+)2 (Wu et al. 2020). Therefore, the problem is transformed as follows:

mina(d), P(d),zd=1Da(d)Xtr(d)P(d)-YF2+d=1D12(|E-Yȯ(Xtr(d)P(d))|+)21+λ1d=1DP(d)22+λ2z+γa22s.t. a(d)0,d=1Da(d)=1  (8)

where ERn×l consists entirely of elements equal to 1. The second part of our proposed method introduces a nonlinear loss term that specifically penalizes under-prediction, ensuring more accurate alignment between the predicted labels and the true labels for each individual label.

Furthermore, we utilized a low-rank constraint on P(d) to harness the correlations among high-order labels. The final object function MLMVTLowRankBin turns out to be

mina(d), P(d),zd=1Da(d)Xtr(d)P(d)-YF2+d=1D12(|E-Yȯ(Xtr(d)P(d))|+)21+λ1d=1DP(d)22+λ2z+λ3d=1DP(d)*+γa22s.t. a(d)0,d=1Da(d)=1  (9)

Once we obtained the transformation matrix P(d), the multi-label TXP’s functional types ytt of the test peptide sequences xtt(d) with D views is calculated as follows:

ytt=signd=1Da(d)xtt(d)P(d)>0 (10)

where sign(x) is 1 when x>0 and −1 otherwise. The optimization solution of the MLMVTLowRankBin model is presented in Supplementary Material 2, available as supplementary data at Bioinformatics online.

Therefore, we proposed the MLMVTLowRankBin framework based on the elastic net regularized (Zhang et al. 2017) for TXP’s multi-functional types classification. In summary, the second stage sub-predictor MLMVTLowRankBin of ToxPre-2L has the following merits: (i) our model integrated a thresholding step into the AdaptMVTL framework by leveraging the BR approach, designed to achieve optimal data fitting and effectively address under-prediction; and (ii) our model utilized the trace norm to harness the correlations among high-order labels across diverse TXP’s multi-functional types. This enhancement is particularly crucial for improving performance in the multi-functional classification of TXP.

2.5 Performance evaluation

In this study, the 10 cross-validation strategy was used to assess the performance of ToxPre-2L. ToxPre-2L is a two-level stage predictor. The first-level stage of ToxPre-2L, which is dedicated to the classification of a test peptide as a TXP or Non-TXP, falls within the domain of single-label classification. The five metrics are commonly employed for assessing the efficacy of the single-label predictor, including the accuracy (ACC), sensitivity (SN), specificity (SP), Mathew’s correlation coefficient (MCC) (Chicco and Jurman 2020, Zhang et al. 2024), and area under curve (AUC). The AUC represents the area under the receiver operating characteristic (ROC) curve, evaluating a model’s ability to distinguish between classes (Bradley 1997, Tang et al. 2022, Zhu et al. 2023, Xie et al. 2024, Yan et al. 2024, Zhang et al. 2024).

{SN=TPTP+FNSP=TNFP+TNACC=TP+TNTP+TN+FN+FPMCC=TP*TN-FP*FN(TP+FN)(TP+FP)(TN+FP)(TN+FN) (11)

where TP is the true positive, TN is the true negative, FN is the false negative, and FP is the false positive.

For the second-level stage, identifying a query TXP among nine functional types [cf. Equation (2)] falls within the domain of multi-label classification. Then we utilize six metrics, including hamming loss (HL), average precision (AP), coverage (CV), and accuracy (ACC). Assume that {(xi,Di)i[1,N]} is the test set, Hi is the predicted label of the test sequence xi, Di is the true label of the test sequence xi, and Di={l1,,lq}. q is the number of the category label

{HL=1Ni=1N(HiDi-HiDiq)AP=1N i=1N(1Dilj,lkDi{lj|rank(xi,lj)rank(xi,lk)}rank(xi,lk))CV=1Ni=1N(HiDiDi)ACC=1Ni=1N(HiDiHiDi) (12)

where is the union of the labels in the two label vectors, is the intersection of the labels in the two label vectors. is the number of 1s in the vector.

3 Results and discussion

3.1 Comparing with existing predictors in identifying TXP and Non-TXP on the benchmark dataset S

In this section, we conducted a comparative assessment of the performance of the proposed method against several state-of-the-art approaches in the discrimination of TXP and Non-TXP peptides. The single-label prediction performance achieved by the first-level stage of ToxPre-2L and the other compared methods for identifying TXP and Non-TXP in the benchmark dataset S is measured by the six metrics [cf. Equation (11)]. The compared methods include ToxinPred2 (Sharma et al. 2022), ToxDL (Pan et al. 2020), and ToxIBTL (Wei et al. 2022). The predicted results of different methods are shown in Fig. 2 and Supplementary Material 1, Fig. 2, available as supplementary data at Bioinformatics online, via 10-fold cross-validation on the benchmark dataset S. These parameters of the first-level stage subpredictor AdaptMVTL of ToxPre-2L were optimized on the validation set, which is dependent with the training set and test set. For fair comparison, we strictly adhered to the parameter settings recommended in the respective original papers of all compared methods. The hyper-parameter sets are documented in Supplementary Material 1, Table 1, available as supplementary data at Bioinformatics online, with corresponding literature citations. Based on the observations from Fig. 2 and Supplementary Material 1, Fig. 2, available as supplementary data at Bioinformatics online, it is evident that ToxPre-2L demonstrates superior performance compared to all other methods considered, exhibiting enhancements ranging from 11.33% to 22.65% in terms of AUC and ACC, and improvements spanning from 0.11 to 0.58 in MCC and F1 scores.

Figure 2.

Figure 2.

The prediction results of ToxPre-2L and other methods in terms of ACC and AUC metrics on the benchmark dataset S. The different methods are used to identify TXP and Non-TXP. The results of the compared methods were obtained through the stand-alone packages reported in Pan et al. (2020), Sharma et al. (2022), and Wei et al. (2022).

The results above discussed demonstrated that the proposed method utilized more discriminative and effective TXP specific features for distinguishing both TXP and Non-TXP peptide sequences. There are two reasons for the outstanding performance of ToxPre-2L. (i) Compared with the machine learning methods based on the hand-crafted features, which are constructed through the knowledge of peptides, the proposed method not only automatically extracts the latent semantic information from each view based on the MLSI method, but also constructs the common subspace information shared with the multiple views, which is critical for the sequence analysis. (ii) Compared with the ToxIBTL and its variants, which capture the high-latent features of the peptide sequences through the deep learning framework, the proposed method adopts tensor learning to learn the correlation information among different views that interfere with the TXP prediction. Therefore, ToxPre-2L combines the MLSI feature extraction strategy and tensor learning strategy to achieve optimal performance.

3.2 Comparing with the multi-label classification algorithms in prediction TXP’s functional types on the benchmark dataset STXP

In this section, we assessed the multi-label prediction performance of the second-level stage sub-predictor of ToxPre-2L on the benchmark dataset STXP. To our knowledge, ToxPre-2L is the first predictor in identifying the multi-functional types of TXP with various subgroup. Then we take several widely used multi-label predictors and multi-view predictors as comparison, including multi-instance multi-label learning K-nearest neighbor (MIML-KNN) (Zhang 2010), multi-instance multi-label learning radial basis function (MIML-RBF) (Zhang and Wang 2009), multi-label K-nearest neighbor (ML-KNN) (Zhang and Zhou 2007), Rank-SVM (Tayal et al. 2018, Wang et al. 2023b), multi-label with label-specific features (LIFT) (Zhang and Wu 2015), ETFC (Fan et al. 2023), IMFP (Luo et al. 2024), TPpred-LE (Lv et al. 2023), and TPpred-SC (Yan et al. 2024). The predictive outcomes for four metrics [cf. Equation (12)] are presented in Table 1 via 10-fold cross-validation strategy, demonstrating the superior performance of the proposed method compared to the alternative methods across these four metrics. All deep learning-based comparison methods (iMFP-LG, TPpred-LE, and TPpred-SC) were implemented using their original published hyper-parameter configurations. The hyper-parameter specifications are provided in Supplementary Material 1, Table 2, available as supplementary data at Bioinformatics online.

Table 1.

Performance evaluation of methods using 10-fold cross-validation on the benchmark dataset STXP.a

Method Acca HLa APa CVa
MIML-KNN 0.77 0.08 0.98 2.4
MIML-RBF 0.80 0.07 0.90 2.05
ML-KNN 0.81 0.07 0.98 2.03
RankSVM 0.58 0.12 0.94 1.7
LIFT 0.77 0.08 0.98 2.38
ETFC 0.62 0.09 0.71 2.64
IMFP 0.60 0.10 0.70 2.62
TPpred-LE 0.55 0.09 0.68 2.56
TPpred-SC 0.57 0.10 0.71 2.60
ToxPre-2L 0.86 0.06 0.94 1.65
a

An upward arrow () denotes that a higher metric value indicates superior performance, whereas a downward arrow () signifies that a lower metric value corresponds to enhanced performance. The bold values represent the best results.

These results demonstrated that the ToxPre-2L can achieve more accurate and effective correlation information of labels between different Toxic’s functional types. In contrast to first-order strategy (Zhang and Zhou 2014) like ML-KNN, ToxPre-2L demonstrates superior performance concerning classification-based metrics (e.g. ACC, HL) primarily due to its incorporation of a low-rank constraint term. This term enables ToxPre-2L to effectively leverage label correlations across various function types, thereby enhancing its predictive capabilities in multi-functional prediction tasks. In contrast to the second-order strategy (Zhang and Zhou 2014) and multi-label therapeutic peptide predictions, including Rank-SVM, ETFC, iMFP-LG, TPpred-LE, and TPpred-SC, ToxPre-2L attains superior performance across classification-based metrics (i.e. ACC, HL) and ranking-based metrics (i.e. CV and AP). This is primarily attributed to the BR method, which addresses the low prediction problem caused by the imbalanced long-tail distribution of different subgroup function types of TXP. The proposed method places greater emphasis on under-predicted labels during the model optimization process by leveraging the BR method, thereby improving the prediction for these labels and ultimately reducing under-prediction. Therefore, the optimal performance of ToxPre-2L is achieved by integrating the BR into the MVTL framework, allowing for the training of the model in a single step. Additionally, the utilization of the low-rank constraint enables the exploitation of label correlations across diverse functional types of TXP, further enhancing the model’s effectiveness.

3.3 Analysis the latent semantic information

ToxPre-2L utilizes the MLSI to capture the latent semantic information from each view feature. MLSI considers the unique information posed by single-label classification and reduces the spatial dimensions by eliminating irrelevant features. Then we utilized several feature selection compared methods to evaluate the performance of MLSI, including principal component analysis (PCA) (Wright et al. 2009), MDDM (Zhang and Zhou 2010), MDS (Anowar et al. 2021). The dimension of the four feature extraction methods is taken from 5 to 25 with an interval of 5. Through 10-fold cross-validation, the feature subset constructed by each method is fed into AdaptMVTL. The results of four feature methods on the benchmark dataset S are shown in Fig. 3.

Figure 3.

Figure 3.

Comparison results based on different feature selection algorithms on benchmark dataset S. The error bars represent the mean ± standard deviation from the cross-validation experiments. When the 10-dimensional feature subset of each view is obtained by MLSI, the results of AUC and ACC achieve the best.

The results demonstrated that the MLSI algorithm achieves optimal prediction performance across all evaluation metrics when the dimension is set to 10. The comparative analysis reveals that the control methods (PCA, MVMD, MDS) exhibit notable instability across different feature dimensions, as evidenced by their larger standard deviations in both ACC and AUC metrics. For instance, PCA shows significant fluctuations in ACC (±2.47 SD at dimension 20) and AUC (±0.04 SD at dimension 5), likely due to its linearity assumption failing to capture complex data structures. MVMD demonstrates parameter sensitivity, with unstable performance in low dimensions (ACC: ±2.57 SD at dimension 5) and inconsistent AUC trends. While MDS achieves high ACC at dimension 5 (76.29 ± 1.6), its performance varies substantially (±2.4 SD at dimension 10), reflecting sensitivity to distance metrics and outliers. In contrast, MLSI maintains consistently lower variability (e.g. AUC ±0.02 SD vs. PCA’s ±0.04 SD), underscoring its robustness through adaptive feature integration and noise resistance. The MLSI takes into consideration the pertinent information associated with the interrelationships among labels and captures the latent semantic information to remove the redundant information in each view features, which greatly improve the prediction performance. This instability in control methods highlights their limitations in handling data variability, whereas MLSI’s stability validates its reliability for practical applications. Therefore, we utilize the MLSI as feature selection algorithm to capture the latent semantic information.

3.4 Interpretability of ToxPre-2L

In this section, we perform the interpretability analysis of ToxPre-2L by visualizing the feature representation ability. We visualized all four features, the linearly concatenated features obtained by fuzing the four individual features, and the latent feature based on the proposed method. The construction of the latent feature subspace, which is shared with the multi-view features, is achieved through the employment of the MVTL algorithm. Subsequently, t-SNE (Van der Maaten and Hinton 2008) is applied to reduce the dimensions of diverse view features to two dimensions, facilitating the visualization of their distribution. The results, as depicted in Fig. 4, reveal the distinguishable information of latent features between TXP sequences and Non-TXP sequences. This observation suggests that the latent features automatically generated by ToxPre-2L possess discriminative properties conducive to peptide toxicity prediction.

Figure 4.

Figure 4.

Feature visualization of ToxPre-2L and other four manual features. (A–F) Feature visualization of DR, DP, Kmer, Pse-AAC, linear concatenated feature, and latent features, respectively.

3.5 Sensitivity to hyper-parameters

In this section, we evaluated the first-level stage subpredictor AdaptMVTL of ToxPre-2L and the second-level stage subpredictor MLMVTLowRankBin of ToxPre-2L to conduct sensitivity analysis to the hyperparameters (λ1,λ2) and (λ1,λ2,λ3), respectively. The parameters λ1,λ2, and λ3 are in the range of {10-6,10-5,10-4,10-3,10-2,10-1,1,101}. We adopted the grid search strategy to optimize the regularization parameters through a 10-fold cross-validation strategy on the datasets S and STXP. The parameters of the proposed method were optimized on the validation set, which is independent with the training set and test set. Because of page limitation, Fig. 5 shows the ACC values for our model across different combinations of λ1,λ2, and λ3. From Fig. 5A and B, we selected two candidate sets {10-6,10-5} and {10-1,1,101} for parameters λ1 and λ2, respectively. We perform the proposed method with various values of the two parameters, λ1 and λ2 in order to identify the best combination of these parameters in the 2D space formed by their candidate values. After determining the optimal combination of λ1 and λ2, we can fix these values and test different values of λ3 to find its optimal value. As shown in Fig. 5C, the proposed method exhibits a relatively low sensitivity to the selection of the parameter λ3 within the range [10-1,101] to some extent. This process allows us to achieve the best combination of all three parameters. Finally, we conduct experiments using the selected parameters and present the results for comparison.

Figure 5.

Figure 5.

Sensitivity analysis to the hyper-parameters λ1,λ2, and λ3 of the ToxPre-2L on the datasets S and STXP. (A) ACC values of the AdaptMVTL model with different combinations λ1 and λ2 on the dataset S. (B) ACC values of MLMVTLowRankBin model with different combinations λ1 and λ2 by fixing λ3 on the dataset STXP. (C) ACC values of MLMVTLowRankBin model with different λ3 by fixing λ1 and λ2 on the dataset STXP.

3.6 Complexity and computational time cost

The computational complexity of ToxPre-2L is analyzed as follows: for the first-level subpredictor AdaptMVTL [Equation (5)], the dominant operations involve updating P(d) with Ο(Dn2s) complexity and tensor z through 3D fast Fourier transformation (FFT)/inverse FFT computations requiring Ο(Dn2log(n)+D2n2) operations, while subsequent basic matrix operations are negligible, yielding an overall complexity of Ο(τDn2(D+log(n)+s)) where τ represents iteration count. For the second-level subpredictor MLMVTLowRankBin [Equation (9)], each iteration’s computational cost is primarily determined by the gradient function update f(P(d)) with Ο(D(ns2+s2l+nsl)) and SVD decomposition of P(d) which, under typical multi-label classification conditions (sl), reduces to Ο(Dsl2), along with an Ο(Dn2log(n)+D2n2) cost for updating tensor z, resulting in a total complexity of Ο(τD(ns2+s2l+nsl+sl2+n2log(n)+Dn2)), with other basic operations contributing insignificantly to the overall computational burden.

We compared the computational time cost for all the comparing methods on the benchmark dataset S and STXP. Tables 2 and 3 present the average computation times of different methods, clearly showing that our approach achieves the fastest processing speed among all compared methods on the two benchmark datasets. Specifically, on dataset S, our method completes the task in just 1.1 s, outperforming the second-best method ToxIBTL by 0.36 s. The ToxDL’s computational time exceeds 72 s, further highlighting the superior efficiency of our proposed method.

Table 2.

The computational time (in seconds) of different methods on the benchmark dataset S.

Method ToxinPred2 ToxDL ToxIBTL ToxPre-2L
Time 11.76 72.10 1.46 1.15

Table 3.

The computational times (in seconds) of various methods on the benchmark dataset STXP.

Method Time
MIML-KNN 0.10
MIML-RBF 0.01
ML-KNN 0.08
RankSVM 0.02
LIFT 0.08
ETFC 2.68
IMFP 9.32
TPpred-LE 6.84
TPpred-SC 19.76
ToxPre-2L 0.02

4 Conclusion

We present a novel two-level predictor, ToxPre-2L, for predicting the function of TXP. To our knowledge, ToxPre-2L is the first computation predictor that adopts MVTL to predict TXPs and their multi-functional types. Its first level is to predict whether a query peptide is of TXP or not. If the output is yes, then the second level will predict its functional types. The proposed method utilized the MVTL and BR with robust low-rank constraint learning, which utilized the essential semantic information among different views and the label correlations between different multi-label information. The objective performance shows that ToxPre-2L outperforms the existing TXP prediction methods and several widely used multi-label predictors almost in all metrics. The experiential results confirmed that the proposed method has better performance for predicting toxicity peptides and their subfunction types. ToxPre-2L is a distinctive in silico method to predict the toxicity of peptides, offering valuable predictive insights. We anticipate that the advancement of ToxPre-2L will catalyze peptide-based drug discovery efforts.

Supplementary Material

btaf489_Supplementary_Data

Acknowledgements

We are very much indebted to the three anonymous reviewers, whose constructive comments are very helpful for strengthening the presentation of this article.

Contributor Information

Ke Yan, School of Computer Science and Technology, Beijing Institute of Technology, Beijing 100081, China; Zhongguancun Academy, Beijing 100094, China.

Shutao Chen, School of Computer Science and Technology, Beijing Institute of Technology, Beijing 100081, China.

Bin Liu, School of Computer Science and Technology, Beijing Institute of Technology, Beijing 100081, China; Zhongguancun Academy, Beijing 100094, China; SMBU-MSU-BIT Joint Laboratory on Bioinformatics and Engineering Biology, Shenzhen MSU-BIT University, Shenzhen, Guangdong 518172, China.

Hao Wu, School of Computer Science and Technology, Beijing Institute of Technology, Beijing 100081, China.

Author contributions

Ke Yan (Formal analysis [lead], Methodology [lead], Software [lead], Visualization [lead], Writing—original draft [lead]), Shutao Chen (Resources [supporting], Software [supporting], Visualization [supporting]), Bin Liu (Project administration [supporting], Supervision [supporting], Writing—review & editing [supporting]), and Hao Wu (Conceptualization [supporting], Project administration [supporting])

Supplementary data

Supplementary data are available at Bioinformatics online.

Conflict of interest: None declared.

Funding

This work was supported by the Beijing Natural Science Foundation [L232067 and L248013], the National Natural Science Foundation of China [62325202, U22A2039, 62473049, and 62372041], and Zhongguancun Academy [Project No. 20240101].

Data availability

The data is available at http://bliulab.net/ToxPre-2L.

References

  1. Altschul SF, Madden TL, Schäffer AA  et al.  Gapped BLAST and PSI-BLAST: a new generation of protein database search programs. Nucleic Acids Res  1997;25:3389–402. [DOI] [PMC free article] [PubMed] [Google Scholar]
  2. Anowar F, Sadaoui S, Selim B.  Conceptual and empirical comparison of dimensionality reduction algorithms (PCA, KPCA, LDA, MDS, SVD, LLE, ISOMAP, LE, ICA, t-SNE). Comput Sci Rev  2021;40:100378. [Google Scholar]
  3. Bairoch A, Apweiler R.  The SWISS-PROT protein sequence data bank and its new supplement TREMBL. Nucleic Acids Res  1996;24:21–5. [DOI] [PMC free article] [PubMed] [Google Scholar]
  4. Bradley AP.  The use of the area under the ROC curve in the evaluation of machine learning algorithms. Pattern Recognit  1997;30:1145–59. [Google Scholar]
  5. Burdukiewicz M, Sidorczuk K, Rafacz D  et al.  Proteomic screening for prediction and design of antimicrobial peptides with AmpGram. Int J Mol Sci  2020;21:4310. [DOI] [PMC free article] [PubMed] [Google Scholar]
  6. Chen S, Yan K, Li X et al.  Protein language pragmatic analysis and progressive transfer learning for profiling peptide–protein interactions. IEEE Trans Neural Netw Learn Syst  202536:15385–99. [DOI] [PubMed] [Google Scholar]
  7. Chicco D, Jurman G.  The advantages of the Matthews correlation coefficient (MCC) over F1 score and accuracy in binary classification evaluation. BMC Genomics  2020;21:6–13. [DOI] [PMC free article] [PubMed] [Google Scholar]
  8. de Ullivarri MF, Arbulu S, Garcia-Gutierrez E et al.  Antifungal peptides as therapeutic agents. Front Cell Infect Mi  2020;10:105. [DOI] [PMC free article] [PubMed] [Google Scholar]
  9. Fan H, Yan W, Wang L  et al.  Deep learning-based multi-functional therapeutic peptides prediction with a multi-label focal dice loss function. Bioinformatics  2023;39:btad334. [DOI] [PMC free article] [PubMed] [Google Scholar]
  10. Fang Y, Xu F, Wei L  et al.  AFP-MFL: accurate identification of antifungal peptides using multi-view feature learning. Brief Bioinform  2023;24:bbac606. [DOI] [PubMed] [Google Scholar]
  11. Gohil D, Thirugnanasambandan T. Nanocarriers in protein and peptide drug delivery. In: Nanocarriers: Drug Delivery System: An Evidence Based Approach. Singapore: Springer, 2021, 349–65.
  12. Gupta S, Kapoor P, Chaudhary K  et al. ; Open Source Drug Discovery Consortium. In silico approach for predicting toxicity of peptides and proteins. PLoS One  2013;8:e73957. [DOI] [PMC free article] [PubMed] [Google Scholar]
  13. He Y, Song X, Wan H  et al.  AmpHGT: expanding prediction of antimicrobial activity in peptides containing non-canonical amino acids using multi-view constrained heterogeneous graph transformer. BMC Biol  2025;23:184. [DOI] [PMC free article] [PubMed] [Google Scholar]
  14. Hearst MA, Dumais ST, Osuna E  et al.  Support vector machines. IEEE Intell Syst Their Appl  1998;13:18–28. [Google Scholar]
  15. Huang Y, Niu B, Gao Y  et al.  CD-HIT suite: a web server for clustering and comparing biological sequences. Bioinformatics  2010;26:680–2. [DOI] [PMC free article] [PubMed] [Google Scholar]
  16. Ispano E, Bianca F, Lavezzo E  et al.  An overview of protein function prediction methods: a deep learning perspective. CBIO  2023;18:621–30. [Google Scholar]
  17. Javadpour MM, Juban MM, Lo WC  et al.  De novo antimicrobial peptides with low mammalian cell toxicity. J Med Chem  1996;39:3107–13. [DOI] [PubMed] [Google Scholar]
  18. Jin Q, Huang J, Sun C et al.  Location embedding based pairwise distance learning for fine-grained diagnosis of urinary stones. In: International Conference on Medical Image Computing and Computer-Assisted Intervention, Marrakesh. Cham: Springer Nature Switzerland, 2024, 405–14. [Google Scholar]
  19. Kai Y, Shipeng Y, Tresp V. Multi-label informed latent semantic indexing. In: SIGIR Forum. New York, NY, United State: Association for Computing Machinery, 2005, 258–65.
  20. Kang Y, Peng Y, Zheng D  et al.  Multi-view framework for multi-label bioactive peptide classification based on multi-modal representation learning. Appl Soft Comput  2025;175:113007. [Google Scholar]
  21. Kavousi K, Bagheri M, Behrouzi S  et al.  IAMPE: NMR-assisted computational prediction of antimicrobial peptides. J Chem Inf Model  2020;60:4691–701. [DOI] [PubMed] [Google Scholar]
  22. Khatun MS, Hasan MM, Shoombuatong W  et al.  ProIn-Fuse: improved and robust prediction of proinflammatory peptides by fusing of multiple feature representations. J Comput Aided Mol Des  2020;34:1229–36. [DOI] [PubMed] [Google Scholar]
  23. Khosravian M, Faramarzi FK, Beigi MM  et al.  Predicting antibacterial peptides by the concept of Chou’s pseudo-amino acid composition and machine learning methods. Protein Pept Lett  2013;20:180–6. [DOI] [PubMed] [Google Scholar]
  24. Lemanov VV, Popov SN, Pankova GA.  Protein amino acid crystals: structure, symmetry, physical properties. Ferroelectrics  2003;285:207–16. [Google Scholar]
  25. Li H-L, Pang Y-H, Liu B.  BioSeq-BLM: a platform for analyzing DNA, RNA and protein sequences based on biological language models. Nucleic Acids Res  2021;49:e129. [DOI] [PMC free article] [PubMed] [Google Scholar]
  26. Liu B, Gao X, Zhang H.  BioSeq-Analysis2.0: an updated platform for analyzing DNA, RNA and protein sequences at sequence level and residue level based on machine learning approaches. Nucleic Acids Res  2019;47:e127. [DOI] [PMC free article] [PubMed] [Google Scholar]
  27. Liu B, Zhang D, Xu R  et al.  Combining evolutionary information extracted from frequency profiles with sequence-based kernels for protein remote homology detection. Bioinformatics  2014;30:472–9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  28. Liu M, Li C, Chen R  et al.  Geometric deep learning for drug discovery. Expert Syst Appl  2024:240:122498. [Google Scholar]
  29. Luo J, Zhao K, Chen J  et al.  iMFP-LG: identification of novel multi-functional peptides by using protein language models and graph-based deep learning. Genom Proteom Bioinform  2024;22:qzae084. [DOI] [PMC free article] [PubMed] [Google Scholar]
  30. Lv H, Yan K, Liu B.  TPpred-LE: therapeutic peptide function prediction based on label embedding. BMC Biol  2023;21:238. [DOI] [PMC free article] [PubMed] [Google Scholar]
  31. Lv Z, Jin S, Ding H  et al.  A random forest sub-golgi protein classifier optimized via dipeptide and amino acid composition features. Front Bioeng Biotechnol  2019;7:215. [DOI] [PMC free article] [PubMed] [Google Scholar]
  32. Naamati G, Askenazi M, Linial M.  ClanTox: a classifier of short animal toxins. Nucleic Acids Res  2009;37:W363–8. [DOI] [PMC free article] [PubMed] [Google Scholar]
  33. Pan X, Zuallaert J, Wang X  et al.  ToxDL: deep learning using primary structure and domain embeddings for assessing protein toxicity. Bioinformatics  2020;36:5159–68. [DOI] [PubMed] [Google Scholar]
  34. Qi R, Zou Q.  Trends and potential of machine learning and deep learning in drug study at single-cell level. Research (Wash D C)  2023;6:0050. [DOI] [PMC free article] [PubMed] [Google Scholar]
  35. Qi Y.  Random Forest for bioinformatics. In: Ensemble Machine Learning: Methods and Applications. New York, NY, United States: Springer, 2012, 307–23. [Google Scholar]
  36. Ren X, Wei J, Luo X  et al.  HydrogelFinder: a foundation model for efficient self‐assembling peptide discovery guided by non‐peptidal small molecules. Adv Sci  2024;11:2400829. [DOI] [PMC free article] [PubMed] [Google Scholar]
  37. Sharma N, Naorem LD, Jain S  et al.  ToxinPred2: an improved method for predicting toxicity of proteins. Brief Bioinform  2022;23:bbac174. [DOI] [PubMed] [Google Scholar]
  38. Shen H-B, Chou K-C.  Ensemble classifier for protein fold pattern recognition. Bioinformatics  2006;22:1717–22. [DOI] [PubMed] [Google Scholar]
  39. Shen HB, Chou KC.  PseAAC: a flexible web server for generating various kinds of protein pseudo amino acid composition. Anal Biochem  2008;373:386–8. [DOI] [PubMed] [Google Scholar]
  40. Sreeraman S, Kannan MP, Singh Kushwah RB  et al.  Drug design and disease diagnosis: the potential of deep learning models in biology. CBIO  2023;18:208–20. [Google Scholar]
  41. Su R, Xiao J, Cui H  et al.  MSKI-Net: towards modality-specific knowledge interaction for glioma survival prediction. In: 2024 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Lisbon, Portugal: IEEE, 2024, 2438–45. [Google Scholar]
  42. Tang W, Dai R, Yan W  et al.  Identifying multi-functional bioactive peptide functions using multi-label deep learning. Brief Bioinform  2022;23:bbab414. [DOI] [PubMed] [Google Scholar]
  43. Tang Y, Pang Y, Liu B.  DeepIDP-2L: protein intrinsically disordered region prediction by combining convolutional attention network and hierarchical attention network. Bioinformatics  2022;38:1252–60. [DOI] [PubMed] [Google Scholar]
  44. Tayal A, Coleman TF, Li Y.  Bounding the difference between RankRC and RankSVM and application to multi-level rare class kernel ranking. Data Min Knowl Disc  2018;32:417–52. [Google Scholar]
  45. Van der Maaten L, Hinton G.  Visualizing data using t-SNE. J Mach Learn Res  2008;9:2579–605. [Google Scholar]
  46. Vaswani A, Shazeer N, Parmar N  et al. Attention is all you need. In: Advances in Neural Information Processing Systems. Red Hook, NY, United States: Curran Associates, Inc.57 Morehouse Lane, 2017, 5998–6008.
  47. Veltri D, Kamath U, Shehu A.  Deep learning improves antimicrobial peptide recognition. Bioinformatics  2018;34:2740–7. [DOI] [PMC free article] [PubMed] [Google Scholar]
  48. Wang JH, Sung TY.  ToxTeller: predicting peptide toxicity using four different machine learning approaches. ACS Omega  2024;9:32116–23. [DOI] [PMC free article] [PubMed] [Google Scholar]
  49. Wang Y, Xu L, Zou Q.  Deep learning methods for bioinformatics and biomedicine. Methods  2023. a;216:1–2. [DOI] [PubMed] [Google Scholar]
  50. Wang Y, Zhai Y, Ding Y  et al. SBSM-Pro: support bio-sequence machine for proteins. Sci China Inf Sci  2024;67:212106. [Google Scholar]
  51. Wei L, He W, Malik A  et al.  Computational prediction and interpretation of cell-specific replication origin sites from multiple eukaryotes by exploiting stacking framework. Brief Bioinform  2021. a:22:bbaa275. [DOI] [PubMed] [Google Scholar]
  52. Wei L, Ye X, Sakurai T  et al.  ToxIBTL: prediction of peptide toxicity based on information bottleneck and transfer learning. Bioinformatics  2022;38:1514–24. [DOI] [PubMed] [Google Scholar]
  53. Wei L, Ye X, Xue Y  et al.  ATSE: a peptide toxicity predictor by exploiting structural and evolutionary information based on graph neural network and attention mechanism. Brief Bioinform  2021. b;22:bbab041. [DOI] [PubMed] [Google Scholar]
  54. Wei L, Zhou C, Chen H  et al.  ACPred-FL: a sequence-based predictor using effective feature representation to improve the prediction of anti-cancer peptides. Bioinformatics  2018;34:4007–16. [DOI] [PMC free article] [PubMed] [Google Scholar]
  55. Weisberg S.  Applied Linear Regression. Hoboken, New Jersey, United States of America: John Wiley & Sons, Inc., ; 2005. [Google Scholar]
  56. Wright J, Ganesh A, Rao S  et al. Robust principal component analysis: exact recovery of corrupted low-rank matrices via convex optimization. In: Advances in Neural Information Processing Systems. Red Hook, NY, United States: Curran Associates, Inc.57 Morehouse Lane, 2009, 2080–8.
  57. Wu G, Zheng R, Tian Y  et al.  Joint ranking SVM and binary relevance with robust low-rank learning for multi-label classification. Neural Netw  2020;122:24–39. [DOI] [PubMed] [Google Scholar]
  58. Wu JL, Lin ZC, Zha HB.  Essential tensor learning for multi-view spectral clustering. IEEE Trans Image Process  2019;28:5910–22. [DOI] [PubMed] [Google Scholar]
  59. Xiao S, Li J, Lu J  et al.  Graph neural networks for multi-view learning: a taxonomic review. Artif Intell Rev  2024;57:341. [Google Scholar]
  60. Xie H, Ding Y, Qian Y  et al.  Structured sparse regularization based random vector functional link networks for DNA N4-methylcytosine sites prediction. Expert Syst Appl  2024;235:121157. [Google Scholar]
  61. Yan K, Fang X, Xu Y  et al.  Protein fold recognition based on multi-view modeling. Bioinformatics  2019;35:2982–90. [DOI] [PubMed] [Google Scholar]
  62. Yan K, Guo Y, Liu B.  PreTP-2L: identification of therapeutic peptides and their types using two-layer ensemble learning framework. Bioinformatics  2023. a;39:btad125. [DOI] [PMC free article] [PubMed] [Google Scholar]
  63. Yan K, Lv H, Guo Y  et al.  TPpred-ATMV: therapeutic peptides prediction by adaptive multi-view tensor learning model. Bioinformatics  2022;38:2712–8. [DOI] [PubMed] [Google Scholar]
  64. Yan K, Lv H, Guo Y  et al.  sAMPpred-GAT: prediction of antimicrobial peptide by graph attention network and predicted peptide structure. Bioinformatics  2023. b;39:btac715. [DOI] [PMC free article] [PubMed] [Google Scholar]
  65. Yan K, Lv H, Shao J  et al.  TPpred-SC: multi-functional therapeutic peptide prediction based on multi-label supervised contrastive learning. Sci China Inf Sci  2024;67:212105. [Google Scholar]
  66. Yan K, Lv H, Wen J  et al.  TP-MV: therapeutic peptides prediction by multi-view learning. CBIO  2022;17:174–83. [Google Scholar]
  67. Yan W, Tang W, Wang L  et al.  PrMFTP: multi-functional therapeutic peptides prediction based on multi-head self-attention mechanism and class weight optimization. PLoS Comput Biol  2022;18:e1010511. [DOI] [PMC free article] [PubMed] [Google Scholar]
  68. Yang C, Chen M, Yuan Q.  The application of XGBoost and SHAP to examining the factors in freight truck-related crashes: an exploratory analysis. Accid Anal Prev  2021;158:106153. [DOI] [PubMed] [Google Scholar]
  69. Yu HT, Ma Z, Meng S  et al.  A novel nanohybrid antimicrobial based on chitosan nanoparticles and antimicrobial peptide microcin J25 with low toxicity. Carbohyd Polym  2021;253:117309. [DOI] [PubMed] [Google Scholar]
  70. Yu Q, Zhang Z, Liu G  et al.  ToxGIN: an in silico prediction model for peptide toxicity via graph isomorphism networks integrating peptide sequence and structure information. Brief Bioinform  2024;25:bbae583. [DOI] [PMC free article] [PubMed] [Google Scholar]
  71. Zhang H-Q, Liu S-H, Li R  et al.  MIBPred: ensemble learning-based metal ion-binding protein classifier. ACS Omega  2024;9:8439–47. [DOI] [PMC free article] [PubMed] [Google Scholar]
  72. Zhang J, Zhang Z, Pu L  et al.  AIEpred: an ensemble predictive model of classifier chain to identify anti-inflammatory peptides. IEEE/ACM Trans Comput Biol Bioinform  2021;18:1831–40. [DOI] [PubMed] [Google Scholar]
  73. Zhang M-L.  A k-nearest neighbor based multi-instance multi-label learning algorithm. In: 2010 22nd IEEE International Conference on Tools with Artificial Intelligence, Arras, France: IEEE, 2010, 207–12. [Google Scholar]
  74. Zhang ML, Wang ZJ.  MIMLRBF: RBF neural networks for multi-instance multi-label learning. Neurocomputing (Amst)  2009;72:3951–6. [Google Scholar]
  75. Zhang ML, Wu L.  LIFT: multi-label learning with label-specific features. IEEE Trans Pattern Anal Mach Intell  2015;37:107–20. [DOI] [PubMed] [Google Scholar]
  76. Zhang ML, Zhou ZH.  ML-KNN: a lazy learning approach to multi-label learning. Pattern Recogn  2007;40:2038–48. [Google Scholar]
  77. Zhang ML, Zhou ZH.  A review on multi-label learning algorithms. IEEE Trans Knowl Data Eng  2014;26:1819–37. [Google Scholar]
  78. Zhang W, Wei H, Zhang W  et al.  Multiple types of disease-associated RNAs identification for disease prognosis and therapy using heterogeneous graph learning. Sci China Inform Sci  2024;67:189103. [Google Scholar]
  79. Zhang Y, Zhou Z-H.  Multilabel dimensionality reduction via dependence maximization. ACM Trans Knowl Discov Data (TKDD)  2010;4:1–21. [Google Scholar]
  80. Zhang Z, Lai Z, Xu Y  et al.  Discriminative elastic-net regularized linear regression. IEEE Trans Image Process  2017;26:1466–81. [DOI] [PubMed] [Google Scholar]
  81. Zhao Y, Wang S, Fei W  et al.  Prediction of anticancer peptides with high efficacy and low toxicity by hybrid model based on 3D structure of peptides. Int J Mol Sci  2021;22:5630. [DOI] [PMC free article] [PubMed] [Google Scholar]
  82. Zhu H, Hao H, Yu L.  Identifying disease-related microbes based on multi-scale variational graph autoencoder embedding Wasserstein distance. BMC Biol  2023;21:294. [DOI] [PMC free article] [PubMed] [Google Scholar]
  83. Zhu W, Yuan S-S, Li J  et al.  A first computational frame for recognizing heparin-binding protein. Diagnostics (Basel)  2023;13:2465. [DOI] [PMC free article] [PubMed] [Google Scholar]
  84. Zou X, Ren L, Cai P  et al.  Accurately identifying hemagglutinin using sequence information and machine learning methods. Front Med (Lausanne)  2023;10:1281880. [DOI] [PMC free article] [PubMed] [Google Scholar]
  85. Zulfiqar H, Guo Z, Ahmad RM  et al.  Deep-STP: a deep learning-based approach to predict snake toxin proteins by using word embeddings. Front Med (Lausanne)  2024;10:1291352. [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

btaf489_Supplementary_Data

Data Availability Statement

The data is available at http://bliulab.net/ToxPre-2L.


Articles from Bioinformatics are provided here courtesy of Oxford University Press

RESOURCES