Skip to main content
BMC Genomics logoLink to BMC Genomics
. 2024 May 9;25:406. doi: 10.1186/s12864-024-10299-x

GNNGL-PPI: multi-category prediction of protein-protein interactions using graph neural networks based on global graphs and local subgraphs

Xin Zeng 1, Fan-Fang Meng 1, Meng-Liang Wen 2, Shu-Juan Li 3, Yi Li 1,
PMCID: PMC11080243  PMID: 38724906

Abstract

Most proteins exert their functions by interacting with other proteins, making the identification of protein-protein interactions (PPI) crucial for understanding biological activities, pathological mechanisms, and clinical therapies. Developing effective and reliable computational methods for predicting PPI can significantly reduce the time-consuming and labor-intensive associated traditional biological experiments. However, accurately identifying the specific categories of protein-protein interactions and improving the prediction accuracy of the computational methods remain dual challenges. To tackle these challenges, we proposed a novel graph neural network method called GNNGL-PPI for multi-category prediction of PPI based on global graphs and local subgraphs. GNNGL-PPI consisted of two main components: using Graph Isomorphism Network (GIN) to extract global graph features from PPI network graph, and employing GIN As Kernel (GIN-AK) to extract local subgraph features from the subgraphs of protein vertices. Additionally, considering the imbalanced distribution of samples in each category within the benchmark datasets, we introduced an Asymmetric Loss (ASL) function to further enhance the predictive performance of the method. Through evaluations on six benchmark test sets formed by three different dataset partitioning algorithms (Random, BFS, DFS), GNNGL-PPI outperformed the state-of-the-art multi-category prediction methods of PPI, as measured by the comprehensive performance evaluation metric F1-measure. Furthermore, interpretability analysis confirmed the effectiveness of GNNGL-PPI as a reliable multi-category prediction method for predicting protein-protein interactions.

Keywords: Multi-category prediction of protein-protein interactions, Graph neural network, Global graphs, Local subgraphs, Asymmetric loss function

Introduction

Protein-protein interactions (PPI) play a crucial role in various biological processes within cells. Identifying PPI is of great significance in advancing research across multiple life science fields, including medical diagnosis, drug design, and disease treatment [1]. Currently, the methods for identifying PPI can be broadly categorized into traditional biological experimental methods and computational methods. Traditional biological experimental methods primarily involve techniques such as yeast two-hybrid [2], protein chips [3], and synthetic lethal analysis [4]. However, these methods suffer from several disadvantages, including being time-consuming, labor-intensive, and requiring high financial resources [5]. To overcome these disadvantages, the computational methods for PPI prediction have developed rapidly. Nevertheless, these computational methods face dual challenges. Firstly, they need to accurately identify multiple specific categories of PPI. Secondly, they must achieve the desired predictive performance.

In recent years, the computational methods for predicting PPI have transitioned from docking-based methods to machine learning and deep learning-based methods. Docking-based methods [6] are capable of effectively predicting PPI, but they require high-quality protein 3D structures and significant computational resources. Additionally, they operate at a slower prediction speed, which cannot keep pace with the demands of processing massive data. Machine learning and deep learning methods, on the other hand, have exhibited better performance in handling large amounts of data [710]. These methods can utilize both protein sequence and structural data. Machine learning-based methods [9, 11, 12] rely on protein sequence or structural features, utilizing models like SVM [10, 13, 14] and random forest [15] to predict PPI. Although these methods have displayed good predictive performance, they cannot automatically extract deep-level features of PPI from the original sequences or structures of proteins. This creates bottlenecks that hamper improvements in predictive performance. However, the emergence of deep learning models, such as multi-layer neural networks, has provided better model prediction performance and pointed researchers towards breakthroughs in addressing the performance bottlenecks encountered by machine learning techniques [16].

Methods for predicting PPI using deep learning techniques leverage protein sequences, structures, and PPI networks. Deep learning techniques such as deep neural networks (DNN) [17], convolutional neural networks (CNN) [18], recurrent neural networks (RNN) [19], attention mechanisms [20], and graph neural networks (GNN) [21] are employed to extract deep-level features from proteins and PPI networks. DNN-based methods extract protein features through multi-layer neural networks to directly predict PPI [22, 23] or employ machine learning models for PPI prediction [24]. CNN and RNN-based methods focus on extracting local features and long-range dependency features from protein sequences, respectively. For instance, LSTM-PHV method [8] and other related approaches [25] leveraged the LSTM (Long Short-Term Memory) model to capture long-range dependency features in protein sequences. Methods like ADH-PPI [26] and DCSE [27] integrated CNN and RNN to extract both local and long-range features from protein sequences, which were then combined to predict PPI. Attention mechanisms have also been widely utilized to identify key sequence features in protein sequences [23, 26, 28]. While these DNN, CNN, RNN, and attention-based methods primarily focused on protein sequence features, they often overlooked the structural features of proteins and the hidden interaction features present in PPI networks. To address these limitations, GNN-based methods have emerged. However, incorporating protein structures into these methods has been challenging due to the slow exploration of protein 3D structures. With the advent of protein 3D structure prediction tools like AlphaFold [29] and ColabFold [30], obtaining monomer protein 3D structures has become easier. This has led to rapid advancements in research that utilizes GNN to extract protein structural features for predicting PPI, either independently or in combination with protein sequence features. For instance, the method proposed in reference [31] directly employed the GCN/GAT model to extract the structural features from two interacting proteins. These features were then concatenated and used to predict PPI. TAGPPI method [32] combined TextCNN and GAT to extract sequence and structural features from protein sequences and contact maps, respectively. The extracted features were fused before being passed through a fully connected (FC) layer to predict PPI. Other methods introduced interaction features by constructing a PPI network graph where proteins served as vertices and interactions served as edges. GNN was then used to extract interaction features from this graph, ultimately predicting PPI. Methods such as S-VGAE [33], HIGH-PPI [34], and Topsy-Turvy [35] utilized PPI network graphs along with protein sequences or structures as vertex features. GNN was applied for binary classification to determine whether there was an interaction between vertices in the graph. The aforementioned methods treated PPI as a binary classification task and did not identify specific interaction categories between proteins. However, methods like PIPR [36], GNN-PPI [37], SemiGNN-PPI [38], and AFTGAN [39] have been developed to predict PPI into multi-category. PIPR utilized a Siemens residual current convolutional neural network (RCNN) to extract local features and contextual information from protein sequences. GNN-PPI, SemiGNN, and AFTGAN leveraged GNN to learn deep-level features from PPI networks, enabling multi-category prediction. Although these methods extracted protein sequence features and PPI network graph features for multi-category PPI prediction, research in this area was still in its early stages. The performance of the models was somewhat below user expectations. Therefore, proposing an efficient and effective model for multi-category prediction of PPI presents significant challenges.

In this study, we proposed a novel graph neural network method, called GNNGL-PPI, for predicting multi-category of protein-protein interactions. GNNGL-PPI utilized a combination of global graph and local subgraph features. Specifically, we used Graph Isomorphism Network (GIN) to extract global graph features from PPI network graphs. Additionally, we employed GIN-AK to extract local subgraph features from subgraphs containing protein vertices in the PPI network graphs. Simultaneously, to address the issue of imbalanced samples for each category in the benchmark dataset, we introduced an Asymmetric Loss (ASL) Function. This loss function helped enhance the predictive performance of the model by assigning different weights to different categories based on their prevalence. We evaluated the model on six standard test sets created using three different dataset partitioning algorithms (Random, BFS, DFS). GNNGL-PPI consistently outperformed the state-of-the-art methods for multi-category prediction of PPI, as measured by the comprehensive evaluation metric F1-measure. Furthermore, we conducted interpretability analysis to validate the effectiveness of GNNGL-PPI. Experimental results confirmed that GNNGL-PPI was a reliable method for predicting multi-category of PPI.

Materials and methods

Datasets

In this study, we approached protein-protein interactions prediction as a multi-category task, encompassing seven PPI categories: Reaction, Binding, Post-translational modification (Ptmod), Activation, Inhibition, Catalysis, and Expression. Each protein-protein interaction pair is assigned to at least one of these categories. For example, the interaction category between protein 9606.ENSP00000005257 (RalA) and protein 9606.ENSP00000202677 (RalGAP A2) is Inhibition (PMID: 34767674), while the interaction category protein 9606.ENSP00000003100 (Cyp51) and protein 9606.ENSP00000240055 (NF-YB) is Activation (PMID:27438,727). To evaluate the performance of the model, we utilized two benchmark datasets, SHS27k and SHS148k, which were consistent with the dataset used in the GNN-PPI [37] method and exhibited sequence consistency of less than 40% in each dataset. These datasets consisted of 7624 (1690 proteins) and 44,488 (5189 proteins) protein-protein interaction pairs, respectively. The number and occupation ratio of samples for the seven PPI categories generated by these protein-protein interaction pairs were shown in Table 1. It was evident from Table 1 that the number of samples for the seven PPI categories was imbalanced. During the training and testing of the model, we also employed the Random, Breath First Search (BFS), and Depth First Search (DFS) algorithms proposed by the GNN-PPI method to partition the SHS27k and SHS148k datasets. For detailed information on the data partitioning algorithms for the Random, BFS, and DFS, please refer to the GNN-PPI method.

Table 1.

Number and occupation ratio of samples for seven categories of protein-protein interactions

PPI Categories SHS27K SHS148K
Number of Samples Occupation Ratio Number of Samples Occupation Ratio
Reaction 3164 18.22% 18,067 17.71%
Binding 4017 23.13% 23,448 22.98%
Ptmod 1303 7.50% 9336 9.15%
Activation 3297 18.98% 18,910 18.53%
Inhibition 1407 8.10% 8987 8.81%
Catalysis 3492 20.11% 19,871 19.47%
Expression 687 3.96% 3419 3.35%

PPI network graph formation and protein features

PPI network graph formation

Assuming a set of proteins P={p_1,p_2,…,p_n } with n as the number of proteins, each protein acted as a vertex in the PPI network graph. The interaction category between p_i and p_j was represented as edge e_ij (1≤i,j≤n). Different protein-protein interaction categories made up the category space D={D_1,…,D_t } (t=7) of the dataset, where t is the number of PPI categories. If there was an interaction between p_i and p_j of a certain category, the corresponding position in the adjacent matrix representing the PPI network graph was assigned a value of 1, otherwise, it was assigned a value of 0.

Protein features

We employed the pre-trained model MASSA [40] to capture high-level and more fine-grained features. This pre-trained model leveraged multi-modal protein data, including protein sequences, structures, gene ontology annotations, motifs, and region positions, to derive comprehensive protein features. Unlike previous research, which mainly focused on protein features derived from protein sequences such as Position-Specific Scoring Matrix (PSSM) [41] or Hidden Markov Models (HMM) [42] matrix, or treated protein sequences as natural language processing (NLP) [43] tasks to extract protein sequence features. However, these researches yielded relatively singular protein features and did not fully encompass other biochemical information pertaining to proteins.

Proposed model

To begin with, we input the multimodal data of proteins, such as sequences, structures, and Go annotations, into the pre-trained model MASSA and obtained the 512-dimensional protein pre-training features. Next, we embedded the pre-trained features using a layer of linear transformation. Finally, we used the embedded features as the protein vertex features in the PPI network (Fig. 1A).

Fig. 1.

Fig. 1

The architecture review of GNNGL-PPI. (A) We utilized the pre-training model MASSA to obtain the comprehensive protein features based on multimodal data such as protein structures, sequences, and gene ontology annotations. (B) GIN-AK extracted the global features and centroid features of subgraphs through two different processes, and then added these extracted features to obtain the final local subgraph features. (C) The extracted global features from the PPI network graph through GIN and the extracted local subgraph features by GIN-AK were concatenated to obtain the high-level features of vertices in the PPI network, i.e., protein features. We multiplied the features of proteins themselves and their interacting proteins and sent them into the multilayer perceptron (MLP) to complete multi-category prediction of protein-protein interactions

The extraction of high-level features in the PPI network graph can be divided into two main parts, as shown in Fig. 1B and C: global graph features extraction and local subgraph features extraction. In the global graph features extraction part, we utilized one layer of GIN followed by two FC layers. ReLU activation function [44] was applied in the FC layers. To ensure stable model training, batch normalization function was employed in the last layer to normalize the extracted features and obtained the global graph features of the PPI network graph. Moving on to the local subgraph features extraction section: firstly, we selected a protein vertex as the center and included all vertices within a path distance of K from that the center vertex. This formed a subgraph of the protein. Secondly, we used the GIN-AK to extract the features from the subgraph. GIN-AK incorporated two distinct processes to extract global features and centroid features of the subgraph, respectively. These features were then combined to obtain the final local subgraph features. To extract the global features of the subgraph, we input both the subgraph and protein features into a single layer of GIN. The output features of GIN were then fed into a FC layer (gating unit) with a sigmoid activation function, resulting in the global features of the subgraph. For the centroid features of the subgraph, we first calculated the path distance between the vertices in the subgraph. The obtained path distance matrix was then processed using the same gating unit to obtain the centroid features. Finally, we concatenated the global features and centroid features of the subgraph, applied batch normalization, and obtained the local subgraph features. The global graph features and local subgraph features were concatenated to form the vertex features in the PPI network graph. It was important to note that ReLU activation function was used in the gating unit during the subgraph features extraction process. Furthermore, to prevent gradient vanishing and enhance model stability during training, a dropout layer was added to the gating unit.

In order to predict multi-category of PPI, we first multiplied the features of the protein itself and its interacting protein. The resulting multiplied features were then fed into a FC layer. The output of the FC layer was a 7-dimensional matrix. Finally, by applying the sigmoid function to the 7-dimensional matrix, we completed the multi-category prediction of PPI.

GIN-AK

GNN have proven to be an effective framework for topology representation learning. GIN is widely regarded as the most repressive GNN [45], but it still faces limitation in breaking through the first-order Weisfeiler-Leman isomorphism testing [46, 47]. To overcome this limitation, a subgraph operation had been introduced based on GIN. This approach updated vertex information by using the feature of the subgraph of vertex, rather than relying solely on the neighboring vertex feature information. This transformation reduced the complexity of the graph feature problem to a smaller and simpler subgraph feature problem. The resulting GIN model was called GIN-AK (GIN As Kernel). In the process of updating vertices using graph convolutional operations, each vertex aggregates information directly from neighboring vertices in a star operation. The star operation

Inline graphic forms a graph that can be defined as formular (1). The specific update method for vertices in GIN is shown in formula 2.

graphic file with name M2.gif 1
graphic file with name M3.gif 2

The limitation of graph convolutional operations lies in their inability to differentiate between graphs that possess the same vertex degree but exhibit different structures. In order to address this issue, we proposed the utilization of a subgraph encoder, denoted as Inline graphic, to substitute the Inline graphic operation. This replacement significantly enhanced the expressive capabilities of the graph convolutional operation. The subgraph encoder Inline graphic facilitated the update of vertex features by constructing subgraph features within a Inline graphic -hop egonet centered around the vertex Inline graphic. This update process was displayed in formula (3), which specifically outlined the procedure for updating vertices in GIN-AK.

graphic file with name M9.gif 3

In GIN-AK, a Inline graphic -hop propagation mechanism was employed to acquire the subgraph surrounding each vertex. Additionally, the path distance between each vertex and the centroid of its corresponding subgraph was computed. This information enhanced the vertex features and contributed to the overall improvement of graph convolutional expressiveness. To capture the nonlinear transformations of global features and centroid features of the subgraph, gating units were introduced. Finally, the global features and centroid features of the subgraph were obtained using formula (4) and (5), respectively.

graphic file with name M11.gif 4
graphic file with name M12.gif 5

Among them, Inline graphicrepresented the path distance matrix from vertex Inline graphic to vertex Inline graphic in the Inline graphic+1-layer, and Inline graphic represented element wise multiplication. Combining the global features and centroid features of the subgraph, the update process of vertex Inline graphic was shown in formula (6).

graphic file with name M19.gif 6

GIN-AK improved the expressiveness of graph convolutional operation through the use of subgraphs instead of star graphs. This enhancement allowed GIN-AK to capture underlying structural features in PPI network, leading to the improved performance in predicting different categories of protein-protein interactions.

ASL

In this study, PPI prediction was regarded as a multi-category classification task. However, the dataset (Table 1) used was imbalanced, which means that there were more samples corresponding to some categories compared to others. This can affect the performance evaluation of the model since it might focus more on the majority classes and ignore the minority classes. To address this problem, we introduced the Asymmetric Loss (ASL) [48] function.

In multi-category imbalanced datasets, using symmetric loss functions like Focal Loss [49], BCE Loss [50], or Cross Entropy [51] may not effectively learn some features from positive samples. These loss functions tend to focus more on negative samples than positive samples [48], which can be suboptimal. For example, in Focal Loss function (formula 8), when using the same Inline graphic for multi-category training, it may eliminate the gradient of sparse positive samples. Among them, Inline graphic is the output probability of the model, Inline graphic is the focusing parameter, Inline graphic and Inline graphic represent the positive and negative loss parts, respectively.

graphic file with name M25.gif 8

Therefore, in ASL, the focusing parameters of the loss function are decoupled into positive focusing parameter Inline graphic and negative focusing parameter Inline graphic. This allows for asymmetric focusing, enabling better control the effect of positive and negative samples on the loss function. In addition, considering the asymmetric focusing parameters in ASL, there are shortcomings in some cases where there are not enough negative samples. In ASL, another asymmetric mechanism of probability translation is further proposed, as shown in formula (9).

graphic file with name M28.gif 9

Inline graphic represents an adjustable probability margin. Through the above two adjustments, the final definition of ASL is shown in formula (10).

graphic file with name M30.gif 10

By employing ASL to dynamically regulate the degree of asymmetry throughout the entire training process, the selection of hyperparameters is simplified. This effectively balances the focus of the network on positive and negative samples, ultimately improving the accuracy of multi-category prediction.

Model training

We trained our model for 400 epochs using the Adam optimizer [52] with a batch size of 1024 and an initial learning rate of 0.01. To optimize the training process, we implemented a learning rate decay function, set the patience to 20, and stopped training when the model did not reduce its loss after 20 epochs. To prevent overfitting, we used a dropout rate of 0.2 in the local subgraph features extraction part and 0.5 in the global graph features extraction section. In the local subgraph features extraction part, we set Inline graphic in Inline graphic-hop to 1, extracted the 1-hop subgraph as the input graph of GIN-AK, and used ASL with a probability margin Inline graphic of 0.05. For asymmetric focusing, we set Inline graphic to 1 and Inline graphic to 0, as recommended by ASL.

Evaluation metrics

In this study, we regarded PPI prediction as a multi-category classification task. The dataset used in our study had an imbalanced distribution of samples across different categories. To effectively evaluate the model’s performance on this imbalanced dataset, we employed F1-measure as the evaluation metric. F1-measure is a widely-used evaluation metric for imbalanced datasets because it takes into account both precision and recall, providing a balanced measure of model performance [53].

graphic file with name M36.gif
graphic file with name M37.gif
graphic file with name M38.gif

Among them, TP, FP, TN, and FN represent the number of predicted true positive, false positive, true negative, and false negative samples, respectively.

Results and discussion

Multimodal pre-training features offer a better representation of protein

The predictive performance of a model is directly influenced by the quality of its input features. In this study, we aimed to select high-quality features for protein vertices in the PPI network. To achieve this, we compared the sequence features of proteins with features obtained from the latest multimodal pre-training model of proteins. Experimental results presented in Table 2 exhibited that multimodal pre-training features can more effectively represent proteins. The sequence features used in our study were derived from the GNN-PPI method, a classic approach for predicting PPI. These sequence features consisted of 13 dimensions, with the first 5 dimensions capturing the co-occurrence similarity features of amino acids, and the remaining 8 dimensions representing the similarity features of electrostatic and hydrophobic interactions between amino acids. Through a combination of one layer of CNN and linear transformation, these 13-dimensional sequence features were transformed into 512-dimensional sequence features, which were utilized in our study. For the multimodal pre-training features, we employed the latest and more refined pre-training model called MASSA. This model leverages protein data from various modalities including protein sequences, structures, gene ontology annotations, motifs, and region positions to extract comprehensive protein features at a higher level. Subsequently, we constructed a GIN model and used the protein sequence features as well as the multimodal pre-training features as vertex features in the PPI network graph to predict PPI.

Table 2.

Performance comparison between multimodal pre-training features and sequence features

Feature SHS27k SHS148K
Random (%) BFS (%) DFS (%) Random (%) BFS (%) DFS (%)
Sequence features 88.21Inline graphic0.47 63.26Inline graphic3.70 74.31Inline graphic3.44 91.98Inline graphic0.20 65.50Inline graphic4.59 82.24Inline graphic0.73
MASSA 88.99Inline graphic0.80 70.34Inline graphic0.67 78.32Inline graphic1.60 92.52Inline graphic0.17 69.99Inline graphic4.60 83.19Inline graphic1.26

In the case of the SHS27K dataset, the GIN model based on multimodal pre-trained features exhibited the improved performance under three different dataset partitioning algorithms (Random, BFS, and DFS). Specifically, the F1-measure increased by approximately 0.7%, 7%, and 4%, respectively. As for the SHS148K dataset, the GIN network based on multimodal pre-trained features performed improvements of 1%, 4%, and 1% under the Random, BFS, and DFS algorithms, respectively. These experimental findings indicated that when using graph neural networks for predicting PPI, multimodal pre-trained features offered a better representation of protein compared to sequence features alone.

Local subgraph features can enhance the performance of the model

In this study, we proposed a novel operation for updating vertex information in GIN. Instead of directly aggregating information from neighboring vertices in a star operation, we updated the vertex information through the features of the vertex’s subgraph. To assess the contribution of local subgraph features to the model’s performance, we conducted ablation experiments based on global graph features, local subgraph features, and their combined features. The global graph features were extracted directly using the GIN model, while the local subgraph features were extracted using the GIN-AK. The combined features were obtained by fusing global graph features and local subgraph features. Experimental results presented in Table 3 indicated that, based solely on global graph features and local subgraph features, the latter performed approximately 5% better than the former for the BFS partitioning algorithm on both SHS27K and SHS148K datasets, while they exhibited similar performance under the other two partitioning algorithms. However, after combining global graph features and local subgraph features, the combined model showed better performance than the global graph features-based and local subgraph features-based models under the two partitioning algorithms (Random and DFS) of the SHS148K dataset and all partitioning algorithms of the SHS27K dataset. Although the F1-measure of the combined model was 1% lower than that of the local subgraph features-based model under the BFS partitioning algorithm of the SHS148K dataset, these findings still showed that incorporating local subgraph features on the basis of global graph features could enhance the model’s performance.

Table 3.

Performance comparison of global graph, subgraph, and their combined features

Feature SHS27k SHS148K
Random (%) BFS (%) DFS (%) Random (%) BFS (%) DFS (%)
Subgraph 89.15Inline graphic0.69 74.08Inline graphic0.32 77.76Inline graphic2.55 90.84Inline graphic0.32 75.97Inline graphic3.64 82.87Inline graphic0.52
Global Graph 88.99Inline graphic0.80 70.34Inline graphic0.67 78.32Inline graphic1.60 92.52Inline graphic0.17 69.99Inline graphic4.60 83.19Inline graphic1.26
Combination 90.04Inline graphic0.55 76.08Inline graphic1.22 79.67Inline graphic4.05 92.76Inline graphic0.17 74.62Inline graphic4.35 84.47Inline graphic1.20

Selection of parameter k in k-hop

To select the most effective parameter k in k-hop, we conducted experiments on the SHS27K dataset using both k values of 1 and 2. Experiment results, as presented in Table 4, showed that a k value of 1 outperformed the alternative. Due to the significant computational cost associated with calculating subgraphs for all vertices, we decided against conducting further experiments on the SHS27K dataset where the k value was greater than or equal to 3. Additionally, given the large size of the SHS148K dataset, it was not feasible to perform subgraph calculation experiments with a k value greater than or equal to 2 on our device. As a result, we ultimately settled on a k value of 1.

Table 4.

Performance comparison of different k values based on SHS27K

k-hop SHS27K
Random (%) BFS (%) DFS (%)
1-hop 90.23Inline graphic0.31 78.51Inline graphic5.25 79.81Inline graphic3.43
2-hop 90.25Inline graphic0.25 68.29Inline graphic5.58 76.18Inline graphic1.19

The impact of ASL on the performance of the model

In the previous experiments, we utilized the BCE loss function, which is a type of symmetric loss function that does not address the issue of imbalanced sample sizes with different interaction categories in the dataset. To address this limitation, we introduced the ASL and conducted some experiments using it. Experimental results presented in Table 5 revealed that, under the three partitioning algorithms (Random, BFS, and DFS) on both SHS27K and SHS148K datasets, the model based on ASL achieved higher F1-measure than the model based on the BCE loss function. Specifically, in the SHS27K dataset, F1-measure increased by approximately 0.20%, 2%, and 0.15%, respectively, while in the SHS148K dataset, F1-measure increased by approximately 1%, 1%, and 2%, respectively. These results illustrated that utilizing the ASL can effectively enhance the performance of the model on imbalanced datasets.

Table 5.

Performance comparison of asymmetric (ASL) and symmetric (BCE) loss functions

Loss Function SHS27k SHS148K
Random (%) BFS (%) DFS (%) Random (%) BFS (%) DFS (%)
BCE 90.04Inline graphic0.55 76.08Inline graphic1.22 79.67Inline graphic4.05 92.76Inline graphic0.17 74.62Inline graphic4.35 84.47Inline graphic1.20
ASL 90.23Inline graphic0.31 78.51Inline graphic5.25 79.81Inline graphic3.43 93.34Inline graphic0.12 75.14Inline graphic2.63 86.07Inline graphic1.05

Performance comparison of GNNGL-PPI with the state-of-the-art methods

In this study, GNNGL-PPI exhibited good performance across three different partitioning algorithms on the SHS27K and SHS148K datasets. To comprehensively assess its performance, we conducted a comparison with two machine learning methods, including Random Forest (RF) and Logistic Regression (LR), as well as four deep learning methods like PIPR, GNN-PPI, LDMGNN [54], and SemiGNN-PPI. Experimental results (Table 6) showed that GNNGL-PPI consistently outperformed these state-of-the-art methods in terms of F1-measure under all three partitioning algorithms on both datasets. These findings affirmed the reliability of GNNGL-PPI as a PPI predictor.

Table 6.

Performance comparison of GNNGL-PPI with the state-of-the-art methods

Method SHS27k SHS148K
Random (%) BFS (%) DFS (%) Random (%) BFS (%) DFS (%)
RF 78.45Inline graphic0.08 37.67Inline graphic1.57 35.55Inline graphic2.22 82.10Inline graphic0.20 38.96Inline graphic1.94 43.26Inline graphic3.43
LR 71.55Inline graphic0.93 43.06Inline graphic5.05 48.51Inline graphic1.87 67.00Inline graphic0.07 47.45Inline graphic1.42 51.09Inline graphic2.09
PIPR 83.31Inline graphic0.75 44.48Inline graphic4.44 57.80Inline graphic3.24 90.05Inline graphic2.59 61.83Inline graphic10.23 63.98Inline graphic0.76
GNN-PPI 87.91Inline graphic0.39 63.81Inline graphic1.79 74.72Inline graphic5.26 92.26Inline graphic0.10 71.37Inline graphic5.33 82.67Inline graphic0.85
LDMGNN 89.34Inline graphic0.44 74.56Inline graphic3.03 78.20Inline graphic2.69 92.38Inline graphic0.08 73.98Inline graphic5.51 83.79Inline graphic0.95
SemiGNN-PPI 89.51Inline graphic0.46 72.15Inline graphic2.87 78.32Inline graphic3.15 92.40Inline graphic0.22 71.78Inline graphic3.56 85.45Inline graphic1.17
GNNGL-PPI 90.23Inline graphic0.31 78.51Inline graphic5.25 79.81Inline graphic3.43 93.34Inline graphic0.12 75.14Inline graphic2.63 86.07Inline graphic1.05

Performance comparison between GNNGL-PPI and state-of-the-art methods under the multi-classification evaluation metrics

In addition to using binary classification evaluation metrics, we further utilized multi-classification evaluation metrics such as Weighted precision, Weighted recall, Weighted f1-score, Macro precision, Macro recall, Macro f1-score, Micro precision, Micro recall, and Micro f1-score to evaluate the model’s performance. Experimental results (Table 7) showed that GNNGL-PPI exhibited superior performance compared to other state-of-the-art methods. These outcomes highlighted the effectiveness of GNNGL-PPI in predicting the multi-category of PPI.

Table 7.

Performance comparison between GNNGL-PPI and other state-of-the-art methods under the multi-classification evaluation metrics

Method Multi-classification evaluation metrics SHS27K SHS148K
Random (%) BFS (%) DFS (%) Random (%) BFS (%) DFS (%)
GNNGL-PPI Weighted precision 89.49Inline graphic0.43 76.41Inline graphic1.14 76.49Inline graphic2.22 92.96Inline graphic0.20 75.37Inline graphic2.71 84.85Inline graphic0.65
LDMGNN 89.13Inline graphic0.55 73.14Inline graphic2.32 76.11Inline graphic0.08 92.89Inline graphic0.33 72.56Inline graphic3.32 83.35Inline graphic1.35
GNN-PPI 88.97Inline graphic0.17 69.75Inline graphic9.24 70.92Inline graphic1.94 92.87Inline graphic0.07 70.91Inline graphic2.55 83.08Inline graphic0.56
GNNGL-PPI Weighted recall 90.81Inline graphic0.43 79.38Inline graphic2.02 82.92Inline graphic2.98 94.00Inline graphic0.07 78.88Inline graphic6.02 88.74Inline graphic0.38
LDMGNN 88.41Inline graphic0.22 74.06Inline graphic3.47 81.15Inline graphic2.79 92.04Inline graphic0.14 65.23Inline graphic2.76 80.88Inline graphic1.67
GNN-PPI 88.10Inline graphic0.32 71.68Inline graphic9.55 73.67Inline graphic2.40 91.63Inline graphic0.25 59.12Inline graphic6.67 82.60Inline graphic1.01
GNNGL-PPI Weighted f1-score 90.13Inline graphic0.41 77.42Inline graphic0.86 79.39Inline graphic2.57 93.47Inline graphic0.07 76.66Inline graphic4.37 86.69Inline graphic0.60
LDMGNN 88.73Inline graphic0.39 72.81Inline graphic2.39 78.16Inline graphic1.30 92.45Inline graphic0.20 67.32Inline graphic2.71 81.86Inline graphic1.04
GNN-PPI 88.50Inline graphic0.25 68.30Inline graphic1.91 71.87Inline graphic1.64 92.24Inline graphic0.10 62.53Inline graphic4.21 82.72Inline graphic0.77
GNNGL-PPI Macro precision 86.36Inline graphic0.66 70.00Inline graphic5.68 72.63Inline graphic2.65 89.76Inline graphic0.47 72.83Inline graphic2.70 78.44Inline graphic0.77
LDMGNN 85.92Inline graphic0.40 70.08Inline graphic2.81 73.16Inline graphic0.47 89.60Inline graphic0.30 71.60Inline graphic3.26 78.26Inline graphic0.87
GNN-PPI 85.88Inline graphic0.40 57.88Inline graphic20.34 67.37Inline graphic1.51 89.64Inline graphic0.15 69.18Inline graphic3.32 77.65Inline graphic0.57
GNNGL-PPI Macro recall 86.52Inline graphic0.60 72.22Inline graphic4.15 78.09Inline graphic2.74 90.20Inline graphic0.06 75.55Inline graphic5.97 83.71Inline graphic0.31
LDMGNN 83.89Inline graphic0.67 64.32Inline graphic2.12 74.34Inline graphic1.63 87.32Inline graphic0.07 58.13Inline graphic1.70 74.14Inline graphic1.84
GNN-PPI 83.43Inline graphic0.45 52.88Inline graphic8.32 69.44Inline graphic1.37 87.07Inline graphic0.40 54.24Inline graphic8.07 72.37Inline graphic0.88
GNNGL-PPI Macro f1-score 86.40Inline graphic0.33 70.37Inline graphic4.35 74.99Inline graphic2.72 89.97Inline graphic0.26 73.70Inline graphic3.83 80.83Inline graphic0.55
LDMGNN 84.80Inline graphic0.52 65.85Inline graphic2.10 72.80Inline graphic0.35 88.47Inline graphic0.16 62.11Inline graphic3.11 75.77Inline graphic1.47
GNN-PPI 84.55Inline graphic0.13 52.96Inline graphic14.06 68.00Inline graphic1.34 88.30Inline graphic0.18 58.62Inline graphic5.55 73.82Inline graphic0.66
GNNGL-PPI Micro precision 89.49Inline graphic0.49 75.74Inline graphic1.09 76.23Inline graphic2.14 92.95Inline graphic0.23 74.77Inline graphic3.26 84.35Inline graphic0.77
LDMGNN 89.24Inline graphic0.48 73.52Inline graphic2.64 76.10Inline graphic0.18 93.05Inline graphic0.32 72.28Inline graphic3.12 82.90Inline graphic1.39
GNN-PPI 89.11Inline graphic0.25 71.54Inline graphic5.70 70.44Inline graphic2.11 92.98Inline graphic0.09 70.03Inline graphic1.74 83.27Inline graphic0.57
GNNGL-PPI Micro recall 90.82Inline graphic0.43 79.39Inline graphic2.03 82.93Inline graphic2.98 94.01Inline graphic0.07 78.88Inline graphic6.03 88.76Inline graphic0.36
LDMGNN 88.42Inline graphic0.22 74.06Inline graphic3.47 81.15Inline graphic2.79 92.05Inline graphic0.15 65.23Inline graphic2.77 80.88Inline graphic1.67
GNN-PPI 88.11Inline graphic0.32 71.69Inline graphic9.55 73.34Inline graphic2.26 91.64Inline graphic0.25 59.12Inline graphic6.68 82.60Inline graphic1.02
GNNGL-PPI Micro f1-score 90.15Inline graphic0.39 77.49Inline graphic0.56 79.43Inline graphic2.53 93.48Inline graphic0.09 76.72Inline graphic4.29 86.50Inline graphic0.56
LDMGNN 88.82Inline graphic0.32 73.74Inline graphic2.48 78.12Inline graphic1.10 92.54Inline graphic0.20 68.52Inline graphic2.26 81.86Inline graphic0.94
GNN-PPI 88.60Inline graphic0.29 70.81Inline graphic1.72 71.83Inline graphic1.61 92.31Inline graphic0.09 63.86Inline graphic3.43 82.87Inline graphic0.70

Performance comparison of some statistical tests between GNNGL-PPI and state-of-the-art methods

In this study, we dealt with an imbalanced dataset where the number of samples across different categories varied. To accurately evaluate the model’s performance on this imbalanced dataset, we opted for the F1 metric as it provided a more accurate reflection of the model’s true performance. Consequently, we conducted some statistical tests on the F1 metric of three methods including GNNPPI, LDMGN, and GNNGL-PPI using three distinct partitioning algorithms on both SHS27K and SHS148K datasets. Experiment results, as shown in Table 8, indicated that GNNGL-PPI exhibited slightly superior performance across two statistical testing metrics compared to two other state-of-the-art methods.

Table 8.

Performance comparison of some statistical tests between GNNGL-PPI and state-of-the-art methods

Rank Number of cases Rank Mean Sum of Ranks
GNNPPI-GNNGLPPI Negative Rank 18a 9.50 171.00
Positive Rank 0b 0.00 0.00
Bind Value 0c
Total 18
LDMGN-GNNGLPPI Negative Rank 17d 9.53 162.00
Positive Rank 1e 9.00 9.00
Bind Value 0f
Total 18

(a) GNNPPI < GNNGLPPI (b) GNNPPI > GNNGLPPI (c) GNNPPI = GNNGLPPI

(d) LDMGN < GNNGLPPI (e) LDMGN > GNNGLPPI (f) LDMGN = GNNGLPPI

Statistic tests (a) Z Progressive significance (two tailed) p-value
GNNPPI-GNNGLPPI -3.724b 0.000 0.000196
LDMGN-GNNGLPPI -3.332b 0.001 0.000863
(a) Wilconxon sign rank test (b) Based on positive rank

Statistical analysis of prediction accuracy of GNNGL-PPI and GNN-PPI under different interaction categories

Table 6 showed that the GNNGL-PPI method outperformed other state-of-the-art methods for comparison under three different partitioning algorithms on both the SHS27K and SHS148K datasets. To further evaluate the predictive performance of GNNGL-PPI across different interaction categories, we compared its prediction accuracy with the GNN-PPI method under the BFS partitioning algorithm using the SHS27K dataset (Fig. 2). Statistical analysis revealed that for the interaction category Reaction with 3164 samples (Table 1), after being divided by the BFS algorithm, 613 test samples were obtained. GNNGL-PPI correctly predicted 496 of the 569 samples it identified, yielding an accuracy of 80.91% (Fig. 2a). In comparison, GNN-PPI correctly predicted 434 of the 510 samples, achieving an accuracy of 70.79%. Similarly, for the Binding, Ptmod, Activation, Inhibition, Catalysis, and Expression categories, GNNGL-PPI achieved a higher accuracy (within parentheses) than that of GNN-PPI (85.19%, 71.55%), (86.02%, 71.55%), (84.35%, 81.42%), (86.67%, 85.74%), (87.47%, 74.94%), and (42.55%, 28.72%), respectively (Fig. 2b). In all interaction categories except for Action and Inhibition, GNNGL-PPI exhibited classification accuracy that was more than 10% higher than that of GNN-PPI. Notably, under the Expression category, there were only 94 test samples. Despite this, GNNGL-PPI and GNN-PPI correctly predicted 27 and 40 samples, respectively, with an accuracy improvement of 13.83%. This suggested that our proposed GNNGL-PPI method could learn features from small sample sizes and made accurate predictions, mitigating the problem of low prediction accuracy due to imbalanced dataset categories with small sample sizes.

Fig. 2.

Fig. 2

Performance comparison between GNNGL-PPI and GNN-PPI across different interaction categories using the BFS partitioning algorithm on the SHS27K dataset. (a) In each interaction category, GNNGL-PPI outperformed GNN-PPI by predicting more correct samples (green color) as compared to the latter (red color). Particularly in the Expression category, GNNGL-PPI correctly predicted 40 positive samples, while GNN-PPI only achieved 27. (b) With the exception of the Action and Inhibition interaction categories, the prediction accuracy of both methods was similar. However, in the remaining five interaction categories, GNNGL-PPI exhibited a prediction accuracy that was more than 10% higher than that of GNN-PPI.

Interpretability analysis of the effectiveness of GNNGL-PPI

To thoroughly evaluate the effectiveness of GNNGL-PPI, we applied the widely recognized dimensionality reduction algorithm t-SNE [55] to the SHS27K dataset. This dataset was partitioned using the Random alogrithm. t-SNE proved to be an ideal choice for our analysis as it preserved the proximity of closely positioned data points even after reducing their dimensions. At the same time, it effectively maintained the separation between originally distant data points. We utilized t-SNE to gain insights into the effectiveness of GNNGL-PPI by visualizing the clustered representations resulting from dimensionality reduction at different epochs of model training. Due to the rapid convergence of GNNGL-PPI in the early stages of training, we applied t-SNE to reduce the dimensionality of features learned in the 1st, 5th, 20th, 50th, 100th, and 200th epochs. The clustering visualization obtained after one epoch of model training (Fig. 3a) revealed that data points representing different interaction categories were intertwined, with no distinct separation between them. However, as the number of training epochs increased, the clustering visualizations after the 5th, 20th, 50th, and 100th epochs (Fig. 3b to e) clearly showed that data points representing different interaction categories gradually moved away from each other, while data points representing the same interaction category started to cluster together. By the 200th epoch, clear boundaries between data points of different interaction categories were observed (Fig. 3f). It’s important to note that since this study treated PPI as a multi-category classification task, some data points from different interaction categories may still appear in the same cluster. Nevertheless, the clustering visualization results from different training epochs indicated that GNNGL-PPI was an effective method for predicting multi-category of PPI.

Fig. 3.

Fig. 3

Interpretability analysis of the effectiveness of GNNGL-PPI. We used t-SNE algorithm to elucidate GNNGL-PPI’s effectiveness by contrasting the clustered visualizations resulting from dimensionality reduction for multiple different epochs of the model training of GNNGL-PPI. The clustering visualization results, represented as (a) to (f), corresponded to the features learned in the first, fifth, 20th, 50th, 100th, and 200th epochs of model training, respectively. Upon examining these results, a clear pattern emerged: data points representing different interaction categories gradually moved apart as the number of training epochs increased, while data points representing the same interaction category gradually clustered together. The clustering visualization results showed that GNNGL-PPI was an effective PPI multi-category predictor

Discussions

Although GNNGL-PPI has exhibited a fine performance in predicting multiple categories of protein-protein interactions and can provide explanatory analysis for its effectiveness, it also has certain shortcomings:

  1. The training of GNNGL-PPI was conducted on two publicly available benchmark datasets, and the model’s learned deep features were limited. This limitation may result in a gap between the generalization performance of GNNGL-PPI and users’ expected thresholds.

  2. GNNGL-PPI relied on sequence features and GO annotations of proteins, while overlooking protein structural features. However, protein structural features often play a crucial role in the performance of PPI models.

  3. The utilization of GNNGL-PPI requires users to input protein-protein interaction networks, which undoubtedly increases the complexity for non-professionals and restricts the application and promotion of GNNGL-PPI.

In response to these shortcomings, we have carefully considered the issues and proposed potential solutions:

  1. To address the limited deep features learned by GNNGL-PPI, we suggest crawling a more comprehensive PPI interaction dataset from databases. By constructing a more extensive PPI interaction network for training GNNGL-PPI, we can gradually enhance its generalization performance.

  2. In order to incorporate protein structural features into GNNGL-PPI, we recommend extracting such features from protein topology graphs or three-dimensional grids using graph neural networks or 3D-CNN. This approach will enrich the protein features and improve the model’s performance.

  3. To alleviate the challenges faced by non-professionals in utilizing GNNGL-PPI, we propose developing an online service tool. With this tool, users would only need to input protein pairs, and the tool would automatically construct a protein-protein interaction network and predict the interaction categories between proteins.

In order to further improve the performance of GNNGL-PPI, we believe that in-depth research should be conducted in the following areas in the future:

  1. Introducing self-supervised contrastive learning into the GNNGL-PPI method enhances its feature extraction ability, enabling the method to learn more key implicit features that affect multi-category prediction of protein-protein interactions.

  2. At present, the protein features used are still relatively single, and multiple modal information such as protein sequences and structures should be integrated to promote the model to achieve better performance based on the rich features of proteins.

  3. Deepening the interpretive analysis of the effectiveness of protein-protein interaction prediction methods will not only promote the application of this method, but also provide support for the research of new methods.

GNNGL-PPI not only provides PPI prediction services to support users in understanding the mechanism of protein-protein interaction, but also the predicted PPI categories or PPI networks constructed from them can provide deep and rich features for drug-target interaction prediction and other related tasks, which is helpful for drug screening, repositioning, and target recognition research. We will briefly provide relevant work in Appendix A.

Conclusions

GNNGL-PPI is a novel method for predicting multi-category of PPI based on graph neural networks. This innovative approach leverages the power of GIN and GIN-AK to extract the features from both the global graph and local subgraph of the PPI network. Additionally, the use of ASL helps enhance the model’s performance on imbalanced datasets. Experimental results have showed that GNNGL-PPI surpasses the performance of existing state-of-the-art methods for PPI prediction on two standard datasets. The effectiveness of GNNGL-PPI is further supported by the t-SNE algorithm, which provides visual evidence of the model’s capability. Therefore, GNNGL-PPI is a reliable multi-category prediction tool for protein-protein interactions.

Appendix A

Here, we will briefly mention that predicting protein-protein interactions can significantly contribute to drug-target interactions, target identification, and other related research in drug discovery, ultimately advancing drug development.

TripletMultiDTI [56] utilized protein-protein interaction (PPI) and drug-drug interaction (DDI) networks as supplementary knowledge. It integrated multimodal information, including drugs, proteins, DDI networks, and PPI networks, as inputs to predict the affinity of drug-target pairs. Furthermore, a review article [57] elaborated on the crucial role of protein-protein interactions in performing essential cellular functions. These interactions have been pivotal drug targets for the past two decades and are fundamental to drug development and design. Additionally, reference [58] provided a detailed exploration of the vital role of protein-protein interactions in structural biology and drug discovery.

Author contributions

All authors contributed to the study conception and design. X.Z.: Ideas, writingoriginal manuscript, writingreview and editing, writingrevised manuscript. F.-F.M.: Creation of models, data curation, implementation of the computer code and supporting algorithms. M.-L.W.: Supervision. S.-J.L.: Data analysis. Y.L.: Development of methodology, formal analysis, modification guidance. All authors read and approved the final manuscript.

Funding

This work was supported by the National Natural Sciences Foundation of China (No. 62366002), Yunnan Fundamental Research Projects (No. 202101BA070001-227), and a grant (No. 2023KF005) from State Key Laboratory for Conservation and Utilization of Bio-Resources in Yunnan, Yunnan University.

Data availability

The source data and code repository can be accessed at https://github.com/dldxzx/GNNGL-PPI.

Declarations

Ethics approval and consent to participate

Not applicable.

Consent for publication

Not applicable.

Competing interests

The authors declare no competing interests.

Footnotes

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

References

  • 1.Raman K. Construction and analysis of protein–protein interaction networks. Autom Exp. 2010;2:2. doi: 10.1186/1759-4499-2-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Fields S, Sternglanz R. The two-hybrid system: an assay for protein-protein interactions. Trends Genet. 1994;10:286–92. doi: 10.1016/0168-9525(90)90012-U. [DOI] [PubMed] [Google Scholar]
  • 3.Zhu H, Bilgin M, Bangham R, et al. Global analysis of protein activities using proteome chips. Science. 2001;293:2101–5. doi: 10.1126/science.1062191. [DOI] [PubMed] [Google Scholar]
  • 4.Tong AHY, Evangelista M, Parsons AB, et al. Systematic Genetic Analysis with ordered arrays of yeast deletion mutants. Science. 2001;294:2364–8. doi: 10.1126/science.1065810. [DOI] [PubMed] [Google Scholar]
  • 5.Hu L, Wang X, Huang Y-A, et al. A survey on computational models for predicting protein–protein interactions. Brief Bioinform. 2021;22:bbab036. doi: 10.1093/bib/bbab036. [DOI] [PubMed] [Google Scholar]
  • 6.Hayashi T, Matsuzaki Y, Yanagisawa K, et al. MEGADOCK-Web: an integrated database of high-throughput structure-based protein-protein interaction predictions. BMC Bioinformatics. 2018;19:62. doi: 10.1186/s12859-018-2073-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Wu J, Liu B, Zhang J, et al. DL-PPI: a method on prediction of sequenced protein–protein interaction based on deep learning. BMC Bioinformatics. 2023;24:473. doi: 10.1186/s12859-023-05594-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Tsukiyama S, Hasan MM, Fujii S, et al. LSTM-PHV: prediction of human-virus protein–protein interactions by LSTM with word2vec. Brief Bioinform. 2021;22:bbab228. doi: 10.1093/bib/bbab228. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Kibar G, Vingron M. Prediction of protein–protein interactions using sequences of intrinsically disordered regions. Proteins. 2023;91:980–90. doi: 10.1002/prot.26486. [DOI] [PubMed] [Google Scholar]
  • 10.Romero-Molina S, Ruiz‐Blanco YB, Harms M, et al. PPI‐Detect: a support vector machine model for sequence‐based prediction of protein–protein interactions. J Comput Chem. 2019;40:1233–42. doi: 10.1002/jcc.25780. [DOI] [PubMed] [Google Scholar]
  • 11.Zhang M, Su Q, Lu Y et al. Application of machine learning approaches for protein-protein interactions prediction. MC 2017; 13. [DOI] [PubMed]
  • 12.Sze-To A, Fung S, Lee E-SA, et al. Prediction of protein–protein Interaction via co-occurring aligned pattern clusters. Methods. 2016;110:26–34. doi: 10.1016/j.ymeth.2016.07.018. [DOI] [PubMed] [Google Scholar]
  • 13.Chatterjee P, Basu S, Kundu M et al. PPI_SVM: prediction of protein-protein interactions using machine learning, domain-domain affinities and frequency tables. Cell Mol Biology Lett 2011; 16. [DOI] [PMC free article] [PubMed]
  • 14.Xu D, Xu H, Zhang Y, et al. Protein-protein interactions Prediction based on Graph Energy and protein sequence information. Molecules. 2020;25:1841. doi: 10.3390/molecules25081841. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Su X-R, Hu L, You Z-H, et al. Multi-view heterogeneous molecular network representation learning for protein–protein interaction prediction. BMC Bioinformatics. 2022;23:234. doi: 10.1186/s12859-022-04766-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Ahmed I, Witbooi P, Christoffels A. Prediction of human- Bacillus anthracis protein–protein interactions using multi-layer neural network. Bioinformatics. 2018;34:4159–64. doi: 10.1093/bioinformatics/bty504. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Canziani A, Paszke A, Culurciello E. An Analysis of Deep Neural Network Models for Practical Applications. 2017.
  • 18.He K, Zhang X, Ren S et al. Deep Residual Learning for Image Recognition. 2015.
  • 19.Zaremba W, Sutskever I, Vinyals O. Recurr Neural Netw Regularization. 2015.
  • 20.Vaswani A, Shazeer N, Parmar N et al. Atten Is all You Need. 2017.
  • 21.Xu K, Hu W, Leskovec J et al. HOW POWERFUL ARE GRAPH NEURAL NETWORKS? international conference on learning representations. 2019.
  • 22.Sun T, Zhou B, Lai L, et al. Sequence-based prediction of protein protein interaction using a deep-learning algorithm. BMC Bioinformatics. 2017;18:277. doi: 10.1186/s12859-017-1700-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Li X, Han P, Wang G, et al. SDNN-PPI: self-attention with deep neural network effect on protein-protein interaction prediction. BMC Genomics. 2022;23:474. doi: 10.1186/s12864-022-08687-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.Mahapatra S, Gupta VR, Sahu SS, et al. Deep neural network and Extreme Gradient boosting based hybrid classifier for Improved Prediction of Protein-Protein Interaction. IEEE/ACM Trans Comput Biol Bioinf. 2022;19:155–65. doi: 10.1109/TCBB.2021.3061300. [DOI] [PubMed] [Google Scholar]
  • 25.Zhou X, Song H, Li J. Residue-frustration-based prediction of protein–protein interactions using machine learning. J Phys Chem B. 2022;126:1719–27. doi: 10.1021/acs.jpcb.1c10525. [DOI] [PubMed] [Google Scholar]
  • 26.Asim MN, Ibrahim MA, Malik MI, et al. ADH-PPI: an attention-based deep hybrid model for protein-protein interaction prediction. iScience. 2022;25:105169. doi: 10.1016/j.isci.2022.105169. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Chen W, Wang S, Song T, et al. DCSE:Double-Channel-Siamese-Ensemble model for protein protein interaction prediction. BMC Genomics. 2022;23:555. doi: 10.1186/s12864-022-08772-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Soleymani F, Paquet E, Viktor H, et al. Protein–protein interaction prediction with deep learning: a comprehensive review. Comput Struct Biotechnol J. 2022;20:5316–41. doi: 10.1016/j.csbj.2022.08.070. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.Jumper J, Evans R, Pritzel A, et al. Highly accurate protein structure prediction with AlphaFold. Nature. 2021;596:583–9. doi: 10.1038/s41586-021-03819-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.Kim G, Lee S, Karin EL et al. Easy and accurate protein structure prediction using ColabFold. 2023. [DOI] [PubMed]
  • 31.Jha K, Saha S, Singh H. Prediction of protein–protein interaction using graph neural networks. Sci Rep. 2022;12:8360. doi: 10.1038/s41598-022-12201-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Song B, Luo X, Luo X, et al. Learning spatial structures of proteins improves protein–protein interaction prediction. Brief Bioinform. 2022;23:bbab558. doi: 10.1093/bib/bbab558. [DOI] [PubMed] [Google Scholar]
  • 33.Yang F, Fan K, Song D, et al. Graph-based prediction of protein-protein interactions with attributed signed graph embedding. BMC Bioinformatics. 2020;21:323. doi: 10.1186/s12859-020-03646-8. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Gao Z, Jiang C, Zhang J, et al. Hierarchical graph learning for protein–protein interaction. Nat Commun. 2023;14:1093. doi: 10.1038/s41467-023-36736-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35.Singh R, Devkota K, Sledzieski S, et al. Topsy-Turvy: integrating a global view into sequence-based PPI prediction. Bioinformatics. 2022;38:i264–72. doi: 10.1093/bioinformatics/btac258. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36.Chen M, Ju CJ-T, Zhou G, et al. Multifaceted protein–protein interaction prediction based on siamese residual RCNN. Bioinformatics. 2019;35:i305–14. doi: 10.1093/bioinformatics/btz328. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37.Lv G, Hu Z, Bi Y, et al. Learning unknown from correlations. Graph Neural Network for Inter-novel-protein Interaction Prediction; 2021.
  • 38.Zhao Z, Qian P, Yang X, et al. SemiGNN-PPI: self-ensembling multi-graph neural network for efficient. and Generalizable Protein-Protein Interaction Prediction; 2023.
  • 39.Kang Y, Elofsson A, Jiang Y, et al. AFTGAN: prediction of multi-type PPI based on attention free transformer and graph attention network. Bioinformatics. 2023;39:btad052. doi: 10.1093/bioinformatics/btad052. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40.Hu F, Hu Y, Zhang W, et al. A Multimodal Protein Representation Framework for Quantifying Transferability across Biochemical Downstream Tasks. Adv Sci. 2023;10:2301223. doi: 10.1002/advs.202301223. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41.Jeong JC, Lin X, Chen X-W. On position-specific Scoring Matrix for protein function prediction. IEEE/ACM Trans Comput Biol Bioinf. 2011;8:308–15. doi: 10.1109/TCBB.2010.93. [DOI] [PubMed] [Google Scholar]
  • 42.Remmert M, Biegert A, Hauser A, et al. HHblits: lightning-fast iterative protein sequence searching by HMM-HMM alignment. Nat Methods. 2012;9:173–5. doi: 10.1038/nmeth.1818. [DOI] [PubMed] [Google Scholar]
  • 43.Li H. Deep learning for natural language processing: advantages and challenges. Natl Sci Rev. 2018;5:24–6. doi: 10.1093/nsr/nwx110. [DOI] [Google Scholar]
  • 44.Maas AL, Hannun AY, Ng AY et al. Rectifier nonlinearities improve neural network acoustic models. Proc. icml. 2013; 30:3.
  • 45.Xu K, Hu W, Leskovec J et al. How Powerful are Graph Neural Networks? 2019.
  • 46.Chen Z, Villar S, Chen L, et al. On the equivalence between graph isomorphism testing and function approximation with gnns. Advances in neural information processing systems 2019; 32.
  • 47.Weisfeiler B, Leman A. The reduction of a graph to canonical form and the algebra which appears therein. nti, Series 1968; 2:12–16.
  • 48.Ridnik T, Ben-Baruch E, Zamir N et al. Asymmetric Loss For Multi-Label Classification. 2021 IEEE/CVF International Conference on Computer Vision (ICCV) 2021; 82–91.
  • 49.Lin T-Y, Goyal P, Girshick R et al. Focal loss for dense object detection. Proceedings of the IEEE international conference on computer vision. 2017; 2980–2988.
  • 50.Jadon S. A survey of loss functions for semantic segmentation. 2020 IEEE Conference on Computational Intelligence in Bioinformatics and Computational Biology (CIBCB) 2020; 1–7.
  • 51.Li L, Doroslovacki M, Loew MH. Approximating the gradient of cross-entropy loss function. IEEE Access. 2020;8:111626–35. doi: 10.1109/ACCESS.2020.3001531. [DOI] [Google Scholar]
  • 52.Kingma DP, Ba J, Adam. A Method for Stochastic Optimization. 2017.
  • 53.Zeng M, Zou B, Wei F et al. Effective prediction of three common diseases by combining SMOTE with Tomek links technique for imbalanced medical data. 2016; 225–8.
  • 54.Zhong W, He C, Xiao C, et al. Long-distance dependency combined multi-hop graph neural networks for protein–protein interactions prediction. BMC Bioinformatics. 2022;23:521. doi: 10.1186/s12859-022-05062-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 55.Linderman GC, Rachh M, Hoskins JG, et al. Fast interpolation-based t-SNE for improved visualization of single-cell RNA-seq data. Nat Methods. 2019;16:243–5. doi: 10.1038/s41592-018-0308-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 56.Dehghan A, Razzaghi P, Abbasi K, et al. TripletMultiDTI: Multimodal representation learning in drug-target interaction prediction with triplet loss function. Expert Syst Appl. 2023;232:120754. doi: 10.1016/j.eswa.2023.120754. [DOI] [Google Scholar]
  • 57.Lee AC-L, Harris JL, Khanna KK, et al. A Comprehensive Review on current advances in peptide Drug Development and Design. IJMS. 2019;20:2383. doi: 10.3390/ijms20102383. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 58.Jubb H, Higueruelo AP, Winter A, et al. Structural biology and drug discovery for protein–protein interactions. Trends Pharmacol Sci. 2012;33:241–8. doi: 10.1016/j.tips.2012.03.006. [DOI] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

The source data and code repository can be accessed at https://github.com/dldxzx/GNNGL-PPI.


Articles from BMC Genomics are provided here courtesy of BMC

RESOURCES