Skip to main content
Bioinformatics and Biology Insights logoLink to Bioinformatics and Biology Insights
. 2025 Aug 29;19:11779322251366087. doi: 10.1177/11779322251366087

R2eGIN: Residual Reconstruction Enhanced Graph Isomorphism Network for Accurate Prediction of Poly (ADP–Ribose) Polymerase Inhibitors

Candra Zonyfar 1, Soualihou Ngnamsie Njimbouom 1, Sophia Mosalla 1, Jeong-Dong Kim 1,2,3,
PMCID: PMC12397607  PMID: 40894331

Abstract

An advanced graph neural network (GNN) is of great promise to facilitate predicting Poly ADPribose polymerase inhibitors (PARPi). Recent studies design models by leveraging graph representations and molecular descriptor representations, unfortunately, still face challenges in comprehensively capturing spatial relationships and contextual information between atoms. Moreover, combining molecular descriptors with graph representations may introduce information redundancy or lead to the loss of intrinsic molecular structures. To this end, we proposed a novel Residual Reconstruction Enhanced Graph Isomorphism Network (R2eGIN) learning model. Specifically, we first designed a residual GIN to learn molecular representations, reduced the impact of vanishing gradients, and enabled the model to capture long-range dependencies. Then, the reconstruction block, by predicting adjacency matrices and node features, was adopted to reconstruct the input graph. To prove the effectiveness of the proposed model, extensive experiments were conducted on 4 data sets of PARPi and compared with 7 existing models. Our evaluation of R2eGIN, conducted using 4 PARPi data sets, shows that the proposed model is comparable to or even outperforms other state-of-the-art models for PARPi prediction. Furthermore, R2eGIN can revolutionize the drug repurposing process through a substantial reduction in the time and costs commonly encountered in traditional drug development methods.

Keywords: Poly ADP-ribose polymerase, PARP inhibitor prediction, graph neural network, graph isomorphism network, drug development

Introduction

Recent discoveries regarding Deoxyribonucleic Acid (DNA) damage repair processes have led to increased attention from researchers and medicinal chemists toward novel antitumor PARPi.1-4 Poly(ADP-ribose) polymerase (PARP), a protein superfamily comprising 17 enzymes, plays a crucial role in DNA damage repair and genome stability maintenance, as well as regulating various cellular processes, including DNA damage signaling, chromatin remodeling, transcription, replication fork stabilization, detection of unligated Okazaki fragments during replication, inflammation, and metabolism.5,6 PARP-1 has garnered attention in the therapeutic field due to its overexpression observed in various cancers, including ovarian, breast, skin, colorectal, lung, and others. 7 PARP-1 plays a key role in cell differentiation, gene transcription, inflammation, mitosis, and cell death, contributing to the antitumor activity of PARP-1 inhibitors. 8 The interaction between PARPi and the β-NAD + binding site on PARP1 results in the inhibition of the enzyme’s function. Consequently, DNA repair mechanisms are disrupted, leading to the accumulation of genetic errors and ultimately cell death.9,10 PARP-2 plays a significant role in regulating epigenetic modifications, promoting cell proliferation, and mediating inflammatory processes. According to existing studies, PARP-2 inhibition is strongly associated with hematological toxicity. 1 In addition to PARP-1 and PARP-2, whose roles are associated with DNA damage repair, there are also PARP-5a (TNKS1) and PARP-5b (TNKS2), which are members of the Tankyrase (TNKS) family. PARP-5b is recognized as a key factor involved in maintaining genetic stability, as it controls DNA damage repair and cell cycle regulation. More specifically, PARP-5B regulates various mitotic functions, including centrosome function, mitotic spindle assembly, mitotic checkpoints, telomere length, and telomere cohesion. 6 The important role of PARP-5a and PARP-5b also includes inhibiting their DNA-binding capacity and inducing Wnt/β-catenin signaling pathway regulation through PARylation-mediated proteasomal degradation. 11 Most importantly, a growing body of research indicates that PARP inhibitors have significant potential in treating noncancer conditions, such as ischemic stroke, 12 cardiovascular disease, 13 and diabetic retinopathy. 14

Recent research highlights the significant potential of PARP inhibitors, such as CVL218, which is undergoing phase I clinical trials, in inhibiting SARS-CoV-2 replication 15 and combating the harmful effects of COVID-19 through various mechanisms.15,16 Nevertheless, approved PARP inhibitors face challenges in clinical use, including toxicity, selectivity issues, and drug resistance.17-19 Therefore, the development of novel PARP inhibitors for the treatment of both tumor and nontumor diseases is highly necessary.

Traditional drug discovery often necessitates numerous experiments that are time-consuming and costly. Accordingly, computational approaches serve as a viable option for early-stage drug discovery, aiming to reduce wet-lab experiments and minimize costs. In the field of bioinformatics and molecular properties, GNN demonstrate their success and effectiveness.20,21 GNN are capable of exploring intrinsic information embedded within nodes, as well as inter-node relationships, and transforming them into valuable features for modeling. This is particularly suitable for data such as molecules, where molecular structures and functional relationships between molecules correlate with significant biological information.

Although the development of GNN models in molecular property prediction research, including PARP inhibitor prediction, has attracted research attention in recent years, there remains a need to enhance information propagation between layers, enabling models to capture deeper molecular features by leveraging key inter-atomic relationships for PARP inhibitor molecule prediction. Furthermore, as GNN become more complex, models face the vanishing gradient problem due to excessive layer depth. We argue that implementing residual learning and a reconstruction network can improve the model’s capacity to predict PARP inhibitor molecules. Applying residual learning to the GNN backbone helps address the vanishing gradient problem and enhances information propagation between layers, allowing models to capture deeper molecular features. Meanwhile, a reconstruction network enables models to refine molecular representations by preserving spatial structures and topological relationships, thereby improving generalization and molecular property prediction, including in PARP inhibitor identification.

In this study, we introduce the R2eGIN model for the prediction of PARPi. The proposed model is an extension of the graph isomorphism network (GIN), incorporating residual connections and reconstruction blocks to improve PARPi prediction accuracy. This study represents the first application of the GIN approach for PARPi prediction, where the integration of residual connections and reconstruction mechanisms enhances the model’s ability to capture molecular structural relationships and improves the stability of information propagation within the network, thereby enabling more accurate predictions. An ablation study was conducted to analyze the impact of each additional component on prediction performance. Finally, the effectiveness of the proposed model was validated by comparing it with existing methods on the same data sets.

Methods

R2eGIN model

The overview of the R2eGIN model is illustrated in Figure 1, which comprises 3 main parts: a GNN backbone, the residual connection, and a reconstruction block. Each of the parts is discussed comprehensively in the following subsections, respectively.

Figure 1.

Explain R2eGIN’s architecture with residual connections, detailing each component for molecule property prediction. Generate an alt text summary of this.

Overview of R2eGIN.

Graph neural networks

In this study, a GNN-based approach was employed to aggregate nodes and edges, thereby generating graph-level representations. The GNN model employed in this paper was an extended version of the GIN. We considered a set of molecules {G1,,GN} represented by graphs G=(V,E) . Here, V is the set of atoms (nodes) and EV×V is the set of chemical bonds (edges) between atoms. Each molecule Gi is associated with a label yiY , indicating whether the molecule is an active or inactive PARP inhibitor. The objective was to perform classification, specifically predicting the label yi for each molecule Gi . To achieve this, the model was developed based on the GIN backbone, aimed at learning effective molecular vector representations.22-24 The proposed model maps each molecular graph Gi to a label through a learning fθ:GY . As illustrated in Figure 1, X,X and X1 represent distinct molecular feature representations extracted from graph G. X is the initial representation, X is the representation after passing through the first GIN layer, and X1 is the representation after an additional step that enhances the model’s understanding of the molecular structure. These GIN layers were used to improve the model’s ability to differentiate graph structures more expressively using a sum-based aggregation function, which has higher discriminative power compared with the average or maximization aggregation methods used in standard GCN. Then, X was processed by the second GIN layer, producing the X1 representation. This additional GIN layer aimed to further refine the feature representation, allowing the model to learn more complex molecular patterns and structures.

The first GIN layer generated an updated feature representation, which was then passed to the second GIN layer. Here, we introduced residual connections, which added the output of the first GIN to the output of the second GIN. These residual connections helped address the vanishing gradient problem and enabled the network to learn more complex features. The output of the second GIN was then passed to the final GIN layer through a second residual connection, which added the output of the initial input X to the output of the second GAT. Before entering the final GIN layer, a ReLU activation function was applied. The third GIN layer generates the final representation of the molecule, which is then globally pooled using global mean pooling (GMP) to produce a single vector representation of the entire molecule. Finally, we introduced a reconstruction component as a regularization mechanism to improve prediction performance. It is important to note that the reconstruction task is not the end goal of our model, but rather a means to enhance predictive capabilities. We consider it a form of regularization that helps prevent overfitting and improve model generalization. The implementation of residual connections and the addition of a reconstruction block to the GIN layers aimed to enable the model to effectively learn rich and informative molecular feature representations, which are crucial for our targeted PARPi prediction task.

The R2eGIN was constructed with 3 message-passing layers to obtain graph-level embeddings hG . These GIN layers computed the AGGREGATE and COMBINE steps using a multilayer perceptron (MLP). As a universal function approximator, MLP nonlinearly transforms input features into discriminative representations. The AGGREGATE operation is a neighborhood summarizer. In this process, each atom (node) collects and combines data from its directly connected atoms and bonds. 25 The COMBINE operation is a critical step, It integrates the aggregated neighborhood information with the node’s own features to update its embedding24,26

aul=AGGREGATEl({hul1uN(v)},{hel1:e=(v,u)}) (1)
hvl=σ(BN(MLPl(COMBINEl(hvl1,aul))))+hvl1 (2)

Here, aul denotes the single aggregated representation for node v at layer l , calculated from all neighbors uN(v) . The subscript u explicitly indicates that the aggregation source is the neighbor set of v . hvl represents the embedding of node v , hul represents the embedding of neighbor u , and hel is the embedding of edge e=(u,v) . MLP performed the UPDATE operation followed by batch normalization (BN), σ (ReLU activation), and the addition of a residual connection from the previous layer hvl1 . Specifically, these operations were given by

hvl=σ(BN(MLPl(uΝ(v)vhul1)))+hvl1 (3)

The node embeddings hv , in the final message-passing iteration, served to determine the graph-level representation hG

hG=1|V|vVhvL (4)

where |V| is the number of nodes in the graph, and hvL is the embedding of node v at the final layer L .

Figure 2 illustrates how information was propagated and aggregated within the R2eGIN layers. 22 Information from the nodes was gathered through a message-passing process, where this neighbor information, represented as hul1 , was aggregated to produce aul . In the Message Aggregation process, the updated node representation hul was computed by combining the previous node representation hul1 and the aggregated neighbor information aul . Residual connections were added by summing hul1 with the combination result, enabling information from previous layers to directly contribute to the current layer.

Figure 2.

R2eGIN leverages node-edge aggregation and neighborhood combination to enhance message passing.

Node-edge neighbor aggregation and node-neighborhood combination within R2eGIN.

Residual learning

In a GNN architecture designed for molecular tasks, shallow models typically failed to effectively capture key information and interactions within molecular data graphs. Meanwhile, as depth increased, models encountered the vanishing gradient problem, leading to performance degradation and suboptimal results.27,28 To mitigate these issues, a residual approach was used. R2eGIN extended the GIN layer with 2 stacked layers by applying residual connections.

Figure 3 shows the block structure of R2eGIN, where refers to the GIN block followed by an activation layer, and P represents a projection function to project the input into the same feature space as the output of the GIN block. The projection function of the residual block P(x) was implemented as a convolution, which introduced additional parameters and increased computational complexity. Meanwhile, maintaining the projection function could enhance network effectiveness, thereby maximizing model performance.

Figure 3.

The structure of the R2eGIN consists of multiple inputs and outputs, with “P” being an input to a “GIN” block that processes “x” into “Activation”, then to “F(x)”. This repeats through multiple “GIN” blocks, with a circular intersection element, and outputs to “y”.

Structure of the R2eGIN.

y=(x,{GIN}+P(x)) (5)

Reconstruction blocks

The implementation of the reconstruction block aimed to compel the model to learn rich and informative representations that could be used to reconstruct the original node features, thereby enhancing the model’s understanding of molecular structures. We incorporated a reconstruction block, consisting of 2 fully connected layers, to reconstruct the original node features from the graph representations. In addition, we included a prediction block, comprising a single fully connected layer followed by a sigmoid activation function, to predict the target. Thus, enabling the model to learn improved node feature representations and enhance prediction performance. The total loss was calculated as follows

Losstotal=Lossrecon+α×Losspred (6)

Where α is a parameter that controls the contribution of the prediction loss to the total loss. We set α to 0.5 for the prediction loss component. The reconstruction loss was calculated using Mean Squared Error (MSE) between the original features of each node in the graph, represented as X , and the reconstructed node features. The reconstruction was obtained by aggregating the reconstruction results from the entire batch during the training process

Lossrecon=1Ni=1NXiX^i2 (7)

With N as the number of nodes in the batch, Xi as the original features of the i -th node, and X^i as the reconstructed features of that node.

To predict PARPi, R2eGIN used a prediction block consisting of a single fully connected layer followed by a sigmoid activation function. Similar to the reconstruction layer, the graph-level molecular representation X was pooled using GMP based on the batch. The resulting representation was then processed by the prediction layer, which mapped this representation to a scalar value. This scalar value was subsequently transformed into a probability using a sigmoid activation function

y^=sigmoid(Wpred×1Ni=1NXi+bpred) (8)

where W represents the weights and b represents the bias. In this study, the binary classification task between PARPi actives and inactives was addressed using the BCEWithLogitsLoss function. This function combined the sigmoid function and the Binary Cross-Entropy (BCE) loss into a single operation, which is numerically more stable.

Molecular graph embedding

The transformation of molecular representations began by converting Simplified Molecular Input Line Entry Specification (SMILES) strings into molecular graph structures, where nodes represent atoms and edges represent chemical bonds. To obtain rich and informative molecular representations, we used crucial features, as employed by literatures29-32 to construct an accurate model. As presented in Table 1, we incorporated 7 atomic features and 2 edge features, including Bond type (BT), Bond direction (BD), Atom type (AT), Atom Formal charge (AF), Atom Chirality (AC), Atom Hybridization (AH), Number of hydrogen atoms bonded (HA), Implicit valence (AV), and Atom Degree (AD). Molecular structures inherently involve complex atomic interactions and electronic structures, and bond features encapsulate rich information regarding molecular frameworks and conformational isomers. Embedded molecular graphs can implicitly capture essential molecular information and inter-atomic interactions, while providing insights into the side characteristics of molecular bonds. The input edges and nodes for aggregation can be formalized by (hv0,he0) , described by these molecular attributes with

Table 1.

Details of atom and edge features.

Bond type Single, Double, Triple, Aromatic
Bond direction None, Endupright, Enddownright
Atom type 1 ~ 118
Formal charge −5 ~ +5
Chirality Unspecified, Tetrahedral_Cw, Tetrahedral_Ccw, Other
Hybridization S, Sp, Sp2, Sp3, Sp3d, Sp3d2, Unspecified
Number of hydrogen atoms bonded 0 ~ 8
Implicit valence 0 ~ 6
Atom degree number of connected bonds, 0 ~ 10
hv0={vAT,vAF,vAC,vAH,vHA,vAV,vAD}andhe0={eBT,eBD}

Experimental

Data sets

The data set employed for the PARPi study was initially procured from various publicly accessible sources, BindingDB, 33 PubChem,34,35 and ChEMBL. 36 In this study, we accessed the experimental data set published by Ai et al. 37 The data set comprises 4 PARP isoform data sets: PARP-1 (UniProt ID: P09874), PARP-2 (Q9UGN5), PARP-5A (O95271), and PARP-5B (Q9H2 K2). The detailed distribution of activity labels within each data set is as follows: PARP-1 (3119 active, 658 inactive), PARP-2 (271 active, 141 inactive), PARP-5A (702 active, 147 inactive), and PARP-5B (628 active, 104 inactive), as depicted in Figure 4A. Overall, the distribution of compound activity labels exhibited an imbalance, with a total of 4720 active and 1050 inactive compounds across all data sets. The average percentage of active and inactive labels for the entire data set is 20.8% and 79.2%, respectively (Figure 4B). In addition, chemical space analysis was conducted on the compounds within the PARP-1 (Figure 4C), PARP-2 (Figure 4D), PARP-5A (Figure 4E), and PARP-5B (Figure 4F) data sets, using molecular weight (MW) and AlogP as defining parameters. The analysis revealed a broad molecular weight range, spanning from 121.139 to 725.683, and the AlogP range of −1.946 to 8.700. These findings indicated that the compounds present in the modeling data sets encompassed a diverse chemical space.

Figure 4.

“Data sets analysis, compound number, molecular weight.”

Compound distribution and chemical space analysis of PARP datasets.

Evaluation metrics

We evaluated the performance of R2eGIN on PARPi prediction using accuracy (ACC.), balanced accuracy (BA) Matthews correlation coefficient (MCC), F1-score (F1), precision (Prec.), and area under the ROC curve (AUC). Given false positive (FP), true positive (TP), false negative (FN), and true negative (TN), their formulas are as follows:

ACC=TP+TN(TP+TN+FP+FN) (9)
BA=12(TPTP+FN+TNTN+FP) (10)
MCC=TP×TNFP×FN(TP+FP)(TP+FN)(TN+FP)(TN+FN) (11)
F1=2TP2TP+FP+FN (12)
PRE=TPTP+FP (13)

Experimental settings

Experiments were conducted using PyTorch on Ubuntu 18.0.4 64bit with NVIDIA GTX 1080Ti × 4 and Intel Core i9-9900K (3.60 GHz). We trained R2eGIN in 200 epochs using the Adam optimizer with a learning rate of 0.001. A specific model with a 3-layer GIN acted as the R2eGIN backbone. The dropout rate range was (0.2, 0.5), and batch sizes were (64, 128, 512) as detailed in Table 2.

Table 2.

Hyperparameter settings.

Parameters Range
Batch size 64, 128, 512
Dropout rate 0.2, 0.5
Number of Layers 3
Optimizer Adam
Learning rate 0.001, 0.0001

Results

R2eGIN for PARPi prediction

R2eGIN achieved the highest accuracy (0.926), MCC (0.728), F1 score (0.956), and BA (0.836) on the PARP-1 data set as presented in Table 3. These results highlighted the model’s ability to effectively capture both local and global structural information within molecular graphs. In addition, the use of residual connections and reconstruction block integrated background information from the initial embeddings into the propagation layers, enabling the model to learn effectively without losing critical information necessary for PARPi classification.

Table 3.

Performance of R2eGIN.

Data sets ACC F1 BA MCC AUC PRE
PARP-1 0.926 0.956 0.836 0.728 0.842 0.938
PARP-2 0.781 0.842 0.730 0.493 0.870 0.800
PARP-5A 0.882 0.932 0.693 0.535 0.844 0.885
PARP-5B 0.904 0.946 0.734 0.547 0.767 0.924

Comparison with existing models

To evaluate the effectiveness of R2eGIN, we compared diverse graph learning methods, including Chemprop, 38 MPNN, 26 FPGNN, 37 AttentiveFP, 39 GCN, 40 GAT, 25 and GIN. 24 Overall, the proposed model demonstrated superior performance, exhibiting average scores of 0.873 for ACC, 0.576 for MCC, 0.919 for F1, and 0.748 for BA, as shown in Figure 5A. Concurrently, Chemprop secured the second-highest position, achieving an average ACC of 0.848 with a standard deviation (STD) of 0.041 and an average F1-score of 0.904 with a STD of 0.039. FPGNN registered the second-best average BA of 0.710, while GIN ranked as the second highest in average MCC, with a score of 0.466 (Table 4).

Figure 5.

The performance of the R2eGIN model is comparable with other top methods such as Chemprop, MPNN, and GAT while it surpasses AttentiveFP or MGAT in performance.

Comparative analysis of performance between the R2eGIN and the other methods.

Table 4.

Comparison performance on average scores between R2eGIN and the other models on 4 data sets PARPi.

Models ACC MCC F1 BA
Chemprop 0.848 ± 0.041 0.462 ± 0.124 0.904 ± 0.039 0.664 ± 0.100
MPNN 0.826 ± 0.072 0.404 ± 0.122 0.893 ± 0.054 0.630 ± 0.092
FPGNN 0.845 ± 0.049 0.464 ± 0.106 0.899 ± 0.046 0.710 ± 0.057
AttentiveFP 0.819 ± 0.067 0.421 ± 0.104 0.883 ± 0.056 0.649 ± 0.080
GCN 0.822 ± 0.083 0.398 ± 0.133 0.881 ± 0.076 0.667 ± 0.067
GAT 0.801 ± 0.078 0.399 ± 0.112 0.861 ± 0.075 0.673 ± 0.057
GIN 0.839 ± 0.049 0.466 ± 0.126 0.900 ± 0.042 0.677 ± 0.056
R2eGIN 0.873 ± 0.056 0.576 ± 0.090 0.919 ± 0.045 0.748 ± 0.053

The best results are highlighted in bold, and the runner-up results are highlighted in underline. Higher values indicate better performance.

The comparative evaluation of R2eGIN and eight baseline models, performed on the test sets of PARP-1, PARP-2, PARP-5A, and PARP-5B, is visualized in Figure 5B. The FPGNN model combined the strengths of traditional molecular fingerprint representations with GNN. While this model performed well on specific tasks, 41 the use of 2 distinct representations could lead to feature redundancy. Certain atomic properties already represented by molecular descriptors might be redundantly incorporated into the graph’s feature representation, potentially diminishing the overall efficiency and clarity of the representation. GAT and AttentiveFP implemented attention mechanisms, enabling the models to dynamically assign importance weights to neighboring nodes within the graph representation. Although this approach effectively enhanced the model’s ability to focus on relevant graph components, it is crucial to acknowledge that no single model is universally optimal across all tasks. Meanwhile, Chemprop model’s superior performance on the PARP-2 data set could be attributed to the data set’s unique characteristics, which might exhibit a molecular structure distribution and activity patterns particularly well-suited to Chemprop capabilities, as detailed in Table 5. Conversely, for the PARP-5A data set, FPGNN (BA 0.681) and GIN (ACC 0.871; MCC 0.595; F1 0.927) outperformed Chemprop. Similarly, on the PARP-5B data set, AttentiveFP (MCC 0.420) and GCN (ACC 0.879; F1 0.933; BA 0.644) ranked second only to the proposed R2eGIN model. Overall, a comparative analysis of the eight models for predicting PARP inhibitor activity revealed intriguing patterns across different PARP isoform data sets. The findings demonstrated that the R2eGIN model consistently surpassed other models in terms of ACC, MCC, F1, and BA on the PARP-1, PARP-5A, and PARP-5B data sets, highlighting the robust performance of the proposed model in PARPi prediction tasks.

Table 5.

Overall performance comparisons of different PARPi prediction models.

Models Metrics PARP-1 PARP-2 PARP-5A PARP-5B
Chemprop ACC 0.899 0.788 0.839 0.868
MCC 0.618 0.526 0.421 0.283
F1 0.940 0.839 0.909 0.927
BA 0.775 0.750 0.576 0.553
MPNN ACC 0.895 0.707 0.834 0.867
MCC 0.612 0.325 0.370 0.308
F1 0.938 0.801 0.907 0.928
BA 0.783 0.616 0.561 0.557
FPGNN ACC 0.897 0.767 0.846 0.872
MCC 0.618 0.498 0.409 0.332
F1 0.939 0.822 0.909 0.928
BA 0.785 0.738 0.681 0.634
AttentiveFP ACC 0.890 0.724 0.789 0.874
MCC 0.589 0.364 0.312 0.420
F1 0.935 0.799 0.868 0.932
BA 0.769 0.670 0.600 0.557
GCN ACC 0.897 0.685 0.826 0.879
MCC 0.617 0.307 0.279 0.390
F1 0.939 0.753 0.900 0.933
BA 0.779 0.642 0.601 0.644
GAT ACC 0.889 0.685 0.775 0.852
MCC 0.588 0.352 0.293 0.364
F1 0.934 0.743 0.853 0.915
BA 0.769 0.660 0.623 0.641
GIN ACC 0.881 0.756 0.871 0.849
MCC 0.561 0.429 0.595 0.276
F1 0.929 0.828 0.927 0.915
BA 0.761 0.694 0.633 0.618
R2eGIN ACC 0.926 0.781 0.882 0.904
MCC 0.728 0.493 0.535 0.547
F1 0.956 0.842 0.932 0.946
BA 0.836 0.730 0.693 0.734

The best results are highlighted in bold, and the runner-up results are highlighted in underline. Higher values indicate better performance.

Ablation study

We performed an ablation study to assess the performance improvement gained by incorporating 2 components, residual connection and the reconstruction block, in the proposed method. Four experimental scenarios were designed: M1, M2, M3, and M4, as presented in Table 6. In scenario M1, the GIN model was used without residual learning or the reconstruction block. Scenario M2 features the GIN backbone with residual connections but without the reconstruction block. In scenario M3, the GIN backbone is enhanced with the reconstruction block but without residual connections. Finally, scenario M4 corresponded to the proposed R2eGIN model. The ablation experiment demonstrated that R2eGIN outperformed the other scenarios across evaluation metrics as shown in Figure 6.

Table 6.

Scenario and abbreviations used in ablation study.

Models GIN Res Rec
M1
M2
M3
M4
Figure 6.

Chart shows ablation experiment results: VariantParp1: M1: 0.881, M2: 0.918, M3: 0.902, M4: 0.926; VariantParp2: M1: 0.849, M2: 0.847, M3: 0.871, M4: 0.882; VariantParp 5A: M1: 0.756, M2: 0.789, M3: 0.781, M4: 0.781; Variant Parp 5B: M1: 0.764

Performances of different model variants in ablation experiment.

Conclusion

In this study, we present the R2eGIN model, which is GIN by incorporating residual connection and reconstruction network. The experimental results on 4 PARPi data sets (PARP-1, PARP-2, PARP-5A, and PARP-5B) demonstrate that the R2eGIN model derives superior performance over state-of-the-art models on the task of PARPi prediction. This study is the first to demonstrate that a GIN-based model enhanced with residual and reconstruction mechanisms is well-suited for predicting PARP inhibitors. In addition, it reveals key strategies for screening potential molecules and accelerating virtual screening methods in drug discovery, particularly for cancer therapy development. Although R2eGIN demonstrates good performance and potential in predicting PARP inhibition activity, niche for improvement for future research is available. The performance of R2eGIN is constrained by the quality and quantity of available data, particularly in the PARP-2 data set, consisting of only 412 samples, with 271 actives and 141 inactives. Overall, across the 4 data sets, inactive molecular data accounts for an average of only 20.8%, while active molecular data constitutes 79.2%. Therefore, to enhance the reliability and robustness of PARPi predictions, we will continue refining our methodology by advancing molecular representation through next-generation deep learning models, integrating GNN with complementary computational approaches to optimize performance, 42 and meticulous collection and curation of high-quality data. In addition, we prioritize improving model interpretability, as it is essential for gaining deeper insights into model behavior.

Footnotes

ORCID iDs: Candra Zonyfar Inline graphic https://orcid.org/0000-0003-0697-882X

Soualihou Ngnamsie Njimbouom Inline graphic https://orcid.org/0000-0003-3256-0021

Jeong-Dong Kim Inline graphic https://orcid.org/0000-0002-5113-221X

Author Contributions: Conceptualization: Candra Zonyfar, and Jeong-Dong Kim; Methodology: Candra Zonyfar, and Jeong-Dong Kim; Software: Candra Zonyfar, and Soualihou Ngnamsie Njimbouom; Validation: Soualihou Ngnamsie Njimbouom, Sophia Mosalla, and Jeong-Dong Kim; Formal analysis: Candra Zonyfar, Soualihou Ngnamsie Njimbouom, and Jeong-Dong Kim; Investigation: Candra Zonyfar, Sophia Mosalla, and Jeong-Dong Kim; Resources: Candra Zonyfar, Soualihou Ngnamsie Njimbouom, Sophia Mosalla, and Jeong-Dong Kim; Data curation: Candra Zonyfar, Soualihou Ngnamsie Njimbouom, and Sophia Mosalla; Writing – original draft: Candra Zonyfar; Writing – review & editing: Candra Zonyfar, Soualihou Ngnamsie Njimbouom, and Jeong-Dong Kim; Visualization: Candra Zonyfar, Soualihou Ngnamsie Njimbouom, and Jeong-Dong Kim; Supervision: Jeong-Dong Kim; Project administration: Jeong-Dong Kim; Funding acquisition: Jeong-Dong Kim. All authors have read and agreed to the published version of the manuscript.

Funding: The authors disclosed receipt of the following financial support for the research, authorship, and/or publication of this article: This research was supported by the MSIT (Ministry of Science ICT), Korea, under the National Program for Excellence in SW, supervised by the IITP (Institute of Information & Communications Technology Planning & Evaluation) in 2025 (No. 2024-0-00023).

The author(s) declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article.

References

  • 1. Bai YR, Yang WG, Jia R, et al. The recent advance and prospect of poly(ADP-ribose) polymerase inhibitors for the treatment of cancer. Med Res Rev. 2025;45:214-273. doi: 10.1002/med.22069 [DOI] [PubMed] [Google Scholar]
  • 2. Pommier Y, Huang SH, Das BB, et al. 284 differential trapping of PARP1 and PARP2 by clinical PARP inhibitors. Eur J Cancer. 2012;48:87. doi: 10.1016/S0959-8049(12)72082-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3. Slade D. PARP and PARG inhibitors in cancer treatment. Genes Dev. 2020;34:360-394. doi: 10.1101/gad.334516.119 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4. Wu S, Li X, Gao F, De Groot JF, Koul D, Yung WKA. PARP-mediated PARylation of MGMT is critical to promote repair of temozolomide-induced O6-methylguanine DNA damage in glioblastoma. Neuro-Oncology. 2021;23:920-931. doi: 10.1093/neuonc/noab003 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5. Padella A, Ghelli Luserna Di Rorà A, Marconi G, Ghetti M, Martinelli G, Simonetti G. Targeting PARP proteins in acute leukemia: DNA damage response inhibition and therapeutic strategies. J Hematol Oncol. 2022;15:10. doi: 10.1186/s13045-022-01228-0 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6. Slade D. Mitotic functions of poly(ADP-ribose) polymerases. Biochem Pharmacol. 2019;167:33-43. doi: 10.1016/j.bcp.2019.03.028 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7. Al-Sanea MM, Chilingaryan G, Abelyan N, et al. Combination of ligand and structure based virtual screening approaches for the discovery of potential PARP1 inhibitors. PLOS ONE. 2022;17:e0272065. doi: 10.1371/journal.pone.0272065 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8. Ramadan SK, Elrazaz EZ, Abouzid KAM, El-Naggar AM. Design, synthesis and in silico studies of new quinazolinone derivatives as antitumor PARP-1 inhibitors. RSC Adv. 2020;10:29475-29492. doi: 10.1039/D0RA05943A [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9. Purnell MR, Whish WJ. Novel inhibitors of poly(ADP-ribose) synthetase. Biochem J. 1980;185:775-777. doi: 10.1042/bj1850775 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10. Shen Y, Rehman FL, Feng Y, et al. BMN 673, a novel and highly potent PARP1/2 Inhibitor for the treatment of human cancers with DNA repair deficiency. Clin Cancer Res. 2013;19:5003-5015. doi: 10.1158/1078-0432.CCR-13-1391. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11. Mukai T, Fujita S, Morita Y. Tankyrase (PARP5) inhibition induces bone loss through accumulation of its substrate SH3BP2. Cells. 2019;8:195. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12. Liu S, Luo W, Wang Y. Emerging role of PARP-1 and PARthanatos in ischemic stroke. J Neurochem. 2022;160:74-87. doi: 10.1111/jnc.15464 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13. Horvath EM, Szabó C. Poly(ADP-ribose) polymerase as a drug target for cardiovascular disease and cancer: an update. Drug News Perspect. 2007;20:171-181. doi: 10.1358/dnp.2007.20.3.1092098 [DOI] [PubMed] [Google Scholar]
  • 14. Pöstyéni E, Gábriel R, Kovács-Valasek A. Poly (ADP-Ribose) polymerase-1 (PARP-1) inhibitors in diabetic retinopathy: an attractive but elusive choice for drug development. Pharmaceutics. 2024;16:1320. doi: 10.3390/pharmaceutics16101320 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15. Ge Y, Tian T, Huang S, et al. An integrative drug repositioning framework discovered a potential therapeutic agent targeting COVID-19. Sig Transduct Target Ther. 2021;6:165. doi: 10.1038/s41392-021-00568-6 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16. Curtin N, Bányai K, Thaventhiran J, Le Quesne J, Helyes Z, Bai P. Repositioning PARP inhibitors for SARS-CoV-2 infection(COVID-19); a new multi-pronged therapy for acute respiratory distress syndrome? Br J Pharmacol. 2020;177:3635-3645. doi: 10.1111/bph.15137. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17. Antolín AA, Mestres J. Linking off-target kinase pharmacology to the differential cellular effects observed among PARP inhibitors. Oncotarget. 2014;5:3023-3028. doi: 10.18632/oncotarget.1814 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18. LaFargue CJ, Dal Molin GZ, Sood AK, Coleman RL. Exploring and comparing adverse events between PARP inhibitors. Lancet Oncol. 2019;20:e15-e28. doi: 10.1016/S1470-2045(18)30786-1 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19. Mateo J, Lord CJ, Serra V, et al. A decade of clinical development of PARP inhibitors in perspective. Ann Oncol. 2019;30:1437-1447. doi: 10.1093/annonc/mdz192 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20. Zhang Y, Hu Y, Han N, Yang A, Liu X, Cai H. A survey of drug-target interaction and affinity prediction methods via graph neural networks. Comput Biol Med. 2023;163:107136. doi: 10.1016/j.compbiomed.2023.107136 [DOI] [PubMed] [Google Scholar]
  • 21. Zhu W, Zhang Y, Zhao D, Xu J, Wang L. HiGNN: a hierarchical informative graph neural network for molecular property prediction equipped with feature-wise attention. J Chem Inf Model. 2023;63:43-55. doi: 10.1021/acs.jcim.2c01099. [DOI] [PubMed] [Google Scholar]
  • 22. Torres LHM, Ribeiro B, Arrais JP. Few-shot learning with transformers via graph embeddings for molecular property prediction. Expert Syst Appl. 2023;225:120005. doi: 10.1016/j.eswa.2023.120005. [DOI] [Google Scholar]
  • 23. Zheng K, Zhao H, Zhao Q, Wang B, Gao X, Wang J. NASMDR: a framework for miRNA-drug resistance prediction using efficient neural architecture search and graph isomorphism networks. Brief Bioinform. 2022;23:bbac338. doi: 10.1093/bib/bbac338. [DOI] [PubMed] [Google Scholar]
  • 24. Xu K, Hu W, Leskovec J, Jegelka S. HOW powerful are graph neural networks? 2019. Accessed July 30, 2025. https://arxiv.org/abs/1810.00826
  • 25. Veličković P, Cucurull G, Casanova A, Romero A, Liò P, Bengio Y. Graph attention networks, 2018. doi: 10.48550/arXiv.1710.10903 [DOI] [Google Scholar]
  • 26. Gilmer J, Schoenholz SS, Riley PF, Vinyals O, Dahl GE. Neural message passing for quantum chemistry, 2017. Accessed July 30, 2025. https://arxiv.org/abs/1704.01212
  • 27. Wang J, Wu HY, Chen JC, Shuai HH, Cheng WH. Residual graph attention network and expression-respect data augmentation aided visual grounding. In: 2022 IEEE international conference on image processing (ICIP), Bordeaux, France, October 16-19, 2022. doi: 10.1109/ICIP46576.2022.9897564 [DOI] [Google Scholar]
  • 28. Liu G, Liu Q, Cao W. Hybrid attention deep adaptive residual graph convolution network for few-shot classification. In: 2023 International Conference on Pattern Recognition, Machine Vision and Intelligent Algorithms (PRMVIA), Beihai, China, March 24-26, 2023. doi: 10.1109/PRMVIA58252.2023.00007 [DOI] [Google Scholar]
  • 29. Lv Q, Chen G, Yang Z, Zhong W, Chen CY. meta learning with graph attention networks for low-data drug discovery. IEEE Trans Neural Netw Learn Syst. 2024;35:11218-11230. doi: 10.1109/TNNLS.2023.3250324 [DOI] [PubMed] [Google Scholar]
  • 30. Guo Z, Zhang C, Yu W, et al. Few-shot graph learning for molecular property prediction. In: Proceedings of the web conference 2021, Ljubljana, Slovenia, April 19-23, 2021. doi: 10.1145/3442381.3450112 [DOI] [Google Scholar]
  • 31. Kearnes S, McCloskey K, Berndl M, Pande V, Riley P. Molecular graph convolutions: moving beyond fingerprints. J Comput Aided Mol Des. 2016;30:595-608. doi: 10.1007/s10822-016-9938-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32. Pavlidis N, Nikolaidis CC, Perifanis V, Papadopoulou A, Efraimidis P, Arampatzis A. An extensive overview of feature representation techniques for molecule classification. In: Proceedings of the 27th Pan-Hellenic conference on progress in computing and informatics. ACM, Lamia, Greece, November 24-26, 2023. doi: 10.1145/3635059.3635083 [DOI] [Google Scholar]
  • 33. Liu T, Lin Y, Wen X, Jorissen RN, Gilson MK. BindingDB: a web-accessible database of experimentally determined protein-ligand binding affinities. Nucleic Acid Res. 2007;35:d198-d201. doi: 10.1093/nar/gkl999 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34. Bolton EE, Wang Y, Thiessen PA, Bryant SH. PubChem: integrated platform of small molecules and biological activities. Ann Rep Comput Chem 2008;4:217-241. doi: 10.1016/S1574-1400(08)00012-1 [DOI] [Google Scholar]
  • 35. Wang Y, Bryant SH, Cheng T, et al. PubChem BioAssay: 2017 update. Nucleic Acids Res. 2017;45:D955-D963. doi: 10.1093/nar/gkw1118 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36. Gaulton A, Bellis LJ, Bento AP, et al. ChEMBL: a large-scale bioactivity database for drug discovery. Nucleic Acids Res. 2012;40:D1100-D1107. doi: 10.1093/nar/gkr777 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37. Ai D, Wu J, Cai H, et al. A multi-task FP-GNN framework enables accurate prediction of selective PARP inhibitors. Front Pharmacol. 2022;13:971369. doi: 10.3389/fphar.2022.971369 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 38. Heid E, Greenman KP, Chung Y, et al. Chemprop: a machine learning package for chemical property prediction. J Chem Inf Model. 2024;64:9-17. doi: 10.1021/acs.jcim.3c01250 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39. Xiong Z, Wang D, Liu X, et al. Pushing the boundaries of molecular representation for drug discovery with the graph attention mechanism. J Med Chem. 2020;63:8749-8760. doi: 10.1021/acs.jmedchem.9b00959 [DOI] [PubMed] [Google Scholar]
  • 40. Kipf TN, Welling M. Semi-supervised classification with graph convolutional networks, 2017. doi: 10.48550/arXiv.1609.02907 [DOI] [Google Scholar]
  • 41. Ai D, Cai H, Wei J, Zhao D, Chen Y, Wang L. DEEPCYPs: a deep learning platform for enhanced cytochrome P450 activity prediction. Front Pharmacol. 2023;14:1099093. doi: 10.3389/fphar.2023.1099093. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42. Yang Y, Li G, Li D, Zhang J, Hu P, Hu L. Integrating fuzzy clustering and graph convolution network to accurately identify clusters from attributed graph. IEEE Trans Netw Sci Eng. 2025;12:1112-1125. doi: 10.1109/TNSE.2024.3524077 [DOI] [Google Scholar]

Articles from Bioinformatics and Biology Insights are provided here courtesy of SAGE Publications

RESOURCES