Abstract
Enzyme kinetic parameters are fundamental for quantifying catalytic efficiency and substrate specificity in biochemistry and drug discovery. However, experimental determination is resource intensive, and accurate prediction remains a persistent challenge due to the complex spatial nature of catalysis. In this paper, we present UniKineG, a novel deep learning framework that redefines kinetic prediction by modeling the explicit spatial state of enzyme–substrate complexes. Unlike conventional methods that treat proteins and ligands as isolated modalities, UniKineG employs molecular docking to embed both entities into a unified 3D coordinate system. Within this shared geometric context, we utilize a heterogeneous graph neural network integrated with geometric vector perceptrons (GVPs) to capture intricate vector-based interactions, such as directional hydrogen bonds, hydrophobic contacts, and electrostatic complementarity. This structure-based approach confers exceptional robustness: UniKineG effectively overcomes the dependency on high-sequence homology, demonstrating superior generalization on out-of-distribution (OOD) datasets encompassing both unseen enzyme sequences and diverse substrate scaffolds. Consistently outperforming state-of-the-art predictors, UniKineG achieves high-precision predictions. This work establishes a solid foundation for understanding enzyme–small molecule interactions in 3D space and offers a transformative tool for computational enzymology.
Keywords: enzyme kinetic parameters, deep learning, graph convolutional networks, heterogeneous graph, molecular docking
1. Introduction
Enzymes serve as the catalytic engines of life, orchestrating biological processes ranging from metabolic flux regulation to signal transduction and drug metabolism [1]. The directionality and rate of these biochemical reactions are governed by fundamental kinetic parameters: the turnover number (), the Michaelis constant (), and the catalytic efficiency () [2]. The correct identification of such parameters plays important roles in elucidating reaction mechanisms as well as in guiding rational enzyme design and rebuilding genome-scale metabolic models [3]. Thus, the Michaelis–Menten framework that is created can form a mathematical basis to measure substrate affinity and catalytic capacity, both having an impact on physiological homeostasis [4]. However, experimental determination using in vitro assays is labor intensive, time consuming, and often constrained by substrate availability and assay sensitivity. Therefore, databases as comprehensive as BRENDA and SABIO-RK only cover a portion of the known enzymatic space [5,6]. Such terrible data poverty amidst the rapidly expanding numbers of sequenced genomes is a problem for computational enzymology [7,8,9].
Artificial intelligence (AI) has emerged as a powerful tool to bridge this gap. Early machine-learning attempts combined structural active site information with flux balance analysis to predict in specific organisms but had low transferability [10]. Next came deep learning. Li et al. were the ones who started this movement with DLKcat, which uses deep convolutional neural networks (CNNs) and graph neural networks (GNNs) to pull features from protein sequences and substrate graphs, respectively, in the first large-scale prediction [11]. TurNup and UniKP also used integrated pretrained protein language models (pLMs), such as ESM-1b, to encode the rich evolutionary semantics of amino acid sequences and conduct multitask prediction [12,13]. In more recent years, efforts have increasingly focused on improving dataset curation, reliability, and out-of-distribution (OOD) evaluation. CataPro and CatPred emphasize more rigorous data standardization and benchmarking [14,15]; notably, CatPred provides prediction-specific uncertainty quantification and explicitly evaluates OOD enzymes using rich sequence and structure representations. In parallel, DEKP [16] integrates pretrained sequence encoders with structure-aware graph modules to mitigate performance degradation under low-sequence similarity and reports sensitivity to mutation-induced changes in catalytic efficiency. Beyond enzyme kinetics, graph neural networks have also been applied to other biochemical prediction settings, such as classifying whether small molecules are depleted by the human gut microbiome [17]. Related studies in drug metabolism/toxicology (e.g., P-gp inhibition prediction) further suggest that model performance can depend strongly on the representation choice and data regime [18].
Although there are architectural innovations, there is still an important limit: Most of the existing frameworks consider enzymes and substrates to be topologically isolated modalities. Even when pretrained representations, structure-aware encoders, or uncertainty modeling are introduced, the enzyme and substrate are often still encoded separately and fused late, leaving the interfacial 3D interaction geometry underutilized.
To represent 1D sequences and 2D molecular graphs by themselves and to concatenate them at the late stage, which cannot properly model the spatial state of the catalytic microenvironment [19], it is very hard to obtain all those hard and important geometry-dependent interactions that help things, like the precise alignment of catalytic residues and substrate functional groups, that govern the reaction turnover number () [20,21].
This shortcoming is the most severe when predicting . Biochemical mechanisms suggest that while is mostly governed by the binding affinity, a property mostly driven by the static geometric fit and surface interactions, like electrostatics and hydrophobicity, reflects the maximal reaction rate, which is governed by the stabilization of the transition state, electron transfer, and dynamic conformational fluctuations [22,23]. These processes are strictly regulated by vector-based geometric constraints, such as the directional alignment of hydrogen bonds and the precise orientation of dipoles. These small vector-level features are usually omitted from the scalar representation or sequence-based embedding used by the previous methods [24]. For this, this sort of a predictive framework needs to have knowledge that it is geometry aware and that it can use different multimodal information within a physically rigorous spatial context.
And now for our UniKineG, which is another deep learning framework, but this is to be able to predict enzymes’ kinetic parameters on the basis of the current atom-level interaction state. Unlike conventional methods that handle proteins and ligands separately, UniKineG integrates both into a unified 3D Euclidean coordinate system via molecular docking. This representation is then constructed into a multimodal heterogeneous graph to preserve the relative spatial orientation of the complex [25]. Crucially, we add a geometric vector perceptron (GVP) module [24] that adds rotation-equivariant processing to be consistent with the complex, no matter the orientation, and captures the directional geometric patterns at the interface. Based on the explicit 3D geometry of the enzyme–substrate complex, UniKineG can greatly reduce the heavy dependency on sequence homology. From our comprehensive benchmarking, UniKineG always outperformed all the current best of the most advanced method, with excellent generalization to out-of-distribution (OOD) datasets, which is an excellent basis for designing an enzyme to catalyze novel substrate specificity.
2. Results
2.1. Overview of the UniKineG Deep Learning Framework
UniKineG creates a geometry-aware prediction pipeline, which is created to be able to predict enzymes’ kinetics through geometry-aware message passing. The pipeline starts by aligning isolated enzyme and substrate structures into a unified 3D Euclidean coordinate system via molecular docking, providing a possible 3D pose for interaction modeling, even if no co-crystal structure is available (Figure 1). Within this unified geometric context, the architecture comprises two tightly coupled subsystems (Figure 2):
Figure 1.
Docked enzyme–substrate complex used to construct 3D interaction graphs. The overall structure of enzyme Q9Y223 with UDP-N-acetyl-α-D-glucosamine docked in the catalytic cleft (protein, ribbon; ligand, sticks), highlighting key contacts between the ligand and surrounding residues.
Figure 2.
Architecture of UniKineG for predicting enzyme kinetic parameters from docked complexes. (a) Protein branch: residue embeddings from ESM-2 are combined with residue-level geometric features derived from the protein structure and encoded by stacked GVPConv layers. (b) Ligand branch: atomic features from the molecular graph and 3D coordinates from the docked pose are encoded by GVPConv layers. (c) Cross-modal fusion: a protein–ligand interaction graph is constructed from spatial contacts and processed by GVP-CrossConv for bidirectional message passing, followed by pooling and an MLP regressor. (d) GVPConv updates scalar and vector channels while preserving SE(3) equivariance.
Multimodal representation and intramolecular modeling (Figure 2a,b,d). UniKineG builds a dual branch input enriched with semantics and physics priors. The enzyme branch is based on ESM-2 evolutionary embeddings, which are concatenated with 30-dimensional physicochemical features (e.g., charge and hydrophobicity), and the substrate branch encodes SMILES using NovoMolGen, which encodes SMILES-derived semantic features as well as 45-dimensional atomic features. The local geometric context is encoded using GVPConv layers in both branches. This module uses coordinate-derived directional vectors to update representations in a direction that can capture directional geometric patterns at the interface. Intermolecular geometric interaction and readout (Figure 2c). For modeling cross-molecular interactions in the docked complex, a heterogeneous graph is built to enable bidirectional geometric message passing with GVPCrossConv. By incorporating relative distances and directional vectors, the model tightly couples the active pocket environment with the substrate’s physicochemical characteristics. Finally, we obtain global scalar and vector representations using weighted pooling and fuse them via an MLP to regress the target kinetic parameters .
2.2. Predictive Performance of UniKineG for Enzyme Kinetic Parameters
To verify that UniKineG generalizes beyond mere sequence and scaffold similarities and avoids learning spurious correlations, we adopted the partitioning protocol from CatPred, with minor modifications. Specifically, we established a benchmark setup with a rigorous OOD constraint based on sequence gradients. We constructed two additional OOD scenarios to mimic real-world discovery challenges beyond standard held-out tests. In enzyme generalization, multilevel sequence clustering by CD-HIT was carried out at a gradient of thresholds, and the stringent 40% identity threshold (CD-HIT-40) was chosen as the main reference. At the same time, in order to perform substrate generalization, we added a constraint according to the ECFP4 fingerprint similarity, making the strict Tanimoto coefficient cutoff threshold as the main metric to block information leakage from chemical neighbors (Figure 3).
Figure 3.
Performances of UniKineG and ablated variants under in-distribution and OOD splits. Coefficient of determination (R2, computed based on log-transformed targets) for predicting kcat (a), Km (b), and kcat/Km (c) in held-out (in-distribution), enzyme-unseen (maximum sequence identity ≤ 40%; CD-HIT-40), and substrate-unseen (maximum ECFP4 Tanimoto similarity ≤ 0.2) test sets.
Compared against this full evaluation framework, the full UniKineG model achieved strong results for all three prediction tasks on both the held-out split as well as the strict OOD splits (Figure 4d–f). When the held-out testing was performed for UniKineG, the values were all above 0.65, demonstrating exceptional precision. Crucially, the performance was still decent under the harshest enzyme-unseen split (CD-HIT-40), with respectable values of 0.538 for and 0.567 for , outperforming standard sequence-based baselines. Similarly, in the strict substrate-unseen setting, stable performance was also observed, indicating that the model is also generalizable to novel unexplored biochemical spaces (Supplementary Table S1).
Figure 4.
Predictive accuracy and ablation analysis of UniKineG. (a–c) Density scatter plots comparing predicted versus experimental values in the test set for (a) , (b) , and (c) . The color gradient represents the data point density, and Pearson correlation coefficients (PCCs) are annotated in the top left corner. (d–f) Ablation study results illustrating the contributions of geometric components across varying enzyme sequence identity thresholds (CD-HIT) and substrate similarity constraints (Tanimoto). The line charts compare the performance of the full UniKineG model against those of four ablated variants: scalar only (without vector features), distance only (without directional vectors), no cross-graph (without intermolecular message passing), and no geometry (sequence-only baseline) for (d) , (e), and (f) .
UniKineG consistently secured the highest value and Pearson correlation coefficient (PCC) in challenging OOD settings (Figure 4). For instance, on the CD-HIT-40 split, UniKineG outperformed the top-performing baseline by substantial margins, achieving relative improvements of about 7–10% in correlation metrics for both and . A similar trend was observed on the substrate-unseen split, where UniKineG demonstrated superior robustness against structural divergence compared to those of models relying on 1D descriptors.
Finally, ablation studies on the CD-HIT-40 split revealed a distinct hierarchy of geometric contributions (Figure 3). First, structural guidance proved to be foundational: Removing the geometry completely (“no geometry”) caused a performance collapse, with dropping by over 50%, confirming that sequence semantics alone cannot reconstruct the catalytic microenvironment. Second, cross-molecular message passing is essential: The “no cross-graph” variant, which lacks explicit enzyme–substrate interaction modeling, suffered significant degradation, demonstrating that independent molecular encoding (“late fusion”) is insufficient for capturing interfacial dynamics. Most critically, a profound gap persisted between the “distance-only” variant and the full model. The inclusion of vector features yielded a substantial performance boost (raising from 0.398 to 0.538). This gain of approximately 35% underscores the pivotal role of “vector perception”: Unlike scalar distances, which merely define the atomic proximity, the directional vectors within UniKineG capture spatial orientation and angular constraints. This geometric completeness empowers the model to recognize precise orientation-dependent interaction patterns, driving the breakthrough performance in OOD scenarios.
2.3. UniKineG Outperforms Existing Deep Learning Frameworks in Benchmark Assessments
To position UniKineG within the current landscape of computational enzymology, we conducted a systematic benchmarking analysis against four representative models, retraining and evaluating all the frameworks under unified data-splitting protocols to ensure rigorous comparability. The baselines represent distinct evolutionary stages of the field: DLKcat, the pioneer in cross-species deep learning; UniKP, a multitask-learning framework; and the recent CataPro and CatPred, which integrate protein language models with graph neural networks.
UniKineG established a clear performance lead across all three tasks (, , and ), despite the generally high fidelity of all the baselines (Figure 5a–c). For prediction, UniKineG achieved high precision (), significantly outperforming the strongest baseline, CataPro (), while earlier models (like DLKcat and UniKP) plateaued at 0.525 and 0.603, respectively. Furthermore, UniKineG reduced the RMSE to 0.972, significantly surpassing the baseline average () (Supplementary Table S2). These results suggest that introducing a docking-driven coordinate system can improve accuracy without degrading the distribution performance.
Figure 5.
Model performances across different tasks and data splits. (a,d) prediction. Performance evaluations on enzyme sequence identity splits (CD-HIT) and substrate structural similarity splits (Tanimoto). (b,e) prediction; (c,f) prediction.
The most distinguishing advantages emerged in rigorous OOD settings. Whether traversing the CD-HIT sequence identity gradient or the Tanimoto chemical similarity gradient , the trend of the fading heatmap intensity reveals a consistent fact: As the distance from the training distribution increases, baseline models suffer from a pronounced stepwise degradation (Figure 6). Specifically, under the stringent CD-HIT-40 condition, DLKcat experienced a sharp degradation, with its value plummeting to 0.151, and even structure-informed models (like CatPred) struggled to maintain fidelity (). In contrast, UniKineG remained robust, maintaining an value of 0.538, significantly outperforming the runner-up CataPro () (Figure 5a). This robustness extends to chemical space: in the setting for novel substrates, UniKineG maintained scores of 0.495 and 0.563 for and , respectively, whereas CataPro dropped to 0.283 and 0.431, and DLKcat fell as low as 0.149 and 0.272 (Figure 5d–f). This consistent value superiority across dual OOD scenarios is visualized by the high-intensity bottom row in (Figure 6). It confirms that UniKineG’s generalization stems not from memorizing outliers but from learning geometry-aware interaction patterns that extend beyond sequence and scaffold similarities.
Figure 6.
Summary of generalization boundaries across different partitions and tasks. The PCC heatmaps demonstrate the evaluation results of five models in retention sets and OOD partitions, covering three modes: retention dataset (CD-HIT consistency from 99% to 40%), enzyme unseen (ECFP4 Tanimoto from 0.4 to 0.2), and substrate unseen. Rows represent partition/task combinations, and columns indicate models.
This systematic superiority can be attributed to fundamental differences in inductive bias. While baselines, like DLKcat and UniKP, rely on “late-stage fusion” to learn statistical correlations, and even CataPro and CatPred integrate structures only implicitly, UniKineG leverages GVP modules to execute geometric vector message passing within a unified coordinate system. By encoding orientation-dependent geometric interaction patterns directly into the representation, UniKineG ensures that predictions are driven by physically rational 3D representations. This marks a shift toward geometry-grounded representations, ensuring robustness, even when sequence homology and scaffold similarity fail.
2.4. Robust Generalization of UniKineG in Independent External Datasets
To assess the model’s generalization beyond our curated benchmark, we evaluated the frozen model in three task-specific external datasets derived from DLKcat (kcat) [11], CatPred-DB (Km) [15], and IECata (kcat/Km) [26].
To ensure the integrity of the evaluation, we implemented a multistage-filtering pipeline to prevent data leakage. Enzyme–substrate pairs overlapping with our benchmark (training/validation/test) were identified and excluded. Specifically, enzymes were matched using sequence hashes (MD5) generated after sequence normalization (whitespace removal and uppercasing). Simultaneously, substrates were standardized using RDKit identifiers, primarily InChIKey, or canonical SMILES, when unavailable, to ensure consistent chemical matching. After filtering and within-set deduplication (median aggregation for repeated pairs), the resulting external benchmarks contained 10,630 kcat, 5023 Km, and 4386 kcat/Km samples.
As summarized in Table 1, the model retains stable predictive performance under expected dataset shifts. For Km prediction, the model achieved PCC = 0.825, R2 = 0.608, and RMSE = 1.106, indicating that the learned representations remain consistent with kinetic trends across unseen data. For kcat and kcat/Km, the model achieved R2 = 0.545 and 0.523, respectively, and PCC remained robust (>0.78 across all the tasks). A modest attenuation relative to the internal benchmark is expected, likely reflecting heterogeneous experimental conditions (e.g., the pH, temperature, cofactors, and buffer composition) that are not explicitly modeled. Overall, these external results support the model’s utility for ranking enzyme candidates in silico; comparisons with prior work are provided in the discussion.
Table 1.
Predictive performances of UniKineG in independent external validation sets.
| Parameter | Dataset Source | Nsamples | R2 | RMSE | PCC |
|---|---|---|---|---|---|
| Kcat | DLKcat | 10,630 | 0.545 | 1.188 | 0.796 |
| Km | CatPred-DB | 5023 | 0.608 | 1.106 | 0.825 |
| Kcat/Km | IECata | 4386 | 0.523 | 1.217 | 0.783 |
3. Discussion
Predicting enzyme kinetics is also another part of knowing biocatalytic mechanisms, as well as developing other novel synthesis routes, discovering metabolic routes, and even finding new types of enzymes [27,28]. The UniKineG framework shown in this paper has integrated molecular-docking-derived 3D structural information with GVP neural networks [24] to improve the accuracy and generalization of , , and . The success of UniKineG signifies more than a performance increment: It shows the value of structure-aware modeling over that of descriptor engineering based on sequences or simple features by explicitly representing enzyme–substrate interaction geometries [29,30]. It could reduce our dependence on sequence homology and improve generalization to less-represented regions of the enzyme–substrate space. Importantly, the use of GVP-based vector features enables orientation-sensitive message passing, and the resulting representations are equivariant to rigid-body rotations and translations of the complex, making the predictor less sensitive to arbitrary coordinate-frame choices and pose orientations. This geometric equivariance improves robustness to coordinate perturbations, but it does not, by itself, constitute the explicit modeling of time-dependent conformational dynamics.
UniKineG’s slightly better results correspond to the 3D interaction context’s importance for kinetic predictions. The strong Km performance indicates that we may have learned something useful about the interfacial geometry and physicochemical context, which is informative for predicting the correct Km value for us in our benchmark setting. More significantly, for , a frequently used measure of catalytic efficiency [22], UniKineG achieves an value of 0.655, which is better than the strongest baseline on the held-out split. This result suggests that to predict catalytic efficiency, one needs to capture not just that atoms are near each other but also the precise spatial orientation. It also shows that directions of directional geometric features can add information beyond distances alone, possibly even orientation-sensitive interactions [31]. At the same time, we note that kcat is often more coupled to transition-state stabilization and dynamic catalytic events than Km, so a single static conformation may only partially capture the factors governing turnover. UniKineG can be used to prioritize candidates, with small structural differences resulting in large changes in predicted kinetics, and reduce experimental effort. On the contrary, multitask frameworks, such as UniKP, are mostly based on 1D sequence representations [13], whereas CataPro [14], which, although has structural information, does not have any explicit cross-molecular modeling, strengthens our idea that including some kind of explicit cross-molecular message passing is good for our model to be able to model enzyme–substrate interactions.
UniKineG has some new developments, but its limitations cannot be ignored as well. This structure-first approach is a model, so its performance ceiling is naturally limited by the fidelity of the antecedent molecular docking. For docking that is not accurate, especially with very flexible or complex substrates, it may introduce structural noise [32]. In particular, we expect the framework to be more challenged using metalloenzymes and cofactor-dependent enzymes (where catalysis depends on metal ions/cofactors that may be missing or simplified in the docked complex), as well as using enzymes with highly flexible active-site loops or pronounced induced-fit behaviors. But it also means that UniKineG is a model that might be improved upon by advancements in complex structure prediction (such as next-generation approaches, like AlphaFold3 [29]) that can produce better-quality complex geometries and, hence, reduce pose-related noise.
Another important limitation of the current structure-first pipeline is the absence of explicit solvent molecules in docked enzyme–substrate complexes. In many enzymes, structured water molecules are not merely a background solvent but can form conserved hydrogen bond networks and even participate directly in catalytic proton transfers (e.g., “water wires”). Notably, a recent large-scale survey of 1013 high-resolution enzyme crystal structures (<1.5 Å) reported that >99% contain continuous chains of water molecules linking active-site residues to the bulk solvent, highlighting the prevalence of active-site water networks in enzymatic systems. In contrast, standard AutoDock Vina (v1.2.6) docking typically treats solvation implicitly and does not represent conserved/bound water explicitly, which can alter the effective binding-site topology and local hydrogen bond geometry and may contribute to pose or scoring errors. Prior docking benchmarks have shown that incorporating carefully selected key water molecules may improve pose predictions, whereas including all the binding-site water molecules indiscriminately can reduce performance. Therefore, incorporating explicit “bridging” water molecules is a promising direction for future extensions of UniKineG, for example, by leveraging hydrated docking protocols available in the AutoDock suite Vina and further representing retained water molecules as additional nodes/edges in the heterogeneous interaction graph.
Although the AlphaFold structures used herein are generally high confidence at the dataset level (as reflected by Avg. pLDDT statistics), we do not yet incorporate the residue-level pLDDT or PAE into the docking and graph construction steps, nor do we explicitly propagate structural uncertainty into the downstream predictor. Future work could integrate confidence-weighted node/edge features (e.g., down-weighting low-confidence regions), PAE-informed uncertainty in inter-residue geometries, and uncertainty-aware message passing to improve robustness when structural priors are less reliable.
Furthermore, we also use this architecture, which relies heavily on static docked conformations and does not explicitly account for “induced-fit” effects or conformational ensembles [33]. More broadly, dynamic effects relevant to enzyme catalysis, such as conformational flexibility, induced-fit rearrangements, and transition-state stabilization, are not explicitly modeled in the current pipeline, which is particularly relevant for kcat prediction. Since catalysis is inherently dynamic, it is therefore promising to model these transient catalytic states through incorporating MD-derived ensembles (or multiple alternative docked poses as a lightweight ensemble) or expanding SE(3)-equivariant architectures [34] to better capture the variability of conformations. Additionally, because explicit geometric modeling is more computationally expensive, the next work will try to use model distillation to find a balance between being precise and the throughput that is required to screen a library scale [35].
Another limitation is that experimental assay conditions (e.g., the pH, temperature, buffer composition, and cofactors) are not incorporated, primarily because such metadata are inconsistently reported across sources. As a result, we treat kinetic labels as condition agnostic, which may introduce heterogeneous measurement noise and contribute to dataset shifts when evaluating across studies from different laboratories.
This point may partially explain why performance can attenuate on external benchmarks and motivates future work to curate condition metadata when available, include assay covariates as model inputs, or adopt uncertainty-aware objectives that account for label heterogeneity.
Finally, while UniKineG improves predictions using explicit geometry, we did not include a dedicated interpretability and attribution analysis to connect learned geometric features to specific catalytic residues or mechanistic interactions. This is mainly due to the lack of a consistent residue-level mechanistic ground truth at our dataset scale and the effects of the high computational cost and potential instability of perturbation-based explanations (e.g., SHAP-style methods) [36] on residue-level geometric graphs. Therefore, we position the current work as a predictive framework rather than a mechanistic inference tool, and we avoid making residue-level mechanistic claims. In future work, we will explore attribution methods tailored to geometric GNNs (e.g., saliency/IG and attention/edge attribution) on smaller curated benchmarks, where catalytic residues or key interactions are experimentally characterized.
Looking forward, UniKineG opens some promising research directions. We want to try combining QM-derived descriptors with geometric deep learning to be able to obtain reaction energetics beyond purely geometric cues [37]. Second, with the expansion of the graph definition, it can also be applied to multisubstrate reactions, allosteric regulation, and inhibitor design [38]. In the end, we hope for a long-term view of developing a generative paradigm that can reduce our reliance on experimentally derived complex structures and allow for more direct links from sequences to predicted functions [39]. To summarize, with UniKineG, by leveraging structural and geometric priors with geometric deep learning, a robust kinetic predictor is given, as well as the practical scalability of structure–kinetic relationship exploration.
4. Methods
4.1. Dataset Curation and Structural Annotation
Data Source and Initial Filtering. In order to create a high-quality benchmark for the structure-aware kinetic predictor, our data were sourced from the publicly available CataPro repository [14], which contains enzymes’ amino acid sequences, substrates’ SMILES strings, and corresponding kinetic parameters (). We implemented a multistage curation pipeline to remove incomplete records and standardize inputs. At first, we filtered the entries to keep only the ones that had complete multimodal information: full-length amino acid sequences, validatable SMILES strings, and experimentally determined kinetic values. We do not have a model for the assay conditions (e.g., the pH and temperature) because they are not uniformly reported in different sources, so we treated the reported kinetics as condition-agnostic labels.
Three-Dimensional Structural Generation and Pocket Profiling. Since experimental structures for enzyme–substrate complexes are sparse, we generated reliable 3D structural priors via computational docking. For each enzyme–substrate pair, the substrate’s SMILES was converted to a 3D conformer. For each enzyme, we retrieved the predicted 3D structure from the AlphaFold Protein Structure Database [40,41], using its UniProt [42] accession.
To assess the reliability of AlphaFold-derived structural inputs, we extracted per-residue pLDDT scores from the AlphaFold models (stored in the PDB/mmCIF temperature factor (B-factor) field provided by AlphaFold) and computed a protein-level average pLDDT (Avg. pLDDT) for each unique enzyme structure (N = 6003). The Avg. pLDDT distribution suggests that the AlphaFold-derived enzyme backbones are predominantly high confidence at the dataset level (mean = 93.52, median = 94.91, P5/P95 = 84.08/97.94, and min/max = 37.22/98.72; 84.61% of the enzymes have an Avg. pLDDT of ≥90, and 99.45% have and Avg. pLDDT of ≥70).
We identified candidate binding pockets using FPocket [41] and defined the docking search box (center coordinates and box dimensions) based on the top-ranked pocket. Docking was performed using AutoDock Vina (v1.2.6) [25,43], a rigid receptor, and a flexible ligand within this FPocket-derived search space, and the top-scoring pose was selected to obtain the enzyme–substrate complex geometry (atomic coordinates) for downstream featurization. At present, these confidence metrics (including the PAE) are not used as model features, and structural uncertainty is not explicitly propagated through docking or complex graph construction.
Wild-type Verification and Noise Elimination. A crucial quality-control step attended to implicit vagueness in the initial repository: Wild-type enzymes and their point mutants frequently possessed the same UniProt IDs, thus generating “label noise” sequence variations, which became decoupled from kinetic changes. Given that an individual point mutant was capable of seriously skewing its catalytic efficiency, we conducted a rigorous reannotation process. We used sequence alignment against the UniProt reference database [42] to make it explicit which were wild-type sequences and which were variants. To guarantee that structure–function relationships learned by the model were clean from any contamination from the mutant structures, only verified wild-type enzyme samples were retained for the final training set.
Final Dataset Statistics. We ended up with a dataset of 15,591 high-confidence enzyme–substrate pairs with an abundant tensor of multimodal features, such as 1D amino acid sequences, 2D molecular graphs (from SMILES), 3D pocket coordinates, and the experimental kinetic labels. This curated dataset served as the ground truth for training and evaluating the UniKineG framework.
4.2. Dataset-Splitting Strategies
Standard random splitting. For the baseline performance assessment, we adopted a random split, partitioning the dataset into training (80%), validation (10%), and held-out testing (10%) sets. This split evaluates the model’s ability to interpolate within the known distribution of enzyme–substrate interactions, serving as a reference point for the idealized performance.
Enzyme-unseen OOD evaluation. To assess generalization to novel enzymes, we constructed OOD splits by clustering enzyme sequences with CD-HIT at sequence identity thresholds of 99%, 80%, 60%, and 40%. For each threshold, we performed cluster-level splitting, such that sequences from the same CD-HIT cluster were assigned to only one split, ensuring that every test sequence had the same or a smaller specified identity compared to those of all the training sequences. The 40% cutoff threshold probes the low-homology regime, where sequence similarity provides limited guidance.
Substrate-unseen OOD evaluation. To evaluate the model’s generalization to novel chemical structures, we constructed OOD splits based on ECFP4 fingerprint similarity and Tanimoto coefficients. Specifically, we required each test substrate to have a maximum Tanimoto similarity to any training substrate of τ ≤ 0.4, 0.3, or 0.2. These splits emulate scenarios where predictions are required for compounds substantially dissimilar to those seen during training.
4.3. Representation Learning of Enzymes and Substrates
Enzyme and substrate feature encoding. We encoded enzymes and substrates using pretrained sequence models together with physicochemical descriptors. For enzyme representation, amino acid sequences were processed with ESM-2 (checkpoint: esm2_t33_650M_UR50D) to obtain 1280-dimensional per-residue embeddings, which were extracted from the final transformer layer [44]. We concatenated each residue embedding with a 30-dimensional vector of residue-level physicochemical features (e.g., charge, hydrophobicity, side chain volume) computed from a fixed feature table provided with our code. For substrate representation, we encoded SMILES strings into 768-dimensional embeddings using the pretrained NovoMolGen model [45]. In addition, we computed 45 global molecular descriptors (e.g., logP [46] and the topological polar surface area) from a fixed descriptor table provided with our code and concatenated them with the NovoMolGen embedding.
4.4. Graph Structure Construction
To comprehensively characterize the complex spatial and chemical dependencies within the enzyme–substrate complex, we constructed three interconnected graph structures:
Protein residue graph construction. To balance the computational cost and structural context between computational efficiency and structural information retention, we constructed a residue-level graph for the enzyme. Nodes represent amino acid residues, initialized with the concatenated features of ESM-2 embeddings [44] and 30-dimensional physicochemical vectors. Edges were established between residues, based on spatial proximity: An edge exists if the Euclidean distance between their atoms in 3D space is <. This topology captures the enzyme’s tertiary structure and local pocket geometry. This cutoff threshold is used during graph preconstruction; distance and direction features are computed from node coordinates during message passing.
Substrate molecular graph construction. The substrate was represented as a 2D molecular graph encoding its chemical topology. Nodes correspond to atoms, each featurized by a 45-dimensional vector capturing atomic chemical properties (e.g., atom type, degree, and hybridization state). Edges are defined based on covalent bonds parsed from SMILES strings using RDKit, representing the intrinsic connectivity of the small molecule.
Enzyme–substrate heterogeneous interaction graph. Serving as the core component for explicit intermolecular modeling, we constructed a heterogeneous graph that unifies the enzyme and substrate. The node set comprises the union of enzyme residues and substrate atoms. Intermolecular edges were constructed based on the 3D conformation generated using molecular docking: A connection is established if the spatial distance between any heavy atom of an enzyme residue and a substrate atom is <. Each intermolecular edge is annotated with geometric features computed from 3D coordinates, including the Euclidean distance and the unit direction vector, which are used as scalar and vector-edge features for geometric message passing.
4.5. UniKineG Network Architecture and Implementation Details
The UniKineG framework was implemented using Python (3.9.19) and PyTorch Geometric (2.6.1) [47]. To initialize the graph with rich semantic priors, we employed a dual-encoder strategy, where enzyme sequences were encoded using the pretrained ESM-2 model (specifically, ESM-2-650M, generating 1280-dimensional embeddings), and substrates were encoded using NovoMolGen (768 dimensions) fused with discrete physicochemical attributes. The core of UniKineG lies in its explicit handling of 3D geometry within a unified coordinate system. We constructed a heterogeneous graph where intraprotein edges represent the geometric/backbone proximity, intraligand edges represent covalent bonds, and bidirectional intermolecular edges are established based on the docking geometry. To explicitly preserve rotational and translational equivariances, node features were split into scalar and vector channels. Vector features were initialized as displacement vectors relative to the sub-graph’s centroid (pos-centroid), replicated across vector channels, while edge features incorporated Euclidean distances (utilized directly as scalars without radial basis function expansion) and unit direction vectors.
The network architecture stacks multiple layers of GVPConv (geometric vector perceptron convolution) to extract intragraph structural features, followed by GVPCrossConv layers that execute bidirectional message passing (protein-to-ligand and ligand-to-protein) to model the interface. The readout phase applies weighted mean/max/add pooling operations to both scalar channels and vector norms, followed by a fusion multilayer perceptron (MLP) to output the target kinetic parameter. We trained independent regression models for , , and using the AdamW optimizer [48] with a batch size of 32. The optimization hyperparameters were strictly configured with an initial learning rate of , a weight decay of , and coefficients , , and . The learning rate followed a warmup cosine annealing schedule, comprising a linear warmup for the first five epochs, followed by cosine decay to near zero over one hundred fifteen epochs. The objective function was a weighted combination of the MSE and Huber loss () in a ratio of 0.7:0.3. To mitigate overfitting and ensure robust generalization, we implemented a suite of regularization techniques: feature-level noise augmentation (with a probability of 0.2, generating a perturbed copy of each graph by adding Gaussian noise with σ = 0.005 to randomly selected node feature elements), gradient clipping (with a maximum norm of 0.8), and an early stopping mechanism (with a patience of 25 epochs). Computational cost and scalability. All the experiments were conducted on a single NVIDIA GeForce RTX 4090 GPU (24 GB). With early stopping, a typical training run per task takes approximately 6 h. For scalability, once complex geometries are prepared, the predictor supports batched inference and can be applied to large enzyme–substrate libraries for high-throughput candidate ranking.
4.6. Deep Learning Pipeline for Turnover Number Prediction in UniKineG
To initialize the heterogeneous graph (), we extract raw features from pretrained models. For protein nodes, sequence embeddings () are derived from ESM-2 (model esm2_t33_650M_UR50D); for substrate nodes, molecular embeddings () are obtained from NovoMolGen. These features are projected into a unified latent dimension () via linear transformation and layer normalization (LN). To incorporate 3D geometry, let denote the coordinate of node k. We compute the centroid () for each graph and construct translation-invariant vector features () as follows:
| (1) |
| (2) |
For any edge connecting node to (intramolecular or intermolecular), we compute the relative coordinate vector () and the Euclidean distance () on the fly. To ensure numerical stability, the normalized unit direction vector () is derived by adding a small constant () to the denominator. These geometric attributes are fed into the model as scalar edge features () and vector edge features (), explicitly driving the geometric information flow while preserving equivariance as follows:
| (3) |
The GVP module processes scalar and vector features to update node representations. To preserve 3D structural information, the vector channel undergoes a linear combination, followed by L2 norm extraction. Let and denote the sigmoid and ReLU activation functions, respectively. The vector-gating mechanism is formulated as follows:
| (4) |
| (5) |
We stack three layers of GVPConv independently for the protein and substrate graphs. At layer , the update rule for node aggregates geometric messages from its neighbors (). The message aggregation step sums the scalar messages () and vector messages () derived from incident edges as follows:
| (6) |
| (7) |
| (8) |
To model the enzyme–substrate interface, we perform bidirectional cross-graph convolutions. This process allows information to flow across bipartite interface edges (), updating substrate features based on the enzyme context and vice versa, utilizing the same GVP-based message-passing mechanism as follows:
| (9) |
| (10) |
Global representations for the protein and substrate are obtained using a weighted multiview-pooling strategy. We fuse , , and pooling operations using learnable scalar weights (). The final graph representation () is constructed by concatenating the pooled scalar features and vector norms. This unified embedding is then passed through an regression head to predict the target kinetic parameter () as follows:
| (11) |
| (12) |
4.7. Statistical Evaluation Metrics
We rigorously evaluated the regression performance of UniKineG across distinct testing scenarios (including held-out, enzyme-unseen, and substrate-unseen splits). For each kinetic target (, , and ), we calculated the following metrics independently. The coefficient of determination () measures the proportion of the variance in the dependent variable explained by the model relative to a mean-only baseline as follows:
| (13) |
The root-mean-square error () quantifies the magnitude of the prediction error, assigning higher penalties to large deviations. The Pearson correlation coefficient () evaluates the linear correlation between predicted and experimental values, independent of the scale, as follows:
| (14) |
| (15) |
where n denotes the total number of samples, represents the log10-transformed experimental value for the th sample, and is the corresponding predicted value. Additionally, and denote the means of the experimental and predicted values, respectively.
4.8. Benchmarking Against Representative Deep Learning Models
We benchmarked UniKineG against four representative models (DLKcat, UniKP, CataPro, and CatPred) by retraining their original implementations on identical data splits with fixed random seeds to ensure equitable comparisons. Model generalization was evaluated across three distinct scenarios: a standard held-out test set created by randomly sampling 10% of the dataset; an enzyme-unseen split, where proteins were clustered via CD-HIT at a strict 40% identity threshold (with sensitivity analyses at relaxed thresholds of 60%, 80%, and 99%) to prevent homologous leakage; and a substrate-unseen split based on Bemis–Murcko scaffolds, where we enforced rigorous structural separation by constraining the maximum Tanimoto similarity (using ECFP4 fingerprints) between training and testing molecules to , alongside robustness checks at looser thresholds of 0.3 and 0.4.
5. Conclusions
In this study, we developed UniKineG, a structure-guided deep learning framework for predicting enzyme kinetic parameters by integrating enzyme structures and substrate molecules into a shared 3D coordinate system through molecular docking. By representing the docking-aligned enzyme–substrate spatial context as a heterogeneous graph and performing rotation-aware geometric message passing with GVPs, UniKineG explicitly leverages intermolecular geometry to learn interaction patterns relevant to .
Using a curated benchmark of 15,591 wild-type enzyme–substrate pairs, UniKineG achieved strong accuracy in held-out evaluations and remained robust in stringent OOD settings that excluded close enzyme homologs (CD-HIT) and chemically similar substrates (ECFP4, Tanimoto). Ablation analyses indicate that docking-derived geometry, explicit cross-molecular message passing, and directional (vector) geometric features each contribute substantially beyond distance-only proximity, supporting the importance of orientation-sensitive representations for structure-grounded kinetic prediction in diverse biochemical spaces.
The framework depends on the fidelity of docking poses and currently uses static conformations; moreover, assay conditions, such as pH and temperature, were not modeled due to inconsistent reporting across data sources. Future work may incorporate improved complex geometry estimation, conformational ensembles, and complementary energetic descriptors to better capture catalysis-relevant effects. Overall, UniKineG provides a practical computational route for structure-aware kinetic annotation and offers a scalable foundation for exploring structure–kinetic relationships in enzyme discovery, pathway design, and enzyme engineering.
Acknowledgments
We are deeply grateful to Suyi Zhang, Hui Qin, Haipo Liu, and Ting Zhao and to the National Treasure Ecological Research Synergetic Innovation Center, National Engineering Research Center of Solid-State Brewing, and China Alcoholic Drinks Association.
Supplementary Materials
The following supporting information can be downloaded at https://www.mdpi.com/article/10.3390/ijms27041731/s1.
Author Contributions
Conceptualization, X.W. and P.S.; methodology, X.W.; data curation, X.W.; validation, P.S., K.L. and S.L.; writing—original draft preparation, X.W.; writing—review and editing, P.S. and S.L.; funding acquisition, S.L. and J.M. All authors have read and agreed to the published version of the manuscript.
Institutional Review Board Statement
Not applicable.
Informed Consent Statement
Not applicable.
Data Availability Statement
The data supporting the results of this study can be found in the article and Supplementary Materials. The raw data of the enzyme kinetic parameters used for curating UniKineG are available at BRENDA (https://www.brenda-enzymes.org/ (accessed on 11 September 2025)) and SABIO-RK (http://sabio.h-its.org/ (accessed on 11 September 2025)). The enzyme sequence data and AlphaFold-predicted enzyme structures are available at UniProt (https://www.uniprot.org/ (accessed on 11 September 2025)). The processed UniKineG datasets are available at Zenodo (https://doi.org/10.5281/zenodo.18125285 (accessed on 2 January 2026)).
Conflicts of Interest
The authors declare no conflicts of interest.
Funding Statement
This work was financially supported by the National Natural Science Foundation of China (22138004), Zhejiang Province Local Collaborative Innovation Project (2024SDXT001-3), and National Engineering Research Center of Solid-State Brewing.
Footnotes
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
References
- 1.Nielsen J. Systems Biology of Metabolism: A Driver for Developing Personalized and Precision Medicine. Cell Metab. 2017;25:572–579. doi: 10.1016/j.cmet.2017.02.002. [DOI] [PubMed] [Google Scholar]
- 2.Bar-Even A., Noor E., Savir Y., Liebermeister W., Davidi D., Tawfik D.S., Milo R. The moderately efficient enzyme: Evolutionary and physicochemical trends shaping enzyme parameters. Biochemistry. 2011;50:4402–4410. doi: 10.1021/bi2002289. [DOI] [PubMed] [Google Scholar]
- 3.Gu C., Kim G.B., Kim W.J., Kim H.U., Lee S.Y. Current status and applications of genome-scale metabolic models. Genome Biol. 2019;20:121. doi: 10.1186/s13059-019-1730-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.Johnson K.A., Goody R.S. The Original Michaelis Constant: Translation of the 1913 Michaelis–Menten Paper. Biochemistry. 2011;50:8264–8269. doi: 10.1021/bi201284u. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.Chang A., Jeske L., Ulbrich S., Hofmann J., Koblitz J., Schomburg I., Neumann-Schaal M., Jahn D., Schomburg D. BRENDA, the ELIXIR core data resource in 2021: New developments and updates. Nucleic Acids Res. 2021;49:D498–D508. doi: 10.1093/nar/gkaa1025. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.Wittig U., Rey M., Weidemann A., Kania R., Müller W. SABIO-RK: An updated resource for manually curated biochemical reaction kinetics. Nucleic Acids Res. 2018;46:D656–D660. doi: 10.1093/nar/gkx1065. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Coates T.L., Young N., Jarrett A.J., Morris C.J., Moody J.D., Corte D.D. Current computational methods for enzyme design. Mod. Phys. Lett. B. 2021;35:2150155. doi: 10.1142/S0217984921501554. [DOI] [Google Scholar]
- 8.Zhang D., Liu M., Yu Z., Xu H., Pfister S., Menichetti G., Kou X., Zhu J., Fan D., Rao P. Domain knowledge, just evaluation, and robust data standards are required to advance AI in food science. Trends Food Sci. Technol. 2025;164:105272. [Google Scholar]
- 9.Shi P., Liu S., Mao J., Liu X., Tu R., Qin H., Sun A., Zhang D., Mao J. AI-driven exploration of microbial resources in fermented foods. Trends Food Sci. Technol. 2026;167:105450. [Google Scholar]
- 10.Heckmann D., Lloyd C.J., Mih N., Ha Y., Zielinski D.C., Haiman Z.B., Desouki A.A., Lercher M.J., Palsson B.O. Machine learning applied to enzyme turnover numbers reveals protein structural correlates and improves metabolic models. Nat. Commun. 2018;9:5252. doi: 10.1038/s41467-018-07652-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Li F., Yuan L., Lu H., Li G., Chen Y., Engqvist M.K.M., Kerkhoven E.J., Nielsen J. Deep learning-based kcat prediction enables improved enzyme-constrained model reconstruction. Nat. Catal. 2022;5:662–672. [Google Scholar]
- 12.Kroll A., Rousset Y., Hu X.-P., Liebrand N.A., Lercher M.J. Turnover number predictions for kinetically uncharacterized enzymes using machine and deep learning. Nat. Commun. 2023;14:4139. doi: 10.1038/s41467-023-39840-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Yu H., Deng H., He J., Keasling J.D., Luo X. UniKP: A unified framework for the prediction of enzyme kinetic parameters. Nat. Commun. 2023;14:8211. doi: 10.1038/s41467-023-44113-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Wang Z., Xie D., Wu D., Luo X., Wang S., Li Y., Yang Y., Li W., Zheng L. Robust enzyme discovery and engineering with deep learning using CataPro. Nat. Commun. 2025;16:2736. doi: 10.1038/s41467-025-58038-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Boorla V.S., Maranas C.D. CatPred: A comprehensive framework for deep learning in vitro enzyme kinetic parameters. Nat. Commun. 2025;16:2072. doi: 10.1038/s41467-025-57215-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Wang Y., Cheng L., Zhang Y., Cao Y., Alghazzawi D. DEKP: A deep learning model for enzyme kinetic parameter prediction based on pretrained models and graph neural networks. Brief. Bioinform. 2025;26:bbaf187. doi: 10.1093/bib/bbaf187. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Maryam, Rehman M.U., Hussain I., Tayara H., Chong K.T. A graph neural network approach for predicting drug susceptibility in the human microbiome. Comput. Biol. Med. 2024;179:108729. doi: 10.1016/j.compbiomed.2024.108729. [DOI] [PubMed] [Google Scholar]
- 18.Maryam, Rehman M.U., Chong K.T., Tayara H. Drugs inhibition prediction in P-gp enzyme: A comparative study of machine learning and graph neural network. Comput. Toxicol. 2025;34:100344. doi: 10.1016/j.comtox.2025.100344. [DOI] [Google Scholar]
- 19.Tian W., Skolnick J. How Well is Enzyme Function Conserved as a Function of Pairwise Sequence Identity? J. Mol. Biol. 2003;333:863–882. doi: 10.1016/j.jmb.2003.08.057. [DOI] [PubMed] [Google Scholar]
- 20.Warshel A., Sharma P.K., Kato M., Xiang Y., Liu H., Olsson M.H. Electrostatic basis for enzyme catalysis. Chem. Rev. 2006;106:3210–3235. doi: 10.1021/cr0503106. [DOI] [PubMed] [Google Scholar]
- 21.Kraut J. How Do Enzymes Work? Science. 1988;242:533–540. doi: 10.1126/science.3051385. [DOI] [PubMed] [Google Scholar]
- 22.Fersht A. In: Structure and Mechanism in Protein Science: A Guide to Enzyme Catalysis and Protein Folding. Freeman W.H., editor. Macmilla; New York, NY, USA: 1999. [Google Scholar]
- 23.Henzler-Wildman K., Kern D. Dynamic personalities of proteins. Nature. 2007;450:964–972. doi: 10.1038/nature06522. [DOI] [PubMed] [Google Scholar]
- 24.Jing B., Eismann S., Suriana P., Townshend R.J.L., Dror R.O. Learning from Protein Structure with Geometric Vector Perceptrons. arXiv. 20202009.01411 [Google Scholar]
- 25.Trott O., Olson A.J. AutoDock Vina: Improving the speed and accuracy of docking with a new scoring function, efficient optimization, and multithreading. J. Comput. Chem. 2010;31:455–461. doi: 10.1002/jcc.21334. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Wang J., Zhao Y., Yang Z., Yao G., Han P., Liu J., Chen C., Zan P., Wan X., Bo X., et al. IECata: Interpretable bilinear attention network and evidential deep learning improve the catalytic efficiency prediction of enzymes. Brief. Bioinform. 2025;26:bbaf283. doi: 10.1093/bib/bbaf283. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Nielsen J., Keasling J.D. Engineering Cellular Metabolism. Cell. 2016;164:1185–1197. doi: 10.1016/j.cell.2016.02.004. [DOI] [PubMed] [Google Scholar]
- 28.Lawson C.E., Martí J.M., Radivojevic T., Jonnalagadda S.V.R., Gentz R., Hillson N.J., Peisert S., Kim J., Simmons B.A., Petzold C.J., et al. Machine learning for metabolic engineering: A review. Metab. Eng. 2021;63:34–60. doi: 10.1016/j.ymben.2020.10.005. [DOI] [PubMed] [Google Scholar]
- 29.Abramson J., Adler J., Dunger J., Evans R., Green T., Pritzel A., Ronneberger O., Willmore L., Ballard A.J., Bambrick J., et al. Accurate structure prediction of biomolecular interactions with AlphaFold 3. Nature. 2024;630:493–500. doi: 10.1038/s41586-024-07487-w. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30.Watson J.L., Juergens D., Bennett N.R., Trippe B.L., Yim J., Eisenach H.E., Ahern W., Borst A.J., Ragotte R.J., Milles L.F., et al. De novo design of protein structure and function with RFdiffusion. Nature. 2023;620:1089–1100. doi: 10.1038/s41586-023-06415-8. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 31.Stumpfe D., Bajorath J. Exploring Activity Cliffs in Medicinal Chemistry. J. Med. Chem. 2012;55:2932–2942. doi: 10.1021/jm201706b. [DOI] [PubMed] [Google Scholar]
- 32.Pinzi L., Rastelli G. Molecular Docking: Shifting Paradigms in Drug Discovery. Int. J. Mol. Sci. 2019;20:4331. doi: 10.3390/ijms20184331. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33.Amaro R.E., Baudry J., Chodera J., Demir Ö., McCammon J.A., Miao Y., Smith J.C. Ensemble Docking in Drug Discovery. Biophys. J. 2018;114:2271–2278. doi: 10.1016/j.bpj.2018.02.038. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34.Chatzipantazis E., Pertigkiozoglou S., Dobriban E., Daniilidis K. SE(3)-Equivariant Attention Networks for Shape Reconstruction in Function Space. arXiv. 20222204.02394 [Google Scholar]
- 35.Zhou J., Rizzo K., Tang Z., Koo P.K. Uncertainty-aware genomic deep learning with knowledge distillation. bioRxiv. 2024 doi: 10.1038/s44387-025-00053-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 36.Salih A.M.A., Raisi-Estabragh Z., Galazzo I.B., Radeva P., Petersen S.E., Lekadir K., Menegaz G. A Perspective on Explainable Artificial Intelligence Methods: SHAP and LIME. Adv. Intell. Syst. 2023;7:2400304. doi: 10.1002/aisy.202400304. [DOI] [Google Scholar]
- 37.Wang X., Mao Y., Wang Z. Machine learning approaches for transition state prediction. Chem Catal. 2025;5:101458. doi: 10.1016/j.checat.2025.101458. [DOI] [Google Scholar]
- 38.Yang J., Li F.Z., Arnold F.H. Opportunities and Challenges for Machine Learning-Assisted Enzyme Engineering. ACS Cent. Sci. 2024;10:226–241. doi: 10.1021/acscentsci.3c01275. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39.Yeh A.H.-W., Norn C., Kipnis Y., Tischer D., Pellock S.J., Evans D., Ma P., Lee G.R., Zhang J.Z., Anishchenko I., et al. De novo design of luciferases using deep learning. Nature. 2023;614:774–780. doi: 10.1038/s41586-023-05696-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40.Varadi M., Anyango S., Deshpande M., Nair S., Natassia C., Yordanova G., Yuan D., Stroe O., Wood G., Laydon A., et al. AlphaFold Protein Structure Database: Massively expanding the structural coverage of protein-sequence space with high-accuracy models. Nucleic Acids Res. 2022;50:D439–D444. doi: 10.1093/nar/gkab1061. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 41.Jumper J., Evans R., Pritzel A., Green T., Figurnov M., Ronneberger O., Tunyasuvunakool K., Bates R., Žídek A., Potapenko A., et al. Highly accurate protein structure prediction with AlphaFold. Nature. 2021;596:583–589. doi: 10.1038/s41586-021-03819-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 42.Apweiler R., Bairoch A., Wu C.H., Barker W.C., Boeckmann B., Ferro S., Gasteiger E., Huang H., Lopez R., Magrane M., et al. UniProt: The Universal Protein Knowledgebase in 2023. Nucleic Acids Res. 2023;51:D523–D531. doi: 10.1093/nar/gkac1052. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43.Morris G.M., Huey R., Lindstrom W., Sanner M.F., Belew R.K., Goodsell D.S., Olson A.J. AutoDock4 and AutoDockTools4: Automated docking with selective receptor flexibility. J. Comput. Chem. 2009;30:2785–2791. doi: 10.1002/jcc.21256. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 44.Lin Z., Akin H., Rao R., Hie B., Zhu Z., Lu W., Smetanin N., Verkuil R., Kabeli O., Shmueli Y., et al. Evolutionary-scale prediction of atomic-level protein structure with a language model. Science. 2023;379:1123–1130. doi: 10.1126/science.ade2574. [DOI] [PubMed] [Google Scholar]
- 45.Chitsaz K., Balaji R., Fournier Q., Bhatt N.P., Chandar S. NovoMolGen: Rethinking Molecular Language Model Pretraining. arXiv. 20252508.13408 [Google Scholar]
- 46.Wildman S., Crippen G. Prediction of Physicochemical Parameters by Atomic Contributions. J. Chem. Inf. Comput. Sci. 1999;39:868–873. doi: 10.1021/ci990307l. [DOI] [Google Scholar]
- 47.Fey M., Lenssen J.E. Fast Graph Representation Learning with PyTorch Geometric. arXiv. 2019 doi: 10.48550/arXiv.1903.02428.1903.02428 [DOI] [Google Scholar]
- 48.Loshchilov I., Hutter F. Fixing Weight Decay Regularization in Adam. arXiv. 20171711.05101 [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
The data supporting the results of this study can be found in the article and Supplementary Materials. The raw data of the enzyme kinetic parameters used for curating UniKineG are available at BRENDA (https://www.brenda-enzymes.org/ (accessed on 11 September 2025)) and SABIO-RK (http://sabio.h-its.org/ (accessed on 11 September 2025)). The enzyme sequence data and AlphaFold-predicted enzyme structures are available at UniProt (https://www.uniprot.org/ (accessed on 11 September 2025)). The processed UniKineG datasets are available at Zenodo (https://doi.org/10.5281/zenodo.18125285 (accessed on 2 January 2026)).






