Abstract
Enzyme kinetic parameters, including k cat, K m, k cat/K m, and K i, are critical for guiding applications in enzyme engineering, metabolic modeling, and synthetic biology by providing quantitative information on enzyme activity under various conditions. Experimental determination of these parameters is often costly and time-consuming. Moreover, traditional computational methods are not well-suited to estimating these parameters. This motivated the development of machine learning (ML) models for in silico predictions. Here, we review recent advances in ML-based prediction of enzyme kinetic parameters by highlighting global models trained on diverse enzyme classes and local models catered toward specific enzyme families. These models have been applied in myriads of applications including predicting mutation effects, accelerating enzyme mining, and parametrizing genome-scale metabolic models. While data scarcity remains the main limitation for these models, we outline emerging opportunities such as high-throughput data generation and semisupervised learning as means to overcome this issue. In summary, this Review provides a roadmap for leveraging ML to enhance the performance, robustness, and scope of enzyme kinetic parameter prediction, leading to the accurate annotation of protein sequences for target functions.
Keywords: artificial intelligence, catalytic activity, enzyme engineering, enzyme evolution, enzyme mining, kinetic parameters, machine learning


1. Introduction
Enzymes are natural catalysts that selectively and efficiently transform chemical compounds by accelerating chemical reactions. , Advances in recombinant DNA and gene cloning technologies have facilitated the scalable expression of enzymes in microbial hosts, making them attractive candidates for the synthesis of value-added chemicals. , In the context of the transition toward a circular economy, these bioprocesses provide sustainable pathways for the production of materials and energy. Nevertheless, the effectiveness of these processes depends on a set of kinetic parameters that govern the rate and efficiency at which enzymes catalyze substrates into products. Hence, optimizing enzymatic systems for industrial applications becomes essential as it demands meticulous engineering and screening to achieve catalytic activity and stability compatible with process conditions. ,
Though enzymes are usually built from only 20 amino acids, they exhibit a great diversity. For instance, there are 20100 combinatorial possibilities for a typical peptide chain of 100 residues, which exceeds the number of known particles in the universe. However, it is estimated that only 1 of 1077 of these sequences folds into a stable and functional structure. , Traditionally, navigating these sequences is facilitated by experimental approaches such as directed evolution to find enhanced mutants of known enzymes , or metagenomic mining to discover new enzymes. While these strategies narrow the search space, their dependence on costly, time-consuming, and iterative assays poses an obstacle to scaling. , This leaves a large portion of enzymes uncharacterized, given the vastness of the sequence space. ,
Despite the limitations discussed above, the experimental efforts to search the sequence space resulted in a surge in the size of the available data. This permits the use of statistical methods to reveal hidden patterns and relationships. Hence, a promising alternative that has been gaining a lot of traction in recent years relies on advances in artificial intelligence to fit models to the available data and make in silico predictions regarding enzymatic properties. , To this extent, machine learning (ML) and deep learning (DL) models have been successfully used for predicting enzyme structure, − function, , and fitness. − The fitness landscape is composed of multiple features, such as activity, stability, and expression levels, that contribute to the overall performance. , However, fitness scores do not directly quantify catalytic performance under reaction conditions. , Also, computational tools to predict activity are still lacking compared to those developed for predicting stability and expression. Hence, a shift from generic fitness proxies to parameter-specific predictions represents a more interpretable approach for engineering enzymes with the desired catalytic properties.
When it comes to quantifying activity, ML models should make predictions for kinetic parameters, such as the turnover number (k cat), the Michaelis constant (K m), the catalytic efficiency (k cat/K m), and the inhibition constant (K i). Each parameter represents enzyme activity under different conditions, such as varying substrate concentrations and the presence or absence of inhibitors. Briefly, k cat represents the maximal number of substrate molecules converted to products per active site of enzymes per unit time, while K m is equivalent to the concentration of the substrate at which the enzyme functions at half of its maximal catalytic rate. The ratio of these two constants represents the efficiency of the enzyme in converting substrate to product, considering the enzyme’s speed and its affinity toward the substrate. Meanwhile, when an inhibitor is present, it is important to account for K i, which quantifies the strength of the inhibitor in blocking the activity of the enzyme. In some cases, k cat and K m can be theoretically determined from the activation free energy and substrate binding energy, respectively. However, in practice, estimating transition state energies at atomic accuracy is computationally challenging for most enzymes. Moreover, K m does not always directly reflect substrate–enzyme binding affinity, as it can also be influenced by other kinetic steps. These challenges make it difficult to reliably estimate enzyme kinetic parameters directly from physical principles, highlighting the potential of ML methods trained on experimental data to bridge this gap.
Early work in this space applied ML algorithms to small curated data sets, , but the field is rapidly evolving to cover large data sets of enzyme–substrate pairs (Figure ). Fast and accurate predictions of these parameters can accelerate the discovery of efficient enzymes suitable for desired biochemical reactions by avoiding sequences with poor k cat/K m or that suffer from strong product inhibition. Moreover, they can assist in genome-scale metabolic modeling and biosynthetic pathway prescreening by providing quantitative information on constituting reactions.
1.
Trend in the number of ML and DL models published for enzyme kinetic parameter estimation. This illustrates the increasing attention that these models are receiving as a useful tool in enzyme engineering.
This Review aims to synthesize the current landscape of ML- and DL-based predictions of enzyme kinetic parameters. We begin by providing the reader with a background on ML basics in the context of enzyme kinetics (Section ). Then, we highlight the common data sets available for kinetic parameter data (Section ). Thereafter, we provide a roadmap of existing models summarizing their inputs, architecture, capabilities, and performance (Section ). We also shed light on several applications in which these models have proven successful (Section ) and discuss local models designed for specific enzymatic families (Section ). Finally, we discuss the major challenges that these models face and propose directions for future research (Section ).
2. Basic Elements of Machine Learning in Enzyme Kinetics
To gain a general view of the basic principles of ML in protein engineering, we refer the readers to a comprehensive review by Kouba and colleagues (2023) and Tables S1–S3 of the Supporting Information. Here, we provide a brief description of the tools and model architectures developed for enzyme kinetic parameters. ML models for enzyme kinetics prediction typically take two main inputs: the enzyme, represented by its amino acid sequence or 3D-structure, and the substrate, represented by chemical structure, simplified molecular input line entry system (SMILES) strings, or molecular fingerprints. These representations convert molecular information about enzymes and substrates to mathematically tractable formats while minimizing information loss (Figure a). Various neural network architectures excel at encoding each modality, but usually require relatively large data sets for training in order to achieve high performance.
2.
Overview of ML workflows for enzyme kinetic parameter prediction and enzyme kinetic data sets. (a) Schematic of typical model architectures and (b) number of available kinetic parameter data in major public databases as of July 2025.
A simple method for extracting enzyme sequence features is to use one-hot encoding, where each amino acid is represented as a binary vector with “1” indicating its presence and “0” indicating its absence at residue position i. However, this leads to sparse and high dimensional feature vectors compromising computational efficiency. Meanwhile, convolutional neural networks (CNNs) extract local motifs using sliding kernels to output n-gram frequencies or catalytic site patterns. The number of features from n-grams can grow extremely large, potentially becoming a high dimensional feature space as n increases. While CNNs are efficient for detecting short functional motifs, they fail to capture long-range dependencies, for example intramolecular interactions between amino acids that are distant in the primary sequence. This issue can be overcome using protein language models (pLMs). In natural language processing (NLP), language models treat sentences as sequences of words and learn their contextual relationships via a form of self-supervised learning called transformer networks. , Similarly, enzyme sequences can be interpreted as sentences and tokenized using the individual amino acids of which they are comprised of. This parallelism opened the door to creating pLMs that are trained to predict masked amino acids in enzyme sequences. Ultimately, the outputs from pLMs, called embeddings, hold essential functional and structural information about the enzyme sequences, encompassing both short- and long-range relationships. To make predictions about enzyme kinetic parameters, enzyme sequences can be numerically encoded using pretrained pLMs such as Evolutionary Scale Modeling 2 (ESM2), ProtT5, and UniRep. These models are trained on millions of unlabeled protein sequences to learn biologically relevant features without requiring functional labels. These models generate a high-dimensional embedding for each amino-acid residue. However, since ML models usually require one vector per enzyme rather than one per residue, these embeddings are compressed into a fixed-length enzyme-level representation by applying pooling operations over the final transformer layer. On average, these features are considered high-dimensional, as the length of a pLM vector based on the last transformer layer is ∼1000–1300.
When structural data are used, a variety of information encoding strategies could be utilized. A simple method for capturing structural information would be to count the number of specific residues within the radius of a point of interest (e.g., active site). Alternatively, enzyme structures can be converted to residue–residue graphs, called contact maps, where the nodes represent amino acids and the edges depict spatial proximity and interactions. This information can be encoded by using graph convolutional networks (GCNs) or graph attention networks (GANs). GCNs aggregate information from neighboring residues in the graph and propagate local structural features across the enzyme. Meanwhile, GANs extend GCNs by introducing attention weights that allow the model to focus on important neighbors, such as catalytic residues and substrate binding sites, rather than treating all neighbors equally.
As for substrates, they can be encoded as fingerprints using various simple molecular descriptors that were historically used by the chemical community to describe small molecules like MACCS keys that tabulate the presence of substructures and functional groups into fixed-length vectors. To encode substrate structures as molecular graphs, GCNs and GANs are used to aggregate atom-level features such as partial charges and aromaticity and weight atom–atom connections attributed to chemically relevant interactions. Moreover, message passing neural networks (MPNNs) extend this by explicitly modeling iterative message exchanges between atoms to depict long-range effects in the molecule. Finally, akin to pLMs, transformer-based networks such as the SMILES transformer and ChemBERTa have been developed to tokenize SMILES strings into numerical features by learning substructure relationships and chemical grammar from large molecular data sets.
3. Databases
ML models used for complex tasks, such as predicting enzyme kinetic parameters, require large sets of training data to generalize well and generate reliable predictions. Such information is publicly available in large databases like BRaunschweig ENzyme DAtabase (BRENDA), System for the Analysis of BIOchemical Pathways - Reaction Kinetics (SABIO-RK), and UNIversal PROTein resource (UniProt).
BRENDA is one of the most comprehensive repositories of experimental enzyme data extracted from more than 100,000 literature references. Besides kinetic parameters, the database contains information about the enzyme’s classification (EC) number, the source organism, and assay conditions, to name a few. Similarly, SABIO-RK is a database containing data about enzymatic reactions and their kinetic parameters manually retrieved from the literature. Whereas BRENDA focuses on enzymes and their kinetic parameters, SABIO-RK is centered around reactions and goes beyond kinetic constants to cover rate laws and experimental conditions. , The number of enzyme kinetic parameters available in these databases is reported in Figure b. It is noteworthy to mention that there are fewer k cat/K m and K i entries in these databases compared to k cat and K m.
As for UniProt, it is the largest database for proteins containing information about protein sequences, structures, and functions, as well as their functional parameters. Enzyme kinetic data in UniProt is scarce, with only around 1% of the enzymes listed in UniProt having an experimentally determined k cat value. − However, its importance lies in the UniProt IDs that are usually used as anchors to align kinetic records from BRENDA and SABIO-RK with their corresponding sequences. For example, Krishnan and colleagues (2025) proposed a structure-oriented kinetic database (SKiD) that integrates k cat and K m with their corresponding 3D-structure data for 13,654 enzyme–substrate pairs spanning six enzyme classes. In this database, kinetic data were retrieved from BRENDA, and the sequences were mapped to their Protein Data Bank (PDB) structures via UniProt IDs. After identifying the catalytic and binding sites for the substrates, docking energies were calculated and stored in the database using GNINA. A similar but smaller database of 1050 enzyme–substrate pairs called IntEnzyDB also exists. Moreover, Boorla & Maranas (2025) compiled CatPred-DB, a comprehensive data set of k cat, K m, and K i values that integrates the UniProt IDs to map entries from BRENDA and SABIO-RK to their corresponding amino acid sequence identifiers and predicted 3D structures in the AlphaFold-2.0 database. , CatPred-DB spans 23197 k cat, 41174 K m, and 11929 K i entries expanding the enzyme sequence space by up to 60% compared to other ML kinetic data sets.
In contrast to the discussed databases that compile experimentally measured values, GotEnzyme and GotEnzyme2 represent a new class of predicted kinetic parameters repositories. The original GotEnzyme database provides k cat values for over 25.7 million enzyme-compound pairs across 8099 organisms predicted from the pretrained DLKcat model (see Section ). , GotEnzyme2 extends this to a broader set of parameters, including K m and k cat/K m, for 59.6 million entries by retraining a suite of ML models (see Section ) on a curated set of experimental data and utilizing the best-performing strategy to make predictions for uncharacterized enzyme–substrate combinations. It is important to note that most of the kinetic constants in these databases lack experimental validation as the values are derived from ML regression models whose predictive performance on held-out experimental measurements reaches R 2 ≈ 0.5–0.7. Consequently, although GotEnzyme and GotEnzyme2 greatly expand the characterized enzymatic sequence space, they should not be treated as ground-truth kinetic data, as the propagation of predicted error can be misleading. Instead, these resources can complement, rather than replace, experimentally curated databases in hypothesis generation and preliminary screening in enzymatic applications.
One caveat regarding the public data sets is that their entries are nonstandardized, where the data are collected under a myriad of experimental conditions. Moreover, many entries have a large portion of missing metadata, such as temperature, pH, and substrate concentration. This heterogeneity complicates the process of detecting anomalies and outliers, especially when the sequence-substrate pair is associated with multiple reported values of kinetic parameters. To overcome this variability, some studies resort to taking the maximum reported value or computing the geometric mean of the available entries as a representative value. While these approaches help with reducing noise, they might overlook biologically meaningful variation and hence highlight the need for more structured and standardized data sets. Moreover, it has been reported that up to 20% of the entries in BRENDA were inconsistent with the results reported in their published references, likely due to human errors and erroneous replacements of units. When it comes to mapping substrate names to their respective SMILES notations, one or more of chemical and biological information databases, such as PubChem, KEGG, or ChEBI, are used. However, the same compounds can have nonidentical common names under different entries, resulting in inaccurate SMILES mapping and highlighting the need for a systematic substrate information retrieval pipeline.
To that extent, recent initiatives to improve the findability, accessibility, interoperability, and reusability (FAIR) of data have been proposed. The Beilstein-Institut proposed a set of guidelines called STandards for Reporting ENzymology DAta (STRENDA) to report the results of enzyme-related measurements, ranging from specifying reaction conditions to linking experimental information with the selected kinetic model and estimated kinetic parameters. These standards are tailored toward enzymes used in synthesis, ensuring that the reported data covers the catalytic parameters of the enzymes which remains optional in general protein databases. While these standards are increasingly being adopted in databases like BRENDA and SABIO-RK, the extent of their usage remains limited and is at the discretion of researchers. To facilitate this transition, validation tools such as EnzymeML and STRENDA DB have been created to automatically check and ensure that enzymology data are complete and valid before being published in a journal or database.
Lastly, it is important to note that most of the kinetic entries in the aforementioned databases are derived from in vitro measurements. As enzyme behavior in vivo significantly differs due to molecular crowding, protein–protein interactions, and post-translational regulations, metabolic models trained on in vitro-based predictions may not fully capture physiological kinetic properties. Hence, this limitation should be considered when using these databases to train models applied to in vivo systems.
4. Global Models
Most ML models developed for enzyme kinetic parameter prediction are inherently predictive in nature. They are designed to quantify catalytic properties within existing biochemical contexts rather than generate de novo enzymes, though generative frameworks are beginning to emerge as complementary tools. Within this predictive landscape, ML models can broadly be categorized as either global or local depending on the scope of their training data and intended applicability. Global models are trained on large data sets that span multiple enzyme families, classes, and organisms. In theory, this enables them to generalize across a wide diversity of sequences and reactions. Their strength lies in capturing extensive trends in sequence-function relationships and being applied to distantly related enzymes. In contrast, local models are tailored to a narrower sequence space, focusing on a single enzyme and its variants or a family of closely related enzymes. By leveraging high-quality data in this restricted domain, local models have the potential to capture fine-grained interactions, albeit at the cost of generalizability. In this section, we highlight the advances in global models to predict k cat, K m, k cat/K m, and K i, summarizing details about the models and their performance in Tables , , , and respectively. The summary of the k cat and K m models not discussed in detail is provided in Tables S4 and S5.
1. Characteristics and Performance of ML Models for Predicting k cat .
| model name | data set size | model architecture | enzyme sequence representation | substrate representation | other features | performance |
|---|---|---|---|---|---|---|
| k cat in E. coli (Section ) | 215 | random forest | enzyme structure, network interactions, biochemistry, assay conditions | R 2 = 0.34 | ||
| DLKcat (Section ) | 16838 | neural network | n-gram (CNN) | molecular graph (GNN) | R 2 = 0.44 | |
| TurNuP (Section ) | 4271 | gradient boosting | ESP | numerical reaction fingerprint | R 2 = 0.44† | |
| CataPro (Section ) | 27658 | neural network | ProtT5-XL | MolT5 and MACCS keys | r = 0.48† | |
| UniKP (Section ) | 16838 | ExtraTrees | ProtT5-XL-UniRef50 | pretrained SMILES transformer | R 2 = 0.68 | |
| EF-UniKP (Section ) | 636 | ExtraTrees | ProtT5-XL-UniRef50 | pretrained SMILES transformer | pH | R 2 = 0.44 |
| 572 | ExtraTrees | ProtT5-XL-UniRef50 | pretrained SMILES transformer | temp | R 2 = 0.38 | |
| DLTKcat (Section ) | 16249 | neural network | n-gram (CNN) | molecular graph (GAN) | temp | R 2 = 0.66 |
| PreTKcat (Section ) | 16249 | ExtraTrees | ProtT5-XL-UniRef50 | molecular graph (MolGNet) | temp | R 2 = 0.69 |
| GELKcat (Section ) | 16838 | neural network | n-gram (word2vec) | molecular graph (graph transformer) | enzyme structure (GCN) | R 2 = 0.56 |
| SAKPE (Section ) | 31507 | gradient boosting | ESM-C | Mole-BERT | catalytic and substrate binding sites from EasIFA | R 2 = 0.49† |
| OmniESI (Section ) | 23197 | neural network | ESM-2 | molecular graph (GCN) | R 2 = 0.41† | |
| DeepEnzyme (Section ) | 11927 | neural network | transformer | molecular fingerprint (GCN) | R 2 = 0.58 | |
| KcatNet (Section ) | 11757 | neural network | ESM-2, ProtT5-XL-UniRef50 | pretrained SMILES transformer | enzyme structure (GCN) | R 2 = 0.69 |
| KinForm (Section ) | 35001 | ExtraTrees | ESMC, ESM-2, ProtT5-XL-UniRef50 | pretrained SMILES transformer | R 2 = 0.68† | |
| CatPred (Section ) | 23197 | neural network ensemble | ESM-2 | D-MPNN | enzyme structure (E-GNN), sequence attention | R 2 = 0.61† |
| ENKIE (Section ) | 25648 | Bayesian multilevel models | MetaNetX reaction identifier, MetaNetX substrate identifier, EC number, enzyme family | R 2 = 0.36† | ||
| CPI-Pred (Section ) | 11834 | neural network | ESM-2 | molecular fingerprint (MPNN) | r = 0.54† | |
| RealKcat (Section ) | 30442 | gradient boosting | ESM-2 | ChemBERTa | accuracy = 0.89 |
Metrics without a superscript are measured after random data split for training and testing. Metrics with the superscript “†” are measured after sequence-aware data splits for training and testing. R 2 is shown as a primary metric to compare the performance of models. When R 2 values were not reported, PCC or accuracy were used instead. The sections where each model appears were listed below their respective names.
2. Characteristics and Performance of ML Models for Predicting K m .
| model name | data set size | model architecture | enzyme sequence representation | substrate representation | other features | performance |
|---|---|---|---|---|---|---|
| Kroll_Km | 11675 | gradient boosting | UniRep | molecular graph (GNN) | substrate molecular weight, substrate octanol–water partition coefficient | R 2 = 0.26† |
| MLAGO | 17151 | random forest | EC number, substrate KEGG ID, organism KEGG ID | R 2 = 0.54 | ||
| GraphKM | 19754 | gradient boosting | ESM-2 | molecular graph (GNN) | r = 0.59 | |
| CatPred | 41174 | neural network ensemble | ESM-2 | D-MPNN | structural features (E-GNN), sequence attention | R 2 = 0.53† |
| MPEK | 24585 | neural network | ProtT5-XL | Mole-BERT | pH, temp, organism name | R 2 = 0.61 |
| DLERKm | 10122 | neural network | ESM-2 | molecular fingerprint (RDKit) | reaction fingerprints (RXNFP) | R 2 = 0.59 |
| UniKP | 11722 | ExtraTrees | ProtT5-XL | pretrained SMILES transformer | temp, pH | R 2 = 0.53 |
Metrics without a superscript are measured after random data split for training and testing; Metrics labeled with “†” are measured after sequence-aware data splits for training and testing. R 2 is shown as a primary metric to compare the performance of models. When R 2 values were not reported, PCC was used instead.
3. Characteristics and Performance of ML Models for Predicting k cat/K m .
| model name | data set size | model architecture | enzyme sequence representation | substrate representation | other features | performance |
|---|---|---|---|---|---|---|
| UniKP | 910 | ExtraTrees | ProtT5-XL | pretrained SMILES transformer | temp, pH | R 2 = 0.65 |
| EITLEM-Kinetics | 13388 | neural network | ESM1v | MACCS Keys (RDKit) | R 2 = 0.68† | |
| CataPro | 25831 | neural network | ProtT5-XL | MolT5 and MACCS Keys | r = 0.41† | |
| CPI-Pred | 8151 | neural network | ESM-2 | molecular fingerprint (MPNN) | r = 0.39† |
Metrics without a superscript are measured after random data split for training and testing; Metrics labeled with “†” are measured after sequence-aware data splits for training and testing. R 2 is shown as a primary metric to compare the performance of models. When R 2 values were not reported, PCC was used instead.
4. Characteristics and Performance of ML Models for Predicting K i .
| model name | data set size | model architecture | enzyme sequence representation | substrate representation | other features | performance |
|---|---|---|---|---|---|---|
| CatPred | 11929 | neural network ensemble | D-MPNN | enzyme structure (E-GNN), sequence attention | R 2 = 0.45† | |
| OmniESI | 11929 | neural network | ESM-2 | molecular graph (GCN) | R 2 = 0.54† | |
| SAKPE | 10841 | gradient boosting | ESM-C | Mole-BERT | catalytic and substrate binding sites from EasIFA | R 2 = 0.36† |
| CPI-Pred | 4341 | neural network | ESM-2 | molecular fingerprint (MPNN) | r = 0.66† |
Metrics without a superscript are measured after random data split for training and testing; Metrics labeled with “†” are measured after sequence-aware data splits for training and testing. R 2 is shown as a primary metric to compare the performance of models. When R 2 values were not reported, PCC was used instead.
4.1. Models for k cat Prediction
4.1.1. Early Approaches (Before 2023)
One of the first ML models to predict k cat was developed by Heckmann and colleagues in 2018. They employed a random forest model to predict in vitro k cat values for various enzyme reactions in Escherichia coli by leveraging diverse structural and biochemical features. With available k cat values for enzymatic reactions in E. coli accounting for only 10% of all catalytic reactions, the training data consisted of 172 k cat values for various endogenous E. coli enzymes and the model achieved a coefficient of determination R 2 = 0.34 on an independent test set. The authors found that the most important feature for predicting k cat was the reaction flux calculated by parsimonious flux balance analyses. Nevertheless, the model’s applicability was limited as the input features used are available for only a small subset of enzymatic reactions from well-studied model organisms such as E. coli, S. cerevisiae, and H. sapiens.
In 2022, DLKcat, an organism-independent DL model that requires more accessible inputs, was developed by Li and colleagues. The model solely relies on the enzyme’s amino acid sequence and one of the reaction substrates to make its prediction across the space of all possible enzymatic reactions. The model predicted logarithmic k cat values that were, on average, within 1 order of magnitude from the experimental values and achieved an R 2 = 0.44 on a random subset of the full data set used as a test set. , However, the data set contains several entries for wild type enzymes and their mutants paired with the same substrate as well as the same enzyme sequence paired with different substrates. Since the data was randomly split without accounting for shared sequence identity between test and training sequences, the model suffered from data leakage between sets where 67.9% of the enzyme sequences in the test set were also present in the training set and 90% shared >99% sequence identity with those in the training set. Accordingly, DLKcat struggles to generalize to unseen sequences that share <60% identity with sequences in the training set: it achieves a negative R 2 value, implying that the prediction for a given enzyme–substrate pair is worse than using the average k cat value of all reactions.
4.1.2. Dealing with Data Leakage
The data leakage observed in DLKcat can be avoided by splitting the data sets into training and test data in a way where enzymes with identical or highly similar amino acid sequences would not occur in both training and test sets (Figure a). One of the first models to do this was TurNuP, published by Kroll and colleagues in 2023. Its key innovation was combining numerical reaction fingerprints and pLM embeddings of enzyme sequences to predict k cat for wild-type enzymes. Though trained on a smaller data set (n = 4271), TurNuP achieved better performance compared to DLKcat (n = 16838) on a test set with sequences dissimilar to training sequences (R 2 = 0.44), including an R 2 = 0.33 for sequences with <40% sequence identity to training sequences. Unlike the model published by Heckmann and colleagues (2018), reaction fluxes did not improve the performance of TurNuP. , Similar to the sequence-aware split used in TurNuP, Wang and colleagues (2025) trained a model called CataPro on 10-fold cross-validation data sets created by clustering their k cat entries such that no two sequences in the same cluster share more than 40% identity. The model achieved a Pearson’s correlation coefficient (PCC) r = 0.48 at low sequence identities between training and test sequences, further confirming that generalization to unseen enzymes critically depends on leakage-free splits.
3.
Strategies for improving predictive performance of ML enzyme kinetic parameter models. (a) Data splitting approaches to mitigate data leakage by controlling sequence identity between training and test sets, (b) addressing class imbalance in kinetic parameter distributions through reweighting schemes, (c) incorporation of interaction-aware features such as attention blocks to capture dependencies between enzymes and substrates, and (d) an uncertainty-aware modeling framework that distinguishes between aleatoric and epistemic uncertainty.
4.1.3. Handling Data Imbalances
One limitation of TurNuP was its tendency to overestimate very low and underestimate very high k cat values, which could be attributed to imbalances and noise in the data. The data imbalance issue was addressed by Yu and colleagues (2023) in their model UniKP using reweighting methods. The model achieved a decent performance owing to the use of an ensemble model which reduces variance, making the model more robust against the training data. However, upon analyzing the k cat value distribution of the data set, the authors found that the absolute error is larger at both ends compared to the entries in the middle. Hence, they used class-balanced reweighting to improve the model’s ability to predict high values of k cat by reducing the relative weight of the densely populated middle range of k cat values. The reweighting was motivated by the argument that additional data points have diminishing returns to their corresponding class due to the overlap of information in the data set (Figure b). The RMSE of the reweighted model for high k cat entries was 6.5% lower than that of the unweighted model.
4.1.4. Accounting for Environmental Factors
The noise in the data set could originate from the fact that the values of k cat in the data set were measured under different experimental conditions such as temperature and pH. These environmental factors were not used as inputs to the previous models mostly due to their unavailability in existing data sets. Accordingly, Yu and colleagues (2023) created Revised UniKP, a retrained UniKP on a smaller data set containing either the temperature or pH at which the k cat was measured, after retrieving the experimental conditions from UniProt. UniKP and Revised UniKP acted as base layers, whose outputs were fed to a linear regression layer to yield a final k cat prediction at the pH or temperature of interest. This two-layer framework was called EF-UniKP and achieved an R 2 = 0.45 on the pH data set and an R 2 = 0.31 on the temperature data set when evaluated on a strict test set where either the enzyme or the substrate was not included in the training set.
While EF-UniKP did not assess the feature importance of environmental factors on their model, it provided a more realistic prediction of k cat as it strongly depends on temperature. , Another model that accounted for the temperature effects was DLTKcat. The model incorporated a bidirectional attention block to depict the interactions between substrate atoms and enzyme residues by computing attention weights in atom-to-residue and residue-to-atom directions as opposed to simply concatenating the vectors. Temperature values were added to the weighted vectors and fed into a set of dense layers to generate a k cat prediction. Even though the model outperformed EF-UniKP achieving an R 2 = 0.66, it suffers from data leakage due to oversampling entries with low (<20 °C) and high (>40 °C) temperatures during data preparation. Since the test set was randomly chosen, some entries were susceptible to having identical matches seen during training, leading to the inflation of the performance metrics for DLTKcat. A more robust model to predict temperature-dependent k cat is PreTKcat. PreTKcat uses an ExtraTrees ensemble model that achieved an R 2 = 0.69 on a 10-fold test set generated by a random split, a 2.98% improvement over UniKP. The authors reported that adding pH did not yield any significant improvement to the PreTKcat’s performance.
4.1.5. Using Interaction-Aware Features
Akin to DLTKcat, Du and colleagues (2025) argued that simply concatenating the substrate and sequence encoding vectors limits the ability to capture the complex feature interactions between the substrates and the enzymes (Figure c). Hence, they developed GELKcat that adopts a gate network that assigns weights to the enzyme and substrate encodings before fusing the vectors together and passing them to a fully connected neural network. Ablation studies showed that the gate network had a 3.92% improvement in R 2 compared to the model without it. Despite the gate network’s role, GELKcat did not outperform ML models that simply concatenated their feature vectors like UniKP and PreTKcat, likely due to their use of n-grams to represent enzyme sequences rather than pLM embeddings. In lieu of the attention blocks and gate networks, Qiu and colleagues (2025) opted to directly incorporate important sites for enzymes (e.g., catalytic and substrate binding sites) as features in their model SAKPE. This was performed using EasIFA, an algorithm that integrates pLMs with structural encodings to determine and assign weights to important residues in the enzyme sequence. Ablation studies showed that additional site features accounted for a 0.02 increase in the model’s R 2, and the model outperformed DLKcat and UniKP.
More recently, OmniESI was introduced as a framework that predicts enzyme–substrate interactions using a conditional DL approach. Unlike the previous models that use concatenated static embeddings for their feature vectors, OmniESI relies on novel feature modulation strategies, bidirectional conditional feature modulation (BCFM) and catalysis-aware conditional feature modulation (CCFM), to refine the joint enzyme–substrate representation toward a catalysis-relevant latent space. Embeddings for enzyme sequences and substrates are fed to the BCFM where the enzyme acts as a condition for the substrate-side network, and the substrate acts as a condition for the enzyme-side network. Then, the modulated features enter the CCFM in parallel where their joint representation acts as a condition to reweight the features based on fine-grained biochemical dependencies. Ablation studies showed that these two modules increase OmniESI’s R 2 from 0.57 to 0.64.
4.1.6. Incorporating Enzyme Structural Features
To a large extent, enzyme function is governed by its 3D-structure, which none of the aforementioned models considered. Structure-based ML models have been widely used in other areas of protein science such as predicting protein function and ligand binding sites, − yet few efforts have extended this approach to enzymatic kinetics. Wang and colleagues (2024) developed DeepEnzyme that, in addition to protein sequence and substrate structure, leverages protein structural features to make predictions for k cat. Since a large portion of kinetically characterized enzymes lack an experimentally determined 3D-structure, the authors used ColabFold to predict the structures of all enzymes in their data set. The model achieved an R 2 = 0.58 on its test set and maintained an impressive R 2 = 0.42 when the test sequences shared <50% sequence identity with the training sequences. The authors attribute this remarkable performance to the features learnt from the 3D-structures claiming that they closely correlate with function. Another model that accounts for structural features is KcatNet. It also incorporates an attention mechanism to capture interactions between the enzyme and the substrates in the feature encodings. The model outperformed UniKP by 18% on the same data set. Even though these models were tested for data leakage by considering sequence identity between test and training sequences, data leakage due to the structural similarity embedded in the contact maps was not considered. For example, two homologous enzymes can display highly conserved 3D-structures despite sharing low sequence similarity from diverging evolutionary paths. ,
4.1.7. Model Stability under High Dimensional pLMs
Some models may overfit the training data and appear deceptively strong due to the mismatch between the data size and the dimensions of the feature vectors. Here, we highlight strategies for improving pLM-based representations, either by selecting more informative layers or by reducing the dimensionality of the embeddings to mitigate overfitting. Alwer and Fleming (2025) built KinForm to address common issues in using pLMs to represent enzyme sequences. Since catalytic activity is often dictated by a handful of residues, uniform pooling of the pLM outputs might dilute relevant signals. Moreover, most of the models discussed thus far use the last transformer layer of the pLM which might not be the most informative for kinetic parameter predictions. Instead, the residue-level information was extracted from an intermediate layer of pLMs (e.g., ESMC and ESM-2) in KinForm. Moreover, the authors substituted uniform pooling with binding-site-weighted pooling to represent the more substrate binding relevant embeddings and concatenated them with the substrate representation vector. The substrate binding site weights were calculated separately using the Pseq2Sites model. To address the high dimensionality of the pLM vectors, principal component analysis (PCA) was used to reduce the vectors to 200–400 components which stabilized the model against overfitting. It is worth mentioning that KinForm outperformed UniKP, achieving R 2 roughly twice as high, especially in the low-similarity sequence bins.
4.1.8. Uncertainty-Aware Models
All of the previous models are deterministic in nature, meaning that they use traditional regression approaches to output a single value k cat prediction. Recently, Boorla and Maranas (2025) developed CatPred, a DL model that utilizes probabilistic regression to add a confidence metric to the k cat predictions by estimating the associated uncertainties. They break down predictive uncertainties into aleatoric and epistemic uncertainty. The former is associated with any intrinsic noise in the training data while the latter results from the scarcity of training samples in a certain region of the latent space (Figure d). The authors imposed more lenient constraints during their data set preparation to end up with 23197 k cat entries denoted as CatPred-DB. The model inputs were passed into a probabilistic regressor that outputs a Gaussian distribution of a k cat predictions with a mean and a variance (Figure d). Since this variance only depicts the aleatoric uncertainty, the authors trained an ensemble of ten models under different initial weights where the average of the means is the final prediction and the variance of the means captures epistemic uncertainty. , The model achieved an R 2 = 0.61 and R 2 = 0.39 on a held-out test set (which does not contain any enzyme substrate pair used for training) and an out of distribution test set (which does not contain any enzyme that shares >99% identity to a sequence used in training) respectively. Moreover, 76% of the predicted k cat values fell within an order of magnitude from the actual value. Most of their predictions had a higher aleatoric uncertainty, which indicates high noise levels in the experimental data. Interestingly, unlike DeepEnzyme, structural features did not improve CatPred’s predictive performance. The authors attribute this finding to the fact that their pLM features encode not only sequence but also structural information. , However, another model called DEKP showed through ablation studies that structural features enhanced their R 2 by 0.08, despite using a pLM to encode enzyme sequences. Overall, CatPred outperformed models like DLKcat and UniKP on the same out of the distribution set and showed a comparable performance to TurNuP under similar settings. This is likely due to the ensemble architecture of CatPred that provides robustness to the model’s predictions.
Building on probabilistic predictions, Gollub and colleagues (2024) introduced ENKIE, a Python package based on hierarchical Bayesian multilevel models (BMMs) to predict k cat along with a calibrated confidence interval. Unlike the previous approaches that rely on enzyme sequence and substrate structures, ENKIE utilizes MetaNetX chemical identifiers for substrates and reactions, EC numbers for enzymes identifiers, and, optionally, the environmental conditions of the reaction. The model achieved an R 2 = 0.36 which is comparable to models developed by Heckmann and colleagues (2018) despite the simpler features. They also showed that the ENKIE estimates remain reliable even for unseen reactions. Lastly, Xu and colleagues (2025) integrated an error predictor that assesses the confidence level of k cat predictions into their model CPI-Pred. The error predictor calculates Euclidean distances between test enzymes and structures against all entries in CPI-Pred’s training set using the embedding vectors and classifies the prediction as low or high confidence based on a threshold of 0.2 in absolute distance to a point in the training set. It achieved a 77% accuracy.
4.1.9. Reformulating the k cat Prediction Task As a Classification Problem
Recently, RealKcat was introduced as a novel ML framework to predict k cat using sequence and substrate embeddings. Unlike the previous models, the prediction task was treated as a classification problem clustering k cat values into biologically meaningful order-of-magnitude bins. This approach aligns with industrial practices where the concern is largely about the order of magnitude of enzymatic activity rather than the exact numerical value. , Due to the class imbalance caused by the lack of data on inefficient or inactive enzymes, the authors resort to a synthetic minority oversampling technique (SMOTE) in which the active sites of available sequences were substituted with alanine, creating a “zero activity” class. The model achieved an accuracy of 89% with 95% of the predictions falling within 1 order of magnitude from their true value.
4.2. Models for K m Prediction
Kroll and colleagues (2021) initiated the first endeavor to create an ML model to predict K m on a large-scale data set. The model achieved an R 2 = 0.53 on the test set from BRENDA and an R 2 = 0.49 on an independent test set obtained from SABIO-RK. Moreover, it was robust against data leakage where the performance dropped minimally to R 2 = 0.45 when either the substrate or the enzyme in the test set was not present in the training data and R 2 = 0.26 when neither of them existed. Overall, the reported relative prediction error of the model was 4.1 on average, implying that the predictions deviate from the experimental values by around 4-folds.
Meanwhile, Maeda and colleagues (2022) integrated ML into the traditional global optimization approach that optimizes kinetic model parameters to best fit experimental data. Since this traditional approach is time-consuming and often leads to biologically unreasonable solutions, they developed an ML-assisted global optimization approach (MLAGO) for K m estimation. Similar to Kroll’s model, there was only a 4-fold difference between the measured K m values and those predicted by MLAGO despite using simplified inputs (e.g., one-hot encodings). However, these predictions resulted in a high badness-of-fit to the experimental data (BOF > 0.02) when used for the carbon and nitrogen metabolism models. Accordingly, the predicted K m values were utilized as a reference in the global optimization parameter estimation, and subsequently converged to values that fit the experimental data better (BOF < 0.02). In terms of limitations, MLAGO has low predictive power for extremely small or large K m values, likely due to data imbalance.
More recently, He and Yan (2024) trained GraphKM on 19,754 sequences using a gradient boosting algorithm to predict K m. The model achieved an r = 0.59 outperforming the K m model developed by Kroll and colleagues (2021) with an r = 0.23. Meanwhile, GraphKM’s performance against MLAGO was comparable on another independent test set. One downside of this model is that it relies on the maximum value of K m reported in the literature for a given sequence rather than using the geometric mean of all observations. Hence, it runs the risk of consistently overestimating K m and skewing the model’s predictions, especially if the maximum value reflected an outlier, error, or nonstandard assay conditions. Moreover, during data preparation, all sequences longer than 1000 amino acids were excluded, increasing the likelihood of the model generalizing poorly on long enzyme sequences. Finally, the data was split randomly even though 43% of their data set is mutant enzymes. Since the authors did not account for sequence identity between training and test sequences, the performance of GraphKM might be inflated and biased toward their specific data split.
Other models covered in Section can also make predictions on K m with the same feature representations and model architectures. Their performances are shown in Table and S5 as well. For CatPred, it is worth mentioning that substrate features were solely able to explain 46.5% of the variance in the data when predicting K m compared to 26% when predicting k cat. This is expected as K m is related to binding affinity, which depends on how the substrate physically and chemically interacts with the enzyme. Meanwhile, k cat is often governed more by the enzyme’s catalytic residues, transition states, and conformational flexibility rather than the substrate structure.
Wang and colleagues (2024) created MPEK, a multitask DL framework based on pLMs that predicts k cat and K m simultaneously, capturing the intrinsic connection between them due to its unique architecture. MPEK has a customized gate control framework composed of two expert layers (for k cat and K m) and one shared expertise layer responsible for capturing this connection. The multitask learning in MPEK showed a 2.4% and 4.5% improvement in R 2 compared to single learning when predicting k cat and K m, respectively.
Most of these models rely on enzyme–substrate pairs as inputs. However, akin to what was applied in TurNuP for k cat prediction, Li and Wang (2025) considered the entire reaction, including both the substrates and products for K m predictions, even though K m is not directly related to the products. They trained DLERKm, a neural network with an attention channel on a database retrieved from SABIO-RK and UniProt that achieved R 2 = 0.59. Their model outperformed UniKP by 14.9%.
4.3. Models for k cat/K m Prediction
k cat/K m serves as a fundamental kinetic parameter to measure the catalytic efficiency. However, ML models for k cat/K m have been rare compared to those for k cat and K m (Table ) mostly due to the lack of large available data sets for training. The first effort to integrate the task of k cat/K m prediction to an ML model was in UniKP. Yu and colleagues (2023) retrieved a small data set of 910 entries to train and validate their model achieving an R 2 = 0.65 and an r = 0.81 with the experimental results. They compared the results to simply dividing the predicted k cat and K m values from their individual models (Tables and ) which achieved a low r = −0.02 with the experimental results. This is because taking the ratio of the independent predictions compounds the error associated with each model and ignores any correlation or codependency between both parameters. Hence, this stresses the need for a unified and task-specific model to predict k cat/K m. While UniKP demonstrates a good performance when predicting k cat/K m, it was not tested on any out-of-distribution sequences.
More recently, Shen and colleagues (2024) presented EITLEM-Kinetics, a novel approach for predicting k cat, K m, and k cat/K m for both wild-type and mutant enzymes based on their sequences and one of the substrates involved in the reaction. They fine-tuned the pLM used to encode enzyme sequences by training attention blocks. This helped EITLEM-Kinetics learn which task-specific information is essential to be stored in a single protein representation rather than storing general protein information. Initially, the performance of the k cat/K m neural network ensemble was significantly lower than that of the k cat and K m models. Accordingly, the authors enhanced the performance of the k cat/K m model by leveraging transfer learning. To elaborate, since the DL models for k cat, K m, and k cat/K m are related, it should be possible to transfer information and capabilities between the different prediction tasks. Moreover, training a model directly on k cat/K m ignores the intricate relationships among the three parameters. Therefore, the authors iteratively readjusted the weight parameters of the k cat and K m networks based on the k cat/K m model, retrained and fine-tuned the networks on their respective data sets, and then fed the new parameters into the k cat/K m model again. After eight iterations, the performance of the k cat/K m model increased from 0.61 to 0.83. Likewise, the performance of the k cat and K m models increased to 0.75 and 0.72 respectively. To ensure that the enhancement was not due to data leakage, the ensemble model was evaluated on a stricter data set excluding samples that overlap with the training set. Its performance for k cat/K m still increased from 0.52 to 0.68. A likely reason iterative transfer learning improves model performance is that it provides a well-informed initial point for optimization. When the model is pretrained on a larger data set of enzymes, its parameters are guided toward a relevant region of the broad sequence space. Subsequent fine-tuning allows the model to gradually refine the parameters, making it less likely to be confined to a local minimum. Similarly, CataPro achieved an r = 0.41 when using transfer learning from their k cat and K m models to predict k cat/K m.
4.4. Models for K i Prediction
Relatively few ML models have been developed for K i prediction for enzyme–inhibitor pairs whereas most ML research has focused on half-maximal inhibitory concentrations (IC50) and drug-target binding affinities (DTBA). Unlike IC50, which depends not only on inhibitor binding but also on downstream biological responses, K i is a direct measure of enzyme–inhibitor binding affinity under different modes of action. By contrast, DTBA predictions typically address a wide set of drug–protein interactions, many of which are not enzymes, and thus offer limited insight into enzyme-specific inhibition. This distinction underscores the added value of ML approaches for K i prediction. Four of the models discussed earlier have a K i module: CatPred, SAKPE, CPI-Pred and OmniESI.
Similar to the K m model, substrate features alone in CatPred could explain 52.5% of the variance. It is noteworthy to mention that the performance of CatPred deteriorated upon adding pLM features when evaluating an out of distribution K i test set. This is a sign that the model is overfitting under high-dimensional pLM feature vectors due to the smaller size of the K i data set. Accordingly, the authors replaced the pLM features with structural features that are lower in dimensions to achieve a good performance across both held out and out-of-distribution sets. When OmniESI was trained on the CatPred data set, it outperformed CatPred on all out-of-distribution test sets, surpassing its R2 by ∼11% at a < 40% sequence similarity cutoff. As for the K i prediction task using SAKPE, the model achieved an R 2 = 0.77 on its test set. However, it shows signs of data leakage as the R 2 dropped to 0.36 and 0.29 when the test sequences shared <99% and <40% identity with the training sequences, respectively. In contrast, CPI-Pred was more robust as the PCC decreased from 0.76 on a randomly split data set to 0.61 when the test sequences shared <60% identity with the training sequences.
5. Applications of Global Models
5.1. Predicting Effects of Mutations
The catalytic efficiency of natural enzymes often falls short of the requirements needed in industrial processes. Therefore, optimizing enzymatic activity becomes essential to reducing production and operation costs. Kinetic parameters of enzymatic reactions can be mapped to enzyme sequence by a conceptual landscape that is navigated through mutational walks. The effect of mutations on function is not additive, so the ability of a model to predict the impact of multiple mutations on enzymatic activity is a direct indicator of whether it understands residue–residue interactions and how they relate to the target kinetic parameters. DLKcat was able to capture the effects of amino acid substitutions on k cat values by assessing their attention weights using a neural attention mechanism and achieved an r = 0.78 on the mutants in the test set. However, it has been proven that the model displayed inflated performance, as it recorded an R 2 = −0.18 when tested against sequences not present but still sharing more than 99% similarity with the training sequences. The authors of DLTKcat reported that the beneficial mutations for the enhancement of k cat values were distributed near high peaks of attention weights from their GAN. However, since their model also suffers from data leakage, it is probable that the residues with high attention weights correspond to those of identical sequences in the training set. Furthermore, some high attention sites were observed at residues where no mutations were introduced, highlighting noise in the data.
More robust models have demonstrated the ability to capture trends in kinetic parameter changes across enzyme mutants, underlining the varying degrees of success in predicting the effects of mutations. For instance, DeepEnzyme predicted a median k cat value for highly active phosphate alkaline phosphatase mutants that was 15% higher than that of the low activity mutants. MPEK split their mutant data into wild-type-like, increased, and decreased k cat or K m categories. The model achieved PCC values for the prediction of mutant classes ranging between 0.8 and 0.9 for all categories across both parameters. Meanwhile, the authors of CataPro showed that their model is capable of ranking mutants based on the favorability of the mutations for a given substrate reaction (r > 0.7). However, it failed to capture the effect of these mutations quantitatively when a comprehensive evaluation was performed across the entire data set (r ≈ 0). Similarly, OmniESI offered a qualitative demonstration of their model’s strength against mutants by classifying single and double mutations in β-lactamase as beneficial or detrimental, achieving >85% accuracy against numerous substrates. EITLEM-Kinetics demonstrated a consistent predictive performance across most mutants with varying numbers of mutations, achieving an R 2 = 0.85 for up to six mutations on the k cat data set. It also achieved an R 2 = 0.66 on mutants that showed more than a 10-fold increase in k cat. The model’s high performance in predicting the effect of mutations could be attributed to the fact that the features of each amino acid in EITLEM-Kinetics were represented individually rather than collectively by feature pooling. This, along with the attention networks used, makes the model more sensitive to mutations by capturing both enzyme-specific residue effects and broader sequence–substrate interactions.
5.2. Enzyme Engineering and Mining
Engineering mutant enzymes with enhanced activity for specific biochemical reactions is a pivotal and typical objective in the fields of protein engineering and synthetic biology. However, identifying effective evolutionary pathways demands an exceptional understanding of reaction mechanisms to navigate the sequence space under biological and physical constraints, such as protein folding and expression. Likewise, resorting to laboratory directed evolution is costly, time-consuming, and labor-intensive and often yields only marginal success. For example, the directed evolution of a tyrosine ammonia lyase (TAL) library from Rhodotorula glutinis through the construction and screening of 4,800 mutants led to the identification of one variant with a k cat of 142 s–1, only a slight improvement from the wild type with a k cat of 114 s–1. To address the limited success of the experimental approach, the top 1000 homologues of the wild-type sequence were identified through a BLAST search and UniKP was used to predict their k cat values for in silico enzyme mining. The top five predictions were experimentally validated, in which two sequences surpassed the wild type k cat value by ∼4-fold. In parallel, UniKP was utilized to predict k cat/K m values for all possible single-point variants of TAL from R. glutinis for in silico enzyme evolution. Ultimately, the authors identified and experimentally characterized two mutants that were up to 3.5-fold more efficient than the wild type. This demonstrates that UniKP learnt deep-level information about the enzymes, as sequences from the TAL mutant library shared <35% identity with the training sequences. Likewise, KcatNet was used for in silico evolution of α-glucosidase by screening all of its possible single-point mutants, the highest showing a 47% improvement in k cat over the wild type. Lastly, CataPro was leveraged in the enzyme mining of more efficient carotenoid cleavage oxygenases. The authors started with the carotenoid cleavage oxygenase from Caulobacter segnis (CSO2) and identified 1500 homologues using BLAST. Experimental validation confirmed that the top hit among the CataPro predictions, Sphingobium sp. CSO (SsCSO), was 19.53 times more active than CSO2. Also, two rounds of in silico directed evolution on SsCSO using CataPro led to the identification of a double-point mutant that showed a 65-fold increase in activity compared to CSO2. However, it is critical to note that these models are more reliable for identifying relative trends, such as which mutants are likely to be more active, rather than for predicting the precise magnitude of mutation effects.
While these ML models have shown promise in predicting kinetic parameters for enzyme mutants, they are not specifically designed to suggest mutations that enhance the enzymatic activity. Yu and colleagues (2024) addressed this limitation by building a diffusion model that proposes multiple amino acid substitutions to optimize activity. They formulate this objective as an inverse folding task combined with a regressor-guided diffusion model denoted as k catDiffuser. In other words, k catDiffuser generates several enzyme sequences compatible with a given backbone structure while being guided by a sampling process favoring amino acid combinations that lead to higher k cat values. It was trained on 15603 protein structures of BRENDA and CATH database entries generated using ESMFold. k catDiffuser outperformed other diffusion models such as ProteinMPNN, PiFold, and GraDe-IF, generating mutants with an overall improvement of 0.21 in log k cat. For example, k catDiffuser enhanced log k cat for an undecaprenyl pyrophosphate synthetase activity by 0.486 while the aforementioned models generated less active mutants when prompted with the same task. The generated structures align well with the original enzyme structure, highlighting k catDiffuser’s advantage in enhancing k cat without compromising structural integrity.
Despite these successes, most of these models show limited interpolation of enzyme families that are not well represented in their training sets. For example, when tasked with predicting k cat/K m for 260 β-glucosidases, UniKP merely predicted the same value for all enzyme sequences (R 2 = 0.05) while EITLEM-Kinetics and CataPro achieved R 2 values of 0.19 and 0.27, respectively. Similarly, when DLKcat was tasked with predicting k cat values for 175 adenylate kinases, it achieved a Spearman’s coefficient of −0.09. These outcomes are well below the performance reported on the models’ respective test sets, and such inaccurate models can significantly hinder sequence exploration for enzyme mining and evolution, especially when targeting specific catalytic function. Accordingly, more realistic metrics are needed to assess whether ML models meaningfully support enzyme screening. Taken literally, regression metrics such as R2 and RMSE can lead to incorrect ranking of variants or failing to identify the true top performers. For applications such as enzyme mining or mutational scanning, metrics that evaluate enrichment in the top-performing region of the sequence space such as the enrichment factor (EF), precision@k, and recall@k offer a more actionable assessment, , although they have primarily been used for evaluating functional fitness rather than predicting kinetic parameters. Notably, models with high R 2 can exhibit poor ranking fidelity or enrichment, rendering them less effective for enzyme engineering workflows. This is evident for models like UniKP where enzyme mining and evolution each resulted in only modest improvements (<3-fold) in k cat/K m.
Lastly, evolving an enzyme for enhanced catalytic activity at a desired temperature is not feasible with current ML models, and experimental approaches involve timely and labor-intensive characterization steps. Recently, Erkanli and colleagues (2025) introduced a three-module ML framework that predicts the optimal temperature, k cat/K m, and the activity–temperature profile of β-glucosidases. By integrating both intrinsic sequence information and extrinsic temperature effects, the framework provides a powerful tool to traverse the sequence–temperature–activity landscape. Looking forward, similar models have the potential to accelerate enzyme evolution by guiding searches through vast sequence and temperature spaces simultaneously, uncovering novel functional mutants that operate under desired reaction conditions.
5.3. Genome-Scale Metabolic Modeling
Genome-scale metabolic models (GEMs) are mathematical representations of the complete set of metabolic reactions within an organism reconstructed from annotated genome sequences. They are advantageous in simulating the metabolic fluxes under different conditions, guiding metabolic engineering endeavors, and studying proteome allocation. , Typically, GEMs are built on stoichiometric constraints derived from reaction networks and mass balance principles to estimate feasible reaction fluxes using methods such as flux balance analysis. However, their accuracy is limited by a key assumption: enzymes are treated as infinitely fast catalysts or enzymes are present in excess quantities. To address this, enzyme-constrained genome-scale metabolic models (ecGEMs) integrate enzyme capacity constraints, most commonly through k cat values and enzyme abundances, thus linking the maximum achievable flux to the catalytic efficiency of the enzyme catalyzing the metabolic reaction. Despite their promise, ecGEMs remain hindered by incomplete or noisy kinetic data as many enzymes lack experimentally measured k cat values. Moreover, available data often come from different organisms, assay conditions, and substrates, adding additional uncertainty. While ecGEMs have been developed for several well-studied organisms such as E. coli, only 10% of all enzymatic reactions have fully matched k cat values in BRENDA, respectively. , In addition, in vivo estimation of k cat values from in vitro experimentation is difficult, especially due to their noncorrelative relationship. , Without such data, ecGEMs usually adopt kinetic parameter values from similar substrates, reactions, or organisms, increasing the deviation of these models from experimental observations.
A way to mitigate the kinetic data bottleneck in ecGEMs construction is to use predicted k cat from the ML models discussed to expand the coverage in genome-scale reconstructions. The model developed by Heckmann and colleagues (2018) aimed to parametrize GEMs for E. coli iML1515. The integration of ML-derived k cat values instead of median-imputed ones from available data sets resulted in a substantial improvement (i.e., reducing the RMSE for their model by 34%). One limitation of this approach is that the ML-predicted k cat values are derived from in vitro rather than in vivo conditions. Similarly, DLKcat was trained on in vitro data, implying that the model would predict in vitro values as well. DLKcat was used to reconstruct the ecGEMs of 343 yeast/fungi species by predicting k cat values for around three million enzyme–substrate pairs. To resolve the discrepancies between the in vitro-based predictions and in vivo values and ensure the biological relevancy of the predictions, the authors resorted to a Bayesian genome-scale modeling approach in which the DLKcat predictions serve as the mean and the model’s RMSE as the variance for prior k cat distributions. Then, these values are updated iteratively based on experimentally measured phenotype data sets to produce posterior distributions that capture the uncertainty in the k cat prediction. Overall, the DLKcat-based ecGEM’s RMSE was 30% lower than that of the original ecGCM. The authors of KcatNet parametrized the same ecGEM using their model and outperformed DLKcat in reducing the original model’s RMSE in 16 out of 22 carbon sources and oxygen conditions for four different yeast species.
Meanwhile, DLTKcat was used to demonstrate how ML models can be used for temperature-sensitive metabolic modeling. The authors showed that DLTKcat predicted a decrease in the activity of catabolism in Lactococcus lactis MG1363 in response to a temperature increase, which is consistent with experimental observations. Moreover, it captured the increase in k cat values for enzymes from Streptococcus thermophilus LMG18311 when temperature increased. However, when quantitatively estimating the growth rates, the model’s accuracy was extremely low, likely due to error propagation.
Beyond steady state ecGEMs, dynamic kinetic models are governed by full enzymatic rate laws, relying directly on all of the kinetic parameters and equilibrium constants. These models enable time-resolved simulations of metabolite dynamics and transient behavior that cannot be captured under a steady-state assumption. Kinetic models have been limited to small subsystems due to data scarcity, but recent advances have produced near-genome-scale kinetic reconstructions that integrate omics data, thermodynamic constraints, and parameter estimation frameworks to infer large sets of kinetic constants. Notable examples include large-scale kinetic models of E. coli metabolism, , S. cerevisiae metabolism, , and even human cancers. These kinetic models demonstrate the growing feasibility of large-scale parametrized ordinary differentia equation systems and highlight that ML-predicted k cat, K m, and K i could play an increasingly central role in enabling genome-scale dynamic simulations.
To show how ML models can be used to fill in the gaps for missing K m values in kinetic models, Kroll and colleagues (2021) leveraged their model to predict K m for enzymes associated with 47 GEMs spanning E. coli, S. cerevisiae, M. musculus, and H. sapiens. While these organisms belong to different biological domains, their training data was dominated by bacterial data. Accordingly, they partitioned their test set by domain to illustrate that their model can make K m predictions equally well for different domains, achieving R 2 values of 0.37, 0.51, and 0.56 for archaea, bacteria, and eukarya, respectively. Lastly, MLAGO was tasked with parametrizing the carbon and nitrogen metabolism models as the reactions’ corresponding K m values were not included in the training set. While the K m predictions achieved good accuracy (RMSE of 0.62 and 0.73 for carbon and nitrogen metabolisms, respectively), the ML-based metabolic models did not fit the experimental metabolic data with sufficient accuracy. When these K m predictions were used as a reference for the MLAGO approach, the resulting metabolic models fit the experimental data well while still maintaining reasonable accuracy for the K m predictions (RMSE of 0.79 and 0.57 for carbon and nitrogen metabolisms, respectively).
6. Local Models
While many recent ML models aim to generalize parameter predictions across diverse enzyme families, one of the first models was developed for an enzyme family specific prediction task. Yan and colleagues (2012) aimed to predict K m for β-glucosidases against their natural substrate, cellobiose, due to their importance in the biofuel industry. The model was constructed using a feedforward backpropagation neural network that takes the amino acid probability distributions as well as 11 properties from the AAIndex as an input. The network was trained on 24 β-glucosidase sequences and tested on another 12 achieving an R 2 = 0.67. However, this model suffers from overfitting as the choice of a neural network is inadequate with the extremely small size of the available data set. In 2016, Carlin and colleagues trained an ensemble of elastic net regressors on 100 mutants of the β-glucosidase from Paenibacillus polymyxa. They showed the ensemble approach was more robust than a single regressor, with PCC increases from 0.57 to 0.76 for k cat/K m, 0.43 to 0.6 for k cat, and 0.29 to 0.71 for 1/K m. The most important features were the hydrogen bonding energy of the substrate, polar contacts between the enzyme and substrate, and a minimal number of voids in enzymes for k cat/K m, k cat, and K m, respectively. Several features were predictive for k cat/K m but not for k cat or K m; this is likely because there was no significant correlation between k cat and K m in the authors’ data set, making all the parameters independent. One of the limitations of this model was its bias toward low k cat/K m values due to their over-representation in the training data set.
Local models with broader coverage have also been reported. Li and colleagues (2023) built a DL platform, DeepGH, for catalytic activity of glycoside hydrolases and applied this model to identify mutants with enhanced activity. DeepGH was trained on 64057 sequences retrieved from the CAZy database, spanning 119 glycoside hydrolase families and sharing at most 65% sequence similarity between training and test sets to avoid data leakage. The model takes a feature matrix containing information about the positions of the amino acids and their catalytic importance as an input. DeepGH was applied to chitosanase CHIS1754, identifying nine residues as target sites for mutations that could potentially enhance its activity. Experimental validation showed that eight of the nine single-point mutants were more active than the wild type. They also created the CHIS1754-MUT7 variant which includes seven of the nine suggested mutations from DeepGH, exhibiting a k cat/K m that is 24-fold higher than that of the wild type.
Shifting focus to other enzyme families, Muir and colleagues (2024) created a model focusing on adenylate kinases (ADK). Leveraging a high-throughput microfluidic platform, they measured k cat, K m, and k cat/K m for 193 ADK orthologs and demonstrated that the ADK functional landscape is rugged and multipeaked, with the type of LID-domain – one of three domains of ADK – rather than phylogeny or reaction temperature playing the most critical role. Moreover, they fed ∼5000 ADK sequences to ESM-2 and observed that the outputs could be clustered by the LID-domain type as well, albeit not by activity. Taking advantage of pLM’s ability to capture high-level structural organization, the authors trained a random forest regressor on the ESM-2 embeddings of ADK sequences. This model achieved a Spearman’s rank correlation coefficient r = 0.44 on k cat compared to r = −0.09 when they used DLKcat for the same test set. Furthermore, a support vector regressor was trained for K m achieving r = 0.49. Despite using fewer sequences, this local model outperformed DL models trained on large data sets, underscoring the value of high-quality kinetic measurements within relatively narrow sequence spaces for building robust family specific predictors. It also highlights the limitations of the global models when trained on noisy and heterogeneous public data sets like BRENDA.
Meanwhile, Ao and colleagues (2023) incorporated structural features for their transaminase-specific model. They trained a gradient boosting regression tree on data-specific activity data from E. coli transaminase and 40 of its variants against 28 different substrates. They claimed that the enzyme sequences and structures contained a lot of information that was not directly relevant to the activity prediction task, depressing the performance of the ML model. Hence, they used only the electronic and steric properties of 14 amino acids at the mutation sites and 8 substrate descriptors as features. Their model achieved an R 2 = 0.80 for activity prediction, with three amino acids showing the highest feature importance. Using this model, more active mutants were identified with up to 3.5 fold increase in activity compared to previously identified variants.
7. Limitations and Future Directions
7.1. The Dilemma of Limited Data
Despite the value of the existing databases, the sparsity and uneven distribution of kinetic measurements across enzyme classes remain a major challenge, as evidenced by our analysis of BRENDA (Figure ). The majority of the characterized k cat and K m values belong to a small subset of hydrolases, oxidoreductases, and transferases against their natural substrates (Figure a). As a result, most ML models show strong performance in interpolating to these families but struggle to generalize to underrepresented enzyme families and non-natural substrates. Moreover, the main issue with collecting kinetic data sets from published literature is the inherent bias toward parameters with intermediate values, rendering those for inefficient or extremely efficient enzymes scarce (Figure b). This inevitably affects the ability of ML models to generalize to extreme cases. To circumvent this issue, large amounts of evenly distributed high-quality data can be obtained by automating assays through high-throughput systems like biofoundry , and microfluidic platforms. ,− Biofoundries provide a self-driven laboratory where an agent designs enzymes and deploys them into a characterization unit for synthesis, expression, and kinetic measurement. Then, the agent’s sequence-activity relationship is updated via learning cycles until it converges to an enzyme that meets the query requirements. Microfluidic platforms can complement this effort by miniaturizing and multiplexing reactions, allowing parallel measurements of kinetic parameters for thousands of enzyme–substrate combinations. Together, these approaches have the potential to generate large, high-quality kinetic data with minimal lab-to-lab variation, covering both well-studied and underrepresented enzyme classes, thereby contributing to the improvements in the performance and generalization of ML models. These automated approaches also provide a realistic pathway to practical closed-loop design–build–test–learn (DBTL) cycles for kinetic modeling. The ML models may be used to propose candidate sequences, mutations, or enzyme–substrate pairs that are predicted to improve kinetic parameters. The design outputs are then executed by automated strain construction or cell-free synthesis platforms and can be characterized via microfluidic assays or in biofoundries. Active-learning strategies implemented in self-driven laboratories prioritize experiments that reduce uncertainty in target regions and accelerate convergence to high-performance variants that can be used to retrain and fine-tune the original ML predictors. , To make these loops reliable for kinetic parameter learning, several engineering considerations are required such as rigorous uncertainty quantification, standardized assay metadata to control for environmental conditions, calibration steps to reconcile cell-free and in vitro measurements with in vivo behavior, ,, and valid stopping criteria to avoid overfitting to assay artifacts. ,,
4.
Limitations of current data sets and model types and the use of semisupervised learning to address these limitations. Data imbalance of (a) enzyme classes and (b) k cat values in BRENDA. (c) Exploration of functional landscapes using global and local models. Global models are well suited for navigating across diverse enzyme families (represented by different colors), though they provide limited accuracy within the local sequence space (within the same color). In contrast, local models offer more detailed information on a specific functional landscape but do not extend to other enzyme sequence spaces. (d) Schematic of a semisupervised learning approach leveraging labeled and unlabeled enzyme data.
7.2. The Need for Hybrid Models
Considering all of the models reviewed here, it is apparent that most of these ML approaches explore either a local or a global view of the sequence space (Figure c). This distinction reflects the scope of the search: global models aim to cover a broad and diverse range of enzyme families and classes while local models restrict their focus to specific families. Global models trained on diverse enzyme families generalize across a wide region of sequences. However, they often suffer from low accuracy when predicting parameters for sequences that are highly dissimilar from the training data. Conversely, local models trained on high-quality data sets of wild types or mutants can capture nuanced sequence-function or structure–function relationships. Nevertheless, their predictive power is confined to a narrow region of the sequence space. This is manifested by the difference in performance between the models in Sections and , which exemplifies the trade-off between precision and scope. Accordingly, a promising direction would be to develop hybrid models that leverage global protein language models for a broad contextual depiction of the sequence space while simultaneously fine-tuning on family specific kinetic data sets to retain local information.
7.3. Leveraging Semisupervised Learning
Current ML models for kinetic parameter prediction rely almost entirely on supervised learning, which requires large amounts of labeled data for training. Given the limitations discussed above, semisupervised learning offers a promising alternative by leveraging both the limited number of labeled kinetic data and the vast amounts of unlabeled enzymatic sequences in databases like UniProt (Figure d). By exploiting functional patterns in sequence space alongside a limited number of labeled examples, semisupervised frameworks can expand model applicability and improve robustness against data scarcity.
7.4. Incorporating Physics-Based Machine Learning
Another promising approach for enzyme kinetic prediction lies in physics-based ML. Unlike the purely data-driven approaches discussed in this review, physics-based ML embeds biophysical constraints into the learning process to ensure that the predictions remain consistent with the principles of enzyme catalysis. For example, relationships between the activation free energy and k cat can be incorporated as constraints during model training. This can be done by regularizing neural networks using penalty terms that enforce consistency with the transition state theory, requiring k cat predictions to fall within feasible ranges of activation energies. Moreover, coupling DL models with quantum mechanics/molecular mechanics (QM/MM) descriptors can assist in capturing molecular mechanisms that govern enzyme kinetics.
Beyond the regression-based predictive models, a recent development is the emergence of generative frameworks that explicitly incorporate biophysical constraints to ensure the mechanistic plausibility of the predicted kinetic parameters. In their 2022 work, Choudhury and colleagues introduced a conditional generative adversarial network that incorporates biophysical and physiochemical constraints to create biologically relevant kinetic models that satisfy thermodynamic requirements, stability constraints, and experimentally observed time scale limits. Their 2024 work integrates the stoichiometry, regulatory information, flux analysis, and dynamic time scale constraints into the generative process allowing for the estimation of missing kinetic parameters. Overall, these models outline a blueprint for next-generation ML approaches that integrate statistical learning with the fundamentals of enzyme engineering.
8. Conclusion
ML has rapidly emerged as a promising tool for predicting enzyme kinetic parameters. By integrating advances in pLMs and graph-based networks, we are beginning to capture the complex sequence-function relationships that govern catalytic efficiency and enzyme inhibition. These models have the potential to reshape how we explore sequence space, prioritize mutations, and accelerate enzyme discovery. Nevertheless, the field currently suffers from an urgent need for larger and more reliable data sets, which remains the major hurdle to the robustness and generalizability of current predictors. Moving forward, the next generation of models will likely require hybrid strategies that combine the breadth of global pLMs with the precision of family specific fine-tuning. Moreover, incorporating physical constraints and environmental factors will be paramount to bridging the gap between in silico predictions and experimental observations. If these challenges are met, then ML will enable the efficient design of enzymes critical in various applications ranging from sustainability to medicine.
Supplementary Material
Acknowledgments
This work was supported by The New York University Research Catalyst Grant.
The Supporting Information is available free of charge at https://pubs.acs.org/doi/10.1021/acs.jcim.5c02428.
Relevant definitions and background for machine learning and statistical terms (Tables S1–S3); Additional information on models to predict k cat and K m (Tables S4 and S5) (PDF)
A.M. and J.R.K. conceptualized the work. A.M. and D.V. curated data, prepared visualizations, and wrote the original draft. A.M. and J.R.K. reviewed and edited the manuscript. J.R.K. supervised the project.
The authors declare no competing financial interest.
References
- Wu S., Snajdrova R., Moore J. C., Baldenius K., Bornscheuer U. T.. Biocatalysis: Enzymatic Synthesis for Industrial Applications. Angew. Chem., Int. Ed. 2021;60:88–119. doi: 10.1002/anie.202006648. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Bornscheuer U. T.. et al. Engineering the third wave of biocatalysis. Nature. 2012;485:185–194. doi: 10.1038/nature11117. [DOI] [PubMed] [Google Scholar]
- Buller R., Lutz S., Kazlauskas R. J., Snajdrova R., Moore J. C., Bornscheuer U. T.. From nature to industry: Harnessing enzymes for biocatalysis. Science. 2023;382:eadh8615. doi: 10.1126/science.adh8615. [DOI] [PubMed] [Google Scholar]
- Bell E. L.. et al. Biocatalysis. Nat. Rev. Methods Primer. 2021;1:1–21. doi: 10.1038/s43586-021-00044-z. [DOI] [Google Scholar]
- Intasian P.. et al. Enzymes, In Vivo Biocatalysis, and Metabolic Engineering for Enabling a Circular Economy and Sustainability. Chem. Rev. 2021;121:10367–10451. doi: 10.1021/acs.chemrev.1c00121. [DOI] [PubMed] [Google Scholar]
- Arnold F. H.. Directed Evolution: Bringing New Chemistry to Life. Angew. Chem., Int. Ed. 2018;57:4143–4148. doi: 10.1002/anie.201708408. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Steiner K., Schwab H.. Recent Advances in Rational Approaches For Enzyme Engineering. Comput. Struct. Biotechnol. J. 2012;2:e201209010. doi: 10.5936/csbj.201209010. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Viknander, S. Deep generative models for analysis and engineering of functional proteins. Doctoral Thesis, Chalmers University of Technology, Gothenburg, Sweden, 2025. [Google Scholar]
- Axe D. D.. Estimating the Prevalence of Protein Sequences Adopting Functional Enzyme Folds. J. Mol. Biol. 2004;341:1295–1315. doi: 10.1016/j.jmb.2004.06.058. [DOI] [PubMed] [Google Scholar]
- Wang Y.. et al. Directed Evolution: Methodologies and Applications. Chem. Rev. 2021;121:12384–12444. doi: 10.1021/acs.chemrev.1c00260. [DOI] [PubMed] [Google Scholar]
- Packer M. S., Liu D. R.. Methods for the directed evolution of proteins. Nat. Rev. Genet. 2015;16:379–394. doi: 10.1038/nrg3927. [DOI] [PubMed] [Google Scholar]
- Delmont T. O.. et al. Metagenomic mining for microbiologists. ISME J. 2011;5:1837–1843. doi: 10.1038/ismej.2011.61. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Longwell C. K., Labanieh L., Cochran J. R.. High-throughput screening technologies for enzyme engineering. Curr. Opin. Biotechnol. 2017;48:196–202. doi: 10.1016/j.copbio.2017.05.012. [DOI] [PubMed] [Google Scholar]
- Xiao H., Bao Z., Zhao H.. High Throughput Screening and Selection Methods for Directed Enzyme Evolution. Ind. Eng. Chem. Res. 2015;54:4011–4020. doi: 10.1021/ie503060a. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Clifton B. E., Kozome D., Laurino P.. Efficient Exploration of Sequence Space by Sequence-Guided Protein Engineering and Design. Biochemistry. 2023;62:210–220. doi: 10.1021/acs.biochem.1c00757. [DOI] [PubMed] [Google Scholar]
- Dryden D. T. F., Thomson A. R., White J. H.. How much of protein sequence space has been explored by life on Earth? J. R. Soc. Interface. 2008;5:953–956. doi: 10.1098/rsif.2008.0085. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kouba P.. et al. Machine Learning-Guided Protein Engineering. ACS Catal. 2023;13:13863–13895. doi: 10.1021/acscatal.3c02743. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Feehan R., Montezano D., Slusky J. S. G.. Machine learning for enzyme engineering, selection and design. Protein Eng. Des. Sel. 2021;34:gzab019. doi: 10.1093/protein/gzab019. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Jumper J.. et al. Highly accurate protein structure prediction with AlphaFold. Nature. 2021;596:583–589. doi: 10.1038/s41586-021-03819-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Yang Z., Zeng X., Zhao Y., Chen R.. AlphaFold2 and its applications in the fields of biology and medicine. Signal Transduct. Target Ther. 2023;8:1–14. doi: 10.1038/s41392-023-01381-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Abramson J.. et al. Accurate structure prediction of biomolecular interactions with AlphaFold 3. Nature. 2024;630:493–500. doi: 10.1038/s41586-024-07487-w. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Yu T.. et al. Enzyme function prediction using contrastive learning. Science. 2023;379:1358–1363. doi: 10.1126/science.adf2465. [DOI] [PubMed] [Google Scholar]
- Lin Z.. et al. Evolutionary-scale prediction of atomic-level protein structure with a language model. Science. 2023;379:1123–1130. doi: 10.1126/science.ade2574. [DOI] [PubMed] [Google Scholar]
- Wang S., Tang H., Zhao Y., Zuo L.. BayeStab: Predicting effects of mutations on protein stability with uncertainty quantification. Protein Sci. 2022;31:e4467. doi: 10.1002/pro.4467. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Wang C., Zou Q.. Prediction of protein solubility based on sequence physicochemical patterns and distributed representation information with DeepSoluE. BMC Biol. 2023;21:12. doi: 10.1186/s12915-023-01510-8. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Musil M., Konegger H., Hon J., Bednar D., Damborsky J.. Computational Design of Stable and Soluble Biocatalysts. ACS Catal. 2019;9:1033–1054. doi: 10.1021/acscatal.8b03613. [DOI] [Google Scholar]
- Pak M. A.. et al. Using AlphaFold to predict the impact of single mutations on protein stability and function. PLoS One. 2023;18:e0282689. doi: 10.1371/journal.pone.0282689. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Lu H.. et al. Machine learning-aided engineering of hydrolases for PET depolymerization. Nature. 2022;604:662–667. doi: 10.1038/s41586-022-04599-z. [DOI] [PubMed] [Google Scholar]
- Shroff R.. et al. Discovery of Novel Gain-of-Function Mutations Guided by Structure-Based Deep Learning. ACS Synth. Biol. 2020;9:2927–2935. doi: 10.1021/acssynbio.0c00345. [DOI] [PubMed] [Google Scholar]
- Yang J., Li F.-Z., Arnold F. H.. Opportunities and Challenges for Machine Learning-Assisted Enzyme Engineering. ACS Cent. Sci. 2024;10:226–241. doi: 10.1021/acscentsci.3c01275. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Romero P. A., Arnold F. H.. Exploring protein fitness landscapes by directed evolution. Nat. Rev. Mol. Cell Biol. 2009;10:866–876. doi: 10.1038/nrm2805. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Klesmith J. R., Bacik J.-P., Wrenbeck E. E., Michalczyk R., Whitehead T. A.. Trade-offs between enzyme fitness and solubility illuminated by deep mutational scanning. Proc. Natl. Acad. Sci. U. S. A. 2017;114:2265–2270. doi: 10.1073/pnas.1614437114. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Currin A., Swainston N., Day P. J., Kell D. B.. Synthetic biology for the directed evolution of protein biocatalysts: navigating sequence space intelligently. Chem. Soc. Rev. 2015;44:1172–1239. doi: 10.1039/C4CS00351A. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Johnson K. A., Goody R. S.. The Original Michaelis Constant: Translation of the 1913 Michaelis–Menten Paper. Biochemistry. 2011;50:8264–8269. doi: 10.1021/bi201284u. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Waley S. G.. A quick method for the determination of inhibition constants. Biochem. J. 1982;205:631–633. doi: 10.1042/bj2050631. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Sousa S. F.. et al. Activation Free Energy, Substrate Binding Free Energy, and Enzyme Efficiency Fall in a Very Narrow Range of Values for Most Enzymes. ACS Catal. 2020;10:8444–8453. doi: 10.1021/acscatal.0c01947. [DOI] [Google Scholar]
- Warshel A.. et al. Electrostatic Basis for Enzyme Catalysis. Chem. Rev. 2006;106:3210–3235. doi: 10.1021/cr0503106. [DOI] [PubMed] [Google Scholar]
- Yan S.-M., Shi D.-Q., Nong H., Wu G.. Predicting Km values of beta-glucosidases using cellobiose as substrate. Interdiscip. Sci. Comput. Life Sci. 2012;4:46–53. doi: 10.1007/s12539-012-0115-z. [DOI] [PubMed] [Google Scholar]
- Heckmann D., Lloyd C. J., Mih N., Ha Y., Zielinski D. C., Haiman Z. B., Desouki A. A., Lercher M. J., Palsson B. O.. Machine learning applied to enzyme turnover numbers reveals protein structural correlates and improves metabolic models. Nat. Commun. 2018;9:5252. doi: 10.1038/s41467-018-07652-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Boorla V. S., Maranas C. D.. CatPred: a comprehensive framework for deep learning in vitro enzyme kinetic parameters. Nat. Commun. 2025;16:2072. doi: 10.1038/s41467-025-57215-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Carbonell P.. et al. Selenzyme: enzyme selection tool for pathway design. Bioinformatics. 2018;34:2153–2154. doi: 10.1093/bioinformatics/bty065. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Markus B.. et al. Accelerating Biocatalysis Discovery with Machine Learning: A Paradigm Shift in Enzyme Engineering, Discovery, and Design. ACS Catal. 2023;13:14454–14469. doi: 10.1021/acscatal.3c03417. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Harding-Larsen D.. et al. Protein representations: Encoding biological information for machine learning in biocatalysis. Biotechnol. Adv. 2024;77:108459. doi: 10.1016/j.biotechadv.2024.108459. [DOI] [PubMed] [Google Scholar]
- Krichen M.. Convolutional Neural Networks: A Survey. Computers. 2023;12:151. doi: 10.3390/computers12080151. [DOI] [Google Scholar]
- Detlefsen N. S., Hauberg S., Boomsma W.. Learning meaningful representations of protein sequences. Nat. Commun. 2022;13:1914. doi: 10.1038/s41467-022-29443-w. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Vaswani, A. ; et al. Attention is All you Need. Advances in Neural Information Processing Systems; Curran Associates, Inc., 2017; Vol. 30. [Google Scholar]
- Kalyan K. S., Rajasekharan A., Sangeetha S.. AMMUS: A Survey of Transformer-based Pretrained Models in Natural Language Processing. arXiv:2108.05542 [cs.CL] 2021:na. doi: 10.48550/arXiv.2108.05542. [DOI] [Google Scholar]
- Brandes N., Ofer D., Peleg Y., Rappoport N., Linial M.. ProteinBERT: a universal deep-learning model of protein sequence and function. Bioinformatics. 2022;38:2102–2110. doi: 10.1093/bioinformatics/btac020. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Elnaggar A.. et al. ProtTrans: Toward Understanding the Language of Life Through Self-Supervised Learning. IEEE Trans. Pattern Anal. Mach. Intell. 2022;44:7112–7127. doi: 10.1109/TPAMI.2021.3095381. [DOI] [PubMed] [Google Scholar]
- Alley E. C., Khimulya G., Biswas S., AlQuraishi M., Church G. M.. Unified rational protein engineering with sequence-based deep representation learning. Nat. Methods. 2019;16:1315–1322. doi: 10.1038/s41592-019-0598-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Bagley S. C., Altman R. B.. Characterizing the microenvironment surrounding protein sites. Protein Sci. 1995;4:622–635. doi: 10.1002/pro.5560040404. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Emerson I. A., Amala A.. Protein contact maps: A binary depiction of protein 3D structures. Phys. Stat. Mech. Its Appl. 2017;465:782–791. doi: 10.1016/j.physa.2016.08.033. [DOI] [Google Scholar]
- Durant J. L., Leland B. A., Henry D. R., Nourse J. G.. Reoptimization of MDL Keys for Use in Drug Discovery. J. Chem. Inf. Comput. Sci. 2002;42:1273–1280. doi: 10.1021/ci010132r. [DOI] [PubMed] [Google Scholar]
- Gilmer, J. ; Schoenholz, S. S. ; Riley, P. F. ; Vinyals, O. ; Dahl, G. E. . Message Passing Neural Networks. In Machine Learning Meets Quantum Physics, Eds.; Schütt, K. T. , et al. Springer International Publishing, Cham, 2020; pp 199–214. 10.1007/978-3-030-40245-7_10. [DOI] [Google Scholar]
- Honda S., Shi S., Ueda H. R.. SMILES Transformer: Pre-trained Molecular Fingerprint for Low Data Drug Discovery. arXiv:1911.04738v1 [cs.LG] 2019:na. doi: 10.48550/arXiv.1911.04738. [DOI] [Google Scholar]
- Chithrananda S., Grand G., Ramsundar B.. ChemBERTa: Large-Scale Self-Supervised Pretraining for Molecular Property Prediction. arXiv:2010.09885 [cs.LG] 2020:na. doi: 10.48550/arXiv.2010.09885. [DOI] [Google Scholar]
- Chang A.. et al. BRENDA, the ELIXIR core data resource in 2021: new developments and updates. Nucleic Acids Res. 2021;49:D498–D508. doi: 10.1093/nar/gkaa1025. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Wittig U., Rey M., Weidemann A., Kania R., Müller W.. SABIO-RK: an updated resource for manually curated biochemical reaction kinetics. Nucleic Acids Res. 2018;46:D656–D660. doi: 10.1093/nar/gkx1065. [DOI] [PMC free article] [PubMed] [Google Scholar]
- The UniProt Consortium. The UniProt Consortium. UniProt: the Universal Protein Knowledgebase in 2023. Nucleic Acids Res. 2023;51:D523–D531. doi: 10.1093/nar/gkac1052. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Yu, X. ; Yi, K. ; Wang, Y. G. ; Shen, Y. . A Regressor-Guided Graph Diffusion Model for Predicting Enzyme Mutations to Enhance Turnover Number. 2024 IEEE International Conference on Bioinformatics and Biomedicine (BIBM) IEEE, 2024; pp 3943–3948. 10.1109/BIBM62325.2024.10822301. [DOI] [Google Scholar]
- Wang T., Xiang G., He S., Su L., Wang Y., Yan X., Lu H.. DeepEnzyme: a robust deep learning model for improved enzyme turnover number prediction by utilizing features of protein 3D-structures. Brief. Bioinform. 2024;25:bbae409. doi: 10.1093/bib/bbae409. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Krishnan S. R., Pandey N., Srinivasan R., Roy A.. A structure-oriented kinetics dataset of enzyme-substrate interactions. Sci. Data. 2025;12:1489. doi: 10.1038/s41597-025-05829-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
- McNutt A. T., Francoeur P., Aggarwal R., Masuda T., Meli R., Ragoza M., Sunseri J., Koes D. R.. GNINA 1.0: molecular docking with deep learning. J. Cheminformatics. 2021;13:43. doi: 10.1186/s13321-021-00522-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Yan B.. et al. IntEnzyDB: an Integrated Structure–Kinetics Enzymology Database. J. Chem. Inf. Model. 2022;62:5841–5848. doi: 10.1021/acs.jcim.2c01139. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Li F., Chen Y., Anton M., Nielsen J.. GotEnzymes: an extensive database of enzyme parameter predictions. Nucleic Acids Res. 2023;51:D583–D586. doi: 10.1093/nar/gkac831. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Lyu B.. et al. GotEnzymes2: expanding coverage of enzyme kinetics and thermal properties. Nucleic Acids Res. 2025:gkaf1053. doi: 10.1093/nar/gkaf1053. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Li F.. et al. Deep learning-based kcat prediction enables improved enzyme-constrained model reconstruction. Nat. Catal. 2022;5:662–672. doi: 10.1038/s41929-022-00798-z. [DOI] [Google Scholar]
- Kroll A., Rousset Y., Hu X.-P., Liebrand N. A., Lercher M. J.. Turnover number predictions for kinetically uncharacterized enzymes using machine and deep learning. Nat. Commun. 2023;14:4139. doi: 10.1038/s41467-023-39840-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Bar-Even A.. et al. The Moderately Efficient Enzyme: Evolutionary and Physicochemical Trends Shaping Enzyme Parameters. Biochemistry. 2011;50:4402–4410. doi: 10.1021/bi2002289. [DOI] [PubMed] [Google Scholar]
- Kim S.. et al. PubChem 2023 update. Nucleic Acids Res. 2023;51:D1373–D1380. doi: 10.1093/nar/gkac956. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kanehisa M., Furumichi M., Tanabe M., Sato Y., Morishima K.. KEGG: new perspectives on genomes, pathways, diseases and drugs. Nucleic Acids Res. 2017;45:D353–D361. doi: 10.1093/nar/gkw1092. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Degtyarenko K.. et al. ChEBI: a database and ontology for chemical entities of biological interest. Nucleic Acids Res. 2007;36:D344–D350. doi: 10.1093/nar/gkm791. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Wilkinson M. D.. et al. The FAIR Guiding Principles for scientific data management and stewardship. Sci. Data. 2016;3:160018. doi: 10.1038/sdata.2016.18. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Goldberg R. N., Giessmann R. T., Halling P. J., Kettner C., Westerhoff H. V.. Recommendations for performing measurements of apparent equilibrium constants of enzyme-catalyzed reactions and for reporting the results of these measurements. Beilstein J. Org. Chem. 2023;19:303–316. doi: 10.3762/bjoc.19.26. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Ao Y.-F., Dorr M., Menke M. J., Born S., Heuson E., Bornscheuer U. T.. Data-Driven Protein Engineering for Improving Catalytic Activity and Selectivity. ChemBioChem. 2024;25:e202300754. doi: 10.1002/cbic.202300754. [DOI] [PubMed] [Google Scholar]
- Lauterbach S.. et al. EnzymeML: seamless data flow and modeling of enzymatic data. Nat. Methods. 2023;20:400–402. doi: 10.1038/s41592-022-01763-1. [DOI] [PubMed] [Google Scholar]
- Swainston N.. et al. STRENDA DB: enabling the validation and sharing of enzyme kinetics data. FEBS J. 2018;285:2193–2204. doi: 10.1111/febs.14427. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Schnell S., Turner T. E.. Reaction kinetics in intracellular environments with macromolecular crowding: simulations and rate laws. Prog. Biophys. Mol. Biol. 2004;85:235–260. doi: 10.1016/j.pbiomolbio.2004.01.012. [DOI] [PubMed] [Google Scholar]
- Suzuki Y., Kadomatsu K., Sakamoto K.. Towards the in vivo identification of protein–protein interactions. J. Biochem. (Tokyo) 2023;173:413–415. doi: 10.1093/jb/mvad013. [DOI] [PubMed] [Google Scholar]
- Ryšlavá H., Doubnerová V., Kavan D., Vaněk O.. Effect of posttranslational modifications on enzyme function and assembly. J. Proteomics. 2013;92:80–109. doi: 10.1016/j.jprot.2013.03.025. [DOI] [PubMed] [Google Scholar]
- Davidi D.. et al. Global characterization of in vivo enzyme catalytic rates and their correspondence to in vitro kcat measurements. Proc. Natl. Acad. Sci. U. S. A. 2016;113:3401–3406. doi: 10.1073/pnas.1514240113. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kroll A., Lercher M. J.. DLKcat cannot predict meaningful kcat values for mutants and unfamiliar enzymes. Biol. Methods Protoc. 2024;9:bpae061. doi: 10.1093/biomethods/bpae061. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Wang Z., Xie D., Wu D., Luo X., Wang S., Li Y., Yang Y., Li W., Zheng L.. Robust enzyme discovery and engineering with deep learning using CataPro. Nat. Commun. 2025;16:2736. doi: 10.1038/s41467-025-58038-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kroll A., Ranjan S., Engqvist M. K. M., Lercher M. J.. A general model to predict small molecule substrates of enzymes based on machine and deep learning. Nat. Commun. 2023;14:2787. doi: 10.1038/s41467-023-38347-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Yu H., Deng H., He J., Keasling J. D., Luo X.. UniKP: a unified framework for the prediction of enzyme kinetic parameters. Nat. Commun. 2023;14:8211. doi: 10.1038/s41467-023-44113-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Qiu S., Zhao S., Yang A.. DLTKcat: deep learning-based prediction of temperature-dependent enzyme turnover rates. Brief. Bioinform. 2023;25:bbad506. doi: 10.1093/bib/bbad506. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Cai Y.. et al. PreTKcat: A pre-trained representation learning and machine learning framework for predicting enzyme turnover number. Comput. Biol. Chem. 2025;115:108327. doi: 10.1016/j.compbiolchem.2024.108327. [DOI] [PubMed] [Google Scholar]
- Du B.-X.. et al. A novel interpretability framework for enzyme turnover number prediction boosted by pre-trained enzyme embeddings and adaptive gate network. Methods. 2025;237:45–52. doi: 10.1016/j.ymeth.2025.02.010. [DOI] [PubMed] [Google Scholar]
- Qiu J.-H.. SAKPE: A Site Attention Kinetic Parameters Prediction Method for Enzyme Engineering. bioRxiv. 2025:na. doi: 10.1101/2025.04.30.651216. [DOI] [Google Scholar]
- Nie Z.. OmniESI: A unified framework for enzyme-substrate interaction prediction with progressive conditional deep learning. arXiv:2506.17963 [q-bio.BM] 2025:na. doi: 10.48550/arXiv.2506.17963. [DOI] [Google Scholar]
- Pan T.. KcatNet: A Geometric Deep Learning Framework for Genome-Wide Prediction of Enzyme Catalytic Efficiency. bioRxiv. 2025:na. doi: 10.1101/2025.03.09.642294. [DOI] [Google Scholar]
- Alwer S., Fleming R.. KinForm: Kinetics Informed Feature Optimised Representation Models for Enzyme k cat and K M Prediction. arXiv:2507.14639 [q-bio.QM] 2025:na. doi: 10.48550/arXiv.2507.14639. [DOI] [Google Scholar]
- Gollub M. G., Backes T., Kaltenbach H.-M., Stelling J.. ENKIE: a package for predicting enzyme kinetic parameter values and their uncertainties. Bioinformatics. 2024;40:btae652. doi: 10.1093/bioinformatics/btae652. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Xu Z.. CPI-Pred: A deep learning framework for predicting functional parameters of compound-protein interactions. bioRxiv. 2025:na. doi: 10.1101/2025.01.16.633372. [DOI] [Google Scholar]
- Sajeevan K. A.. Robust Prediction of Enzyme Variant Kinetics with RealKcat. bioRxiv. 2025:na. doi: 10.1101/2025.02.10.637555. [DOI] [Google Scholar]
- Dietterich, T. G. Ensemble Methods in Machine Learning. Multiple Classifier Systems; Springer, Berlin, Heidelberg, 2000; pp 1–15. 10.1007/3-540-45014-9_1. [DOI] [Google Scholar]
- Cui, Y. ; Jia, M. ; Lin, T.-Y. ; Song, Y. ; Belongie, S. . Class-Balanced Loss Based on Effective Number of Samples. 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR); IEEE, 2019; pp 9260–9269. 10.1109/CVPR.2019.00949. [DOI] [Google Scholar]
- Daniel R. M., Danson M. J.. A new understanding of how temperature affects the catalytic activity of enzymes. Trends Biochem. Sci. 2010;35:584–591. doi: 10.1016/j.tibs.2010.05.001. [DOI] [PubMed] [Google Scholar]
- Arroyo J. I., Díez B., Kempes C. P., West G. B., Marquet P. A.. A general theory for temperature dependence in biology. Proc. Natl. Acad. Sci. U. S. A. 2022;119:e2119872119. doi: 10.1073/pnas.2119872119. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Wang X.. et al. Multi-modal deep learning enables efficient and accurate annotation of enzymatic active sites. Nat. Commun. 2024;15:7348. doi: 10.1038/s41467-024-51511-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Anfinsen C. B.. Principles that Govern the Folding of Protein Chains. Science. 1973;181:223–230. doi: 10.1126/science.181.4096.223. [DOI] [PubMed] [Google Scholar]
- Durairaj J., de Ridder D., van Dijk A. D. J.. Beyond sequence: Structure-based machine learning. Comput. Struct. Biotechnol. J. 2023;21:630–643. doi: 10.1016/j.csbj.2022.12.039. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Gligorijević V.. et al. Structure-based protein function prediction using graph convolutional networks. Nat. Commun. 2021;12:3168. doi: 10.1038/s41467-021-23303-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Jiménez J., Doerr S., Martínez-Rosell G., Rose A. S., De Fabritiis G.. DeepSite: protein-binding site predictor using 3D-convolutional neural networks. Bioinformatics. 2017;33:3036–3042. doi: 10.1093/bioinformatics/btx350. [DOI] [PubMed] [Google Scholar]
- Mirdita M.. et al. ColabFold: making protein folding accessible to all. Nat. Methods. 2022;19:679–682. doi: 10.1038/s41592-022-01488-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Chothia C., Lesk A. M.. The relation between the divergence of sequence and structure in proteins. EMBO J. 1986;5:823–826. doi: 10.1002/j.1460-2075.1986.tb04288.x. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Illergård K., Ardell D. H., Elofsson A.. Structure is three to ten times more conserved than sequenceA study of structural response in protein cores. Proteins Struct. Funct. Bioinforma. 2009;77:499–508. doi: 10.1002/prot.22458. [DOI] [PubMed] [Google Scholar]
- Seo S.. et al. Pseq2Sites: Enhancing protein sequence-based ligand binding-site prediction accuracy via the deep convolutional network and attention mechanism. Eng. Appl. Artif. Intell. 2024;127:107257. doi: 10.1016/j.engappai.2023.107257. [DOI] [Google Scholar]
- Hirschfeld L., Swanson K., Yang K., Barzilay R., Coley C. W.. Uncertainty Quantification Using Neural Networks for Molecular Property Prediction. J. Chem. Inf. Model. 2020;60:3770–3780. doi: 10.1021/acs.jcim.0c00502. [DOI] [PubMed] [Google Scholar]
- Shen J.. et al. Unbiased organism-agnostic and highly sensitive signal peptide predictor with deep protein language model. Nat. Comput. Sci. 2024;4:29–42. doi: 10.1038/s43588-023-00576-2. [DOI] [PubMed] [Google Scholar]
- Wang Y., Cheng L., Zhang Y., Cao Y., Alghazzawi D.. DEKP: a deep learning model for enzyme kinetic parameter prediction based on pretrained models and graph neural networks. Brief. Bioinform. 2025;26:bbaf187. doi: 10.1093/bib/bbaf187. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Moretti S., Tran V. D. T., Mehl F., Ibberson M., Pagni M.. MetaNetX/MNXref: unified namespace for metabolites and biochemical reactions in the context of metabolic models. Nucleic Acids Res. 2021;49:D570–D574. doi: 10.1093/nar/gkaa992. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Metcalf K. J., Slininger Lee M. F., Jakobson C. M., Tullman-Ercek D.. An estimate is worth about a thousand experiments: using order-of-magnitude estimates to identify cellular engineering targets. Microb. Cell Factories. 2018;17:135. doi: 10.1186/s12934-018-0979-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Dias Gomes M., Woodley J. M.. Considerations when Measuring Biocatalyst Performance. Molecules. 2019;24:3573. doi: 10.3390/molecules24193573. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kroll A., Engqvist M. K. M., Heckmann D., Lercher M. J.. Deep learning allows genome-scale prediction of Michaelis constants from structural features. PLOS Biol. 2021;19:e3001402. doi: 10.1371/journal.pbio.3001402. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Maeda K., Hatae A., Sakai Y., Boogerd F. C., Kurata H.. MLAGO: machine learning-aided global optimization for Michaelis constant estimation of kinetic modeling. BMC Bioinformatics. 2022;23:455. doi: 10.1186/s12859-022-05009-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
- He X., Yan M.. GraphKM: machine and deep learning for KM prediction of wildtype and mutant enzymes. BMC Bioinformatics. 2024;25:135. doi: 10.1186/s12859-024-05746-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Wang J., Yang Z., Chen C., Yao G., Wan X., Bao S., Ding J., Wang L., Jiang H.. MPEK: a multitask deep learning framework based on pretrained language models for enzymatic reaction kinetic parameters prediction. Brief. Bioinform. 2024;25:bbae387. doi: 10.1093/bib/bbae387. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Li Y., Wang K.. A Deep Learning-Based Approach for Predicting Michaelis Constants from Enzymatic Reactions. Appl. Sci. 2025;15:4017. doi: 10.3390/app15074017. [DOI] [Google Scholar]
- Shen X.. et al. EITLEM-Kinetics: A deep-learning framework for kinetic parameter prediction of mutant enzymes. Chem. Catal. 2024;4:101094. doi: 10.1016/j.checat.2024.101094. [DOI] [Google Scholar]
- Kroll A.. Transfer learning improves predictions of enzyme kinetic parameters. Chem. Catal. 2024;4:101121. doi: 10.1016/j.checat.2024.101121. [DOI] [Google Scholar]
- Neyshabur B., Sedghi H., Zhang C.. What is being transferred in transfer learning? arXiv:2008.11687 [cs.LG] 2021:na. doi: 10.48550/arXiv.2008.11687. [DOI] [Google Scholar]
- Badwan B. A.. et al. Machine learning approaches to predict drug efficacy and toxicity in oncology. Cell Rep. Methods. 2023;3:100413. doi: 10.1016/j.crmeth.2023.100413. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Thafar M., Raies A. B., Albaradei S., Essack M., Bajic V. B.. Comparison Study of Computational Prediction Tools for Drug-Target Binding Affinities. Front. Chem. 2019;7:na. doi: 10.3389/fchem.2019.00782. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Miton C. M., Tokuriki N.. How mutational epistasis impairs predictability in protein evolution and design. Protein Sci. 2016;25:1260–1272. doi: 10.1002/pro.2876. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Zhou S.. et al. Characterization of mutants of a tyrosine ammonia-lyase from Rhodotorula glutinis. Appl. Microbiol. Biotechnol. 2016;100:10443–10452. doi: 10.1007/s00253-016-7672-8. [DOI] [PubMed] [Google Scholar]
- Altschul S. F., Gish W., Miller W., Myers E. W., Lipman D. J.. Basic local alignment search tool. J. Mol. Biol. 1990;215:403–410. doi: 10.1016/S0022-2836(05)80360-2. [DOI] [PubMed] [Google Scholar]
- Knudsen M., Wiuf C.. The CATH database. Hum. Genomics. 2010;4:207. doi: 10.1186/1479-7364-4-3-207. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Dauparas J.. et al. Robust deep learning–based protein sequence design using ProteinMPNN. Science. 2022;378:49–56. doi: 10.1126/science.add2187. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Gao Z., Tan C., Chacón P., Li S. Z.. PiFold: Toward effective and efficient protein inverse folding. arXiv:2209.12643 [cs.AI] 2023:na. doi: 10.48550/arXiv.2209.12643. [DOI] [Google Scholar]
- Yi K., Zhou B., Shen Y., Liò P., Wang Y. G.. Graph Denoising Diffusion for Inverse Protein Folding. arXiv:2306.16819 [q-bio.QM] 2023:na. doi: 10.48550/arXiv.2306.16819. [DOI] [Google Scholar]
- Erkanli M. E.. et al. A Three-Module Machine Learning Framework for Protein Sequence- and Temperature-Dependent kcat/Km Prediction in β-Glucosidases. ACS Synth. Biol. 2025;14:3927–3939. doi: 10.1021/acssynbio.5c00257. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Muir D. F., Asper G. P. R., Notin P., Posner J. A., Marks D. S., Keiser M. J., Pinney M. M.. Evolutionary-scale enzymology enables exploration of a rugged catalytic landscape. Science. 2025;388:eadu1058. doi: 10.1126/science.adu1058. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Paton A. E.. et al. Connecting chemical and protein sequence space to predict biocatalytic reactions. Nature. 2025;646:108–116. doi: 10.1038/s41586-025-09519-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Deng J., Yang Z., Wang H., Ojima I., Samaras D., Wang F.. A systematic study of key elements underlying molecular property prediction. Nat. Commun. 2023;14:6395. doi: 10.1038/s41467-023-41948-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Thiele I., Palsson B. Ø.. A protocol for generating a high-quality genome-scale metabolic reconstruction. Nat. Protoc. 2010;5:93–121. doi: 10.1038/nprot.2009.203. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Reed J. L., Famili I., Thiele I., Palsson B. O.. Towards multidimensional genome annotation. Nat. Rev. Genet. 2006;7:130–141. doi: 10.1038/nrg1769. [DOI] [PubMed] [Google Scholar]
- Orth J. D., Thiele I., Palsson B. Ø.. What is flux balance analysis? Nat. Biotechnol. 2010;28:245–248. doi: 10.1038/nbt.1614. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Henson M. A.. Genome-scale modelling of microbial metabolism with temporal and spatial resolution. Biochem. Soc. Trans. 2015;43:1164–1171. doi: 10.1042/BST20150146. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Bekiaris P. S., Klamt S.. Automatic construction of metabolic models with enzyme constraints. BMC Bioinformatics. 2020;21:19. doi: 10.1186/s12859-019-3329-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Sanchez B. J, Zhang C., Nilsson A., Lahtvee P.-J., Kerkhoven E. J, Nielsen J.. Improving the phenotype predictions of a yeast genome-scale metabolic model by incorporating enzymatic constraints. Mol. Syst. Biol. 2017;13:935. doi: 10.15252/msb.20167411. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Toumpe I., Choudhury S., Hatzimanikatis V., Miskovic L.. The Dawn of High-Throughput and Genome-Scale Kinetic Modeling: Recent Advances and Future Directions. ACS Synth. Biol. 2025;14:2906–2919. doi: 10.1021/acssynbio.4c00868. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Khodayari A., Maranas C. D.. A genome-scale Escherichia coli kinetic metabolic model k-ecoli457 satisfying flux data for multiple mutant strains. Nat. Commun. 2016;7:13806. doi: 10.1038/ncomms13806. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Narayanan B., Weilandt D., Masid M., Miskovic L., Hatzimanikatis V.. Rational strain design with minimal phenotype perturbation. Nat. Commun. 2024;15:723. doi: 10.1038/s41467-024-44831-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Hu M.. et al. Comparative study of two Saccharomyces cerevisiae strains with kinetic models at genome-scale. Metab. Eng. 2023;76:1–17. doi: 10.1016/j.ymben.2023.01.001. [DOI] [PubMed] [Google Scholar]
- Narayanan B.. et al. Kinetic-model-guided engineering of multiple S. cerevisiae strains improves p-coumaric acid production. Metab. Eng. 2025;91:430–441. doi: 10.1016/j.ymben.2025.06.008. [DOI] [PubMed] [Google Scholar]
- Toumpe I., Masid M., Hatzimanikatis V., Miskovic L.. Multi-omics-driven kinetic modeling reveals metabolic vulnerabilities and differential drug-response dynamics in ovarian cancer. bioRxiv. 2025:na. doi: 10.1101/2025.10.22.682756. [DOI] [Google Scholar]
- Korotkova O. G.. et al. Isolation and properties of fungal β-glucosidases. Biochem. Mosc. 2009;74:569–577. doi: 10.1134/S0006297909050137. [DOI] [PubMed] [Google Scholar]
- Kawashima S.. et al. AAindex: amino acid index database, progress report 2008. Nucleic Acids Res. 2007;36:D202–D205. doi: 10.1093/nar/gkm998. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Safonova A.. et al. Ten deep learning techniques to address small data problems with remote sensing. Int. J. Appl. Earth Obs. Geoinformation. 2023;125:103569. doi: 10.1016/j.jag.2023.103569. [DOI] [Google Scholar]
- Carlin D. A.. et al. Kinetic Characterization of 100 Glycoside Hydrolase Mutants Enables the Discovery of Structural Features Correlated with Kinetic Constants. PLoS One. 2016;11:e0147596. doi: 10.1371/journal.pone.0147596. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Liu H.. et al. MECE: a method for enhancing the catalytic efficiency of glycoside hydrolase based on deep neural networks and molecular evolution. Sci. Bull. 2023;68:2793–2805. doi: 10.1016/j.scib.2023.09.039. [DOI] [PubMed] [Google Scholar]
- Cantarel B. L.. et al. The Carbohydrate-Active EnZymes database (CAZy): an expert resource for Glycogenomics. Nucleic Acids Res. 2009;37:D233–D238. doi: 10.1093/nar/gkn663. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Ao Y.-F., Pei S., Xiang C., Menke M. J., Shen L., Sun C., Dorr M., Born S., Hohne M., Bornscheuer U. T.. Structure- and Data-Driven Protein Engineering of Transaminases for Improving Activity and Stereoselectivity. Angew. Chem., Int. Ed. 2023;62:e202301660. doi: 10.1002/anie.202301660. [DOI] [PubMed] [Google Scholar]
- Rapp J. T., Bremer B. J., Romero P. A.. Self-driving laboratories to autonomously navigate the protein fitness landscape. Nat. Chem. Eng. 2024;1:97–107. doi: 10.1038/s44286-023-00002-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Shao F., Lee P.-W., Li H., Hsieh K., Wang T.-H.. Emerging platforms for high-throughput enzymatic bioassays. Trends Biotechnol. 2023;41:120–133. doi: 10.1016/j.tibtech.2022.06.006. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Markin C. J., Mokhtari D. A., Sunden F., Appel M. J., Akiva E., Longwell S. A., Sabatti C., Herschlag D., Fordyce P. M.. Revealing enzyme functional architecture via high-throughput microfluidic enzyme kinetics. Science. 2021;373:eabf8761. doi: 10.1126/science.abf8761. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Neun S., van Vliet L., Hollfelder F., Gielen F.. High-Throughput Steady-State Enzyme Kinetics Measured in a Parallel Droplet Generation and Absorbance Detection Platform. Anal. Chem. 2022;94:16701–16710. doi: 10.1021/acs.analchem.2c03164. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Li X., He Z., Li C., Li P.. One-step enzyme kinetics measurement in 3D printed microfluidics devices based on a high-performance single vibrating sharp-tip mixer. Anal. Chim. Acta. 2021;1172:338677. doi: 10.1016/j.aca.2021.338677. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Carbonell P.. et al. An automated Design-Build-Test-Learn pipeline for enhanced microbial production of fine chemicals. Commun. Biol. 2018;1:66. doi: 10.1038/s42003-018-0076-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Astolfi M. C. T.. et al. Automated Strain Construction for Biosynthetic Pathway Screening in Yeast. ACS Synth. Biol. 2025;14:4143–4151. doi: 10.1021/acssynbio.5c00554. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Godin R., Hejazi S., Lange B., Aldamak B., Reuel N. F.. Rapid Cell-Free Combinatorial Mutagenesis Workflow Using Small Oligos Suitable for High-Iteration, Active Learning-Guided Protein Engineering. bioRxiv. 2025:na. doi: 10.1101/2025.06.18.660386. [DOI] [Google Scholar]
- Burger B.. et al. A mobile robotic chemist. Nature. 2020;583:237–241. doi: 10.1038/s41586-020-2442-2. [DOI] [PubMed] [Google Scholar]
- Heckmann D., Lloyd C. J., Mih N., Ha Y., Zielinski D. C., Haiman Z. B., Desouki A. A., Lercher M. J., Palsson B. O.. Machine learning applied to enzyme turnover numbers reveals protein structural correlates and improves metabolic models. Nat. Commun. 2018;9:5252. doi: 10.1038/s41467-018-07652-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
- MacLeod B. P.. et al. Self-driving laboratory for accelerated discovery of thin-film materials. Sci. Adv. 2020;6:eaaz8867. doi: 10.1126/sciadv.aaz8867. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Barbero-Aparicio J. A., Olivares-Gil A., Rodríguez J. J., García-Osorio C., Díez-Pastor J. F.. Addressing data scarcity in protein fitness landscape analysis: A study on semi-supervised and deep transfer learning techniques. Inf. Fusion. 2024;102:102035. doi: 10.1016/j.inffus.2023.102035. [DOI] [Google Scholar]
- Jurich C., Shao Q., Ran X., Yang Z. J.. Physics-based modeling in the new era of enzyme engineering. Nat. Comput. Sci. 2025;5:279–291. doi: 10.1038/s43588-025-00788-8. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Shao Q., Hollenbeak A. C., Jiang Y., Ran X., Bachmann B. O., Yang Z. J.. SubTuner leverages physics-based modeling to complement AI in enzyme engineering toward non-native substrates. Chem. Catal. 2025;5:101334. doi: 10.1016/j.checat.2025.101334. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Platero-Rochart D., Krivobokova T., Gastegger M., Reibnegger G., Sánchez-Murcia P. A.. Prediction of Enzyme Catalysis by Computing Reaction Energy Barriers via Steered QM/MM Molecular Dynamics Simulations and Machine Learning. J. Chem. Inf. Model. 2023;63:4623–4632. doi: 10.1021/acs.jcim.3c00772. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Choudhury S.. et al. Reconstructing Kinetic Models for Dynamical Studies of Metabolism using Generative Adversarial Networks. Nat. Mach. Intell. 2022;4:710–719. doi: 10.1038/s42256-022-00519-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Choudhury S., Narayanan B., Moret M., Hatzimanikatis V., Miskovic L.. Generative machine learning produces kinetic models that accurately characterize intracellular metabolic states. Nat. Catal. 2024;7:1086–1098. doi: 10.1038/s41929-024-01220-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.




