Abstract
Artificial intelligence (AI) and machine learning (ML) are rapidly reshaping the landscape of computational chemistry, offering new opportunities for accelerating catalyst discovery and deepening our understanding of chemical reactivity. This perspective highlights emerging methodologies ranging from machine learning potentials and reinforcement learning to generative AI and large language models that are poised to transform computational catalysis. We discuss challenges in developing robust molecular representations for transition-metal complexes, bridging mechanistic understanding with AI-driven predictions, and constructing reliable data sets that capture both successful and failed reactivity outcomes. By drawing on the authors’ practical experience across computational, experimental, and AI-driven domains, we emphasize the importance of integrating chemical intuition and methodological expertise with data-driven approaches while remaining open to serendipitous discoveries enabled by automation and self-driving laboratories. Ultimately, the future of computational catalysis lies in balancing human intuition with algorithmic power, leveraging AI not as a replacement but as an accelerator of chemical insight, mechanistic understanding, and catalyst design.

1. Introduction
Within less than a decade, machine learning (ML) and artificial intelligence (AI) have become pervasive in chemistry. A new branch of chemistry, often referred to as digital chemistry, has emerged at the intersection of experiment and computation. This field is dedicated to data science applications across chemistry in the broadest sense, extending well beyond the scope of traditional chemoinformatics. While pioneering work in this field is decades old, − efforts during the 2010s facilitated by significant algorithmic and hardware advances, have prepared the ground for a data revolution in chemistry that we see now unfolding. − What is different compared to previous (r)evolutions is the rapid pace of development. This high innovation pressure appears to leave us hardly any time to step aside and evaluate the accomplishments. What have been the most important developments? How are they consolidated, up to the point where they need to be transformed into teaching material to educate the next generation of computational chemists with a focus on data science? What are the next steps to expect? What are the holy grails in the field? Thus, in this article we attempt to provide a perspective on the field and its context.
While AI-driven research on organic molecules and organic synthesis has entered in a more mature stage, ,, largely propelled by pharmaceutical applications − the fields of transition metal chemistry and heavy element chemistry remain comparatively underexplored. − Technologies developed for molecules composed primarily of first- and second-row elements can play a pivotal role in advancing AI workflows for applications in homogeneous catalysis, , spintronics, heavy element separations and capture. However, important challenges remain for transitioning such technologies to molecular complexes. Those have to do with emerging AI/ML methodologies such as generative AI and large language models (LLMs), molecular representations of metal–ligand structures, chemical insights into the AI workflows, and generation of reliable and diverse data sets.
In this perspective, we explore these emerging opportunities and challenges and discuss recent developments in AI/ML techniques and how they might transform molecular modeling, with a particular focus on molecular transition metal systems. We will have an emphasis on opportunities driven by actual practical experience of the authors, which is (in general) not documented in data sets and literature and therefore not accessible to AI-powered assistants. Hence, by contrast to a traditional review article covering recent literature on a specific topic, which is a perfect target for AI-powered robots (as long as they have access to all literature and without hallucinating about scientific facts), we understand this work as being part of an increasingly important push toward perspective and opinion pieces as they draw upon human intelligence and wisdom, not reachable by machines any time soon. Accordingly, we rely on a diverse background that ranges from traditional computational chemistry applications in catalysis and chemical reactivity to the development of chemical LLMs. While molecular catalysis is the primary focus of this perspective, we also highlight recent and particularly compelling examples of AI applications in heterogeneous catalysis whenever they provide meaningful extensions to our discussion. Figure summarizes the topics that are discussed in the next sections.
1.
Schematic workflow of the four main topics explored in this perspective article.
2. Emerging AI/ML Methodologies
New methodologies such as transfer learning, − reinforcement learning, − self-supervised learning, generative AI, , and foundation models − are beginning to influence computational chemistry workflows. While there is still uncertainty about where these trends will lead, active learning, in particular, may soon replace the (manual or automated) generation of training setsone of the most labor-intensive aspects of machine learning model development. For specific domains though, such as the ML-based computation of potential energy surfaces, there are already software suites which assist in this task, including PES-learn, Asparagus, ArcaNN, or DP-Gen to name a few. In this section, we discuss two emerging areas that are increasingly shaping computational chemistry workflows: machine learning interatomic potentials (MLIPs) and generative AI methods.
2.1. Machine Learning Interatomic Potentials vs Established Computational Methods
An increasingly prominent theme in molecular simulations is the emergence and growing influence of MLIPs and their prospect to replace traditional quantum chemical methods such as density functional theory (DFT). − For many routine tasks such as conformer generation, reaction mechanism analysis, and energy profiling of drug-like molecules, MLIPs are beginning to match, and in some cases surpass DFT performance. Their advantage can be measured either in terms of accuracy, when trained on data from post-Hartree–Fock reference data (e.g., coupled-cluster), or in terms of speed, which can approach that of classical simulations. The rapid scaling of data sets and improvements in learning architectures suggest that many simulations currently handled with first-principles methods could soon be delegated to machine-learned models. In this view, DFT may increasingly serve as a data generation engine for training and validating machine learning models. This trend becomes even more evident with the emergence of general-purpose foundation models, such as MACE and UMA (the latter is trained on the OMOL2025 data set). Unlike task-specific MLIPs, these models aim to capture broadly transferable representations of chemical interactions, enabling rapid adaptation to diverse molecular systems and tasks. Their promise lies in reducing the need for system-specific retraining while offering accuracy and scalability that approach, or in some cases exceed, traditional quantum methods.
This optimism can be balanced with an awareness of current limitations, which highlight opportunities for growth. Despite training on large data sets, MLIPs can exhibit “uncharted regions” in their energy landscapes, corresponding to portions of configuration space where the model produces unreliable or unphysical results. − These issues are particularly pronounced in systems that fall outside the training distribution or that require fine control of electronic structure details. For instance, a computational campaign considered complete may need to be extended by months once it becomes clear that parts of the relevant potential energy surface are poorly represented. This points to a key limitation: while machine learning models always return a result, assessing the reliability of that result, particularly in underexplored regions of chemical space, remains a significant challenge. Recent progress in uncertainty estimation has the potential to elevate active learning for the exploration of complex potential energy surfaces, including transition metal complexes, with significant implications for catalyst design. −
We may draw a useful comparison to the early adoption of DFT itself. When DFT methods were first introduced, there was skepticism about their limitations and applicability to complex systems. Similar doubts are now being raised about MLIPs. While there will always be challenging cases such as multireference problems, , or photochemistry, , many everyday chemical simulations may eventually be handled by MLIPs with minimal loss in accuracy and significant gains in efficiency. Before reaching this level of maturity, computational chemists will have access to a rapidly growing array of semiempirical methods, including those based on tight-binding (TB) approximations to Kohn–Sham DFT. , Much of the eventual success of DFT can be traced to the systematic identification of functional-specific failure modes, which provided confidence in their domain of applicability. Whether MLIPs will undergo a comparable process remains to be seen.
Although DFT has played a prominent role for various reasons, theoretical investigations of catalytic mechanisms, particularly in the case of transition metal catalysts, currently employ a range of higher-level electronic structure methods, either single-reference (e.g., coupled-cluster-based methods) or multireference. For systems with complicated electronic structures, multinuclear catalysts, complexes with noninnocent ligands, magnetic exchange interactions, reaction intermediates during bond formation and breaking possibly involving spin-state crossings, these methods may be necessary to reach a certain level of accuracy but they may also be desirable in themselves because they facilitate a level of insight that is not readily accessible by other means (e.g., analyzing the nature of multireference states). Here, a balanced view of the role of ML might be as a facilitator/accelerator of such approaches rather than as an alternative to them. Similarly, for heterogeneous catalysis, post-DFT electronic-structure methods for periodic systems (such as GW, periodic RPA, embedding approaches, and local-correlation methods, , ) fulfill a comparable role by enhancing both the accuracy and interpretability of surface reaction mechanisms.
Protocols where ML is used in conjunction with classic quantum chemistry methods (e.g., in a Δ-ML fashion or through transfer learning), , such as DFT or correlated methods, have shown promise to improve accuracy for challenging problems. Some examples include data-driven quantum chemical methodologies that utilize low-level quantum chemical data for the prediction of higher-level wave functions (e.g., from MP2 → CCSD(T) − or CASSCF → CASPT2), which can be in principle extended to DLPNO–CCSD(T) for larger systems, extrapolation of NEVPT2 dynamic correlation corrections to estimate higher level corrections, such as NEVPT2 → NEVPT4, or on the automated active space selection. , These protocols can also be used in turn to train more computationally efficient models. Δ-ML and transfer learning combined with cluster-based parametrization strategies and/or molecular tailoring provide a roadmap to address problems in homogeneous and even heterogeneous catalysis.
Training of such models is typically limited to data sets of representative systems with few atoms, on which a large percentage of electron correlation is possible to be recovered, but it is questionable whether the results obtained using such data sets can be extrapolated to realistic systems. In contrast, parameters such as orbital entropy, spin populations, and related quantities can instead be used to represent the complexes in the training set. These parameters capture features that can be shared between both small and large systems, such as the local electronic structure around the metal center(s). In this way, the model can learn connections between correlation energies and these parameters, rather than relying directly on molecular geometries.
For computational chemists working with metalloenzymes, a central consideration is how ML can be incorporated into existing workflows of multilevel calculations (e.g., QM/MM) and facilitate improved embedding strategies for treatment of environmental effects and solvation spheres. We do not envisage–in the short term–a complete dismantling of established multicomponent workflows but anticipate ML approaches to provide unique solutions to sensitive or demanding steps. For example, a common bottleneck in QM/MM/MD calculations of metalloenzymes is the generation of appropriate force-field parameters for metallocofactors, which may even have to be changing from step to step when different catalytic intermediates are studied. Here ML/MLIPs could potentially be used to automate this step for metalloenzyme active sites, thus drastically accelerating the overall workflow ofotherwise conventionalmultiscale simulations. −
2.2. Generative AI
Speculation about the future may also touch on more transformative ideas. One may imagine a future in which foundation modelstrained on massive experimental and computational data setscould bypass traditional simulation-based studies of molecular systems. , In such a vision, we could imagine moving directly from a research question to an experimental observable, without explicitly modeling the molecular system. Though still aspirational, the emergence of self-driving laboratories, , real-time experimental data streams, and closed-loop AI systems point in this direction.
AI-driven laboratories will also require synergy between experimental and digital chemists to elevate the strengths of machine learning and develop models easily accessible to nonexperts. Web services or easy-to-use applications will allow chemists to use such models without prior programming experience.
Diffusion and flow-matching models are powerful new generative tools in chemistry that directly operate on 3D molecular geometries, providing a natural interface with traditional quantum chemistry modeling. Unlike autoregressive models, they do not generate structures sequentially but instead evolve entire geometries in continuous space, which improves validity and facilitates modeling of interactions relevant to reactivity. While much of their early success has been in drug discovery, , flow-matching models have been applied to generate transition-states, , in inverse design of organocatalysts and enzymes, with recent extensions demonstrating their applicability to heterogeneous catalysis. These capabilities suggest that generative models based on diffusion or flow matching may open new avenues for interpretable catalyst design and mechanistic discovery.
As MLIPs continue to redefine how we model potential energy surfaces with quantum-level accuracy, a parallel frontier is emerging at a very different level of abstraction: foundation models. In contemporary computational chemistry, the term “foundation model” has come to encompass two complementary developments. The first includes broadly transferable MLIPsdiscussed in the previous sectionwhich serve as universal energy and force predictors across chemical and materials spaces. The second refers to large language models (LLMs) trained on extensive chemical, molecular, or materials corpora. − Although both groups of models share underlying principles such as large-scale training, generalization, and cross-domain transferability, they differ fundamentally in their objectives, data modalities, and modes of interaction with chemical problems. While LLMs themselves still perform poorly in generating valid molecules or reactions from scratch, they are rapidly improving at interpreting, contextualizing, and reasoning about chemical information. These models are increasingly used by students and researchers for coding support, structure interpretation, and scientific ideation. LLMs have the potential to bridge traditional computational tools (such as DFT) with the broader research workflow, enabling more natural and agentic interactions with complex data pipelines. LLMs with specialized tools have shown promise for coding, processing of chemical information, and controlling lab automation. ,
However, are foundation models a panacea? A cautionary analogy might be offered by the story of the Universal Force Field (UFF), which was once promoted as a general solution for all chemistry but ultimately fell short due to its lack of precision and transferability. The same risk applies to universal machine learning models: while the allure of generality is strong, chemistry continues to demand rigor, specificity, and domain expertise. Even as ML tools become more powerful, careful model validation and deep chemical intuition remain essential. But we should keep in mind that foundation models offer, in effect, an implicit universal projection of diverse electronic structures into a compact analytic representation, enabled through careful retraining and fine-tuning. In many ways, they can be viewed as a natural continuation of the vision behind UFF: the pursuit of a broadly applicable analytic model of fundamental physical interactions. What is different today is the nature of the representations themselves. Modern ML representations are implicit rather than explicit, which gives them a surprising capacity for extrapolation and allows them to be rapidly retrained into highly accurate, problem-specific models. In this sense, foundation models extend and refine the goal that UFF originally sought to achieve.
Equally important is the conceptual formulation of the scientific questions we ask. In heterogeneous catalysis, while developing an accurate potential energy surface remains essential, a central difficulty often also lies in selecting an appropriate system and representation to model from the outset. AI/ML tools can help accelerate discovery, but they do not replace the need for human insight in defining the scope and relevance of a problem. However, recent advances on multiagent architectures and literature search agents have shown promise in generating new, original knowledge and formulating self-improving research hypotheses for automating scientific discoveries. ,
Clearly, a sense of both opportunity and caution must be exerted. AI and ML techniques are rapidly expanding the frontier of what is possible in computational chemistry, especially in areas like transition metal chemistry that have traditionally resisted automation. But the road ahead requires thoughtful integration of these tools with existing methodologies, careful attention to their limitations, and a commitment to preserving scientific rigor even in the face of accelerating automation. If pursued responsibly, these emerging technologies could dramatically enhance our ability to model, predict, and ultimately design the chemical systems of the future.
3. Molecular Representations
A complex and evolving challenge is how chemical systems are represented in a form suitable for computational modeling. , Whether designing new machine learning models, interpreting experimental data, or predicting physical properties, the way we encode chemical information into molecular representations plays a central role in determining what questions can be asked and what answers can be trusted.
Choosing the “right” molecular representation remains an open problem, one deeply tied to the nature of the task at hand. In contrast to fields like computer vision or natural language processing, where inputs are relatively standardized (images, text), chemistry offers a puzzling range of possible representations: text-based formats like SMILES, graph-based molecular topologies, three-dimensional geometries, electron densities, and quantum mechanical wave functions, among others. One may use text-based representations to take advantage of LLMs as general-purpose feature extractors. With relatively few training examples, these models can adapt representations on the fly to new tasks or reactions. However, this flexibility comes at a cost, as these representations are often highly abstract, and their interpretability is limited.
The trade-offs between accuracy and interpretability, generality and specificity, are important to understand. One may advocate for dynamic or task-specific representations, learned in tandem with the data set as it is collected. However, such adaptive methods may obscure the underlying chemistry and make it more difficult to rationalize model behavior. Just as in human scientific communities, diversity of perspectives, described as the romantic, the pragmatic, and the artistic, can be a strength, allowing the field to pursue multiple lines of inquiry without prematurely standardizing on a single “best” approach.
The abundance of representations in chemistry is both a challenge and an opportunity. Unlike many areas of machine learning where data are prevectorized, chemistry permitsand requiresinvestigation into how the form of representation affects model outcomes, which makes machine learning in chemistry a particularly rich domain: it opens the door to investigating more complex problems than are typically addressed by off-the-shelf ML benchmarks.
Yet despite the promise of learned or abstract representations, hand-crafted, expert-defined features still play an important role. Especially in data-scarce regimes or for chemically challenging systems, such as transition metal complexes, ,,− or heterogeneous catalysts, domain knowledge remains crucial for designing meaningful inputs. As the field matures and access to experimental and computational data grows, we may see a convergence where learned representations become more powerful, efficient, and easy to use, while still respecting the chemical constraints and principles that underlie molecular behavior.
One should be cautious against the temptation to search for a single unifying representation. The needs of spectroscopy differ from those of reaction prediction; electron density offers different information than molecular graphs or SMILES strings. The representation should not just describe the data but it should also reflect the specific modeling goal. A good representation is one that enables extrapolation, generalization, and problem-specific reasoning. And depending on the task (predicting spin crossover, , reactivity, or spectral lines) different aspects of a molecule’s physical or electronic structure may become relevant.
From a computational chemistry standpoint, structure-based representations remain the default starting point. But even here, nuances arise. For instance, many descriptors assume a single, static conformation, whereas chemical reality often involves ensembles of rotamers, conformers, tautomeric forms or resonance structures. − Thus, representations may need to move beyond idealized geometries to encode chemical states, incorporating information such as charge, spin multiplicity, solvation, or even temperature and concentration. Particularly in open-shell systems or complex electronic environments, these additional layers of description are critical for meaningful predictions.
In this context, electron density emerges as a particularly intriguing modality. − Although it contains all the information needed to reconstruct molecular structure and properties, it is computationally expensive to calculate accurately and difficult to work with directly. In this sense, much of modern representation theory in chemistry can be seen as constructing reduced models that compress the full electronic description into something tractable and application-ready, such as, for example, the smooth overlap of atomic positions (SOAP) representation, which can be understood as a simplified easy-to-evaluate model of molecular electron densities.
Looking ahead, the importance of multimodal representations is to be highlighted. In real chemical workflows, a molecule is rarely defined by a single modality. Structural data may be complemented by spectra, thermodynamic observables, or reaction yields. As such, the ability to combine different forms of informationtext, graphs, images, numerical spectrainto a unified model offers a promising path forward. Multimodal machine learning frameworks, which can integrate and reason over diverse input types, may be key to unlocking more powerful and generalizable models. This approach also reflects experimental practice more accurately: a chemist rarely sees just a molecule but rather a chromatogram, an NMR spectrum, a color change in solution.
Ultimately, representation is not just about how data are encoded. It is also about how information is preserved, transformed, and used to answer meaningful scientific questions. While machine learning today leans heavily on large data sets to learn patterns, traditional quantum chemistry began from minimal assumptions and sought predictive power from first principles. Future methods will likely benefit from bridging these paradigms: integrating the rigor of physics-based models with the flexibility and scalability of data-driven approaches. , This fusion could yield representations that are not only accurate and efficient, but also transparent and chemically meaningful, enabling both prediction and understanding in complex molecular systems.
4. Bridging the Model-Mechanism Gap
Returning to catalysis, the central theme of this perspective, we next consider one of its most enduring challenges: the elucidation of catalytic reaction mechanisms. From a computational perspective, mechanism discovery often demands years of effort, requiring not only the identification of the most probable pathway but also exploration of multiple competing steps. This naturally raises the question: how can we leverage machine learning, in particular physics-informed or knowledge-embedded approaches, to accelerate and enrich mechanistic studies? ,
Experimental practice in mechanistic investigations of homogeneous or enzymatic transition metal catalysts typically rests on correlating spectroscopic observables with structural parameters of isolatable intermediates. Beyond simple structural characterization, these spectroscopic observables exclusively report on electronic structure, which is crucial to rationalize differences in efficiencies or mechanistic pathways between different catalysts. Although a lot of emphasis has been placed on an energy-based view of computational investigations (potential energy surfaces, transition states), a more immediate and still invaluable goal would be the use of ML for proposing correlations between (experimental) spectroscopy such as EPR, XAS, and Mössbauer, and molecular/electronic structure. This can be a “final answer” in itself, but could also be seen as supplanting a major part of the work of computational chemists, the part involving hypothesizing/conceptualizing molecular models that would be subsequently evaluated for their spectroscopic properties via quantum chemical methods.
While AI continues to advance, we must not overlook the enduring importance of domain knowledge. ,, Traditional organometallic catalysis, for instance, benefits from decades of mechanistic insights. Reaction mechanisms are typically described as a succession of established reaction steps, including oxidative addition, ligand dissociation, and transmetalation. Therefore, a way to approach the problem computationally is to determine the order of these steps and the catalyst structure, yielding the lowest energy pathway. Knowledge of these steps often allows domain experts to navigate complex systems more efficiently than those approaching the problem purely from theory or computation and serves as a valuable first filter, particularly when mapping out initial steps or identifying plausible intermediates. Techniques such as ML-assisted transition state searches and microkinetic modeling can then refine this initial scaffold, particularly when augmented by experimental kinetic data.
For researchers less adept at mechanistic intuition, advances in reaction network modeling, such as those pioneered by groups working on automatic reaction exploration, can be transformative. Over time, these tools could evolve into powerful discovery frameworks, uncovering off-cycle pathways, decomposition channels, or unforeseen bifurcations in systems that resist traditional mechanistic analysis.
A productive path forward may lie less in further constraining models with additional physical or chemical principles and more in broadening access to high-quality, diverse data sets derived from both computation and experiment. , For example, if one had infinite kinetic data sets, , or well-characterized molecular simulations accounting for solvation and entropy, a data-driven approach could be sufficiently expressive without human-imposed biases.
However, this raises a provocative counterpoint: does encoding so much prior knowledge into our models limit the discovery potential of AI? Relying too heavily on textbook mechanisms, intuition, or handcrafted features risks reinforcing existing biases, driving models toward familiar solutions and potentially away from more innovative or unexpected ones. If large models are trained only on known chemistry, can they truly generalize or extrapolate?
The role of dynamics and entropy in computational studies of transition metal complexes is an underexplored but critical dimension. Traditional approaches (both computational and human-driven) often emphasize energy landscapes and static structures. Yet, bifurcating transition states, entropic barriers, and nonstatistical dynamics likely play a much larger role than typically appreciated. Capturing such behavior may require novel descriptors or models that treat mechanistic pathways as ensembles rather than linear sequences.
With increasing model complexity, interpretability re-emerges as a central consideration in ensuring scientific validity. , Could a future model such as, for example, an LLM trained on domain-specific data, strike a balance between interpretability and prediction? Natural language interfaces to such models could actually improve transparency, allowing users to query mechanisms or compare alternatives in human-understandable terms. However, the application of LLMs or similar models in catalysis faces two important challenges. First, the data scarcity in this field prevents the development or fine-tuning of LLMs, while lower complexity statistical models such as decision tree ensembles or Gaussian processes are used more frequently. Second, the lack of clarity around the exact structure of the catalystspecifically, how it can be characterized with limited information, typically restricted to reaction conditions, synthesis parameters, and averaged elemental descriptors from nominal loadingsmakes it difficult to develop the kind of precise representation needed for concise embedding in an LLM. This is in contrast to fully defined molecular structures, which can be easily tokenized via SMILES.
Ultimately, while AI will undoubtedly transform how we explore catalytic mechanisms, experimental validation will remain the final arbiter. Efforts in automation and self-driving laboratories, , are a path toward fully integrated platforms for hypothesis generation, validation, and iterative improvement. Yet even here, some studies caution that smart sampling strategies (e.g., active learning) may be less impactful than how we represent the underlying chemical space.
We are best advised to remain thoughtful about the balance between automation and human input. While automation may accelerate exploration, the intuition honed through years of experimentation and theory still plays a pivotal role, especially in under-characterized or ill-defined systems, such as many homogeneous or heterogeneous catalysts. Moving forward, the field may benefit from embracing both extremes: large, expressive models trained on extensive data, and targeted, domain-informed approaches where prior knowledge provides critical context. Machine learning models could also be trained to capture chemists’ intuition in a similar way as has been attempted in the drug discovery space. ,
In terms of catalyst optimization, emphasis is usually placed on the first coordination sphere of a transition metal ion. Bioinorganic catalysis however also encompasses metalloenzymes, and in this case the transferability of the developed approaches is not obvious. It is likely that quite distinct ML tools would have to be combined into multicomponent workflows, borrowing from more “locally focused” domains as well from domains focusing on directed enzyme evolution, recognizing that “textbook”-guided considerations of ligand optimization in standard coordination chemistry can quickly become even irrelevant when the “ligand” is a functionally important protein matrix. These approaches could be integrated in a bottom-up protocol for enzyme discovery acceleration, for example (re)designing enzyme function. , Distinct goals will have to be combined in order to construct a multidimensional surface correlating structural and electronic parameters of the active site with catalytic activity and with amino acid sequence, and ultimately to locate combinations that maximize overall function. Besides conventional conceptions of reactivity, repurposing natural or designing artificial metalloenzymes would need to consider multiple other factors, such as robustness, sensitivity of different parts of the system to specific conditions, optimization of substrate delivery pathways and proton channels, or even allostery for controlling selectivity and function, or for signaling between catalytic components (feedback).
Finally, in the realm of catalyst design, we anticipate significant breakthroughs in the coming years driven by inverse design strategies and generative models. ,, The effectiveness of training generative models depends critically on the benchmark tasks they are designed for. While initial molecular design tasks focused on simplistic objectives such as molecular rediscovery and proxies such as quantitative estimates of drug likeness, benchmarks more tailored toward realistic targets have been developed. Other important aspects are the focus on sample-efficiency, and beating relevant baseline models such as genetic algorithms.
5. Chemical Datasets
Another difficulty in applying machine learning in catalysis research lies in data collection for the development of supervised learning models with predictive power. There is scarcity of computationally refined databases of transition metal complexes that machine learning can capitalize on, ,− although there are ligand data sets that can be used for combinatorial expansions and genetic algorithms. − Another known limitation is the lack of negative reaction outcomes that are not published by chemists, even though they are highly valuable to train machine learning models to avoid a positive reaction bias. , A focus must therefore be put on encouraging and incentivizing chemists to also publish negative reaction outcomes, a culture shift which requires a long community effort. This can be facilitated with electronic lab notebooks, in which all data are stored in a systematic and concise format. Recent studies have shown that the explicit generation of negative reaction data can help the models to better distinguish between reactive and unreactive substrates. Such models can then be used to identify further negative reactions that can be executed with high-throughput experimentation equipment to augment current unbalanced data sets. As awareness of this topic grows within the field, we are confident that the community will continue to move in the right direction. Data from automation platforms will also likely contain a larger percentage of failed reactions. Another important aspect is how to format reaction data once they are recorded since the role of chemical species in reactions is often unclear. We envision that the usage of electronic lab notebooks and open reaction databases will provide more interoperable and reusable reaction data, facilitating machine learning efforts. Such database-oriented format might also be combined with user-friendly spreadsheet formats at the time of input. The development of more sophisticated models and data infrastructure enables the community to move beyond standard data sets, such as the United States Patent and Trademark Office (USPTO), which is dominated by industrially relevant reactions such as the Buchwald–Hartwig or Suzuki coupling. Making efficient use of these sparser data is therefore highly important, e.g. through transfer learning or the previously alluded to chemistry-informed models. Finally, the validation of reaction standardization platforms will critically depend on the involvement of experimental chemists. As the primary data providers, it is essential to maintain a careful balance by offering them a flexible recording format that does not constrain their workflow while ensuring the data remains useable for machine learning applications.
A critical challenge in modern chemical research lies in reconciling experimental and computational data sets, particularly in terms of their reliability, resolution, and mutual alignment, despite several “best practice” rules that have been recommended in the literature. − For example, while data sets of computed activation energies − have helped fuel model development, the end goal is to predict experimental values. For many reaction types, agreement between computation and experiment remains challenging due to many factors. Machine learning models that bridge the computed activation and experimental activation energies through delta or transfer learning is a promising strategy. , Using MLIPs and free energy simulations can also enable the generation of higher quality data sets from computations.
6. Conclusions
A well-known old quote attributed to Eugene Wigner (but probably apocryphal) notes that “It is nice to know that the computer understands the problem, but I would like to understand the problem, too”. It is hard to believe that at Wigner’s time, a computer that delivered some numerical result could be considered to have understood this result. Certainly the quote was referring to the fact that numerical data alone, no matter how accurate and reliable, do not allow us to grasp a scientific problem. However, we have now approached a qualitatively new reality in computational science in general, where machine learning models do recognize patterns invisible to the human eye and maybe even incomprehensible by humans. Now, based on their “insights”, computers may actually take decision to perform new calculations or even experiments autonomously, and therefore, we have now really come to the situation where we can truly say “it is nice that the computer understands a problem, but how and to what degree can we understand it, too?”
In this perspective, we focused on how we can bridge such “insights” from chemical knowledge and intuition with the role of modern AI/ML in accelerating molecular discovery, particularly in catalysis. Machine learning has begun transforming how researchers navigate chemical space and identify candidates for new catalysts, yet significant challenges remain. Model performance often hinges on data quality, and while generative models, graph neural networks, and LLMs show promise, careful representation of chemical knowledge remains critical. Domain expertise may be leveraged to guide model training, but one must not restrict the ability of letting models discover novel patterns without human bias, pointing to trade-offs between interpretability, physical intuition, and empirical discovery, that will need to be balanced depending on the application and availability of data.
In bridging the gap between mechanistic understanding and AI-driven modeling, which is a foundational concern in catalysis, the rich body of mechanistic knowledge found in textbooks and chemical intuition could still serve as valuable inputs when designing machine learning workflows. The rather new promise of automated reaction network exploration with coupled microkinetic modeling, requires us to ponder the trade-offs between incorporating prior knowledge versus enabling models to learn from data alone. As the field moves toward self-driving laboratories and closed-loop discovery, a future can be envisioned where models can generate hypotheses, guide experiments, and build new knowledge autonomously. Yet, without interpretability, domain specificity, and experimental validation, scientific understanding may be compromised. The balance between automation and human insight remains a defining question for the future of AI in chemistry and catalysis. It is therefore essential to emphasize the role of scientific expertise, ensuring that AI remains a powerful tool to augment, rather than replace, human reasoningcontrary to the perception that it offers a “panacea” or an “effortless cure”.
Acknowledgments
K.D.V. acknowledges the Collegium Helveticum International Fellowship Program, which partially supports his Sabbatical visit at ETH Zürich, Switzerland, and the National Science Foundation under grant no. 2143354 (CAREER: CAS-Climate). A.N. acknowledges financial support from the Research Council of Norway through the Centre of Excellence (no. 262695). M.M. gratefully acknowledges financial support from the Swiss National Science Foundation through grants 200020_219779, 200021_215088, and the University of Basel. D.A.P. acknowledges support by the Max Planck Society. M.D. acknowledges support from the Alexander von Humboldt Stiftung (Humboldt Research Fellowship) and from the Fonds der Chemischen Industrie (Liebig Fellowship). This work was also created as part of NCCR Catalysis (grant number 180544), a National Centre of Competence in Research funded by the Swiss National Science Foundation. The authors would also like to thank Prof. Christophe Copéret for fruitful discussions.
The authors declare no competing financial interest.
References
- Bräse S.. Digital chemistry: Navigating the confluence of computation and experimentation–definition, status quo, and future perspective. Digital Discovery. 2024;3(10):1923–1932. doi: 10.1039/D4DD00130C. [DOI] [Google Scholar]
- Varnek A., Baskin I.. Machine learning methods for property prediction in chemoinformatics: Quo Vadis? J. Chem. Inf. Model. 2012;52(6):1413–1437. doi: 10.1021/ci200409x. [DOI] [PubMed] [Google Scholar]
- Muratov E. N., Bajorath J., Sheridan R. P.. et al. QSAR without borders. Chem. Soc. Rev. 2020;49(11):3525–3564. doi: 10.1039/D0CS00098A. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Rodríguez-Pérez R., Miljković F., Bajorath J.. Machine learning in chemoinformatics and medicinal chemistry. Annu. Rev. Biomed. Data Sci. 2022;5(1):43–65. doi: 10.1146/annurev-biodatasci-122120-124216. [DOI] [PubMed] [Google Scholar]
- Noé F., Tkatchenko A., Müller K.-R., Clementi C.. Machine learning for molecular simulation. Annu. Rev. Phys. Chem. 2020;71(1):361–390. doi: 10.1146/annurev-physchem-042018-052331. [DOI] [PubMed] [Google Scholar]
- Deringer V. L., Bartók A. P., Bernstein N., Wilkins D. M., Ceriotti M., Csányi G.. Gaussian process regression for materials and molecules. Chem. Rev. 2021;121(16):10073–10141. doi: 10.1021/acs.chemrev.1c00022. [DOI] [PMC free article] [PubMed] [Google Scholar]
- von Lilienfeld O. A., Müller K.-R., Tkatchenko A.. Exploring chemical compound space with quantum-based machine learning. Nat. Rev. Chem. 2020;4(7):347–358. doi: 10.1038/s41570-020-0189-9. [DOI] [PubMed] [Google Scholar]
- Huang B., von Lilienfeld O. A.. Ab initio machine learning in chemical compound space. Chem. Rev. 2021;121(16):10001–10036. doi: 10.1021/acs.chemrev.0c01303. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Meuwly M.. Machine learning for chemical reactions. Chem. Rev. 2021;121(16):10218–10239. doi: 10.1021/acs.chemrev.1c00033. [DOI] [PubMed] [Google Scholar]
- Keith J. A., Vassilev-Galindo V., Cheng B.. et al. Combining machine learning and computational chemistry for predictive insights into chemical systems. Chem. Rev. 2021;121(16):9816–9872. doi: 10.1021/acs.chemrev.1c00107. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Dou B., Zhu Z., Merkurjev E.. et al. Machine learning methods for small data challenges in molecular science. Chem. Rev. 2023;123(13):8736–8780. doi: 10.1021/acs.chemrev.3c00189. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Anstine D. M., Isayev O.. Generative models as an emerging paradigm in the chemical sciences. J. Am. Chem. Soc. 2023;145(16):8736–8750. doi: 10.1021/jacs.2c13467. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Palos E., Dasgupta S., Lambros E., Paesani F.. Data-driven many-body potentials from density functional theory for aqueous phase chemistry. Chem. Phys. Rev. 2023;4(1):2688–4070. doi: 10.1063/5.0129613. [DOI] [Google Scholar]
- Aspuru-Guzik A., Lindh R., Reiher M.. The matter simulation (r)evolution. ACS Cent. Sci. 2018;4(2):144–152. doi: 10.1021/acscentsci.7b00550. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Elton D. C., Boukouvalas Z., Fuge M. D., Chung P. W.. Deep learning for molecular designa review of the state of the art. Mol. Sys. Des. Eng. 2019;4(4):828–849. doi: 10.1039/C9ME00039A. [DOI] [Google Scholar]
- von Lilienfeld O. A., Burke K.. Retrospective on a decade of machine learning for chemical discovery. Nat. Commun. 2020;11(1):4895. doi: 10.1038/s41467-020-18556-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Zhao L., Ciallella H. L., Aleksunes L. M., Zhu H.. Advancing computer-aided drug discovery (CADD) by big data and data-driven machine learning modeling. Drug Discovery Today. 2020;25(9):1624–1638. doi: 10.1016/j.drudis.2020.07.005. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Stokes J. M., Yang K., Swanson K., Jin W., Cubillos-Ruiz A., Donghia N. M., MacNair C. R., French S., Carfrae L. A., Bloom-Ackermann Z.. et al. A deep learning approach to antibiotic discovery. Cell. 2020;180(4):688–702e13. doi: 10.1016/j.cell.2020.01.021. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Jeon J., Kang S., Kim H. U.. Predicting biochemical and physiological effects of natural products from molecular structures using machine learning. Nat. Prod. Rep. 2021;38(11):1954–1966. doi: 10.1039/D1NP00016K. [DOI] [PubMed] [Google Scholar]
- Tiihonen A., Cox-Vazquez S. J., Liang Q., Ragab M., Ren Z., Hartono N. T. P., Liu Z., Sun S., Zhou C., Incandela N. C.. et al. Predicting antimicrobial activity of conjugated oligoelectrolyte molecules via machine learning. J. Am. Chem. Soc. 2021;143(45):18917–18931. doi: 10.1021/jacs.1c05055. [DOI] [PubMed] [Google Scholar]
- Zeng X., Wang F., Luo Y., Kang S. g., Tang J., Lightstone F. C., Fang E. F., Cornell W., Nussinov R., Cheng F.. Deep generative molecular design reshapes drug discovery. Cell Rep. Med. 2022;3(12):100794. doi: 10.1016/j.xcrm.2022.100794. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Offensperger F., Tin G., Duran-Frigola M.. et al. Large-scale chemoproteomics expedites ligand discovery and predicts ligand behavior in cells. Science. 2024;384(6694):eadk5864. doi: 10.1126/science.adk5864. [DOI] [PubMed] [Google Scholar]
- Nandy A., Duan C., Taylor M. G., Liu F., Steeves A. H., Kulik H. J.. Computational discovery of transition-metal complexes: From high-throughput screening to machine learning. Chem. Rev. 2021;121(16):9927–10000. doi: 10.1021/acs.chemrev.1c00347. [DOI] [PubMed] [Google Scholar]
- Kulik H. J.. Making machine learning a useful tool in the accelerated discovery of transition metal complexes. Wiley Interdiscip. Rev.:Comput. Mol. Sci. 2020;10(1):e1439. doi: 10.1002/wcms.1439. [DOI] [Google Scholar]
- Jones G. M., Smith B. A., Kirkland J. K., Vogiatzis K. D.. Data-driven ligand field exploration of Fe(IV)–oxo sites for C–H activation. Inorg. Chem. Front. 2023;10(4):1062–1075. doi: 10.1039/D2QI01961B. [DOI] [Google Scholar]
- Chernyshov I. Y., Pidko E. A.. Mace: Automated assessment of stereochemistry of transition metal complexes and its applications in computational catalysis. J. Chem. Theory Comput. 2024;20(5):2313–2320. doi: 10.1021/acs.jctc.3c01313. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Watson M. P., Weix D. J.. The once and future catalysts: How the challenges of first-row transition-metal catalysis grew to become strengths. Acc. Chem. Res. 2024;57(17):2451–2452. doi: 10.1021/acs.accounts.4c00496. [DOI] [PubMed] [Google Scholar]
- Morán-González L., Burnage A. L., Nova A., Balcells D.. AI approaches to homogeneous catalysis with transition metal complexes. ACS Catal. 2025;15(11):9089–9105. doi: 10.1021/acscatal.5c01202. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Chilton N. F.. Molecular magnetism. Annu. Rev. Mater. Res. 2022;52(1):79–101. doi: 10.1146/annurev-matsci-081420-042553. [DOI] [Google Scholar]
- Hovey J. L., Dittrich T. M., Allen M. J.. Coordination chemistry of surface-associated ligands for solid–liquid adsorption of rare-earth elements. J. Rare Earths. 2023;41(1):1–18. doi: 10.1016/j.jre.2022.05.012. [DOI] [Google Scholar]
- Käser S., Boittier E. D., Upadhyay M., Meuwly M.. Transfer learning to CCSD(T): Accurate anharmonic frequencies from machine learning models. J. Chem. Theory Comput. 2021;17(6):3687–3699. doi: 10.1021/acs.jctc.1c00249. [DOI] [PubMed] [Google Scholar]
- Käser S., Meuwly M.. Transfer learned potential energy surfaces: Accurate anharmonic vibrational dynamics and dissociation energies for the formic acid monomer and dimer. Phys. Chem. Chem. Phys. 2022;24(9):5269–5281. doi: 10.1039/D1CP04393E. [DOI] [PMC free article] [PubMed] [Google Scholar]
- King-Smith E.. Transfer learning for a foundational chemistry model. Chem. Sci. 2024;15(14):5143–5151. doi: 10.1039/D3SC04928K. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Noto N., Kunisada R., Rohlfs T.. et al. Transfer learning across different photocatalytic organic reactions. Nat. Commun. 2025;16(1):3388. doi: 10.1038/s41467-025-58687-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Lan T., An Q.. Discovering catalytic reaction networks using deep reinforcement learning from first-principles. J. Am. Chem. Soc. 2021;143(40):16804–16812. doi: 10.1021/jacs.1c08794. [DOI] [PubMed] [Google Scholar]
- Margraf J. T., Jung H., Scheurer C., Reuter K.. Exploring catalytic reaction networks with machine learning. Nat. Catal. 2023;6(2):112–121. doi: 10.1038/s41929-022-00896-y. [DOI] [Google Scholar]
- Hoque A., Surve M., Kalyanakrishnan S., Sunoj R. B.. Reinforcement learning for improving chemical reaction performance. J. Am. Chem. Soc. 2024;146:28250–28267. doi: 10.1021/jacs.4c08866. [DOI] [PubMed] [Google Scholar]
- Lan T., Wang H., An Q.. Enabling high throughput deep reinforcement learning with first principles to investigate catalytic reaction mechanisms. Nat. Commun. 2024;15(1):6281. doi: 10.1038/s41467-024-50531-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Xie J., Wang Y., Rao J., Zheng S., Yang Y.. Self-supervised contrastive molecular representation learning with a chemical synthesis knowledge graph. J. Chem. Inf. Model. 2024;64(6):1945–1954. doi: 10.1021/acs.jcim.4c00157. [DOI] [PubMed] [Google Scholar]
- Loeffler H. H., He J., Tibo A.. et al. Reinvent 4: Modern AI–driven generative molecule design. J. Cheminf. 2024;16(1):20. doi: 10.1186/s13321-024-00812-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Guo J., Schwaller P.. Saturn: Sample-efficient generative molecular design using memory manipulation. arXiv. 2024:arXiv:2405.17066. doi: 10.48550/arXiv.2405.17066. [DOI] [Google Scholar]
- Born J., Manica M.. Regression transformer enables concurrent sequence regression and generation for molecular language modelling. Nat. Mach. Intell. 2023;5(4):432–444. doi: 10.1038/s42256-023-00639-z. [DOI] [Google Scholar]
- Su Y., Wang X., Ye Y., Xie Y., Xu Y., Jiang Y., Wang C.. Automation and machine learning augmented by large language models in a catalysis study. Chem. Sci. 2024;15(31):12200–12233. doi: 10.1039/D3SC07012C. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Mok D. H., Back S.. Generative pretrained transformer for heterogeneous catalysts. J. Am. Chem. Soc. 2024;146(49):33712–33722. doi: 10.1021/jacs.4c11504. [DOI] [PubMed] [Google Scholar]
- Morandi S., Loveday O., Renningholtz T., Pablo-García S., Vargas-Hernández R. A., Seemakurthi R. R., Sanz Berman P., García-Muelas R., Aspuru-Guzik A., López N.. A foundational model for reaction networks on metal surfaces. ChemRxiv. 2024:2024-bfv3d. doi: 10.26434/chemrxiv-2024-bfv3d. [DOI] [Google Scholar]
- Choi J., Nam G., Choi J., Jung Y.. A perspective on foundation models in chemistry. JACS Au. 2025;5(4):1499–1518. doi: 10.1021/jacsau.4c01160. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Feng M., Zhao C., Day G. M., Evangelopoulos X., Cooper A. I.. A universal foundation model for transfer learning in molecular crystals. Chem. Sci. 2025;16:12844–12859. doi: 10.1039/d5sc00677e. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Abbott A. S., Turney J. M., Zhang B., Smith D. G. A., Altarawy D., Schaefer H. F.. PES-Learn: An open-source software package for the automated generation of machine learning models of molecular potential energy surfaces. J. Chem. Theory Comput. 2019;15(8):4386–4398. doi: 10.1021/acs.jctc.9b00312. [DOI] [PubMed] [Google Scholar]
- Töpfer K., Vázquez-Salazar L. I., Meuwly M.. Asparagus: A toolkit for autonomous, user-guided construction of machine-learned potential energy surfaces. Comput. Phys. Commun. 2025;308:109446. doi: 10.1016/j.cpc.2024.109446. [DOI] [Google Scholar]
- David R., de la Puente M., Gomez A., Anton O., Stirnemann G., Laage D.. ArcaNN: Automated enhanced sampling generation of training sets for chemically reactive machine learning interatomic potentials. Digital Discovery. 2025;4(1):54–72. doi: 10.1039/D4DD00209A. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Zhang Y., Wang H., Chen W., Zeng J., Zhang L., Wang H., E W.. DP-GEN: A concurrent learning platform for the generation of reliable deep learning based potential energy models. Comput. Phys. Commun. 2020;253:107206. doi: 10.1016/j.cpc.2020.107206. [DOI] [Google Scholar]
- Gkeka P., Stoltz G., Barati Farimani A.. et al. Machine learning force fields and coarse-grained variables in molecular dynamics: Application to materials and biological systems. J. Chem. Theory Comput. 2020;16(8):4757–4775. doi: 10.1021/acs.jctc.0c00355. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Gokcan H., Isayev O.. Learning molecular potentials with neural networks. Wiley Interdiscip. Rev.:Comput. Mol. Sci. 2022;12(2):e1564. doi: 10.1002/wcms.1564. [DOI] [Google Scholar]
- Friederich P., Häse F., Proppe J., Aspuru-Guzik A.. Machine-learned potentials for next-generation matter simulations. Nat. Mater. 2021;20(6):750–761. doi: 10.1038/s41563-020-0777-6. [DOI] [PubMed] [Google Scholar]
- Pinheiro M., Ge F., Ferré N., Dral P. O., Barbatti M.. Choosing the right molecular machine learning potential. Chem. Sci. 2021;12(43):14396–14413. doi: 10.1039/d1sc03564a. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Huang B., von Rudorff G. F., von Lilienfeld O. A.. The central role of density functional theory in the AI age. Science. 2023;381(6654):170–175. doi: 10.1126/science.abn3445. [DOI] [PubMed] [Google Scholar]
- Kulichenko M., Nebgen B., Lubbers N., Smith J. S., Barros K., Allen A. E. A., Habib A., Shinkle E., Fedik N., Li Y. W.. et al. Data generation for machine learning interatomic potentials and beyond. Chem. Rev. 2024;124(24):13681–13714. doi: 10.1021/acs.chemrev.4c00572. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kovács D. P., Moore J. H., Browning N. J.. et al. MACE-OFF: Short-range transferable machine learning force fields for organic molecules. J. Am. Chem. Soc. 2025;147(21):17598–17611. doi: 10.1021/jacs.4c07099. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Levine D. S., Shuaibi M., Spotte-Smith E. W. C.. et al. The open molecules 2025 (OMol25) dataset, evaluations, and models. arXiv. 2025:arXiv:2505.08762. doi: 10.48550/arXiv.2505.08762. [DOI] [Google Scholar]
- Li J., Qu C., Bowman J. M.. Diffusion Monte Carlo with fictitious masses finds holes in potential energy surfaces. Mol. Phys. 2021;119(17–18):e1976426. doi: 10.1080/00268976.2021.1976426. [DOI] [Google Scholar]
- Jiang B., Li J., Guo H.. High-fidelity potential energy surfaces for gas-phase and gas–surface scattering processes from machine learning. J. Phys. Chem. Lett. 2020;11(13):5120–5131. doi: 10.1021/acs.jpclett.0c00989. [DOI] [PubMed] [Google Scholar]
- Westermayr J., Marquetand P.. Machine learning and excited-state molecular dynamics. Mach. Learn.: Sci. Technol. 2020;1(4):043001. doi: 10.1088/2632-2153/ab9c3e. [DOI] [Google Scholar]
- Manzhos S., Carrington T.. Neural network potential energy surfaces for small molecules and reactions. Chem. Rev. 2020;121(16):10187–10217. doi: 10.1021/acs.chemrev.0c00665. [DOI] [PubMed] [Google Scholar]
- Li J., Lopez S. A.. A look inside the black box of machine learning photodynamics simulations. Acc. Chem. Res. 2022;55(14):1972–1984. doi: 10.1021/acs.accounts.2c00288. [DOI] [PubMed] [Google Scholar]
- Kulichenko M., Barros K., Lubbers N.. et al. Uncertainty-driven dynamics for active learning of interatomic potentials. Nat. Comput. Sci. 2023;3(3):230–239. doi: 10.1038/s43588-023-00406-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Mace S., Xu Y., Nguyen B. N.. Automated transition metal catalysts discovery and optimization with AI and machine learning. ChemCatChem. 2024;16(10):e202301475. doi: 10.1002/cctc.202301475. [DOI] [Google Scholar]
- Vazquez-Salazar L. I., Käser S., Meuwly M.. Outlier-detection for reactive machine learned potential energy surfaces. npj Comput. Mater. 2025;11(1):33. doi: 10.1038/s41524-024-01473-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Cohen A. J., Mori-Sánchez P., Yang W.. Challenges for density functional theory. Chem. Rev. 2012;112(1):289–320. doi: 10.1021/cr200107z. [DOI] [PubMed] [Google Scholar]
- Jones, G. M. ; Pathirage, P. S. ; Vogiatzis, K. D. . Data-driven acceleration of coupled-cluster and perturbation theory methods. In Quantum Chemistry in the Age of Machine Learning; Elsevier, 2023; pp 509–529. [Google Scholar]
- Upadhyay M., Töpfer K., Meuwly M.. Molecular simulation for atmospheric reactions: Non-equilibrium dynamics, roaming, and glycolaldehyde formation following photoinduced decomposition of syn-acetaldehyde oxide. J. Phys. Chem. Lett. 2024;15:90–96. doi: 10.1021/acs.jpclett.3c03131. [DOI] [PubMed] [Google Scholar]
- Westermayr J., Marquetand P.. Machine learning for electronically excited states of molecules. Chem. Rev. 2021;121(16):9873–9926. doi: 10.1021/acs.chemrev.0c00749. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Li J., Lopez S. A.. Machine learning accelerated photodynamics simulations. Chem. Phys. Rev. 2023;4(3):2688–4070. doi: 10.1063/5.0159247. [DOI] [Google Scholar]
- Froitzheim T., Müller M., Hansen A., Grimme S.. g-xTB: A general-purpose extended tight-binding electronic structure method for the elements H to Lr (Z = 1–103) ChemRxiv. 2025:2025-bjxvt. doi: 10.26434/chemrxiv-2025-bjxvt. [DOI] [Google Scholar]
- Mermigki M. A., Drosou M., Pantazis D. A.. On the nature of high-spin forms in the S2 state of the oxygen-evolving complex. Chem. Sci. 2025;16(9):4023–4047. doi: 10.1039/D4SC07818G. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Reining L.. The GW approximation: Content, successes and limitations. Wiley Interdiscip. Rev.:Comput. Mol. Sci. 2018;8(3):e1344. doi: 10.1002/wcms.1344. [DOI] [Google Scholar]
- Ren X., Rinke P., Joas C., Scheffler M.. Random-phase approximation and its applications in computational chemistry and materials science. J. Mater. Sci. 2012;47(21):7447–7471. doi: 10.1007/s10853-012-6570-4. [DOI] [Google Scholar]
- Wesolowski T. A., Shedge S., Zhou X.. Frozen-density embedding strategy for multilevel simulations of electronic structure. Chem. Rev. 2015;115(12):5891–5928. doi: 10.1021/cr500502v. [DOI] [PubMed] [Google Scholar]
- Wouters S., Jiménez-Hoyos C. A., Sun Q., Chan G. K.-L.. A practical guide to density matrix embedding theory in quantum chemistry. J. Chem. Theory Comput. 2016;12(6):2706–2719. doi: 10.1021/acs.jctc.6b00316. [DOI] [PubMed] [Google Scholar]
- Hermes M. R., Gagliardi L.. Multiconfigurational self-consistent field theory with density matrix embedding: The localized active space self-consistent field method. J. Chem. Theory Comput. 2019;15(2):972–986. doi: 10.1021/acs.jctc.8b01009. [DOI] [PubMed] [Google Scholar]
- Ramakrishnan R., Dral P. O., Rupp M., von Lilienfeld O. A.. Big data meets quantum chemistry approximations: The Δ-machine learning approach. J. Chem. Theory Comput. 2015;11(5):2087–2096. doi: 10.1021/acs.jctc.5b00099. [DOI] [PubMed] [Google Scholar]
- Townsend J., Vogiatzis K. D.. Data-driven acceleration of the coupled-cluster singles and doubles iterative solver. J. Phys. Chem. Lett. 2019;10(14):4129–4135. doi: 10.1021/acs.jpclett.9b01442. [DOI] [PubMed] [Google Scholar]
- Townsend J., Vogiatzis K. D.. Transferable MP2-based machine learning for accurate coupled-cluster energies. J. Chem. Theory Comput. 2020;16(12):7453–7461. doi: 10.1021/acs.jctc.0c00927. [DOI] [PubMed] [Google Scholar]
- Pathirage P. D. V. S., Phillips J. T., Vogiatzis K. D.. Exploration of the two-electron excitation space with data-driven coupled cluster. J. Phys. Chem. A. 2024;128(10):1938–1947. doi: 10.1021/acs.jpca.3c06600. [DOI] [PubMed] [Google Scholar]
- Pathirage P. D. V. S., Quebedeaux B., Akram S., Vogiatzis K. D.. Transferability across different molecular systems and levels of theory with the data-driven coupled-cluster scheme. J. Phys. Chem. A. 2025;129(13):2988–2997. doi: 10.1021/acs.jpca.4c05718. [DOI] [PubMed] [Google Scholar]
- Jones G. M., Vogiatzis K. D.. Capturing electron correlation with machine learning through a data-driven CASPT2 framework. J. Chem. Theory Comput. 2025;21(21):10879–10892. doi: 10.1021/acs.jctc.5c01333. [DOI] [PubMed] [Google Scholar]
- Jeong W., Stoneburner S. J., King D.. et al. Automation of active space selection for multireference methods via machine learning on chemical bond dissociation. J. Chem. Theory Comput. 2020;16(4):2389–2399. doi: 10.1021/acs.jctc.9b01297. [DOI] [PubMed] [Google Scholar]
- Golub P., Antalik A., Veis L., Brabec J.. Machine learning-assisted selection of active spaces for strongly correlated transition metal systems. J. Chem. Theory Comput. 2021;17(10):6053–6072. doi: 10.1021/acs.jctc.1c00235. [DOI] [PubMed] [Google Scholar]
- Töpfer K., Boittier E., Devereux M., Pasti A., Hamm P., Meuwly M.. Force fields for deep eutectic mixtures: Application to structure, thermodynamics and 2D-infrared spectroscopy. J. Phys. Chem. B. 2024;128(44):10937–10949. doi: 10.1021/acs.jpcb.4c05480. [DOI] [PubMed] [Google Scholar]
- Sahu N., Gadre S. R.. Molecular tailoring approach: A route forab initiotreatment of large clusters. Acc. Chem. Res. 2014;47(9):2739–2747. doi: 10.1021/ar500079b. [DOI] [PubMed] [Google Scholar]
- Hudson P. S., Woodcock H. L., Boresch S.. Use of Nonequilibrium Work Methods to Compute Free Energy Differences Between Molecular Mechanical and Quantum Mechanical Representations of Molecular Systems. J. Phys. Chem. Lett. 2015;6:4850–4856. doi: 10.1021/acs.jpclett.5b02164. [DOI] [PubMed] [Google Scholar]
- Rufa D. A., Bruce Macdonald H. E., Fass J., Wieder M., Grinaway P. B., Roitberg A. E., Isayev O., Chodera J. D.. Towards chemical accuracy for alchemical free energy calculations with hybrid physics-based machine learning/molecular mechanics potentials. bioRxiv. 2020:2020.07.29.227959. doi: 10.1101/2020.07.29.227959. [DOI] [Google Scholar]
- Böselt L., Thürlemann M., Riniker S.. Machine learning in QM/MM molecular dynamics simulations of condensed-phase systems. J. Chem. Theory Comput. 2021;17:2641–2658. doi: 10.1021/acs.jctc.0c01112. [DOI] [PubMed] [Google Scholar]
- Pan X., Yang J., Van R.. et al. Machine-learning-assisted free energy simulation of solution-phase and enzyme reactions. J. Chem. Theory Comput. 2021;17:5745–5758. doi: 10.1021/acs.jctc.1c00565. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Gastegger M., Schütt K. T., Müller K. R.. Machine learning of solvent effects on molecular spectra and reactions. Chem. Sci. 2021;12:11473–11483. doi: 10.1039/d1sc02742e. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Giese T. J., Zeng J., Lerew L.. et al. Software infrastructure for next-generation QM/MM-ΔMLP force fields. J. Phys. Chem. B. 2024;128(26):6257–6271. doi: 10.1021/acs.jpcb.4c01466. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Tao Y., Giese T. J., Ekesan Ş.. et al. Amber free energy tools: Interoperable software for free energy simulations using generalized quantum mechanical/molecular mechanical and machine learning potentials. J. Chem. Phys. 2024;160(22):224104. doi: 10.1063/5.0211276. [DOI] [PubMed] [Google Scholar]
- Tkaczyk S., Karwounopoulos J., Schöller A.. et al. Reweighting from Molecular Mechanics Force Fields to the ANI-2x Neural Network Potential. J. Chem. Theory Comput. 2024;20:2719–2728. doi: 10.1021/acs.jctc.3c01274. [DOI] [PubMed] [Google Scholar]
- Pregeljc D., Hügli R. J. R., Riniker S.. Efficient multistate free-energy calculations with qm/mm accuracy using replica-exchange enveloping distribution sampling. J. Phys. Chem. B. 2025;129(24):5948–5960. doi: 10.1021/acs.jpcb.5c02086. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Gradisteanu V., Chan E. W., Hedges L., Malagarriga M., David R., de la Puente M., Laage D., Tuñón I., van der Kamp M. W., Zinovjev K.. Simulating enzyme catalysis with electrostatically embedded machine learning potentials. ChemRxiv. 2025:2025-nw9lt. doi: 10.26434/chemrxiv-2025-nw9lt. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Bensberg M., Eckhoff M., Thomasen F. E.. et al. Machine learning-enhanced calculation of quantum-classical binding free energies. J. Chem. Theory Comput. 2025;21(16):8182–8198. doi: 10.1021/acs.jctc.5c00388. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Bensberg M., Eckhoff M., Husistein R. T.. et al. Hierarchical quantum embedding by machine learning for large molecular assemblies. J. Chem. Theory Comput. 2025;21(15):7662–7674. doi: 10.1021/acs.jctc.5c00389. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Lu J., Song Z., Zhao Q., Du Y., Cao Y., Jia H., Duan C.. Generative design of functional metal complexes utilizing the internal knowledge and reasoning capability of large language models. J. Am. Chem. Soc. 2025;147(36):32377–32388. doi: 10.1021/jacs.5c02097. [DOI] [PubMed] [Google Scholar]
- Tom G., Schmid S. P., Baird S. G.. et al. Self-driving laboratories for chemistry and materials science. Chem. Rev. 2024;124(16):9633–9732. doi: 10.1021/acs.chemrev.4c00055. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Rafel, F. D. S. ; Rob, M., II ; Ben, M. ; et al. Shaping the Future of Self-Driving Autonomous Laboratories Workshop; Oak Ridge National Laboratory (ORNL): Oak Ridge, TN (United States), 2024, Oak Ridge Leadership Computing Facility (OLCF), Technical Report. https://www.osti.gov/biblio/2481197. [Google Scholar]
- Wang L., Song C., Liu Z.. et al. Diffusion models for molecules: A survey of methods and tasks. arXiv. 2025:arXiv.2502.09511. doi: 10.48550/arXiv.2502.09511. [DOI] [Google Scholar]
- Schneuing A., Harris C., Du Y.. et al. Structure-based drug design with equivariant diffusion models. Nat. Comput. Sci. 2024;4(12):899–909. doi: 10.1038/s43588-024-00737-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Cremer J., Irwin R., Tibo A., Janet J. P., Olsson S., Clevert D.-A.. FLOWR: Flow matching for structure-aware de novo, interaction- and fragment-based ligand generation. arXiv. 2025:arXiv:2504.10564. doi: 10.48550/arXiv.2504.10564. [DOI] [Google Scholar]
- Duan C., Du Y., Jia H., Kulik H. J.. Accurate transition state generation with an object-aware equivariant elementary reaction diffusion model. Nat. Comput. Sci. 2023;3(12):1045–1055. doi: 10.1038/s43588-023-00563-7. [DOI] [PubMed] [Google Scholar]
- Tuo P., Chen J., Li J.. Accurate generation of chemical reaction transition states by conditional flow matching. arXiv. 2025:arXiv:2507.10530. doi: 10.48550/arXiv.2507.10530. [DOI] [Google Scholar]
- Cornet F., Benediktsson B., Hastrup B., Schmidt M. N., Bhowmik A.. OM-Diff: Inverse-design of organometallic catalysts with guided equivariant denoising diffusion. Digital Discovery. 2024;3(9):1793–1811. doi: 10.1039/D4DD00099D. [DOI] [Google Scholar]
- Ahern W., Yim J., Tischer D., Salike S., Woodbury S. M., Kim D., Kalvet I., Kipnis Y., Coventry B., Altae-Tran H. R.. et al. Atom-level enzyme active site scaffolding using RFdiffusion2. Nat. Methods. 2025;23:96–105. doi: 10.1038/s41592-025-02975-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Boiko D. A., MacKnight R., Kline B., Gomes G.. Autonomous chemical research with large language models. Nature. 2023;624(7992):570–578. doi: 10.1038/s41586-023-06792-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Suvarna M., Vaucher A. C., Mitchell S., Laino T., Pérez-Ramírez J.. Language models and protocol standardization guidelines for accelerating synthesis planning in heterogeneous catalysis. Nat. Commun. 2023;14(1):7964. doi: 10.1038/s41467-023-43836-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Jablonka K. M., Ai Q., Al-Feghali A.. et al. 14 examples of how llms can transform materials science and chemistry: A reflection on a large language model hackathon. Digital Discovery. 2023;2(5):1233–1250. doi: 10.1039/D3DD00113J. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Bran A. M., Cox S., Schilter O., Baldassari C., White A. D., Schwaller P.. Augmenting large language models with chemistry tools. Nat. Mach. Intell. 2024;6(5):525–535. doi: 10.1038/s42256-024-00832-8. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Bhattacharya D., Cassady H. J., Hickner M. A., Reinhart W. F.. Large language models as molecular design engines. J. Chem. Inf. Model. 2024;64(18):7086–7096. doi: 10.1021/acs.jcim.4c01396. [DOI] [PubMed] [Google Scholar]
- Kim S., Jung Y., Schrier J.. Large language models for inorganic synthesis predictions. J. Am. Chem. Soc. 2024;146(29):19654–19659. doi: 10.1021/jacs.4c05840. [DOI] [PubMed] [Google Scholar]
- Ramos M. C., Collison C. J., White A. D.. A review of large language models and autonomous agents in chemistry. Chem. Sci. 2025;16(6):2514–2572. doi: 10.1039/D4SC03921A. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Van Herck J., Gil M. V., Jablonka K. M.. et al. Assessment of fine-tuned large language models for real-world chemistry and material science applications. Chem. Sci. 2025;16(2):670–684. doi: 10.1039/D4SC04401K. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Schilling-Wilhelmi M., Ríos-García M., Shabih S.. et al. From text to insight: Large language models for chemical data extraction. Chem. Soc. Rev. 2025;54(3):1125–1150. doi: 10.1039/D4CS00913D. [DOI] [PubMed] [Google Scholar]
- Vogiatzis K. D.. Design of CO2-philic molecular units with large language models. Chem. Commun. 2025;61:10166–10169. doi: 10.1039/D5CC02652K. [DOI] [PubMed] [Google Scholar]
- Narayanan S. M., Braza J. D., Griffiths R.-R.. et al. Training a scientific reasoning model for chemistry. arXiv. 2025:arXiv:2506.17238. doi: 10.48550/arXiv.2506.17238. [DOI] [Google Scholar]
- Balcells D.. Co-intelligent design of catalysis research with large language models: Hype or reality? ACS Catal. 2025;15(18):16412–16420. doi: 10.1021/acscatal.5c04238. [DOI] [Google Scholar]
- Rappe A. K., Casewit C. J., Colwell K. S., Goddard W. A., Skiff W. M.. UFF, a full periodic table force field for molecular mechanics and molecular dynamics simulations. J. Am. Chem. Soc. 1992;114(25):10024–10035. doi: 10.1021/ja00051a040. [DOI] [Google Scholar]
- Gottweis J., Weng W.-H., Daryin A.. et al. Towards an AI co-scientist. arXiv. 2025:arXiv.2502.18864. doi: 10.48550/arXiv.2502.18864. [DOI] [Google Scholar]
- Ghareeb A. E., Chang B., Mitchener L.. et al. Robin: A multi-agent system for automating scientific discovery. arXiv. 2025:arXiv.2505.13400. doi: 10.48550/arXiv.2505.13400. [DOI] [Google Scholar]
- Leeman J., Liu Y., Stiles J., Lee S. B., Bhatt P., Schoop L. M., Palgrave R. G.. Challenges in high-throughput inorganic materials prediction and autonomous synthesis. PRX Energy. 2024;3:011002. doi: 10.1103/PRXEnergy.3.011002. [DOI] [Google Scholar]
- Jones G. M., Story B., Maroulas V., Vogiatzis K. D.. Molecular Representations for Machine Learning. ACS In Focus Series. 2023:7e7006. doi: 10.1021/acsinfocus.7e7006. [DOI] [Google Scholar]
- Wigh D. S., Goodman J. M., Lapkin A. A.. A review of molecular representation in the age of machine learning. Wiley Interdiscip. Rev.:Comput. Mol. Sci. 2022;12:e1603. doi: 10.1002/wcms.1603. [DOI] [Google Scholar]
- Reiser P., Neubert M., Eberhard A.. et al. Graph neural networks for materials science and chemistry. Commun. Mater. 2022;3(1):2662–4443. doi: 10.1038/s43246-022-00315-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Ranković B., Griffiths R.-R., Schwaller P.. GOLLuM: Gaussian process optimized LLMs – reframing LLM finetuning through bayesian optimization. arXiv. 2025:arXiv.2504.06265. doi: 10.48550/arXiv.2504.06265. [DOI] [Google Scholar]
- Gallegos L. C., Luchini G., St John P. C., Kim S., Paton R. S.. Importance of engineered and learned molecular representations in predicting organic reactivity, selectivity, and chemical properties. Acc. Chem. Res. 2021;54(4):827–836. doi: 10.1021/acs.accounts.0c00745. [DOI] [PubMed] [Google Scholar]
- Sinha V., Laan J. J., Pidko E. A.. Accurate and rapid prediction of pKa of transition metal complexes: Semiempirical quantum chemistry with a data-augmented approach. Phys. Chem. Chem. Phys. 2021;23(4):2557–2567. doi: 10.1039/D0CP05281G. [DOI] [PubMed] [Google Scholar]
- Arunachalam N., Gugler S., Taylor M. G.. et al. Ligand additivity relationships enable efficient exploration of transition metal chemical space. J. Chem. Phys. 2022;157(18):184112. doi: 10.1063/5.0125700. [DOI] [PubMed] [Google Scholar]
- Jin H., Merz K. M.. Modeling zinc complexes using neural networks. J. Chem. Inf. Model. 2024;64(8):3140–3148. doi: 10.1021/acs.jcim.4c00095. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Laub J. A., Vogiatzis K. D.. Ligand design for 227Ac extraction by active learning and molecular topology. Digital Discovery. 2025;4(4):1100–1112. doi: 10.1039/D5DD00007F. [DOI] [Google Scholar]
- Nettem C., Mondal A., Rajaraman G.. Leveraging quantum chemistry and machine learning for the design of low-valent transition metal catalysts in nitrogen to ammonia conversion. J. Am. Chem. Soc. 2025;147(20):16948–16963. doi: 10.1021/jacs.5c00099. [DOI] [PubMed] [Google Scholar]
- Morán-González L., Betten J. E., Kneiding H., Balcells D.. AABBA graph kernel: Atom-atom, bond-bond, and bond-atom autocorrelations for machine learning. J. Chem. Inf. Model. 2024;64(23):8756–8769. doi: 10.1021/acs.jcim.4c01583. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kneiding H., Lukin R., Lang L.. et al. Deep learning metal complex properties with natural quantum graphs. Digital Discovery. 2023;2(3):618–633. doi: 10.1039/D2DD00129B. [DOI] [Google Scholar]
- Vennelakanti V., Kilic I. B., Terrones G. G., Duan C., Kulik H. J.. Machine learning prediction of the experimental transition temperature of Fe(II) spin-crossover complexes. J. Phys. Chem. A. 2024;128(1):204–216. doi: 10.1021/acs.jpca.3c07104. [DOI] [PubMed] [Google Scholar]
- de Mendonça J. P. A., Mariano L. A., Devijver E., Jakse N., Poloni R.. Artificial neural network-based density functional approach for adiabatic energy differences in transition metal complexes. J. Chem. Theory Comput. 2023;19(21):7555–7566. doi: 10.1021/acs.jctc.3c00600. [DOI] [PubMed] [Google Scholar]
- Schmid S. P., Rajaonson E. M., Ser C. T.. et al. One set to rule them all: How to obtain general chemical conditions via bayesian optimization over curried functions. arXiv. 2025:arXiv.2502.18966. doi: 10.48550/arXiv.2502.18966. [DOI] [Google Scholar]
- Zhang W., Berthebaud D., Halet J.-F., Mori T.. Electronic configurations of 3d transition-metal compounds using local structure and neural networks. J. Phys. Chem. A. 2022;126(40):7373–7381. doi: 10.1021/acs.jpca.2c03901. [DOI] [PubMed] [Google Scholar]
- Fourches D., Ash J.. 4D-quantitative structure–activity relationship modeling: Making a comeback. Expert Opin. Drug Discovery. 2019;14(12):1227–1235. doi: 10.1080/17460441.2019.1664467. [DOI] [PubMed] [Google Scholar]
- Zankov D., Madzhidov T., Varnek A., Polishchuk P.. Chemical complexity challenge: Is multi-instance machine learning a solution? Wiley Interdiscip. Rev.:Comput. Mol. Sci. 2024;14(1):e1698. doi: 10.1002/wcms.1698. [DOI] [Google Scholar]
- Zalte A. S., Pang H.-W., Doner A. C., Green W. H.. RIGR: Resonance-invariant graph representation for molecular property prediction. J. Chem. Inf. Model. 2025;65(20):10832–10843. doi: 10.1021/acs.jcim.5c00495. [DOI] [PubMed] [Google Scholar]
- Bartók A. P., Kondor R., Csányi G.. On representing chemical environments. Phys. Rev. B:Condens. Matter Mater. Phys. 2013;87(18):184115. doi: 10.1103/PhysRevB.87.184115. [DOI] [Google Scholar]
- Willatt M. J., Musil F., Ceriotti M.. Atom-density representations for machine learning. J. Chem. Phys. 2019;150(15):154110. doi: 10.1063/1.5090481. [DOI] [PubMed] [Google Scholar]
- Fabrizio A., Grisafi A., Meyer B., Ceriotti M., Corminboeuf C.. Electron density learning of non-covalent systems. Chem. Sci. 2019;10(41):9424–9432. doi: 10.1039/C9SC02696G. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Gugler S., Reiher M.. Quantum chemical roots of machine-learning molecular similarity descriptors. J. Chem. Theory Comput. 2022;18(11):6670–6689. doi: 10.1021/acs.jctc.2c00718. [DOI] [PubMed] [Google Scholar]
- Shao X., Paetow L., Tuckerman M. E., Pavanello M.. Machine learning electronic structure methods based on the one-electron reduced density matrix. Nat. Commun. 2023;14(1):6281. doi: 10.1038/s41467-023-41953-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Musil F., Grisafi A., Bartók A. P., Ortner C., Csányi G., Ceriotti M.. Physics-inspired structural representations for molecules and materials. Chem. Rev. 2021;121(16):9759–9815. doi: 10.1021/acs.chemrev.1c00021. [DOI] [PubMed] [Google Scholar]
- Jorner K.. Putting chemical knowledge to work in machine learning for reactivity. Chimia. 2023;77(1/2):22. doi: 10.2533/chimia.2023.22. [DOI] [PubMed] [Google Scholar]
- Mai H., Le T. C., Chen D., Winkler D. A., Caruso R. A.. Machine learning for electrocatalyst and photocatalyst design and discovery. Chem. Rev. 2022;122(16):13478–13515. doi: 10.1021/acs.chemrev.2c00061. [DOI] [PubMed] [Google Scholar]
- Vogiatzis K. D., Polynski M. V., Kirkland J. K.. et al. Computational approach to molecular catalysis by 3d transition metals: Challenges and opportunities. Chem. Rev. 2019;119(4):2453–2523. doi: 10.1021/acs.chemrev.8b00361. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Abraham B. M., Jyothirmai M. V., Sinha P., Viñes F., Singh J. K., Illas F.. Catalysis in the digital age: Unlocking the power of data with machine learning. Wiley Interdiscip. Rev.:Comput. Mol. Sci. 2024;14(5):e1730. doi: 10.1002/wcms.1730. [DOI] [Google Scholar]
- Wellawatte G. P., Seshadri A., White A. D.. Model agnostic generation of counterfactual explanations for molecules. Chem. Sci. 2022;13(13):3697–3705. doi: 10.1039/D1SC05259D. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Wellawatte G. P., Schwaller P.. Human interpretable structure-property relationships in chemistry using explainable machine learning and large language models. Commun. Chem. 2025;8(1):11. doi: 10.1038/s42004-024-01393-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Dewyer A. L., Argüelles A. J., Zimmerman P. M.. Methods for exploring reaction space in molecular systems. Wiley Interdiscip. Rev.:Comput. Mol. Sci. 2018;8(2):e1354. doi: 10.1002/wcms.1354. [DOI] [Google Scholar]
- Liu J., Rama R. J., Cordero-Lanzac T., Safy M. E. A., Franke R., Nova A.. Outer-sphere co release mechanism in the methanol-to-syngas reaction catalyzed by a Ru-PNP pincer complex. ACS Catal. 2025;15(6):5113–5122. doi: 10.1021/acscatal.4c06818. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Steiner M., Reiher M.. Autonomous reaction network exploration in homogeneous and heterogeneous catalysis. Top. Catal. 2022;65(1–4):6–39. doi: 10.1007/s11244-021-01543-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Queen O., McCarver G. A., Thatigotla S.. et al. Polymer graph neural networks for multitask property learning. npj Comput. Mater. 2023;9(1):90. doi: 10.1038/s41524-023-01034-3. [DOI] [Google Scholar]
- Horn K. P., Vazquez-Salazar L. I., Koch C. P., Meuwly M.. Improving potential energy surfaces using measured feshbach resonance states. Sci. Adv. 2024;10(9):eadi6462. doi: 10.1126/sciadv.adi6462. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Hoonakker F., Lachiche N., Varnek A., Wagner A.. A representation to apply usual data mining techniques to chemical reactionsillustration on the rate constant of SN2 reactions in water. Int. J. Artif. Intell. Tools. 2011;20(02):253–270. doi: 10.1142/S0218213011000140. [DOI] [Google Scholar]
- Glavatskikh M., Madzhidov T., Horvath D.. et al. Predictive models for kinetic parameters of cycloaddition reactions. Mol. Inf. 2019;38(1–2):1800077. doi: 10.1002/minf.201800077. [DOI] [PubMed] [Google Scholar]
- Zhang H., Juraskova V., Duarte F.. Modelling chemical processes in explicit solvents with machine learning potentials. Nat. Commun. 2024;15(1):6114. doi: 10.1038/s41467-024-50418-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Lee S., Goodman J. M.. Rapid route-finding for bifurcating organic reactions. J. Am. Chem. Soc. 2020;142(20):9210–9219. doi: 10.1021/jacs.9b13449. [DOI] [PubMed] [Google Scholar]
- Oviedo F., Ferres J. L., Buonassisi T., Butler K. T.. Interpretable and explainable machine learning for materials science and chemistry. Acc. Mater. Res. 2022;3(6):597–607. doi: 10.1021/accountsmr.1c00244. [DOI] [Google Scholar]
- Esterhuizen J. A., Goldsmith B. R., Linic S.. Interpretable machine learning for knowledge generation in heterogeneous catalysis. Nat. Catal. 2022;5(3):175–184. doi: 10.1038/s41929-022-00744-z. [DOI] [Google Scholar]
- Suzuki K., Toyao T., Maeno Z., Takakusagi S., Shimizu K.-i., Takigawa I.. Statistical analysis and discovery of heterogeneous catalysts based on machine learning from diverse published data. ChemCatChem. 2019;11(18):4537–4547. doi: 10.1002/cctc.201900971. [DOI] [Google Scholar]
- Schwaller P., Laino T., Gaudin T.. et al. Molecular transformer: A model for uncertainty-calibrated chemical reaction prediction. ACS Cent. Sci. 2019;5(9):1572–1583. doi: 10.1021/acscentsci.9b00576. [DOI] [PMC free article] [PubMed] [Google Scholar]
- He J., You H., Sandström E.. et al. Molecular optimization by capturing chemist’s intuition using deep neural networks. J. Cheminf. 2021;13(1):26. doi: 10.1186/s13321-021-00497-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Choung O.-H., Vianello R., Segler M., Stiefl N., Jiménez-Luna J.. Extracting medicinal chemistry intuition via preference machine learning. Nat. Commun. 2023;14(1):6651. doi: 10.1038/s41467-023-42242-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Yang J., Li F.-Z., Arnold F. H.. Opportunities and challenges for machine learning-assisted enzyme engineering. ACS Cent. Sci. 2024;10(2):226–241. doi: 10.1021/acscentsci.3c01275. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Merlicek L. P., Neumann J., Lear A.. et al. AI.zymes: A modular platform for evolutionary enzyme design. Angew. Chem., Int. Ed. 2025;64(27):e202507031. doi: 10.1002/anie.202507031. [DOI] [PubMed] [Google Scholar]
- Merz K. M., De Fabritiis G., Wei G.-W.. Generative models for molecular design. J. Chem. Inf. Model. 2020;60(12):5635–5636. doi: 10.1021/acs.jcim.0c01388. [DOI] [PubMed] [Google Scholar]
- Strandgaard M., Linjordet T., Kneiding H.. et al. A deep generative model for the inverse design of transition metal ligands and complexes. JACS Au. 2025;5(5):2294–2308. doi: 10.1021/jacsau.5c00242. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Gómez-Bombarelli R., Wei J. N., Duvenaud D.. et al. Automatic chemical design using a data-driven continuous representation of molecules. ACS Cent. Sci. 2018;4(2):268–276. doi: 10.1021/acscentsci.7b00572. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Nigam A., Pollice R., Tom G.. et al. Tartarus: A benchmarking platform for realistic and practical inverse molecular design. arXiv. 2023:arXiv:2209.12487. doi: 10.48550/arXiv.2209.12487. [DOI] [Google Scholar]
- Balcells D., Skjelstad B. B.. tmQM datasetquantum geometries and properties of 86k transition metal complexes. J. Chem. Inf. Model. 2020;60(12):6135–6146. doi: 10.1021/acs.jcim.0c01041. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Janet J. P., Duan C., Nandy A., Liu F., Kulik H. J.. Navigating transition-metal chemical space: Artificial intelligence for first-principles design. Acc. Chem. Res. 2021;54(3):532–545. doi: 10.1021/acs.accounts.0c00686. [DOI] [PubMed] [Google Scholar]
- Garrison A. G., Heras-Domingo J., Kitchin J. R., dos Passos Gomes G., Ulissi Z. W., Blau S. M.. Applying large graph neural networks to predict transition metal complex energies using the tmQM_wB97MV data set. J. Chem. Inf. Model. 2023;63(24):7642–7654. doi: 10.1021/acs.jcim.3c01226. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Gensch T., dos Passos Gomes G., Friederich P.. et al. A comprehensive discovery platform for organophosphorus ligands for catalysis. J. Am. Chem. Soc. 2022;144(3):1205–1217. doi: 10.1021/jacs.1c09718. [DOI] [PubMed] [Google Scholar]
- Chen S.-S., Meyer Z., Jensen B., Kraus A., Lambert A., Ess D. H.. Realigands: A ligand library cultivated from experiment and intended for molecular computational catalyst design. J. Chem. Inf. Model. 2023;63(23):7412–7422. doi: 10.1021/acs.jcim.3c01310. [DOI] [PubMed] [Google Scholar]
- Kneiding H., Nova A., Balcells D.. Directional multiobjective optimization of metal complexes at the billion-system scale. Nat. Comput. Sci. 2024;4(4):263–273. doi: 10.1038/s43588-024-00616-5. [DOI] [PubMed] [Google Scholar]
- Strieth-Kalthoff F., Sandfort F., Kühnemund M., Schäfer F. R., Kuchen H., Glorius F.. Machine learning for chemical reactivity: The importance of failed experiments. Angew. Chem., Int. Ed. 2022;61(29):e202204647. doi: 10.1002/anie.202204647. [DOI] [PubMed] [Google Scholar]
- Beker W., Roszak R., Wołos A.. et al. Machine learning may sometimes simply capture literature popularity trends: A case study of heterocyclic Suzuki–Miyaura coupling. J. Am. Chem. Soc. 2022;144(11):4819–4827. doi: 10.1021/jacs.1c12005. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Brazil R.. Illuminating ‘the ugly side of science’: Fresh incentives for reporting negative results. Nature. 2024:1476–4687. doi: 10.1038/d41586-024-01389-7. [DOI] [PubMed] [Google Scholar]
- Higgins S. G., Nogiwa-Valdez A. A., Stevens M. M.. Considerations for implementing electronic laboratory notebooks in an academic research environment. Nat. Protoc. 2022;17(2):179–189. doi: 10.1038/s41596-021-00645-8. [DOI] [PubMed] [Google Scholar]
- Sigmund L. M., Seifert T., Halder R.. et al. Predicting reaction feasibility and selectivity of aromatic C–H thianthrenation with a QM–ML hybrid approach. Angew. Chem., Int. Ed. 2025;64(44):1521–3773. doi: 10.1002/anie.202510533. [DOI] [PubMed] [Google Scholar]
- Kearnes S. M., Maser M. R., Wleklinski M., Kast A., Doyle A. G., Dreher S. D., Hawkins J. M., Jensen K. F., Coley C. W.. The open reaction database. J. Am. Chem. Soc. 2021;143(45):18820–18826. doi: 10.1021/jacs.1c09820. [DOI] [PubMed] [Google Scholar]
- Nippa D. F., Müller A. T., Atz K.. et al. Simple user-friendly reaction format. Mol. Inf. 2025;44(1):e202400361. doi: 10.1002/minf.202400361. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Taylor C. J., Felton K. C., Wigh D.. et al. Accelerated chemical reaction optimization using multi-task learning. ACS Cent. Sci. 2023;9(5):957–968. doi: 10.1021/acscentsci.3c00050. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Artrith N., Butler K. T., Coudert F.-X.. et al. Best practices in machine learning for chemistry. Nat. Chem. 2021;13:505–508. doi: 10.1038/s41557-021-00716-z. [DOI] [PubMed] [Google Scholar]
- Maxson T., Soyemi A., Chen B. W. J., Szilvási T.. Enhancing the Quality and Reliability of Machine Learning Interatomic Potentials through Better Reporting Practices. J. Phys. Chem. C. 2024;128(16):6524–6537. doi: 10.1021/acs.jpcc.4c00028. [DOI] [Google Scholar]
- Butler K. T., Choudhary K., Csányi G., Ganose A. M., Kalinin S. V., Morgan D.. Setting standards for data driven materials science. npj Comput. Mater. 2024;10(231):1–3. doi: 10.1038/s41524-024-01411-6. [DOI] [Google Scholar]
- Doney A. C., Rooks B. J., Lu T., Wheeler S. E.. Design of organocatalysts for asymmetric propargylations through computational screening. ACS Catal. 2016;6(11):7948–7955. doi: 10.1021/acscatal.6b02366. [DOI] [Google Scholar]
- Grambow C. A., Pattanaik L., Green W. H.. Reactants, products, and transition states of elementary chemical reactions based on quantum chemistry. Sci. Data. 2020;7(1):137. doi: 10.1038/s41597-020-0460-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
- von Rudorff G. F., Heinen S. N., Bragato M., von Lilienfeld O. A.. Thousands of reactants and transition states for competing E2 and SN2 reactions. Mach. Learn.: Sci. Technol. 2020;1(4):045026. doi: 10.1088/2632-2153/aba822. [DOI] [Google Scholar]
- Stuyver T., Jorner K., Coley C. W.. Reaction profiles for quantum chemistry-computed [3 + 2] cycloaddition reactions. Sci. Data. 2023;10(1):66. doi: 10.1038/s41597-023-01977-8. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Jorner K., Brinck T., Norrby P.-O., Buttar D.. Machine learning meets mechanistic modelling for accurate prediction of experimental activation energies. Chem. Sci. 2021;12(3):1163–1175. doi: 10.1039/D0SC04896H. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Nieminen, R. M. From number crunching to virtual reality: Mathematics, physics and computation. In Mathematics Unlimited2001 and Beyond; Springer Berlin Heidelberg, 2001; pp 937–959. [Google Scholar]

