Abstract
This article reviews the foundations and applications of machine learning force fields (MLFFs) in electrochemistry, highlighting their role as a transformative tool in materials science. We first provide an overview of MLFFs, then discuss their applications in ionics and electrochemical reactions, and finally outline future directions. Most MLFF approaches use invariant or equivariant descriptors derived from body-order expansions to represent many-body atomic interactions. These descriptors feed into linear regression models, kernel methods, or neural networks to construct potential energy surfaces for gases, liquids, solids, and interfaces involving inorganic and organic materials. MLFFs have enabled a wide range of advances, including all-atom molecular dynamics (MD), data extraction from MD, and accelerated materials discovery. In MD simulations, MLFFs allow accurate evaluation of ionic conductivity via the fluctuation–dissipation theorem and nonequilibrium MD under electric fields, applied to both solid and polymer electrolytes. For electrochemical reactions, MLFFs and Δ-ML models have been used to predict redox potentials in homogeneous and interfacial systems through thermodynamic integration. MLFFs also enable the extraction of key thermodynamic and kinetic information–such as free energy landscapes and local transport coefficients–from atomic trajectories, facilitating coarse-grained modeling of mass transport and reactions in complex electrolytes. In materials discovery, MLFFs have allowed high-throughput screening of 107 to 108 crystal structures, leading to the identification of promising Li-ion and Na-ion solid electrolytes. MLFFs are expected to continue evolving as a core technology in computational materials science, spanning a wide range from high-precision calculations to large-scale materials exploration.
Keywords: machine learning force field, first-principles calculation, molecular dynamics, coarse-graining, materials discovery, ionics, electrochemical reaction, conductivity, free energy


Introduction
The global energy consumption has been increasing at an annual rate of 1.7% from 1990 to 2020. , The majority of this demand is met by fossil fuels, raising concerns about global warming due to the large amounts of CO2 emissions. To address this issue, the development of a sustainable energy system based on renewable energy sources is desirable. Renewable energy sources such as solar and wind power provide an almost inexhaustible supply of resources for human activities. However, they are inherently variable and geographically dispersed. To effectively utilize these energy sources, technologies for storing electricity generated from renewable energy and converting stored energy back into electricity are essential.
Secondary batteries, fuel cells, and electrolytic cells have attracted significant attention as promising systems for converting electrical and chemical energy. Extensive fundamental and applied research has been conducted to develop cost-effective and durable systems while improving energy conversion efficiency and power density. − Energy conversion in electrochemical systems is realized through ion-electron transfer reactions at the electrodes and ion transport within the electrolyte. The efficiency and power density of these systems are strongly influenced by the reaction kinetics at the electrodes and the ionic conductivity of the electrolyte. Additionally, the durability of the system depends on the mechanical, thermal, and chemical stability of the materials. Furthermore, system performance and durability are not only determined by the intrinsic properties of individual materials but also by the methods used to assemble these materials into cells. The properties of the materials themselves are significantly affected by the applied potential at the electrodes and the environmental conditions to which they are exposed. Lastly, it goes without saying that cost–largely determined by the choice of materials and manufacturing processes–is a critical factor in the widespread adoption of these systems. Given this background, the primary focus of research and development in electrochemical systems includes the discovery of materials that enable superior performance and durability, the optimization of battery and cell assembly, and the development of control strategies to maximize the performance of materials and systems. Crucially, addressing all these challenges requires a deep understanding of atomic- and molecular-scale structures of materials, as well as mechanisms of charge carrier transport and reactions.
Electrochemistry is a discipline that investigates the structures of electrolytes and electrode surfaces, as well as their transformations. Shortly after its inception, microscopic structural models of electrolytes and electrodes were proposed. With the subsequent advancement of classical mechanics, electromagnetism, statistical mechanics, and quantum mechanics, mathematical models describing ionic conduction and electrode reactions were developed. − Today, electrochemistry has become closely integrated with first-principles (FP) calculations, which explicitly treat atomic nuclei and electrons and solve wave equations to predict material properties. This integration has enabled quantitative predictions of the thermodynamics and kinetics of ionics and reactions at electronic and atomic scales, establishing FP methods as a fundamental tool for mechanistic studies and the discovery of advanced electrochemical materials. − Furthermore, multiscale computational methods − have been developed to bridge the gap between atomic-scale FP calculations and larger-scale simulations.
However, FP calculations and the earlier-mentioned physics-based bottom-up multiscale simulation methods built upon them have inherent limitations. Because of the high computational cost of FP calculations, the typical length- and time-scales that can be handled are restricted to a few nanometers and a few hundred picoseconds. As a result, accurately predicting long-time scale ion transport and reaction dynamics in heterogeneous structures beyond this scale remains a significant challenge. Multiscale simulation techniques have been developed to address these limitations. For example, one approach involves parametrizing the potential energy surface (PES) obtained from FP calculations to construct an interatomic interaction model, which is used in molecular dynamics (MD) simulations. This classical MD approach , can be regarded as a type of multiscale simulation. Another typical method is coarse-grained MD, in which groups of atoms are mapped onto a single particle, and the potential of mean force acting on these particles are modeled based on MD-derived data. ,− Additionally, to extend the length- and time-scales of simulations, local transport coefficients of molecules or ions, obtained from classical MD simulations of heterogeneous structures, can be incorporated into kinetic Monte Carlo (kMC) simulations. ,, Another approach involves constructing energy diagrams for elementary reactions on solid surfaces and determining reaction rate constants based on transition state theory (TST). These rate constants can be incorporated into mean-field models or kMC simulations to comprehensively analyze the complex reactions at heterogeneous surface sites. ,,,, However, applying these methods to complex materials presents significant challenges. For instance, accurately parametrizing the PES of multielement materials using physics-based interaction models is challenging. Moreover, constructing such interaction models often requires extensive trial and error by human experts. This issue becomes particularly severe for phenomena involving formations and scissions of chemical bonds, limiting the range of materials that classical MD can handle and introducing significant errors in the predicted material properties. Challenges also arise when formulating reaction rate constants for elementary processes. In complex reaction networks involving a vast number of elementary steps, essential reaction pathways can be overlooked in modeling. In other words, a fundamental challenge in multiscale simulations is that errors can arise when transferring information from microscopic scales to larger scales.
In recent years, artificial intelligence (AI) and its subfield, machine learning (ML), have gained attention as promising approaches to addressing these issues. The utilization of AI and ML is ubiquitous in everyday life, playing a crucial role in applications such as chatbots, recommendation systems, autonomous driving, voice assistants, image recognition, language translation, and medical diagnosis support. Their impact on natural science is also profound, as evidenced by the 2024 Nobel Prize in Physics being awarded for the foundational discoveries and inventions related to artificial neural networks, and the 2024 Nobel Prize in Chemistry recognizing AI-driven protein structure prediction and design. The application of AI and ML continues to expand in the fields of electrochemistry and energy storage, where they are being utilized for a wide range of tasks, including materials discovery, synthesis, manufacturing process optimization, analysis, diagnostics and degradation prediction, and autonomous experimentation. A few years ago, AI and ML were predicted to drive a paradigm shift toward data-driven science, emerging as the “fourth paradigm” alongside experimental, theoretical, and computational science. Today, we are witnessing this innovation dynamically progress. However, this shift does not imply that data-driven science operates independently from the existing paradigms. Rather, it builds upon the foundations of experimental, theoretical, and computational science, integrating AI and ML into these established domains. Numerous excellent review articles − have already been published on the applications of AI and ML. Therefore, this work will not repeat those discussions.
One of the objectives of this article is to highlight the nature of the aforementioned hybrid paradigm shift by providing an overview of recent advancements in machine learning force fields (MLFFs), a class of supervised ML regression models for interatomic potentials, within bottom-up simulations of electrochemical phenomena. MLFFs can accurately reproduce PESs derived from FP calculations while accelerating finite-temperature MD simulations by orders of magnitude through bypassing explicit FP computations. In addition, variant models of MLFFs can be extended to predict free energy landscapes and transfer coefficients for use in coarse-grained simulations. As in other fields of materials science, AI and ML technologies that predict target properties from material descriptors while treating the underlying relationships as a black box are widely employed in electrochemistry. Compared to such approaches, the construction of bottom-up methodologies requires a significant amount of time, and their applicability may initially appear limited. However, once a bottom-up approach is established, it offers a key advantage: it enables the presentation of extensive information spanning microscopic (atomic and electronic) to mesoscopic and macroscopic scales in a form that is interpretable by humans. The rapid progress of MLFFs in recent years has led to an increasingly comprehensive understanding of diverse electrochemical phenomena.
The second objective of this article is to provide a comprehensive overview of the various forms of MLFF and explain their essential concepts. The development of MLFF has progressed rapidly, resulting in a seemingly overwhelming number of models, creating a somewhat complex landscape. However, the major forms of MLFF follow the physical framework of body-order expansion, as discussed later, and are designed to compactly and accurately represent many-body interactions between atoms. In other words, the development of MLFF formulation itself has been guided by theoretical science.
The third objective of this article is to describe methods for predicting key properties of electrochemical systems, as introduced at the beginning, with the aid of MLFFs and relevant ML technologies. Figure provides an overview of these approaches. The key properties of interest in this article include the transport coefficients of ions and molecules in electrolytes, as well as the thermodynamic and kinetic properties of electrochemical reactions at electrolyte–electrode interfaces. Over the years, research based on theoretical and computational sciences has established bottom-up approaches for predicting these properties. However, because of fundamental limitations in time- and length-scales, applying these methods to complex materials and diverse chemical systems remains challenging. To overcome these limitations and expand the applicability of computational approaches, MLFFs have rapidly gained traction. MLFFs enable the prediction of transport coefficients and reaction free energies at finite temperatures via molecular dynamics simulations. − MLFFs also enable the analysis of complex reaction networks that emerge during MD simulations, facilitating the automatic extraction of key reaction pathways. Additionally, they are utilized for coarse-graining models to extend simulations to larger scales beyond conventional MD calculations. ,,,,,,,, The application of MLFFs requires training data for constructing ML models. In recent years, the development of extensive databases − has accelerated the adoption of MLFFs. As in other fields, several excellent studies have been published on the application of MLFFs in atomic-scale property prediction and multiscale simulations. However, there are still relatively few comprehensive reviews that specifically address key electrochemical properties through the integration of fundamental theoretical and computational sciences with MLFF-based methodologies. Therefore, this article focuses on electrochemical properties and discusses the current state and future prospects of MLFFs and related technologies in this domain.
1.
Schematic of AI- and ML-aided evaluations, predictions and explorations of electrolyte and electrode materials.
The structure of this article is as follows. First, we provide a brief overview of the Hamiltonian and density functional theory (DFT), which serve as the foundation for simulating electrochemical phenomena. While there are many excellent textbooks on these topics, ,,, we revisit them concisely in this section to clarify what MLFFs aim to represent. Additionally, we touch upon the grand canonical DFT, which is essential for describing electrochemical reactions using FP methods. Second, we introduce MLFFs. The field of MLFFs is rapidly evolving, with a vast number of methods continuously being proposed, making it challenging to assess their relationships and relative advantages. This section explains representative approaches, focusing on how energy and materials are represented, and discusses the connections between different methodologies. Third, we outline methods to predict the properties of ionics and electrochemical reactions and introduce practical applications. We describe approaches for predicting ionic conductivity and the thermodynamics and kinetics of reactions by integrating FP calculations, MD, and MLFFs. Furthermore, we provide an overview of statistical analysis methods using MLFFs to extract information from atomic-scale simulations, which is then used in mesoscale and macroscale simulations of heterogeneous systems. Additionally, we introduce recent studies where MLFFs facilitate large-scale materials exploration beyond conventional approaches. Finally, we summarize the current state of the field and future perspectives, concluding this article.
Hamiltonian and Density Functional Theory
Hamiltonian
In this article, we focus on ionics and electrochemical reactions, whose theoretical description relies fundamentally on quantum mechanics. The computational approach that provides predictions for material properties is FP calculations. The starting point for these calculations is the relativistic Dirac equation for Fermions. , However, in many modern FP calculations, relativistic effects are incorporated into effective core potentials, such as pseudopotentials − and the Projector Augmented Wave (PAW) method. , By solving the nonrelativistic Schrödinger equation under these effective external potentials, material properties that account for relativistic effects can be predicted within this framework. − Therefore, in this work, we begin our discussion with the Schrödinger equation as the starting point.
Nonrelativistic quantum mechanics describes the stationary state of a system consisting of N n nuclei and their surrounding electrons by the following many-body wave equation ,
| 1 |
| 2 |
where Φ represents the many-body wave function of the nuclei and electrons, K̂ n denotes the kinetic energy operator of the nuclei, and Ĥ e is the electronic Hamiltonian. The position coordinates of the N e electrons are given by r N e , while R N n represents the position coordinates of the N n nuclei. Even for the lightest hydrogen nucleus, its mass is 1836 times that of an electron, and electrons move approximately 100 times faster than the nuclei. Under such conditions, the kinetic energy of the nuclei is significantly smaller than other terms, and the electrostatic interactions between nuclei can be approximated as constants. In the majority of FP calculations, the Born–Oppenheimer approximation, which is based on this assumption, is applied. As a result, the wave equations for the nuclei and electrons can be separated as
| 3 |
| 4 |
where Ψ n represents the antisymmetric many electron wave function corresponding to the n-th eigenstate of the N e-electron system, and E n is its eigenenergy. The function χ m denotes the nuclear wave function associated with the m-th eigenenergy E m . Using these definitions, the many-body wave function of the nuclei and electrons can be expressed as Φ(r N e , R N n ) = χ m (R N n ) Ψ m (r N e , R N n ). Condequently, the system can be approximated as an ensemble of nuclei moving on the energy surface defined by the eigenenergy E m , which is obtained by solving the many electron wave equation, eq , for electrons surrounding the nuclei fixed at a given set of positions R N n . In most FP calculations, this energy surface is typically taken as the electronic ground-state energy E 0, and a classical approximation is often applied to eq . Under this approximation, the motion of the nuclei is described by the following classical Hamiltonian
| 5 |
where P I is the momentum of the I-th nucleus with a mass of M I . Alternatively, eq can be solved by incorporating quantum effects in nuclear motion based on Feynman’s path integral method. − As described by the following Hamiltonian, the system can be regarded as a canonical ensemble of PN n classical particles moving on an effective potential
| 6 |
where P I represents the momentum of a particle with a fictitious mass M I , and ω P = P/βℏ, where β is the inverse temperature. Equation is exact in the limit P → ∞; however, in practical numerical calculations, P is typically set to a finite value based on the empirical criterion P ≥ 4 ℏβωmax. Solving eq corresponds to the first-principles molecular dynamics (FPMD) method, while solving eq corresponds to the first-principles path integral molecular dynamics (FP-PIMD) method.
Once the PES E 0 is obtained, it becomes possible to formulate equations of motion and partition functions for describing nuclear motion according to eq or eq . By numerically integrating the equations of motion via MD simulations or analyzing the minima and saddle points of E 0, one can derive information on ionic conductivity, reaction rates, and a wide range of other material properties based on statistical mechanics–either exactly or approximately. ,,,,,,− In any case, predicting nuclear motion requires the ground-state PES E 0, which is obtained by solving eq . Since exactly solving the many electron wave equation demands an enormous computational cost, approximations are applied in most calculations. Among these, the most widely used approach is DFT. ,,
Density Functional Theory
Ground State
Solving eq is equivalent to minimizing the energy of the system with respect to an antisymmetric many electron wave function as
| 7 |
where Ψ0 is the ground-state many electron wave function that minimizes E. However, both solving eq and minimizing with respect to this many electron wave function require an enormous computational cost, scaling as a power of the number of electrons N e. The foundation of DFT was established by the Hohenberg–Kohn theorem and Levy’s constrained search. , These works demonstrated that the energy E can be expressed as a functional of the electron density ρ derived from an N-representable and antisymmetric many electron wave function Ψ. By minimizing this functional with respect to the electron density ρ, instead of the wave function Ψ, one can determine the ground-state energy E 0 and obtain all relevant information about the electronic state
| 8 |
where ρ0 is the ground-state electron density.
To determine the ground state, a specific form of the functional is required. In DFT, it is a natural approach to describe the energy as a functional of the electron density. − However, this formulation did not accurately represent material properties. To address this, Kohn and Sham introduced a reference system of noninteracting N e electrons and proposed representing the energy functional as
| 9 |
| 10 |
| 11 |
| 12 |
| 13 |
where T s is the kinetic energy of the reference system, E h is the electrostatic interaction energy between electrons, E xc is the exchange-correlation energy, and E nn is the electrostatic interaction energy between nuclei. The function ψ i represents the i-th one-electron wave function of the reference system, while Z I and R I denote the nuclear charge and position vector, respectively, of the I-th nucleus. For simplicity, spin and Brillouin zone quantum numbers are not explicitly represented in the notation. Additionally, ψ i satisfies the following orthonormality condition
| 14 |
Equation can be interpreted as introducing a reference system in which the kinetic energy functional can be exactly defined as eq , and defining E xc as the difference between this energy and the energy of the interacting system. The energy minimization is performed with respect to ψ i instead of ρ. The variational principle yields the following equation for ψ i
| 15 |
| 16 |
| 17 |
| 18 |
where ϵ i represents the one-electron eigenenergy, and v ext(r) denotes the external potential arising from nuclei and other sources. The term v xc is referred to as the exchange-correlation potential. The eigen equation for the one-electron wave function ψ i , given by eq , is known as the Kohn–Sham equation. To obtain the ground-state energy and electronic structure, the Kohn–Sham equation is solved to determine ψ i . Then, using eq , the electron density ρ is computed from the N e lowest-energy orbitals. The one-electron Hamiltonian in eq is then updated, and this procedure is iterated until convergence is achieved. The Kohn–Sham method reduces the computational cost of solving eq , which originally scales factorially with N e (N e!), to a problem that scales as N e . Initially, in the Kohn–Sham method, the exchange-correlation energy E xc was modeled using the exchange-correlation energy of a uniform electron gas. − Despite its simplicity, this model proved to be surprisingly accurate for describing the electronic structure of metals, attracting significant attention to the method. Later, issues were identified in describing the electronic states of semiconductors and insulators , as well as dispersion interactions. , To address the issues, improvements were made by incorporating nonlocal electronic interactions into the exchange-correlation energy, leading to enhanced accuracy. −
Canonical Ensemble
In practical FP calculations, instead of the ground-state DFT described in the previous section, DFT for electrons in thermal equilibrium under a canonical ensemble with a fixed electron number, temperature, and volume is often used. Here, the canonical ensemble refers to an ensemble defined for a subsystem of electrons moving around fixed nuclei. This finite-temperature DFT was developed shortly after the establishment of the ground-state theory. ,, In finite-temperature DFT, the Helmholtz free energy A of N e electrons at temperature T e can be expressed as a functional of the electron density ρ. By minimizing this functional with respect to ρ, the thermal equilibrium state in the canonical ensemble can be determined
| 19 |
The following form, incorporating the entropy of noninteracting electrons, was proposed as a functional for A
| 20 |
| 21 |
| 22 |
where T e is the electronic temperature, k B is the Boltzmann constant, and f i represents the electron occupancy of the i-th orbital. Similar to the previous section, applying the variational principle with respect to ψ i yields equations of the same form as eqs to (). The key difference lies in the following equation for f i
| 23 |
where μe is the electronic chemical potential (Fermi energy), determined such that the occupancy f i satisfies the following condition
| 24 |
For numerical convenience, instead of using the Fermi–Dirac distribution given in eqs and (), alternative forms of the electronic entropy S e–where the occupancy is expressed using an integral representation of the Gaussian or Hermite function–have been employed. In most FPMD simulations, DFT under a canonical ensemble with a fixed electron number N e is applied. The nuclear motion is simulated by replacing E 0 in eqs and () with A 0. , This method is used in the FPMD simulations of ionic conduction in electrolytes, introduced later.
Grand Canonical Ensemble
In electrochemical reactions, the reaction center is modeled as an open system that allows electron exchange with a reservoir. Consequently, the reaction proceeds under conditions where the electrode potential, which corresponds to the electronic chemical potential, is held constant rather than the electron number. Under such conditions, the equilibrium state of electrons around the nuclei within the subsystem at a given moment follows the grand canonical ensemble. The nuclei are considered to move on the grand potential energy surface in equilibrium. In other words, E 0 in eq is expressed in terms of the grand potential.
The formulation of DFT within the grand canonical ensemble was also developed shortly after the inception of DFT. ,, In grand canonical ensemble DFT, the grand potential is expressed as a functional of the electron density, and the equilibrium state of the system is determined by minimizing this functional with respect to the electron density
| 25 |
Similarly to the formulations for the ground state and the canonical ensemble, the functional Ω is defined as
| 26 |
The Kohn–Sham equation is obtained using the same variational principle as in the canonical ensemble. The key difference from the canonical ensemble is that the occupancy f i is determined by eq using the fixed μe, without the constraint condition in eq .
As indicated by the above equations, in the grand canonical ensemble, the electron number N e within the subsystem varies to maintain a fixed μe. However, adjusting only the electron number is insufficient for FP calculations. The FP method, originally developed in solid-state physics to describe the electronic states of crystals, employs periodic systems to represent solids, liquids, and interfaces. In such a periodic system, if the total charge within the unit cell is not neutral, the energy diverges. Consequently, if only the electron number in the subsystem changes, the subsystem becomes charged, making energy calculations impossible. To resolve this issue, an additional charge must be introduced into the unit cell to compensate for changes in the number of electrons. In real electrochemical reactions, this role is fulfilled by the ion distribution in the electrolyte, which forms the electrical double layer. Within this layer, ions arrange themselves to establish electrochemical potentials in equilibrium with the bulk electrolyte in contact with the subsystem. Therefore, the grand canonical ensemble must also be applied to ions. Consequently, applying grand canonical ensemble DFT to electrochemical reactions requires explicit modeling of the electrical double layer.
As discussed in the previous section, FP calculations have traditionally targeted systems with a fixed N e under the canonical ensemble. Standard FP calculation programs have also been designed to solve the Kohn–Sham equation within the canonical ensemble. However, electrochemical reactions require a different theoretical and numerical framework based on the grand canonical ensemble. Before the year 2000, such numerical methods and programs did not exist, and it was not initially clear how to treat electrochemical reactions using FP methods. However, as the importance of electrochemical systems, such as fuel cells and rechargeable batteries, grew, there emerged a demand to clarify the microscopic mechanisms of reactions, leading to the development of various computational methods. ,,,− Although the developed methods vary widely, they all share a common foundation: the theory of the grand canonical ensemble.
Machine Learning Force Field
Overview of Machine Learning Force Field
DFT has reduced the computational cost required to calculate the ground-state energy surface E 0 from a power-law dependence on the number of electrons N e to a cubic scaling. However, the computational cost remains substantial, limiting the typical time- and length-scales that FPMD can handle to only a few nm and a few hundred ps. To overcome this limitation, MLFFs − have been developed. MLFFs enable the calculation of the ground-state PES E 0(R N n ) without solving the wave equation. Similarly, attempts to represent PES as a function have also been made in classical force fields. − ,,, The key difference between classical force fields and MLFFs lies in their functional forms. Classical force fields rely on functions derived from physical laws, such as Coulomb interactions and Lennard-Jones potentials, whereas MLFFs utilize functions developed in the field of ML, such as neural networks and kernel functions. Classical force fields, based on physical models, can represent reasonable energy surfaces with a small number of parameters. This also means that relatively little training data is required to determine these parameters. However, due to their simple functional forms, classical force fields struggle to accurately describe complex energy surfaces, particularly those involving bond rearrangements. Additionally, determining the functional form and parameters often requires significant human intervention. Since the optimal functional form varies depending on the phenomenon being modeled, developing a universal parametrization algorithm is challenging. On the other hand, MLFFs can accurately describe complex energy surfaces for a wide range of phenomena using a unified functional form, and they allow the generation of accurate interatomic interaction models–without modifying programs or algorithms–as long as sufficient training data is available. As shown in eqs and (), the energy surface is an essential physical property that governs not only ionics and electrochemical reactions but also various physical processes. Consequently, establishing a universal representation of energy surfaces has a profound impact. Furthermore, although FP calculations are computationally demanding, they can generate significantly more training data than experimental measurements. Given their significance and advantages, MLFFs were among the earliest applications of AI and ML in materials science and are now widely used across various areas of the field.
MLFFs can be classified into four representative approaches, which will be discussed in the next section: linear regression, kernel regression, neural networks proposed by Behler and Parrinello − and their variants, and message-passing neural networks (MPNN) − [or convolutional neural networks (CNN) , ]. Figure presents key MLFF performance metrics as a function of publication year, including the errors in energy and atomic forces (σE and σF), the number of elemental species (N el) in the materials for which properties were evaluated using MLFFs, and the number of structures (N st) used to generate the FP training data. Since MLFF performance strongly depends on the choice of parameters, training data, and target materials, quantitatively comparing these data is challenging. However, a time-series overview of the performance of MLFFs applied to practical property evaluations provides valuable insight into their development. The application of ML to energy surfaces began with the introduction of neural networks by Blank et al. However, these early models were limited to representing the energy surfaces of small molecules using low-dimensional variables. A major breakthrough in this field came with the development of the neural network potential (NNP) by Behler and Parrinello, which enabled the representation of energy surfaces for condensed-matter systems composed of multiple atoms. Since then, various MLFFs have been developed, incorporating methods such as kernel regression, linear regression, MPNN, ,, and CNN. Figure (a),(b) show that, even at the time of the first NNP, the prediction error was already an order of magnitude lower compared to conventional classical force fields. , Although the plotted data show some scatter due to dependencies on MLFF parameters, training data, materials, and evaluation conditions, there is a clear trend of improving best accuracy over time. Figure (c) further illustrates the gradual increase in the number of elemental components (N el) in materials studied using MLFFs, indicating the growing capability of these models to handle complex multielement systems. Another noteworthy trend is shown in Figure (d), which tracks the transition in the number of structures required for training (N st). Over time, a bifurcation has emerged, with MLFF development progressing in two directions: one aiming to reduce N st and the other increasing it. The reduction in N st has been driven by the introduction of relatively simple regression methods, such as linear regression ,− and kernel regression, − along with advances in active learning, an efficient training strategy. ,,,, These developments have demonstrated that, when the target properties are well-defined, MLFFs can achieve sufficient accuracy for property evaluations even with a limited number of training structures. With these methods, even when accounting for the time required to train MLFFs, they can accelerate the overall simulation by 2 to 3 orders of magnitude compared to FP calculations. On the other hand, the trend toward increasing N st has been driven by advancements in computing power, the expansion of FP databases, − ,, and the introduction of MPNN and CNN, which facilitate training on large data sets of multielement systems. Recently, training data sets on the order of N st = 106 to 108 have been utilized, accelerating efforts to develop general-purpose MLFFs that cover the entire periodic table. ,,,,, While these general-purpose MLFFs still require fine-tuning for high-precision calculations, they are already being employed for the discovery of advanced materials across extensive search spaces. ,−
2.
Trends over time in the energy error (a) and force error (b) of MLFFs used in condensed matter simulations, the number of constituent elements in the studied materials (c), and the amount of structural data used for training (d). Yellow diamonds, red triangles, blue circles and green squares indicate linear regressions (SNAP, , MTP and ACE , ), kernel regressions (GAP, VASP-MLFF and FLAIR), Behler–Parrinello type NNPs, − and MPNNs (NequIP, Allegro and MACE) and graph-CNNs (M3GNet and CHGNet). Dashed black and gray lines are values reported for EAM and ReaxFF , potentials. Data were collected from refs ,− ,,− .
Table lists several MLFF software packages. MLFFs are rapidly evolving, with their applications continuing to expand and diversify. Dozens to hundreds of new methods have been emerging every six months, and many packages are distributed through repository hosting platforms like GitHub. The aim of this review article is not to overview all of these emerging simulation software packages. However, for readers who are mainly interested in applications to electrochemistry rather than fundamental studies of MLFF methods, the current complex circumstance makes it difficult to judge which software packages are suited for applications. Therefore, in this table, we focus on software that, based on our experience and the information available to us, is relatively easy to use and regularly maintained. In the early stages of MLFF development, packages were primarily written in traditional programming languages, such as C++ and Fortran, and simulations were mainly executed on CPU architectures. However, with the emergence and widespread availability of modern ML libraries such as PyTorch and TensorFlow, many recent implementations have increasingly adopted these frameworks and are commonly optimized for execution on GPU architectures. Several methods are integrated into well-established simulation engines such as LAMMPS, VASP, , CASTEP, and CP2K. Several are also compatible with the Atomic Simulation Environment (ASE). For model training and fine-tuning, FP electronic structure software is required. VASP and CASTEP, which incorporate efficient on-the-fly active-learning algorithms, provide robust platforms for seamlessly conducting data collection, model generation, and production simulations. There are also commercially available platforms featuring user-friendly graphical user interfaces (GUIs), such as MedeA and Materials Studio.
1. Software Packages That May Be Suitable for Users Focusing on Applications in Electrochemistry.
| software | regression method | descriptor | release year and version |
|---|---|---|---|
| RuNNer | Behler–Parrinello type NN | invariant symmetry functions | Since 2007, available upon request from the developers. |
| GAP | Kernel-based method | invariant power spectrum and bispectrum | Since 2010, available via the QUIP package and implemented in both LAMMPS and CASTEP. |
| SNAP | linear regression | invariant bispectrum | Since 2015, available in LAMMPS and MedeA; latest version released in 2025. |
| MTP | linear regression | invariant moment tensor descriptors | Since 2016, available on GitHub and integrated into LAMMPS; the latest version, MLIP-3, was released in 2023. |
| N2P2 | Behler–Parrinello type NN | invariant symmetry functions | Since 2017, available in LAMMPS and MedeA. |
| SchNet | convolutional NN | invariant interatomic distances | Since 2017, compatible with ASE; latest version updated to 2.1.1 in 2024. |
| deep potential | deep NN | invariant local geometory in local frame | Since 2018, available as DeePMD-kit on GitHub; version 3.0 released in 2023. |
| VASP | Kernel-based method | invariant power spectrum | Since 2019, available in VASP (v6.3.0 to 6.5.1); MLFF-generated models also usable in LAMMPS. |
| ACE | linear regression | invariant body-ordered symmetric basis | Since 2019, available via ACE1.jl (Julia) and PACE (C++) on GitHub; integrated into LAMMPS, CP2K, and MedeA. |
| M3GNet | graph CNN | invariant two and three body descriptors | Since 2022, available on GitHub. |
| NequIP | MPNN | equivariant learned atomic representations | Since 2021, available on GitHub and in LAMMPS; latest version 0.7.1. |
| MACE | MPNN | equivariant body-ordered symmetric basis | Since 2022, available on GitHub; integrated into LAMMPS. |
a refs ,, .
refs , .
ref .
refs , .
ref .
ref .
refs , .
refs ,, .
refs ,, .
ref .
ref .
refs ,,, .
In this way, MLFFs are evolving into a core technology in computational science, supporting both high-accuracy property calculations and high-throughput materials exploration. In the following sections, we will describe the representative MLFF frameworks.
Energy Functional
MLFF does not utilize functions based on physical laws but employs functions developed within the field of ML. However, similar to classical force fields, MLFF also emphasizes physical laws to achieve a compact representation of the PES. A fundamental concept underlying many MLFFs is the body-order expansion of energy, as illustrated in Figure . This concept is not exclusive to MLFFs; many classical force fields are also constructed based on it. In the body-order expansion, energy is decomposed into one-body, two-body, three-body, and higher-order interactions, as described below
| 27 |
For simplicity, we denote the ground-state energy as E in this equation. Following the definition by Glielmo et al., we also define the (ν + 1)-th interaction term, E (ν+1), as the term that vanishes upon the (ν + 2)-th differentiation with respect to atomic coordinates
| 28 |
Equation is also referred to as the cluster expansion. Mayer demonstrated that cluster expansion is a low-density perturbative series in which higher-order cluster integrals contribute progressively less as the order ν + 1 increases. Under appropriate conditions, these contributions decay rapidly with increasing ν. This behavior is consistent with the empirical observation that electronic screening diminishes the significance of long-range interactions, resulting in smaller contributions from higher-order terms. The concept of body-order expansion plays a central role in efficiently computing many-body interactions. The computational cost of evaluating (ν + 1)-body interactions scales as N n , making it necessary to truncate the expansion at lower orders for practical simulations. The (ν + 1)-body term can, in principle, be defined recursively as the excess energy of the full (ν + 1)-body system beyond the contributions from all constituent ν-body interactions. Alternatively, and more commonly in the context of MLFFs, the body-order expansion is treated as a global functional form, where the (ν + 1)-body contribution is determined by fitting the total energy expression to a database containing systems of varying sizes and configurations.
3.
Body-order expansion of the PES and related descriptors.
Many MLFFs represent the energy E of a material as the sum of the local energies E I associated with its constituent atoms I ,,,− ,,,−
| 29 |
E I can also be expressed in the form of a body-order expansion, as in eq . The local energy E I is determined by the intrinsic properties of the atom I, which depend on its elemental species, as well as the positions of the surrounding atoms. It is commonly represented as a function f of a descriptor vector x I , which consists of N D-dimensional variables that encode the local structural environment as a function of the atomic positions R N n . The introduction of x I in the representation of E I ensures strict invariance of the energy under a permutation of identical atoms, translation, and rotation–fundamental symmetries that must be preserved. While R N n itself is not invariant under these operations, invariant descriptors can be constructed from R N n . By expressing E I as a function of these descriptors, the required symmetry of the energy can be strictly enforced. For example, in the first MLFF for condensed systems, the Behler–Parrinello NNP, a cutoff radius R cut is introduced. The descriptor x I is then constructed as a function of the atomic positions within this cutoff radius, incorporating two-body and three-body terms. These descriptors are transformed via nonlinear activation functions, thereby enabling the representation of higher-order interactions. The generation of higher-order interaction terms from lower-order interaction terms through a nonlinear function can be understood from the following Taylor expansion
| 30 |
Many MLFFs developed subsequently have adopted a similar framework, albeit with different choices of descriptors and functional forms. This approach emphasizes the physical property that atomic interactions are generally stronger at shorter distances. However, interactions also include long-range effects, such as dispersion interactions and Coulomb interactions. Equation allows for an approximate representation of long-range interactions by distributing their influence into short-range interactions. For instance, an interaction originating from the atom I can influence the position of the atom J if it lies within the cutoff radius R cut of I. In turn, changes in the position of the atom J can affect the position of the atom K if K is within R cut of J, and so on. However, this indirect representation does not physically capture long-range interactions correctly. This issue becomes particularly evident in the failure to capture phenomena such as the LO-TO splitting of optical modes in polar materials and the subtle correlations associated with ionic motion.
To better capture long-range interactions, an approach has been proposed in which the total energy is decomposed into short-range and long-range interactions. The short-range interactions are represented using eq , while the long-range interactions are treated explicitly in the form of electrostatic interactions or dispersion forces ,,−
| 31 |
An accurate and flexible representation of the electrostatic long-range energy functional E long across a wide range of materials remains a challenging problem and is an active area of research in the development of MLFFs. Methods have also been proposed to incorporate not only the local environment of the atom I but also the surrounding environments of other atoms in the representation of E I −
| 32 |
Alternatively, a different approach that uses a descriptor x N D representing the entire molecular or crystalline structure to directly predict the total energy without decomposition has been proposed ,,,,−
| 33 |
In both eqs and (), neural network is predominantly used. It should be noted, however, that the neural network formulations do not explicitly capture the physics of long-range electrostatic interactions. While they enhance the representational capability of the energy landscape by incorporating longer-range information compared to MLFFs that rely solely on the short-range term E short, they may still fail to fully capture the subtle correlations arising from electrostatic interactions.
In the following subsections, we explain three commonly used representations of E I : linear regression, kernel regression, and neural networks.
Linear Regression
In linear regression models, the energy E I is expressed as a linear combination of the descriptor vector x I ,,
| 34 |
where the vector c consists of N D regression coefficients, which are determined to reproduce the training data. In materials composed of multiple elements, c depends on the elemental species. However, for simplicity, this elemental dependence is not explicitly indicated here. The representational capability of a function in linear regression depends on the completeness of the descriptors used as basis functions. Assuming that the descriptor components are linearly independent, increasing the number of basis functions enhances the representational capability, thereby improving the accuracy of the resulting ML model. However, as the number of basis functions increases, the degrees of freedom of the model increase, requiring a larger amount of training data to achieve high accuracy while avoiding overfitting. The feasibility of achieving high accuracy with a limited amount of training data depends on whether the descriptors can compactly represent atomic interactions. Systematic formulations of descriptors for many-body interactions have been proposed, , and it is common to use descriptors that can represent interactions up to four- or five-body terms as basis functions.
A key advantage of linear regression is that the regression coefficients can be determined by solving simple linear equations. Additionally, as shown in Figure (a), Bayesian inference enables the prediction of not only the mean of the regression coefficients but also their variance. Moreover, it is possible to statistically assess whether a given descriptor x I falls within the interpolation region of the training data. , This error estimation capability enables the construction of active learning algorithms that select the structures from which training data should be acquired. ,,,, Furthermore, it allows the generated ML model to determine whether its predictions lie within the interpolation region of the training data. ,
4.
Illustrations of (a) Gaussian process regression, (b) the weight-space view of the kernel regression, (c) the Behler–Parrinello neural network, and (d) the crystal graph convolutional neural network. The inset in (a) shows the predicted and actual force errors (adapted with permission from ref Copyright 2019 American Physical Society). The schematics in (c, d) have been reconstructed based on refs and for high-resolution presentation.
Kernel Regression
In kernel regression, E I in (eq ) is expressed as a linear combination of kernel functions ,− ,,,
| 35 |
where the vector k is an N K-dimensional vector whose elements are positive definite kernel functions k(x I , x I K ). The kernel function k(x I , x I K ) can be interpreted as a measure of similarity between the local environment of the atom I, described by x I , and the local reference environment represented by the reference point x I K in the descriptor space. The reference points are often selected from the local environments of atoms in structures within the training data.
As shown in Figure (b), from the perspective of weight space, eq can be interpreted as defining N K reference points in the descriptor space, placing kernel functions at each of these points, and adjusting their weights to represent the PES. By selecting reference points from the training data of the target system, the energy surface of that system can be efficiently represented.
Equation can also be interpreted from the perspective of function space. Consider expanding the function f in terms of H basis functions ϕ h (x I ), which are nonlinear with respect to the descriptor x I
| 36 |
where ϕ is a feature vector whose elements are ϕ h (x I ). The optimization problem of the coefficients c in ridge regression with Tikhonov regularization is equivalent to the problem of optimizing w in eq , with the kernel function defined as k (x, x′) = ϕT(x)ϕ(x′). This transformation is known as the dual representation. Compared to linear regression, kernel regression offers several advantages. In linear regression, eq assumes a linear dependence of E I on x I ; however, in reality, there is no guarantee that E I can be accurately expressed as a linear function of x I . Therefore, in linear regression, it is necessary to predefine a descriptor x I that allows for an accurate expansion of E I . In kernel regression, a nonlinear mapping is applied to transform the N D-dimensional descriptor space x I into an H-dimensional feature space ϕ, converting a nonlinear problem in x I into a linear problem in ϕ. However, this transformation alone does not fundamentally resolve the issue, as it merely replaces the task of selecting an appropriate representation of x I with the selection of ϕ. The kernel regression addresses this issue by defining kernel functions as inner products of feature vectors rather than explicitly constructing ϕ. Similar to the basis functions used in linear regression, the form of the kernel function must be predefined. However, even if the form of the kernel function is fixed, the kernel functions k(x I , x I K ) used to expand E I vary depending on the N K reference points. As a result, the feature vector implicitly handled in the kernel regression changes according to the reference points as ,
| 37 |
where Λ h and U I K h correspond to the eigenvalues and components of the eigenvectors of the N K × N K matrix K, whose elements are given by the kernel function k(x I K , x J K ). In kernel regression, reference points can be selected based on the training data, allowing for flexible construction of the feature vector. This method represents functions by setting inner products based on the data, without explicitly handling the high-dimensional feature space. This is known as the kernel trick.
From the perspective of interactions, as shown in the Taylor expansion in eq , kernel regression can be interpreted as generating higher-order interaction terms from descriptors corresponding to lower-order interactions through the nonlinear mapping from the descriptor space to the feature space.
While kernel regression provides flexibility in handling nonlinear problems, the coefficients w can still be determined by solving a system of linear equations. Consequently, regression coefficients can be efficiently determined in kernel regression as well. Furthermore, as illustrated in Figure (a), the uncertainty of the prediction can be readily evaluated using Bayesian inference and other statistical methods.
Equation was originally proposed to represent short-range interactions using the descriptor x I , which characterizes the local environment within a cutoff radius from the atom I. Subsequently, methods have been developed to separate short-range and long-range interactions, introduce appropriate descriptors for each, and represent the PES in the form of eq . , This approach enables a flexible representation of long-range interactions by employing descriptors that distinguish structural differences over extended distances. However, eq does not explicitly incorporate the physical laws governing electrostatic interactions. As noted by Faller and co-workers, a major challenge in developing any MLFF–including those based on kernel regression–is the incorporation of physics-based long-range multipole interactions.
Neural Network
Among the various representation methods for MLFFs, neural networks are currently the most widely used, and numerous variations continue to be proposed. As mentioned earlier, the first method proposed for condensed systems, which remains widely used today, is the neural network potential (NNP) developed by Behler and Parrinello. This approach also adopts the form of eq . For instance, when using a network structure with two hidden layers, as illustrated in Figure (c), E I can be expressed as
| 38 |
where G I,μ represents descriptors known as symmetry functions, φ is an activation function, the parameter N ND denotes the number of nodes in the m-th hidden layer, and a kl and b l are learnable parameters optimized to reproduce the training data. The descriptor G I,μ consists of two-body and three-body variables that describe the local environment within a cutoff radius R cut around the atom I. These descriptors are invariant under atomic permutation, translation, and rotation operations. However, as mentioned earlier, this formulation does not explicitly account for long-range interactions beyond the cutoff radius R cut. To explicitly incorporate long-range interactions, methods have been proposed to decompose the total energy into short-range and long-range components, as shown in eq . The short-range interactions are represented by eq , while the long-range interactions are described using Coulomb interactions between point charges or Gaussian charges. According to their treatment of interatomic interactions, Behler and co-workers classified NNPs into different generations. The first generation consists of early models developed for small systems by Blank et al. The second generation includes models that use eq to represent short-range interactions among many particles in condensed systems. The third and fourth generations explicitly incorporate Coulomb interactions. In third-generation NNPs, the point charge of the atom I is expressed as a function of local environment descriptors using a neural network of the form given in eq . However, since charge distributions inherently depend on the long-range structure of the system, this approach suffers from accuracy limitations. Fourth-generation NNPs address this issue by defining the long-range interaction energy E long in eq as the electrostatic interaction energy E elect between Gaussian charges centered at atomic positions, with each Gaussian charge having a total atomic charge Q I
| 39 |
The charge Q I is determined through a process known as charge equilibration, , in which the energy associated with the charges is minimized with respect to Q I
| 40 |
where χ I and γ I represent the electronegativity and hardness of the atom I, respectively, and γ I is a constant specific to the elemental species. The electronegativity χ I is expressed using a neural network, as in eq . The parameters of the neural network are optimized to reproduce the atomic charges obtained from FP calculations, such as Hirshfeld charges. Consequently, the charge distribution is determined in a manner that reflects the influence of the global structure while ensuring the equilibration of chemical potential. This charge determination mechanism is physically well-founded and has been shown to provide a more accurate representation of long-range interactions between organic molecules compared to third-generation NNPs. Charge equilibration was originally proposed to predict charge distributions in molecules and solids within the framework of classical force fields. A conceptually similar model has been employed to describe charge distributions at electrolyte–electrode interfaces, and the method has since been further developed , and integrated into MD simulation packages such as MetalWalls. Therefore, this approach can be a promising direction for representing long-range interactions in MLFFs based on physical laws. However, challenges remain in the representation of electrostatic interactions. FP calculations do not uniquely define point charges, and constructing an ML model for electronegativity requires training data based on arbitrarily assigned charges. Various charge definitions exist, ,− but all involve some degree of arbitrariness, and there is no guarantee that the computed charges accurately capture electrostatic interactions. In addition, long-range interactions in materials involve not only monopole but also higher-order multipole contributions. Therefore, it should be recognized that the Gaussian charge model alone may not fully or accurately capture these interactions. Moreover, it remains uncertain whether the simple functional form of eq can universally capture the partial screening behavior inherent to dielectric materials. In addressing this issue, a promising approach would be to incorporate a more physically realistic charge distribution, uniquely determined from FP, along with its interactions with MLFFs. Further research on such approaches is necessary.
In recent years, methods have been proposed to incorporate long-range interactions within the framework of neural networks, utilizing convolutional neural networks (CNNs) ,,,, and message-passing neural networks (MPNNs). − ,, For instance, in the Crystal Graph Convolutional Neural Network (CGCNN) proposed by Xie and Grossman, which was the first application of CNNs to condensed systems, the crystal structure is represented as graph data, where atoms serve as nodes and atomic bonds act as edges, as shown in Figure (d). This graph representation is used as the network input. The information of each node and edge is represented by feature vectors h I and e IJ , respectively. The feature vector h I consists of discretized numerical values that encode elemental properties such as position in the periodic table (group and period), electronegativity, and covalent radius. The feature vector e IJ consists of discretized values of bond distances. The feature vector of the atom I is updated iteratively in a process known as convolution, which occurs in layers referred to as convolutional layers. During this process, the information from neighboring atomic nodes and edges is incorporated into the updated feature vector. For example, in a simple convolutional operation, the update at the t-th iteration is performed as
| 41 |
where φ is a nonlinear activation function, N(I) represents the set of atoms bonded to the atom I, and ⊕ denotes a concatenation operation. The parameters W c , W s , and b t characterize the convolution process. This formulation shows how the properties of neighboring atoms J and their bonding information contribute to updating the feature vector of the atom I. By iteratively applying this process, information from distant atoms is incorporated into the feature representation of the atom I, resulting in a feature vector that reflects the long-range structural environment. Although more complex functions are used in practice, the essential mechanism remains the same: neighboring atoms exchange information through convolutional layers. The generated feature vectors are then further transformed by pooling layers and subsequently used for the regression of the total energy E. Following CGCNN, various methods utilizing crystal graph representations and CNNs have been proposed and applied to the regression of E and E I . ,,,
MPNNs incorporate information from neighboring atoms in a manner similar to CNNs. ,,,,, In MPNNs, a message function M t is used to generate a message m I from the states of the nodes representing the neighboring atoms of the atom I at the t-th layer, denoted as σ J 1 , σ J 2 , ···, σ J ν
| 42 |
Next, the node state of the atom I, σ I , is updated using the update function U t , based on the message m I and the current node state σ I
| 43 |
If M t and U t are set as the concatenation and activation function, respectively, the formulation takes a form similar to eq . Various forms of MPNNs have also been proposed and applied to the regression of E and E I . − ,
In general, neural networks provide more flexible functional representations compared to linear regression and kernel regression, but they also tend to require larger amounts of training data. However, this characteristic strongly depends on the complexity of the network and the type of descriptors, which will be discussed in the next subsection. To evaluate the uncertainty of predictions made by MLFFs based on neural networks, the Query by Committee (QBC) method is often employed. ,− In this approach, multiple ML models are generated from the same training data by initializing the network parameters differently, and the variance among their predictions is used as a measure of uncertainty. Compared to Bayesian and statistical methods commonly used in linear regression and kernel regression, QBC lacks a rigorous mathematical foundation. While Bayesian approaches for uncertainty estimation also exist for neural networks, QBC is often preferred in practice due to its lower computational cost.
Descriptor
Descriptors should have the ability to systematically and compactly represent many-body interactions, as their selection determines both the amount of training data required for constructing an ML model and the achievable accuracy. Following the proposal of the Behler–Parrinello NNP, various research groups introduced invariant descriptors for two- to four-body interactions that are rotationally invariant. , More recently, efforts have been made to unify these descriptors into a systematic many-body descriptor. ,,, Furthermore, recent advancements have extended invariant descriptors to equivariant descriptors, which transform covariantly under rotations. ,,,,− This extension was initially motivated by the need to predict physical properties such as polarizability and electron density, which exhibit equivariance under rotational transformations. , Shortly thereafter, a neural network architecture composed of nodes described by equivariant descriptors and edges that transform variables preserving equivariance was proposed, known as the equivariant neural network (E3NN). , In this approach, the invariant components of the final network output are used for energy regression. It was later discovered that combining equivariant descriptors, E3NN, and MPNN improves the accuracy of PESs for condensed systems, even with limited training data. This led to MLFFs that incorporate many-body equivariant descriptors through the frameworks E3NN and MPNN. ,,, The following sections provide an explanation of invariant and equivariant descriptors.
Invariant Descriptor
To provide a bird’s-eye view of various descriptors, we first introduce a framework that unifies major descriptors and explain how each individual descriptor corresponds to specific components within this framework. The formulation of descriptors representing the local environment of the atom I begins with the definition of the following local elemental distribution function
| 44 |
where R IJ denotes the distance between atoms I and J, and f cut(R) is a cutoff function that smoothly approaches zero at a radius R cut. The function g is a Dirac delta function or a broadened Gaussian function. For simplicity, the elemental dependence is not explicitly indicated in eq , but the distribution function is constructed separately for each elemental species based on the methodology described below. This approach treats different elemental species as entirely distinct particles without any similarity. While this method accurately distinguishes different elemental species, it has the drawback that the number of descriptors scales exponentially with the number of elemental species, leading to a significant computational cost in multielement systems. To address this issue, several strategies have been proposed, including tensor reduction (or element embedding) methods, in which learnable weights are assigned to elemental species as well as radial and angular channels. These approaches enable the collective treatment of multiple elemental distribution functions and descriptors. In addition, compression techniques have been developed to extract the most relevant descriptors from elemental distributions through data-driven selection. ,,,,,,,−
From the two-body distribution function in eq , the following (ν + 1)-body correlation function is constructed
| 45 |
where the function ρ i represents the probability of finding ν + 1 atoms, including the atom I, in a specific relative configuration, where the first atom is located at a relative position r 1, the second atom at r 2, and so on. This function is invariant under an atomic permutation and a translation but is not invariant under a rotation. A rotationally invariant distribution function is obtained by averaging over the rotational operation
| 46 |
The function ρ̅ I is a descriptor that represents the (ν + 1)-body relative configuration while preserving the required invariance. In practical MLFF constructions, discretized variables of the relative coordinates (r 1, ···, r ν) are used as descriptors. Owing to its analytical convenience, relative coordinates are often expressed in spherical coordinates, and ρ I (r) is expanded in terms of orthogonal basis functions for the radial and angular components, χ nl and Y lm , respectively
| 47 |
where Y lm is the spherical harmonics, χ nl is, for example, chosen as a polynomial or a spherical Bessel function, the index v represents the set of indices (nlm), and ϕ v (r) is defined as ϕ v (r) = χ nl (|r|)Y lm (r̂). The expansion coefficients A I,nlm are used to construct descriptors. ,, The expansion coefficients are computed using the radial functions and spherical harmonics as
| 48 |
| 49 |
In the Smooth Overlap of Atomic Positions (SOAP), which employs a Gaussian function for the function g in eq , the radial function h nl is given by the following integral of the radial basis function χ nl over the radial direction
| 50 |
In the Atomic Cluster Expansion (ACE), which employs a Dirac delta function for the function g, the radial function h nl is identical to the basis function χ nl . Hence, η v = ϕ v . Consequently, the difference in the expansion coefficients between SOAP and ACE arises solely from the choice of the function g.
Using eq , the components of the invariant descriptor are obtained through the following integral
| 51 |
where v is a notation that represents all possible combinations of (v 1, ···, v ν).
For example, the descriptor for ν = 1 is given as
| 52 |
The component A I, v 1 with v 1 = (n 100) corresponds to the expansion coefficient that appears when the radial distribution function is expanded using the basis function χ n 10. Hence, n 1 serves as an index for the interatomic distance between a pair of atoms depicted in Figure . ,, The descriptor using the two-body symmetry functions proposed by Behler and Parrinello can be interpreted as a descriptor that represents g using a Dirac delta function and utilizes the two-body symmetry functions as the basis functions. Similarly, the two-body component of the FCHL descriptor proposed by Faber et al. can be interpreted as a two-body descriptor that discretizes the radial direction in real space.
The descriptor for ν = 2 is given as
| 53 |
The coefficient p I, n 1 n 2 l 1 corresponds to the expansion coefficient of the angular distribution function, which represents the relative configuration of three-body. , It is expanded using the two radial basis functions χ n 1 l 1 and χ n 2 l 1 , along with the Legendre polynomial P l 1 of order l 1, which serves as the angular basis function. The two indices, n 1 and n 2, specify the lengths of the two sides of the triangle formed by three atoms depicted in Figure , while the other index, l 1, represents the angle between these two sides. ,, This descriptor is known as the power spectrum and is used as the descriptor in the SOAP kernel. Similar to the two-body case, the three-body symmetry functions of Behler and Parrinello and the three-body descriptors proposed by Faber et al. are equivalent to this descriptor.
The descriptor for ν = 3 is given as
| 54 |
In the tetrahedral structure formed by four atoms shown in Figure , the three edge lengths are specified by n 1, n 2, and n 3, while the three angles are specified by l 1, l 2, and l 3. The coefficient is known as the bispectrum and has been used as a descriptor in GAP. The four-body descriptor proposed by Faber et al. can be considered a variant of this descriptor.
As demonstrated by Pozdnyakov and co-workers, both the power spectrum p I,n 1 n 2 l 1 and the bispectrum b I,n 1 n 2 n 3 l 1 l 2 l 3 possess inherent limitations in completely capturing the local structural environment. To address this issue, descriptors for ν = 4 and higher have been formulated as follows ,,
| 55 |
| 56 |
where the coefficient C vv′ is a generalized Clebsch–Gordan coefficient that integrates the coefficients appearing in eqs to (). In the Atomic Cluster Expansion (ACE), the elements of C vv′ are formulated using Wigner symbols and are employed as basis functions in linear regression. This formulation in ACE explicitly represents the energy in the form of a body-order expansion, as in eq . Spherical harmonics can be expressed in Cartesian coordinates, providing a representation in which descriptors are given as products of moment tensors. The Moment Tensor Potential (MTP), proposed by Shapeev, utilizes this form as basis functions for linear regression.
The above constitutes an overview of invariant descriptors. Four additional points regarding invariant descriptors are discussed below. The first point concerns the equivalence of descriptors. As mentioned earlier, lower-order invariant descriptors, such as two-body and three-body terms, have been proposed in various forms. , Higher-order interaction terms can be derived from these lower-order descriptors through nonlinear transformations, such as neural network activation functions or kernel functions, as shown in eq . Although the proposed descriptors may appear to have different mathematical forms, the information they encode is fundamentally equivalent. In fact, a study comparing the Behler–Parrinello NNP, which has a completely different formalism, with the power spectrum-based SOAP kernel demonstrated that when trained on the same data set, both methods yielded nearly identical results in predicting the density isobar of liquid water. The equivalence of four-body and higher-order invariant descriptors has not yet been fully established. However, with careful evaluation, it is expected that MLFFs utilizing equivalent descriptors should achieve comparable performance.
The second point concerns the completeness of descriptors. As demonstrated by Glielmo et al., nonlinear transformations of lower-order terms alone cannot fully generate interaction terms of all orders. This leads to an inherent limitation in the accuracy of models due to the incomplete representation of many-body interactions. To overcome this issue, descriptors of arbitrary order must be systematically incorporated in the form of the body-order expansion. This formalization has been provided by the ACE and the MTP. However, since the number of descriptors grows exponentially with ν, it is necessary to truncate the expansion at a lower order in practical computations. For this reason, two-body and three-body nonlinearized methods, such as NNP and SOAP, are still widely used. Similarly, in ACE, the expansion is typically truncated at four- or five-body interactions in actual calculations.
The third point concerns the role of eqs and () in constructing many-body descriptors. If (ν + 1)-body descriptors were computed individually, the computational cost would scale proportionally to N n to the power of ν + 1, making direct computation impractical. The approach described above circumvents this issue by first computing the distribution function in eq through the enumeration of atoms surrounding the central atom I and then constructing higher-order terms by taking their product in eq . This procedure is known as the density trick. ,
The fourth point concerns tensor reduction. While the density trick mitigates the steep scaling of descriptor calculations with respect to the body order ν + 1, the number of many-body descriptors still grows rapidly with the number of element types, scaling as N el . As briefly mentioned at the beginning of this subsection, this scaling has hindered the direct application of higher-order descriptors in MLFFs. The challenge has been addressed through tensor reduction, which embeds chemical element information into a fixed-size latent space using a learnable transformation . This strategy has enabled the practical application of higher-order terms to multielemental materials across various MLFF frameworks. ,,,,,
Equivariant Descriptor
Around 2018, a method was proposed in which descriptors equivariant to a three-dimensional rotational operation were used as inputs to E3NN, while the invariant components of the output layer were employed to represent the PES. ,
A descriptor is defined as equivariant under rotational operations if it satisfies the following properties
| 57 |
where D L (R̂) is a matrix representing a rotational operation, known as the Wigner D-matrix. This descriptor takes different forms depending on L: for L = 0, it is a scalar with a single degree of freedom that remains invariant under the rotation; for L = 1 and L = 2, they are a three-dimensional vector with three degrees of freedom and a second-order symmetric tensor with five degrees of freedom, respectively, both of which transform equivariantly under the rotation. The parameter L is referred to as the rank. One of the orthogonal basis functions of the SO(3) group, the spherical harmonics, satisfies eq . Consequently, A I, v 1 , which is expressed as a linear combination of spherical harmonics, is equivariant. However, the product A I, v 1 A I, v 2 is not equivariant. A three-body equivariant descriptor can be constructed by linearly combining products of spherical harmonics using Clebsch–Gordan coefficients as expansion coefficients
| 58 |
Here, the fact that eq is an equivariant tensor of rank L is proven using the following equation, which the Clebsch–Gordan coefficients satisfy under the rotational operation
| 59 |
where D l (R̂) represents a component of the Wigner D-matrix. A four-body equivariant descriptor can be obtained similarly to eq by taking the previously constructed three-body equivariant descriptor B I, v 1 v 2 , multiplying it by A I, v 3 , and performing a linear combination using the Clebsch–Gordan coefficients. In this manner, an equivariant descriptor for an arbitrary (ν + 1)-body system can be constructed. These equations can be summarized in a similar form to eq as ,,,,
| 60 |
In E3NN, the state of each node is represented using equivariant descriptors, and information is transferred to the next layer while preserving equivariance, as expressed in eq . , Since E3NN is well-suited to the MPNN framework, MPNNs in the E3NN formalism were soon adopted for energy surface regression. ,,,,
At the end of this section, we explain MACE ,,, as an example of a state-of-the-art MLFF that utilizes many-body equivariant descriptors as inputs to E3NN and MPNN. Similar to other MPNNs, MACE performs message-passing through a T-layer neural network, where each node corresponds to one of the N n atoms. The state of a node at layer t is represented by the atomic number z I , position R I , and equivariant descriptor h I, kLM for the atom I. Here, L represents the rank of the equivariant descriptor, as previously explained. In addition to L and M, the descriptor also has an additional index k, referred to as the channel index, which takes values from 1 to N ch. The number of channels, N ch, is typically set between 64 and 256. As described later, in MACE, discretized information about elemental species and their surrounding elemental distributions is mapped to N ch variables using learnable weights through the tensor reduction. Thus, k can be regarded as an index that encodes both elemental species and structural features. Since the state of a node is described by the three indices k, L, and M, the total number of variables per node is given by N ch × (2L+1). The input for the I-th node is computed as
| 61 |
The weight W kz is a learnable parameter determined through training, and the summation over z is taken over all atomic numbers. This technique, commonly referred to as element embedding (or tensor reduction), assigns element-specific weights and compresses the information into channel k, enabling the handling of multielement systems without exponentially increasing the number of descriptors. Messages are generated from the equivariant descriptors h I, kLM . The message construction process begins by linearly mixing the equivariant descriptors of each node with respect to the channel variables
| 62 |
where the weight W k k̃L is also a learnable parameter determined through training. Next, an atomic basis function corresponding to the previously defined A I, v is constructed as
| 63 |
where the function ϕ IJkη1 l 3 m 3 corresponds to the previously defined ϕ v and serves as a one-particle basis function, forming an equivariant tensor of rank l 3, and the index η1 represents one of the radial basis function indices, which will be discussed later. A key characteristic of MPNNs is that they incorporate information from the neighboring atom J into the one-particle basis function, as described below
| 64 |
The function R kη1 l 1 l 2 l 3 (r) corresponds to the previously defined radial basis function χ nl . However, in MACE, the radial basis function has a more flexible form, depending not only on the channel index k but also on the angular indices l 1, l 2 and l 3. This function is represented by transforming spherical Bessel functions through a three-layer multilayer perceptron (MLP). Unlike eq , in eq , the information associated with k̃ and η1, which correspond to elemental species, radial components, and angular components, is compressed into channel k. Another important point to note is that at t = 0, the l 3 = 0 component of eq provides an atomic basis function equivalent to that of ACE. Similar to eq , the product of ν atomic basis functions is taken to construct the product basis
| 65 |
Following the same procedure as in eq , an equivariant basis function of rank L is constructed as
| 66 |
where the summation over l and m denotes a summation over all possible combinations of l ξ and m ξ, and the subscript ην corresponds to the index v′ in eq and serves as a collective index for the nonzero basis functions. The message is constructed by taking a linear combination of the basis functions as
| 67 |
the message m I, kLM also forms an equivariant tensor of rank L similar to the basis functions. The descriptor for the (t + 1)-th layer is obtained by linearly mixing the message with the descriptor h I,kLM from the t-th layer as
| 68 |
The above process can be repeated T times to obtain the invariant descriptor h I,kLM at the T-th layer. In MACE, T is effectively set to 2. As a result, the equivariant descriptors h I,kLM and h I,kLM are generated. The invariant components of these equivariant descriptors are used to compute the energy E I of the atom I as
| 69 |
| 70 |
The first term is a linear combination of the invariant components of the equivariant descriptors output from the first layer. As mentioned above, this is equivalent to the ACE formulation. In other words, the first term reflects only the information within a radius of R cut. In the second term, the invariant components of the equivariant descriptors from the second layer–obtained after a single message-passing step–are nonlinearly transformed using the MLP function . As a result, the second term incorporates information within a radius of 2R cut. In practical calculations, four-body equivariant descriptors (ν = 3) are often used.
Neural networks utilizing equivariant descriptors, such as MACE ,,, and NequIP, have been shown to achieve high accuracy with a comparable or smaller training data compared to conventional MLFFs that rely on invariant descriptors. As a result, they have become one of the mainstream approaches in MLFF. However, the underlying reason for this improved efficiency remains unclear. The original motivation for using equivariant descriptors was based on the idea that interatomic interactions can be expressed as a sum of terms involving the products of multipole moments of the central atom and its neighboring atoms. The MACE framework follows this concept, where the network maintains an equivariant structure internally, takes products of equivariant multipoles (ϕ IJkη1 l 3 m 3 ), and ultimately utilizes the invariant components of the resulting descriptors to represent energy. However, the same applies to ACE and MTP, which also form invariant descriptors by taking products of equivariant multipoles (ϕ I,v ). One possible explanation is that message-passing enables the incorporation of longer-range interactions. Another possible reason for this accuracy is that, in ACE and MTP, invariant descriptors are constructed from equivariant descriptors (spherical harmonics) using a predefined functional form, as in eq . In contrast, MACE and NequIP adopt a more flexible framework, where invariant descriptors are formed through multiple layers including the MPNN with learnable parameters, optimizing the composition of tensor products during training. The ability to flexibly represent diverse many-body interactions may certainly contribute to accuracy. However, MACE and NequIP achieve high-accuracy regression even with small training data, despite their increased model complexity. This suggests that the invariant descriptors obtained in the final network layer provide a compact representation of many-body interactions. However, the underlying physical reason remains unclear.
Applications to Electrochemistry
Once the energy surface E(R N n ) is computed, properties related to ionics and electrochemical reactions can be predicted using analytical statistical mechanics models or simulations based on MD and MC methods, following the Hamiltonian described in eqs and (). Furthermore, the atomic-scale properties and mechanisms obtained from these simulations can be utilized in the modeling of larger-scale inhomogeneous systems. The use of MLFFs is expanding in both microscopic simulations and in bridging the gap from the microscopic to mesoscopic and macroscopic scales. The accessible time and length scales of MLFF-aided microscopic simulations range from 10–10 to 10–6 seconds and from 102 to 1011 atoms, depending on factors such as system complexity, target properties, required accuracy, and available computational resources. ,− ,,− , MLFFs also offer an accurate and flexible framework for coarse-graining methods. Additionally, the acceleration of simulations for various materials through MLFFs has enabled the exploration of a much broader range of materials than was previously possible. This chapter discusses the application of MLFFs to ionics and electrochemical reactions, as well as the associated computational methods.
Ionics
The evaluation of properties related to ionic conduction in electrolytes and active materials based on FP calculations has been conducted using two approaches: dynamic calculations and static calculations. As discussed later, ionic conductivity is rigorously formulated based on the fluctuation–dissipation theorem, , and in principle, it can be computed using MD simulations. However, FPMD is computationally expensive, making it difficult to accurately obtain ionic conductivity or diffusion coefficients at the operating temperature of electrolytes within feasible simulation times. To address this limitation, alternative methods such as extrapolating high-temperature results to estimate properties at operating temperatures have been employed. ,− Even with such extrapolations, applying FPMD to a large number of materials remains challenging. For this reason, static calculations have also been widely used. In this approach, the minimum energy migration path between quasi-minimum sites in an optimized electrolyte structure at absolute zero temperature is determined. , The diffusion coefficient is estimated from the calculated activation energies using dilute diffusion theory or kMC simulations. ,− However, static calculations require significant human effort and computational resources, and there is no guarantee that they fully capture the overall ionic conduction process. Recently, advances in MLFFs have helped overcome the limitations of MD simulations. This section introduces the MD-based approaches.
Based on the Green–Kubo equation, , ionic conductivity is formulated as
| 71 |
where e is the elementary charge, and V represents the system volume. The notation ⟨.⟩ denotes an ensemble average. In principle, ionic conductivity can be obtained by performing equilibrium MD under an appropriate ensemble and averaging the values inside the brackets. However, in practice, this method often fails to achieve convergence within the feasible length- and time-scales of equilibrium MD simulations.
To calculate ionic conductivity in electrolytes, an alternative approach is often adopted. In this approach, only the first term is evaluated while the second term is ignored. ,− ,, The first term corresponds to the contribution from self-diffusions, and the equation that uses only this term is known as the Nernst–Einstein equation. , In experimental studies, the Nernst–Einstein equation has also been frequently employed to estimate diffusion coefficients from conductivities measured via impedance spectroscopy, particularly in cases where direct measurements of the diffusion coefficients of charge carriers are challenging. A typical example is proton transport in aqueous electrolytes. − Proton conduction in aqueous environments is a key ion transport phenomenon in physical chemistry and biochemistry and also plays a critical role in proton exchange membrane (PEM) fuel cells and water electrolysis, where Nafion is used as a representative PEM as shown in Figure (a). Nafion absorbs water and forms clusters consisting of water molecules and sulfonic acid groups, which serve as ion-exchange sites. These clusters, typically a few nanometers in size, are connected to form continuous channels that enable proton transport. The environment within these channels resembles that of aqueous solutions, and the proton transport mechanism in Nafion is considered similar to that in aqueous media. In aqueous solutions, protons chemically bind with water molecules to form hydronium ions (H3O+) and hydrated complexes such as Zundel ions (H5O2 ), where hydronium ions form hydrogen bonds with surrounding water molecules. ,,, However, proton transport does not necessarily occur via the simple diffusion of these complex ions. As shown in Figure (a), protons can transfer from one water molecule to another via hydrogen bonding with H3O+. Furthermore, once an H3O+ ion is formed by accepting a proton, it can subsequently donate a proton to another water molecule through hydrogen bonding. Importantly, the donated proton in this second step is not necessarily the same proton that was initially accepted. This continuous proton transfers between water molecules constitutes the Grotthuss mechanism. ,, In contrast, when protons remain bound to H3O+ or hydrated complexes and move as part of these species, the transport process is referred to as the vehicle mechanism. A key property of the Grotthuss mechanism is that long-range charge transport is achieved through successive proton transfers, even without significant displacements of hydrogen and oxygen nuclei. In reality, however, proton conduction does not occur via an idealized Grotthuss mechanism alone; rather, it results from a complex interplay of both the Grotthuss and vehicle mechanisms. Nevertheless, the fact that the ionic conductivity of protons in aqueous solutions is more than four times higher than that of water molecules or other monovalent cations highlights the significant contribution of the Grotthuss mechanism. The presence of the Grotthuss mechanism complicates the identification of the diffusion coefficient of charge carriers. This is because tracking the trajectories of hydrogen or oxygen nuclei alone does not yield the trajectories of the actual charge carriers. For these reasons, the Nernst–Einstein equation has been used to estimate the diffusion coefficient of charge carriers from experimentally measured conductivity. In aqueous solutions, NMR measurements of the time scale of proton transfers as a function of pH and temperature have been conducted. It has been confirmed that the diffusion coefficient estimated using the measured time scale and an assumed displacement by a proton transfer closely agrees with the diffusion coefficient estimated from conductivity via the Nernst–Einstein equation. However, in complex electrolyte membranes such as Nafion, performing such a precise validation is challenging. ,,
5.
Applications of MD simulations to ionics: (a) diffusion coefficients of protons and water in hydrated Nafion, calculated from mean square displacements (open dots), compared with experimental values (closed dots) (adapted with permission from ref Copyright 2023 American Chemical Society), (b) Li-ion conductivity in LGPS, calculated using the Nernst–Einstein equation (σdilute), the Green–Kubo equation (σEMD), and the color diffusion MD method (σNEMD) (reproduced from ref under CC-BY 4.0), and (c) Li-ion conduction under an electric field, simulated using Behler–Parrinello-type neural network models for the Born effective charge tensor and interatomic potential (reproduced from ref under CC-BY 4.0).
The presence of the Grotthuss mechanism also makes MD simulations of proton transport challenging. Classical MD has been widely used to study ion transport phenomena in aqueous solutions. However, describing the formation and dissociation of chemical bonds in the Grotthuss mechanism requires specialized models. For systems with a limited number of elemental species, such as protons in pristine water, the empirical valence bond (EVB) method has been established and has contributed to elucidating proton transport mechanisms. However, for multielement systems like Nafion, constructing such models is time-consuming. FPMD can accurately describe these phenomena. However, evaluating proton conduction within the heterogeneous structure of Nafion requires an expensive computational cost. ,
Recently, the kernel method described in the previous section was applied to proton transport in Nafion. In this study, a total of 2.4 ns of MD simulations were performed with a 0.5 fs time step on hydrated Nafion, modeled using unit cells containing 832 to 1264 atoms, employing VASP 6.3.0. Prior to the production runs, an MLFF was trained on-the-fly via MD simulations of smaller hydrated Nafion fragments. A total of 278 structures were sampled over 10 days using 32 cores of an Intel Xeon Platinum 8358 processor (2.6 GHz). With the trained MLFF, each MD step took 8.8 s on 32 cores–already a significant speedup over the 1032 s required for full FP calculations. The latest version, VASP 6.5.1, further reduced this to just 0.5 s per step on 16 cores, thanks to algorithmic improvements. Overall, the MLFF approach enabled a three-orders-of-magnitude acceleration in total simulation time compared to the FP method, including the training runs. Figure (a) presents the calculated and experimental diffusion coefficients of charge carriers and water molecules as a function of the water uptake λ (the number of water molecules per sulfonic acid group) in Nafion. The diffusion coefficient of charge carriers was determined by identifying H3O+ at each MD step and tracing its origin from the H3O+ in the preceding step, enabling the construction of charge carrier trajectories. The mean squared displacement (MSD) of these trajectories was computed to determine the diffusion coefficient. For water molecules, the diffusion coefficient was calculated from the MSD of oxygen atoms. For both water molecules and charge carriers, the calculated diffusion coefficients agree well with experimental values. Both simulations and experiments show that as λ increases, the diffusion coefficient of charge carriers eventually exceeds that of water molecules. In the experiments, the diffusion coefficient of water was measured using NMR, while the diffusion coefficient of protons was estimated by converting the conductivity using the Nernst–Einstein equation. Strictly speaking, it was unclear whether these two values could be directly compared. On the other hand, the diffusion coefficients obtained from simulations were derived consistently from MD trajectories. This result suggests that while the Grotthuss mechanism is not as dominant as in aqueous solutions, it still contributes to proton transport in hydrated Nafion.
The method of estimating conductivity using only the self-diffusion contributions in the first term of eq ensures good convergence in simulations. However, the second term can sometimes have a significant impact on ionic conduction. A clear example is when cations and anions move together without dissociating. In this case, the first term contributes additively through the self-diffusions of both cations and anions, yielding a finite conductivity. However, in reality, the first and second terms cancel each other out, resulting in zero ionic conductivity. Thus, when ions exhibit correlated motion with other ions, the Nernst–Einstein equation can lead to large errors. For this reason, nonequilibrium MD simulations are used to directly simulate the behavior of charge carriers under an external electric field. One such method is the color diffusion method. , This approach assigns a color charge C I to each ion I. An external electric field is applied, and the force acting on the atom I is computed as
| 72 |
The current density due to the color charge is calculated as
| 73 |
The same calculation is performed using two additional linearly independent electric fields, and , to obtain the corresponding current densities, J 2 and J 3. From these calculations, the conductivity tensor σ is obtained as
| 74 |
where J is a matrix collecting the current density vectors. The color diffusion method provides a well-converged result within a feasible simulation time. Figure (b) shows the simulation results obtained using the color diffusion method for Li10GeP2S12 (LGPS), a representative solid electrolyte for Li-ion batteries. In this simulation, as shown in the figure, color charges of +1, −3, and −4 were assigned to Li, PS4 tetrahedra, and GeS4 tetrahedra, respectively. This study also compared the conductivity calculated using the Nernst–Einstein equation and the Green–Kubo formula from equilibrium MD simulations. FPMD was used in this study. As shown in the figure, the conductivity obtained from the Nernst–Einstein equation (σdilute) is significantly lower than those obtained using the Green–Kubo formula (σEMD) and the color diffusion method (σNEMD). Additionally, while σEMD and σNEMD are in good agreement, σNEMD exhibits faster convergence.
However, the assignment of color charges involves inherent arbitrariness. If color charges are assigned unrealistically, the resulting ionic conduction mechanism may also become unrealistic. At the same time, as noted in the previous section, there is no completely arbitrary-free method to assign point charges to atoms. As a more rigorous approach to simulating the dynamics of atoms under an electric field within the framework of linear response theory, a method utilizing the Born effective charge tensor has been proposed. The Born effective charge tensor is defined as the second derivative of the total energy of the system with respect to the electric field and atomic positions:
| 75 |
where i and j represent one of the three Cartesian directions: x, y, or z. The term P i denotes the polarization in the i-direction, u I,j represents the displacement of the atom I in the j-direction, and F I,j is the force acting on the atom I in the j-direction. The Born effective charge offers a well-defined framework within linear response theory to compute the force on the atom I under an electric field as
| 76 |
where Z I is a tensor whose elements are given by Z I,ij . The current density J 1 is obtained from Maxwell’s equations under a static electric field as
| 77 |
By performing the same procedure for two additional linearly independent electric fields, and , the conductivity tensor can be obtained. Recently, an ML method has been proposed that predict the Born effective charge tensor using the invariant and equivariant descriptors described in previous section. ,, This approach enables NEMD simulations under an electric field with high accuracy and without arbitrariness. Figure (c) shows the result of the 300 ps NEMD simulation of Li3PO4, modeled with a unit cell containing 127 atoms, conducted by Shimizu et al. as a demonstration. In this simulation, an ML model was constructed to predict the polarizability using a Behler–Parrinello-type neural network. The Born effective charge tensor was computed by taking the partial derivative of the ML model with respect to atomic positions. The Behler–Parrinello NNP was also used to describe interatomic interactions in the absence of an applied electric field. Although conductivity was not explicitly computed in this study, the results clearly show the Li-ion conduction along the direction of the applied electric field. A comparison of computation times between (b) FPMD-based simulations and (c) NNP-based simulations shows that the NNP-based approach in (c) allows for longer simulation times.
Electrochemical Reactions
Key properties of electrochemical reactions are thermodynamics and kinetics of elementary reactions, which are represented by free energy landscapes along reaction coordinates of elementary processes. Thermodynamic properties of an elementary process can be determined from a free energy difference between a reactant and a product along the reaction coordinate. Similarly, a free energy difference between a reactant and a transition state allows for the calculation of the rate constant based on TST. However, as briefly mentioned previously, the treatment of electrochemical reactions requires the theory of the grand canonical ensemble for electrons and ions, and modeling the electric double layer is essential for describing electrochemical reactions within this theoretical framework. However, the motion of ions and solvent molecules is slow, making it challenging for FP calculations to handle. This issue has limited the application of FP methods to electrochemical reactions. Since the early 2000s, research on FP methods for electrochemistry has become increasingly active, leading to the proposal of various approaches. ,,,− , More recently, MLFFs have been utilized in FP-based free energy calculations that incorporate the statistical averaging of ion and solvent motion at finite temperatures, enabling accurate predictions of electrochemical properties. ,,,,− This section discusses electrochemical reactions in homogeneous solutions and at electrode–electrolyte interfaces, introducing both conventional FP methods and finite-temperature MD calculations using MLFFs. The discussion highlights how theoretical, computational, and data-driven sciences are increasingly integrating and advancing in the field of electrochemical reactions. Besides the homogeneous solution and electrolyte–electrode interfaces, redox reactions within bulk solid electrodes are also practically important electrochemical processes. A typical example is redox reactions in Li battery cathode materials, which represent one of major applications of FP calculations. However, the formulas used for bulk electrode reactions are included within those used for interfacial electrochemical reactions, which will be discussed below. Accordingly, they are not discussed in this article.
Electrochemical Reactions in Homogeneous Solutions
The most fundamental thermodynamic property of an electrochemical reaction is the redox potential. Once the redox potential is determined, it becomes possible to predict whether a reaction will proceed in the oxidative or reductive direction at a given electrode potential. Additionally, in a homogeneous solution, the free energy difference (reaction free energy) associated with the reaction at any specified potential can be calculated.
Sprik and co-workers developed a statistically rigorous method for determining the reaction free energy based on thermodynamic integration (TI) , within FPMD. ,,, To illustrate this approach, consider an n-electron transfer reaction, Ox+ne–→ Red. As discussed in the previous section, Red and Ox species in the electrolyte can be regarded as subsystems in contact with an electron reservoir characterized by the chemical potential μe. Nuclei in each state move according to the Hamiltonian defined by eq or (), with the PES corresponding to the grand potential of the respective state. To compute the free energy difference between the two states, Sprik et al. introduced the following grand potential, which linearly combines the grand potentials of Ox and Red using a coupling parameter λ
| 78 |
where N e (λ) represents the number of electrons in the subsystem at λ. The motion of the subsystem containing the redox species is assumed to be described by the Hamiltonian in eq or (), with the grand potential surface Ω acting as the PES. The free energy change associated with the electron transfer reaction is derived using TI as
| 79 |
| 80 |
where ⟨.⟩λ represents the ensemble average at the coupling constant λ. Here, we replace the Gibbs free energy with the Helmholtz free energy, supposing that the changes in volume during the electron-transfer half reaction are negligible. By definition, n = N e(1) – N e(0). It should be noted that the Helmholtz free energy discussed here differs from the Helmholtz free energy A for the electronic subsystem introduced in the previous section, which is ensemble-averaged only over electronic states. In contrast, also includes the ensemble average over nuclear motion. To distinguish between the two, the notation is introduced. At equilibrium, since , the electronic chemical potential μe that equilibrates the electron transfer reaction is given by . Defining the chemical potential of an electron in vacuum as zero, the redox potential U redox relative to vacuum is obtained as
| 81 |
This is the Nernst equation.
Sprik et al. ,,, developed a method to predict U redox using FPMD calculations, where the energy surfaces A Ox(R N n ) and A Red(R N n ) were obtained from canonical ensemble DFT, as described previously. In the FPMD calculations, a system was constructed by placing redox species in liquid water under a periodic boundary condition (PBC). A uniform background charge was introduced to neutralize the solute charge, representing a dilute-limit.
Here, we note the choice of a reference for electronic energy levels in FP calculations using PBC. In general, FP calculations for materials under PBC cannot set the electrostatic potential reference to the vacuum level. In FP calculation programs, the reference is set such that the average value of the local component of the effective nuclear potential in the computational system is zero. Ox is in a state where it has lost one electron compared to Red. The removed electron is assumed to reside in an electron reservoir, whose chemical potential is defined by this artificial reference. As a result, the energy difference obtained from FP calculations, A Red(R N n ) – A Ox(R N n ), is referenced to the energy level of this electron reservoir. This reference is specific to the computational system and depends on the local potential constructed in each program, making direct comparison with experimental values impossible. To obtain a value comparable to the experimental result, alignment is required. Figure (a) illustrates the conceptual framework of a recently proposed alignment method. In this approach, in addition to the calculations of the homogeneous solution, FP calculations are performed for a water slab model in a vacuum, as shown in Figure (a). These calculations enable the determination of the 1s energy levels of oxygen atoms in water molecules near the center of the slab, using the vacuum level within the computational system as the reference. Similarly, from the FP calculations of the homogeneous solution at a coupling constant λ, as illustrated in the same figure, the 1s energy levels of oxygen atoms in water molecules located far from the redox species are also obtained. As mentioned earlier, the 1s energy levels in the homogeneous solution are shifted from those referenced to the vacuum level by an electrostatic potential offset, ⟨Δϕ⟩λ, which is specific to the system and the computational program. If this shift did not exist, the average 1s energy level in one system would match that of the other, as the water molecules in both systems would be in a similar chemical environment. Therefore, by aligning the energy levels so that the average 1s energy level of one system coincides with that of the other, as shown in Figure (a), the shift, ⟨Δϕ⟩λ, can be determined as
| 82 |
where μvac represents the vacuum level. Using ⟨Δϕ⟩λ, the free energy difference referenced to vacuum, , is obtained as
| 83 |
6.
ML-aided FP calculations of redox potentials in homogeneous aqueous systems: (a) model systems and alignment method, (b) three-step procedure for computing redox potentials using thermodynamic integration and thermodynamic perturbation theory (reproduced from ref under CC-BY 4.0), (c) particle insertion and element substitution methods (adapted with permission from ref Copyright 2025 American Institute of Physics), and (d) computed redox potentials of seven redox species in aqueous systems using five exchange-correlation functionals (reproduced from ref under CC-BY 3.0).
There is another important point to note regarding the FP calculation of the redox potential. Due to its relatively low computational cost, the exchange-correlation functional based on the generalized gradient approximation (GGA) ,, is commonly used in FP calculations. However, for some redox species in water, GGA can introduce significant errors exceeding 0.5 V. ,,, This issue arises because GGA underestimates the band gap of water. For instance, GGA predicts the valence band maximum of water to be several eV too shallow. As a result, the redox level in the aqueous solution hybridizes with this incorrect band edge and becomes pinned to an erroneous energy level. To avoid this issue, a hybrid functional that provides a more accurate water band gap is necessary.
Sprik et al. calculated redox potentials using FPMD with Gaussian localized basis sets, following the principles of the aforementioned method, though the details of their alignment procedure differ from those described above. ,,, However, the FPMD-based approach requires substantial computational resources. In particular, when using a complete basis set such as plane-wave basis functions, the computational cost becomes extremely high. For this reason, many studies on redox potentials in homogeneous solutions do not employ FPMD. Instead, alternative semiempirical approaches, such as implicit solvation models, are predominantly used. − However, continuum solvation models involve empirical parameters and may not accurately capture solute–solvent interactions. To address this issue, hybrid models have been utilized, where only the first solvation shell around the solute is explicitly represented as atoms, while the remaining solvent is treated as a continuum. ,− However, the results obtained from such models depend on the choice of the explicit atom region and the continuum solvation region.
Recently, a method for accelerating a series of calculations using MLFF has been proposed. , A conceptual diagram is shown in Figure (b). The calculation consists of three stages. In the first stage, the energy surfaces A Ox(R N n ) and A Red(R N n ) in eq are represented using MLFFs, and TI is performed. Additionally, the MD simulations required for the alignment calculations, which handle the water slab model over several nanoseconds, are accelerated using an MLFF. By replacing these FP calculations with MLFF-based calculations, the overall computational cost can be significantly reduced, even when including the cost of MLFF training. However, while MLFFs can represent the energy surfaces obtained from FP calculations with higher accuracy than classical force fields, they may still introduce non-negligible errors in redox potential calculations. The second stage is designed to correct for this issue. This correction is performed through TI in a manner analogous to eqs to (), transitioning from the MLFF energy surface A ML to the FP energy surface A FP as
| 84 |
| 85 |
This correction requires FPMD; however, since the majority of the free energy difference has already been computed through TI using MLFF, the integrand in eq becomes relatively small, allowing the TI to converge rapidly within a few picoseconds of simulation. Through this approach, the accurate redox potential from FP calculations can be obtained. However, applying this method to the hybrid functional requires substantial computational resources to generate the hundreds of training data points necessary for the generation of MLFFs. To address this issue, the first two stages of the calculation are performed using GGA, and in the third stage, the free energy difference between GGA and the hybrid functional is used to correct the GGA results. At this stage, the Δ-ML method is employed to represent the difference between the energy surfaces obtained from two different exchange-correlation functionals in FP calculations. A key feature of the Δ-ML method is that it requires an order of magnitude fewer training data points than conventional MLFF while achieving an order of magnitude smaller errors. This is because the difference between the energy surfaces of the two exchange-correlation functionals varies much more smoothly with atomic positions than the individual energy surfaces themselves. As a result, an ML model with significantly higher accuracy can be generated with much less training data compared to standard MLFF. The free energy difference between GGA and the hybrid functional can be obtained, for example, using thermodynamic perturbation thoery (TPT) calculations based on the second-order cumulant expansion as
| 86 |
where ΔA ΔML represents the difference between the energy surfaces of the hybrid functional and GGA, as predicted by the Δ-ML method, and HYB denotes the hybrid functional. This approach reduces the number of hybrid functional calculations required for generating the ML model and allows the redox potential of the hybrid functional to be obtained without additional corrections.
In the above method, MLFF and Δ-ML can be regarded as ML models that enable the necessary “electron insertion” for computing redox potentials in electron transfer reactions. A similar concept can also be applied to proton-coupled electron transfer (PCET) reactions, represented as Ox + mH+ + ne– → Red. In principle, the same approach can be followed: TI is performed using MLFFs to represent the energy surfaces of the reactant and product, the difference between the MLFF energy surface and that obtained by the FP method is corrected via TI, and the free energy difference between GGA and the hybrid functional is obtained using the Δ-ML method. However, in the case of PCET, in addition to “electron insertion”, it is also necessary to insert a proton. This so-called “particle insertion” can, in principle, be carried out through TI from a noninteracting state, where the inserted particle does not interact with surrounding particles, to an interacting state, following the same formalism as eqs to (). Such particle insertion methods have been widely used in classical MD simulations. However, caution is needed when implementing this approach with MLFF. In the noninteracting limit, the inserted particle and surrounding particles may overlap. MLFF is often not trained on such unrealistic configurations, and learning energy surfaces where particles experience extremely strong repulsive forces can degrade its accuracy. As a result, performing particle insertion TI solely with MLFF may lead to significant errors. Recently, a TI method using MLFF was developed to overcome this issue. , In this method, a virtual intermediate state is introduced between the noninteracting and fully interacting states, with model interactions defined to prevent unphysical particle overlap. By performing TI in two stages–first from the noninteracting state to the model interaction, and then from the model interaction to the full MLFF-based interaction–a smooth TI pathway is achieved. In addition to this particle insertion method, a TI for “element substitution”, in which an inserted particle in solution is replaced with another element, has also been proposed. Element substitution is performed by TI from the MLFF of the initial element to that of the substituted element. This technique has also been widely used in classical MD simulations. − As shown in Figure (c), inserting an element X using the particle insertion method and subsequently replacing it with another element Y via the element substitution method achieves the insertion of the element Y. A previous study has used both the particle insertion method and a combination of particle insertion and element substitution to evaluate the solvation free energies of proton, alkali cations, and halide anions. The results obtained from both methods were confirmed to be statistically consistent within the margin of error. This consistency indicates that both approaches provide reliable TI pathways.
By combining ML models for the “electron insertion”, “particle insertion” and “element substitution”, it becomes possible to compute the redox potentials of a wide range of electrochemical reactions in homogeneous solution. , Figure (d) presents the computed redox potentials referenced to the vacuum level for seven electrochemical reactions in water, using five functionals: RPBE + D3, − PBE0, PBE0 + D3, − HSE06, , and B3LYP. In these calculations, the TI and TPT MD calculations of eqs , (), and () were performed on unit cells containing 64 water molecules, after confirming system size dependence using unit cells with 32, 64, and 96 molecules. The shift ⟨Δϕ⟩λ in eq was calculated using a water slab with 128 water molecules per unit cell, following a size-dependence check with slab systems containing 96, 128, 192, and 1024 water molecules. For each system, a single MLFF was generated, and MD simulations using the MLFFs were performed for durations ranging from several hundred picoseconds to at most several tens of nanoseconds, to ensure sufficient statistical sampling (see details in refs , ). All calculations were performed using VASP 6.3.0 and 6.5.0. The MLFF for each redox species was generated in approximately 2 days using 16 cores of an Intel Xeon Platinum 8358 processor (2.6 GHz), with each MD step taking just 0.1 s. A highly accurate Δ-ML model was developed for each functional and redox species using only 40 structures. Overall, the ML-assisted approach reduced the computational cost for evaluating redox potentials from 20 million core hours to just 16,800 core hours. The results indicate that PBE0 + D3 accurately reproduces experimental results. This approach is also applicable to methods beyond DFT with even higher accuracy. It provides an FP methodology for obtaining highly accurate redox potentials in homogeneous solutions while determining the solvation structure around redox species at finite temperature via MD simulations without arbitrariness. An important question is whether corrections to the MLFF, which require FPMD calculations, are necessary. Previous studies suggest that these corrections are essential for accurate predictions and error verification when computing redox potentials in a system containing approximately 200 atoms per unit cell, if the MLFF error is on the order of 1 to 10 meV atom–1 or 10 to 100 meV Å–1. In contrast, with Δ-ML, which achieves an order of magnitude smaller error (0.1 to 1 meV atom–1, 1 to 10 meV Å–1), accurate results have been obtained without any corrections. , MLFFs that employ the previously described equivariant descriptors are approaching this level of accuracy for certain systems. ,,,, In the future, MLFF alone is expected to be sufficient for predicting redox potentials without the need for corrections.
Interfacial Electrochemical Reactions
Electrochemical reactions at electrolyte–electrode interfaces are essential phenomena in secondary batteries, fuel cells, and electrolytic cells. A key theoretical framework for describing interfacial electrochemical reactions is the grand canonical ensemble theory for electrons. As mentioned earlier, this theory allows the number of electrons in the subsystem to fluctuate while maintaining a constant electronic chemical potential. On the other hand, the subsystem, which is represented under PBC, must remain electrically neutral. Therefore, to explicitly account for changes in the number of electrons, some form of electric double-layer model is required to compensate for the charge variation. Several models have been proposed, including fully explicit models that represent the entire subsystem using nuclei and electrons, ,,, simplified charge distributions such as Gaussian charges or background charges, ,, continuum electrolyte models, ,,,,,,,, and molecular statistical mechanics models. In all these models, the atoms in the subsystem are assumed to move along the surface of the grand potential Ω.
There are several ways to represent Ω. ,,, To improve the clarity of approximations, the Helmholtz free energy A, which is related to Ω through (eq ), is expanded in a Taylor series around N e,0 with respect to the electron number N e ,,
| 87 |
where A 0 represents the Helmholtz free energy at N e = N e,0. According to Janak’s theorem, the first derivative of the Helmholtz free energy A with respect to the electron number N e is equal to the electronic chemical potential μe, 0 at N e = N e,0
| 88 |
The second derivative of A with respect to N e is defined as the inverse of C:
| 89 |
Using these equations along with eq , Ω can be expressed as
| 90 |
where Δμ e = μe – μe,0. Assuming C to be a constant independent of N e and neglecting terms of O(ΔN e ), the value of ΔN e that minimizes Ω can be determined from the condition ∂Ω/∂N e = 0 as
| 91 |
This equation implies that C corresponds to the capacitance. From eq , Ω can be approximated as
| 92 |
This approach is referred to as the capacitor model. Furthermore, if the terms of order O(ΔN e ) are also neglected, Ω can be approximated as
| 93 |
Here, this is referred to as the first-order approximation.
Equations equivalent to the first-order approximation have been widely used, for example, in calculations of redox potentials for bulk electrodes such as cathode materials in Li-ion batteries. , In bulk electrodes, which are conductors, electrons are delocalized, and it is natural to assume that the interior of the electrode remains electrically neutral. Consequently, setting ΔN e = 0 leads to an equation equivalent to the first-order approximation, which is valid in such cases. However, near the electrolyte, the electrode can become charged. To compensate for this charge, ions in the electrolyte redistribute, forming an electric double layer. This polarized layer generates a potential gap at the interface, which, in turn, adjusts the electrode potential. Before the year 2000, it was not obvious how to incorporate this interfacial electrochemical phenomenon within the framework of FP calculations. However, as electrochemical systems gained increasing significance, the demand for FP calculations of interfacial electrochemical properties increased. In response to this, Nørskov and co-workers proposed this approximation for modeling electrode–electrolyte interfaces. As shown in eq , this approximation allows the determination of Ω at any electrode potential μe based solely on an FP calculation at the reference state with electron number N e,0. As discussed later, this method has successfully explained a wide range of interfacial electrochemical properties and is now the most widely used approach for FP calculations of electrochemical reactions. However, this approximation does not account for how the electron number N e changes when the potential deviates from the reference μe,0. Consequently, it cannot predict properties related to the number of transferred electrons. To address this limitation, methods utilizing electric double-layer models have been proposed. In these methods, FP calculations are performed at multiple potentials (or electron numbers) to determine the potential dependence of Ω and the charge transfer properties. − Applications of these methods indicated that the capacitance C can be well approximated as a constant. Based on this, the capacitor model was developed. In the capacitor model, FP calculation results for a small number of conditions with different electron numbers are extrapolated to predict the potential dependence of Ω. ,,, For further developments beyond the first-order approximation, refer to existing review articles. ,,,
Here, we introduce a method for calculating the redox potential of interfacial reactions using the first-order approximation. As an example of a reaction, we consider the reduction of adsorbed OH on a solid surface in an aqueous solution under standard conditions: OH* + H+ + e– → H2O + *, where * represents a surface site, and OH* denotes the adsorbed OH species. This reaction is not only the rate-limiting step of the oxygen reduction reaction (ORR) on Pt, a representative electrode catalyst in fuel cells, , but its reverse reaction also corresponds to the first elementary step of the oxygen evolution reaction (OER) in water electrolysis. , Furthermore, this reaction serves as a typical example of PCET, and the method described here can be applied to many other interfacial electrochemical reactions. The free energy change associated with this reaction is expressed as
| 94 |
where represents the free energy, obtained by taking the ensemble average over the nuclei and electrons of the chemical species inside the brackets. Here, we consider the hydrogen redox reaction: H+ + e–↔1/2 H2. The electronic chemical potential μe at which this reaction is in equilibrium satisfies the following condition
| 95 |
Using eq , eq can be rewritten as
| 96 |
where U represents the electrode potential relative to the standard hydrogen electrode (SHE), which is defined as
| 97 |
This approach, in which electrode potentials and free energy changes are computed by referencing the computationally determined SHE, is known as the Computational Standard Hydrogen Electrode (CSHE).
In the first-order approximation, nuclei in the subsystem move along the surface of Ω given in eq . Therefore, the free energy change in eq is obtained as
| 98 |
where is the free energy based on the Hamiltonian with Ω as the energy surface, as given in eq , and it does not depend on U. Therefore, once is determined, the reaction free energy at any potential can be predicted as a linear function of U. Using this linear equation, the redox potential of the interfacial electrochemical reaction can be calculated from the condition as
| 99 |
| 100 |
Nørskov and co-workers developed a method for predicting the activity of electrocatalysts by combining this approach with statistical mechanical models, such as the ideal gas model and the harmonic oscillator model, within an analytical framework. For example, in the OH* reduction reaction described above, the free energies of nonadsorbed molecules such as H2O and H2 were calculated using the ideal gas model, based on molecular structures optimized at absolute zero and vibrational frequencies through FP calculations. Similarly, the free energies of the interfacial species * and OH* before and after the reaction were computed via the harmonic oscillator model. In this approach, the interfacial electrolyte was represented by an ice-like model, as shown in Figure (a). This methodology has been applied to a wide range of electrochemical reactions, including the hydrogen evolution reaction (HER), ORR, OER, CO2 reduction reaction (CO2RR), nitrogen reduction reaction (N2RR), and lithium deposition reactions. These studies have shown that a broad range of electrocatalytic reactions follow the Sabatier principle, with their reaction rates represented by volcano plots, as shown in Figure (b–f), where catalytic activities are correlated with the stabilities of reaction intermediates. Owing to its ability to efficiently evaluate both the thermodynamics and kinetics of electrochemical reactions, this method has also been widely used for exploring advanced catalytic materials. − ,,
7.
An example of the ice-like model on a close-packed (111) surface (a) and volcano plots predicted by FP calculations for the hydrogen evolution reaction (HER) (adapted with permission from ref Copyright 2005 Electrochemical Society) (b), oxygen reduction reaction (ORR) (adapted with permission from ref Copyright 2004 American Chemical Society) (c), oxygen evolution reaction (OER) (adapted with permission from ref Copyright 2011 WILEY) (d), CO2 reduction reaction (CO2RR) (reproduced from ref under the CC-BY 4.0) (e), and N2 reduction reaction (N2RR) (adapted with permission from ref Copyright 2014 Royal Society of Chemistry) (f). Experimental data in (f) were taken from refs − .
However, the structure and dynamics of interfacial atoms at finite temperatures cannot simply be described as harmonic vibrations around the structure at absolute zero. This issue becomes particularly evident when dealing with complex interfacial structures that are inconsistent with the structure of ice. A typical example is the Pt surface functionalized with organic molecules, as shown in Figure (a). Recently, it has been confirmed that functionalizing Pt catalysts with organic molecules such as melamine enhances the ORR rate by several times on both nanoparticle catalysts and single-crystal surfaces. Additionally, on nanoparticle catalysts, melamine functionalization has been found to suppress Pt dissolution, thereby improving catalyst durability. This approach has attracted attention as one of the few strategies capable of enhancing both the activity and durability of ORR catalysts. However, determining the structure of interfacial water in this complex system is not straightforward. First, determining the adsorption structure of organic molecules in a solvated environment through structural optimization is challenging. Furthermore, in interfacial structures where symmetry is reduced by organic molecules, numerous quasi-stable configurations with comparable energies can exist, making it challenging to determine the solvent structure that dominates the free energy landscape. To address such challenges, employing the TI calculation using finite-temperature MD, as described in the previous section, is necessary. However, applying FPMD requires an expensive computational cost. Although a few previous studies have performed TI using FPMD, , the range of reaction coordinates that can be explored to map the free energy landscape is limited to the order of a few Å. With such a narrow exploration range, the evaluated free energy differences may be referenced to a system-specific value, making direct comparison with experimental results difficult. Moreover, the time-scale of reorientations of interfacial water molecules has been reported to be on the order of 1 to 10 ns, , making it difficult to obtain sufficient statistical averaging within the typical 10 ps time scale of FPMD simulations.
8.
An ML-aided FP calculation of interfacial electrochemistry: (a) interfacial model of melamine and OH adsorbed on the Pt(111) surface in water, (b) experimental and simulated volcano plot of ORR at an electrode potential of 0.9 V vs SHE on Pt(111) (black), Pt(111) with melamine (red), Pt1–x Cu x (111) (orange), other Pt alloys (purple), and transition metals (gray), where open dots and the solid black line represent calculations, and closed dots represent experimental data, ,− and (c) vibrational density of states of deuterated hydrogen atoms within a water molecule hydrogen-bonded to OH* on Pt(111) (black) and Pt(111) with melamine (red), compared to that in bulk liquid water (gray). Here, the relative change in ORR current density j compared to Pt(111) is plotted as a function of the redox potential of the OH* reduction reaction relative to Pt(111), U OH – U OH (adapted with permission from ref Copyright 2025 American Chemical Society).
Recently, the particle insertion method utilizing the MLFF described in the previous section ,, has been applied to this problem, leading to the elucidation of the mechanism behind the ORR enhancement through the melamine functionalization. In this approach, the free energy difference in eq is expressed as the sum of the reaction free energy for the bond formation between the adsorbed OH* and H atom at the solid–liquid interface, OH* + H → H2O, and the reaction free energy for the dissociation reaction of H2 molecule, 1/2H2 → H
| 101 |
can be readily computed by applying the ideal gas model to an H2 molecule and an H atom in the gas phase. On the other hand, calculating the free energy of the bond formation between the OH* species and H atom at the solid–liquid interface is challenging. To address this, the particle insertion method utilizing MLFF was applied. As in the method described in the previous section, MLFFs were generated for both the initial and final states of the reaction. Using these MLFFs, TI calculation was performed over a total time scale of tens of nanoseconds, following the procedure of eqs to (), to compute the free energy change from an ideal gas-phase H atom, which does not interact with its surroundings, to a state where it is captured by the interfacial OH*. Furthermore, as described in eqs and (), TI was performed from the MLFF energy surface to the FP energy surface, thereby correcting for MLFF errors. From the obtained free energy difference , the redox potential U OH for the OH* reduction reaction was determined. Using this U OH value, the ORR reaction current was predicted based on the ORR reaction model. For comparison, evaluations were conducted not only on Pt(111) surfaces with and without adsorbed melamine but also on Pt monolayers on PtCu alloys [Pt/Pt1–x Cu x /Pt(111)]. The liquid–solid interfaces were modeled using unit cells containing 48 Pt atoms and between 38 and 48 water molecules. A single MLFF was generated for each surface system through on-the-fly training, which was performed over approximately 3 days using 64 cores of the same processor described in the previous section for the Nafion system. The resulting MLFF achieved a simulation speed of 0.1 s per MD step using 8 cores of the same processor. For each system, more than 15 ns of MD simulations were conducted for the TI calculations. Figure (b) presents the computed ORR activity, expressed as the logarithm of the ratio of the reaction current j to the reaction current on Pt(111), j Pt, plotted against the thermodynamic stability of adsorbed OH*, represented by the difference between U OH and the OH* reduction redox potential U OH on Pt(111) without melamine. Previous studies on PtCu alloys , have shown that the introduction of Cu shortens the Pt–Pt bond distance, weakening the interaction between Pt and OH*, which leads to an increase in U OH. This, in turn, accelerates the OH* reduction reaction, enhancing the ORR activity. The TI method incorporating MLFF successfully reproduces these results. Similar to alloying, the melamine adsorption increases U OH and enhances the ORR activity. The impact of the surface functionalization closely matches that observed for the most active alloy, Pt/Pt2Cu/Pt, and results in a substantial enhancement–approximately a 9-fold increase in per-site activity compared to Pt(111). The computational result that melamine reduces OH* production and improves ORR activity is consistent with previous experimental findings on Pt(111). However, further analyses of interfacial electronic states, interfacial hydrogen bonding structures, and vibrational frequencies revealed that melamine destabilizes OH* through a mechanism fundamentally different from alloying. Figure (c) illustrates the effect of surface functionalization on the vibrational density of states for the O–H stretching mode in water molecules hydrogen-bonded to OH*. On the Pt(111) surface without melamine, water molecules form strong hydrogen bonds with OH*, causing a significant redshift in the O–H stretching frequency compared to that in liquid water. In contrast, the O–H stretching frequency is blueshifted on the Pt(111) surface in the presence of melamine. This indicates that melamine disrupts the hydrogen bonds between adsorbed OH* and surrounding water molecules. This study concludes that such interfacial hydroben bond disruption leads to the destabilization of OH*. Capturing this disruption using FP calculations, which impose an ice-like solvent structure at absolute zero, is challenging. By incorporating MLFF, finite-temperature MD simulations exceeding the reorientation time scale of interfacial water (several nanoseconds) , become feasible, enabling free energy calculations that account for fluctuations of interfacial water molecules.
Not only for the first-order approximation but also for capacitor models and even higher-order approximations, MLFF is expected to be applicable. By constructing an MLFF that represents interatomic interactions in the subsystem at each electrochemical potential (μe) or each electron number (N e) under the electric double layer, and performing TI calculations for electron or particle insertion as in the examples discussed above, it should be possible to evaluate reaction free energies while explicitly considering variations in the electrode potential. Moreover, in addition to improving the accuracy of MLFF, incorporating electrostatic interactions and electric field response theories ,,,,, could enhance the transferability of MLFF. This would allow efficient predictions for various potential states based on training with a limited number of potential conditions. Furthermore, leveraging the high computational efficiency of MLFF, explicitly incorporating a large ion reservoir in direct contact with the subsystem would enable highly accurate grand canonical ensemble simulations, not only for electrons but also for ions, using a fully atomistic model. Thus, in the future, attempts to explicitly incorporate electric double layers using MLFF will become a major focus of research.
Knowledge Extractions and Simulations on Inhomogeneous Systems
MLFF enables simulations on larger length- and time-scales than FP calculations by representing the PES, allowing for precise predictions of transport coefficients and reaction properties. However, this alone is insufficient to address complex phenomena involving multiple elementary processes within heterogeneous structures in electrochemical systems. To accomplish this, it is necessary to extract key information from MD simulations and transfer it to simulations at even larger length- and time-scales. As shown in Figure , MLFF has been utilized for this purpose as well. In this section, we present coarse-grained simulations of transport phenomena in polymer electrolytes and detailed analyses of complex surface reactions, explaining the methods used for information extraction and their applications.
Multiscale Molecular Simulations
In the electrolytes and electrodes of fuel cells, electrolytic cells, and secondary batteries, heterogeneous structures larger than a few nanometers are formed, and transport phenomena at submicrometer to larger scales can become the dominant factor influencing performance. − , Experimentally elucidating the locations of chemical species and their transport pathways within such heterogeneous media is extremely challenging. MD simulations can provide rigorous predictions of transport properties based on statistical mechanics. However, even classical MD simulations are typically limited to time-scales of 100 ns to 1 μs and length-scales of 10 to 100 nm, making it difficult to fully address the above issues. An alternative approach is to treat materials as continuous medium and describe transport phenomena by solving diffusion equations or fluid dynamics equations. While this approach can cover the relevant length- and time-scales, the solubility and effective diffusion coefficients of chemical species used in such calculations are often adjusted to reproduce experimental results, making it difficult to consider this a truly predictive simulation method.
Recently, with the advancement of MLFF, systematic methods have emerged that enable bridging the gap between the atomic scale and the mesoscale. One representative technique is coarse-graining, a method in which groups of atoms are combined into a single pseudoparticle, and interactions between these coarse-grained particles are represented as functions of their positions. This approach reduces the number of particles to be handled, increases the MD time step, and expands the accessible length- and time-scales. In general, coarse-graining enables simulations on time scales ranging from 1 μs to 1 ms and length scales from 100 nm to 1 μm. Similar to classical MD, coarse-graining has been extensively studied, and many physics-based interaction models have been proposed. − There are various coarse-graining techniques, and numerous excellent review articles have been published. For further details, we refer the reader to these refs ,− . Coarse-grained models have been developed to represent the free energy landscape obtained from MD simulations
| 102 |
where R N CG is a variable that collects the position vectors of N CG coarse-grained particles. M N CG is a mapping function that transforms the atomic coordinates R N n of N n atoms into the coordinates of the coarse-grained particles, R N CG . In many applications, the linear mapping shown below is used
| 103 |
The relationship between the inputs and outputs of is analogous to that of MLFF, allowing MLFF techniques to be utilized without significant modification. − Figure (a) presents simulation results for a variant of Chignolin, a 10-amino-acid protein, using SchNet, a type of MPNN. In this case, the mean forces obtained from the classical force field CHARMM22 were used as training data. Mean forces can also be obtained from all-atom MD simulations using MLFF, allowing MLFF to provide a seamless framework for hierarchical connections–linking FP calculations to all-atom models and all-atom models to coarse-grained models.
9.
Applications of ML models to multiscale simulations: (a) the structure of the miniprotein chignolin obtained using the coarse-grained SchNet model with MPNN (adapted with permission from ref Copyright 2020 American Institute of Physics), (b) the structure of Nafion at λ = 10, and (c) the global diffusion coefficient D global in Nafion, simulated using the kMC method with local A(r) and D(r) predicted by the SOAP kernel method. Images (b, c) are adapted with permission from ref Copyright 2025 American Chemical Society.
Another notable application of MLFF-related technology is the development of methods that extract not only the free energy landscape but also local transport coefficients from all-atom MD simulations and construct ML models to represent these properties. In this application, the free energy landscape was calculated using the classical force field DREIDING based on the following equation
| 104 |
where ⟨.⟩ M(Rsolt)=r denotes an ensemble average taken with the center of mass of the solute, M solt, fixed at r. The local diffusion coefficient of the solute, D(r), was calculated using the flat-bottom potential method, which estimates the diffusion coefficient from the local mean square displacement of the solute. The method was applied to the diffusion of H2 molecules in the Nafion polymer electrolyte. As shown in Figure (b), the electrolyte was modeled using 200 Nafion molecules, each composed of 10 monomers per polymer chain. All-atom MD simulations were conducted for hydrated Nafion with water uptakes of λ = 6, 10, and 14. A uniform grid r with a resolution of approximately 0.4 to 0.8 nm was set within the unit cell, and A(r) and D(r) of H2 molecules were computed at each grid point. The computed values of A(r) and D(r) were used as training data to generate ML models based on the kernel regression explained previously. In this study, both A(r) and D(r) were assumed to be represented as functions of the elemental distribution function around r. For representing the elemental distribution functions, the three-body invariant descriptor in eq was used, and the SOAP kernel was employed for regression. Using the predicted values from the ML models, kMC simulations were performed based on the following transition probabilities
| 105 |
where i represents the grid position, Δi denotes the displacement of movement, A(i) and D(i) correspond to the free energy and local diffusion coefficient at grid point i, respectively, and D max represents the maximum diffusion coefficient. This kMC simulation is equivalent to solving the following diffusion equation in a heterogeneous field
| 106 |
Figure (c) presents the global diffusion coefficient of H2 molecules obtained from the kMC simulation, plotted against the number of training data points, 3N, where N represents the number of data points for each λ. The ML model successfully reproduces the kMC results obtained using A(i) and D(i) from all-atom MD simulations (shaded region) with approximately 15,000 data points. As a result, the overall computational speed improved by an order of magnitude. The acceleration factor will increase as the system size grows. In this study, not only the global diffusion coefficient but also the spatial distributions of A(i) and D(i) within Nafion were presented. These distributions identified regions with high H2 solubility and diffusivity. The results suggested that H2 molecules primarily diffuse through the boundary regions between the Nafion main chain region and the cluster region. The kMC simulation is highly efficient and is expected to enable the simulation of transport phenomena on length scales beyond 1 μm and time scales of up to 1 s. This advancement indicates that, in the near future, entire electrolyte membranes and electrodes with thicknesses of several micrometers can be simulated.
Simulations of Complex Reactions
Many electrochemical reactions are complex multistep processes composed of a vast number of elementary steps. Even the simple hydrogen evolution reaction (HER), 2H+ + 2e– → H2, consists of three elementary steps: the Volmer reaction (H+ + e– → H*), the Tafel reaction (2H *→ H2), and the Heyrovsky reaction (H* + H+ + e– → H2). Even on highly symmetric single-crystal surfaces, identifying the reaction pathways is not straightforward, as the reaction intermediate H* can adopt various configurations due to interactions with the surrounding solvent molecules. Elucidating these pathways requires extensive and long-time FPMD simulations with significant computational cost. , In more complex reactions such as CO2 reduction and N2 reduction, which can generate numerous reaction intermediates and products, the number of possible elementary steps exceeds tens of thousands, even on single-crystal surfaces, making the selection of reaction pathways extremely challenging. ,, Furthermore, on electrode surfaces with defects or nanoparticle catalysts, surface symmetry is reduced, leading to a diverse range of surface sites. −
Traditionally, dominant reaction pathways have been identified based on experimental observations such as the dependence of the overall reaction rate on reactant concentrations and electrode potential, the types of reaction intermediates and products, as well as the chemical intuition of researchers. ,,, FP calculations have then been performed on a limited number of elementary steps to construct reaction models. ,, A typical procedure for FP calculations involves the following steps: constructing initial structures of reactants and products based on experimental information, identifying local minima and saddle points on the energy surface E(R N n ) through structural optimizations, ,, determining the structures, energies, and vibrational frequencies of reactants, products, and transition states, and finally, calculating reaction rate constants k based on TST. Alternatively, in some cases, a reaction coordinate is predefined, and the activation free energy is determined using a TI method based on FPMD. However, in complex reactions involving tens of thousands of elementary steps, this approach may overlook important reaction pathways and lead to inaccurate predictions. Furthermore, the methods that utilize experimental data for model construction cannot predict the properties of unknown reactants.
To facilitate more systematic predictions, a method leveraging correlations identified in FP calculation data has been developed. An early approach in this direction is the descriptor-based method. Through detailed analyses of surface reactions, FP calculation data on surface reactions have been accumulated, , leading to the discovery of key linear relationships, such as scaling relations, which describe linear correlations between formation energies of reaction intermediates and the Brønsted-Evans–Polanyi (BEP) relationship, which expresses a linear correlation between the reaction energy and the activation energy, as illustrated in Figure (a). By leveraging these relationships, the high-dimensional variables in complex reactions can be effectively reduced to a smaller set of key parameters. Using these relations, reaction rate constants can be determined, allowing the prediction of complex reaction kinetics through chemical reaction models. This approach has made it possible to systematically summarize the activity of various electrocatalytic reactions in the volcano-type correlations shown in Figure . ,,,,,−
10.
ML-aided analyses of complex catalytic reactions: (a) the Brønsted-Evans–Polanyi (BEP) relationship between activation energy (E a) and reaction energy (ΔE) observed in an FP data set of dimer dissociations from ref ; (b) identification of a key subset from the full reaction network of syngas (CO and H2) conversion to CO2 (reproduced from ref under CC-BY 4.0); (c) energy diagrams of the direct decomposition of NO on a cuboctahedral Rh138Au63 alloy nanoparticle and adsorption energies of reaction intermediates (N, O, and NO) on a cuboctahedral Rh578Au133 alloy nanoparticle (adapted with permission from ref Copyright 2017 American Chemical Society); and (d) a contour plot illustrating 14,958 reaction pairs (IS, TS, and FS) and low-energy pathways for methanol synthesis from a CO, CO2, and H2 mixture on Cu(211) (adapted with permission from ref Copyright 2022 American Chemical Society).
The outstanding success of this descriptor-based approach, together with advancements in AI and ML, has driven the development of more systematic methods for exploring reaction pathways. Figure (b) presents an application of this approach by Ulissi and co-workers to the gas-phase synthesis of C1 and C2 compounds from H2 and CO. In this method, the key reaction pathways are extracted through the following process. First, the main reaction intermediates and elementary steps are identified, and all possible reaction pathways connecting these elementary steps are generated. In this study, 99 reaction intermediates and 190 elementary steps were used to generate thousands of possible reaction pathways, as shown in the figure. Next, an ML model is developed to predict the formation energies of reaction intermediates using an FP calculation database for surface reactions. In this study, Gaussian process regression based on the kernel method was employed for training. To construct input features, reaction intermediates were represented using SMILES, and their approximately 50-dimensional descriptors were reduced to around 10 dimensions through principal component analysis. From the predicted formation energies of reaction intermediates, the reaction energies of elementary steps were computed. Using these reaction energies, the activation energies of elementary steps were estimated based on the BEP relationship. From these activation energies, reaction rates for elementary steps were evaluated, allowing dominant reaction pathways to be identified. FP calculations were performed on these dominant reaction pathways, and the resulting data were added to the database. The ML model was subsequently updated using this new database, and the updated model was used again to extract dominant reaction pathways. By repeating this process, the accuracy of the predicted dominant reaction pathways improved, ultimately leading to the identification of the key reaction pathways shown in the figure.
With the advancement of MLFF, descriptors for representing three-dimensional atomic configurations have also evolved, enabling the prediction of the energies of reactants and products on three-dimensional surface sites of alloy nanoparticles. Figure (c) illustrates an application of this approach to the direct NO decomposition reaction, 2NO → N2 + O2, on Rh1–x Au x alloy nanoparticles. , In this study, the power spectrum, one of the three-body descriptors, as defined in eq , was employed to predict the formation energies of three reaction intermediates: N*, O*, and NO*. The nanoparticles were modeled as cuboctahedral structures with diameters ranging from 2 to 5 nm, consisting of 201 to 3355 atoms, as shown in the figure. The atomic positions in the nanoparticles were determined based on the lattice constants of an optimized bulk fcc crystal, following the so-called lattice gas model. From these atomic configurations, the power spectrum was computed and used as input for the ML model. For regression, Bayesian linear regression with a SOAP kernel was employed. The training data consisted of FP calculations for single-crystal surfaces with local structures similar to terraces and edges of the nanoparticles. As with the nanoparticle model, descriptors were computed from atomic configurations determined by the lattice constants of an unrelaxed bulk fcc crystal. However, the formation energies used for training were those obtained after structural relaxation. As a result, the model enabled the prediction of the formation energies for relaxed reaction intermediates based on unrelaxed atomic configurations. Similar to the method of Ulissi et al., the reaction energies of elementary steps were calculated from the predicted formation energies of reaction intermediates. The activation energies of the elementary steps were predicted from the formation energies using the BEP relationship. In addition to the ML model for predicting formation energies, another ML model was developed to evaluate the stability of nanoparticles. A lattice gas model was again used, and the training data consisted of the total energies of bare single-crystal surfaces after structural relaxations. The same SOAP-kernel-based regression method was applied, allowing the ML model to predict the total energies of alloy nanoparticles after relaxation based on their unrelaxed atomic configurations. Using this ML model, MC simulations were performed to determine the stable elemental distributions within the alloy nanoparticles. The NO direct decomposition activity was then evaluated on the identified stable surfaces. An example of the results is shown in Figure (c). The ML model predicted the formation energies of the three reaction intermediates at various sites on the alloy nanoparticle. Based on these predictions, site-specific free energy diagrams were generated, and the overall activity of the nanoparticle was assessed as the sum of reaction rates across all sites. The energy diagrams in the figure demonstrate that the ML model accurately reproduces FP calculation results. On the Au(211) single-crystal surface, the reaction intermediates are too unstable, resulting in large energy barriers for their formation. Conversely, on the Rh(211) single-crystal surface, the intermediates are overly stabilized, leading to large energy barriers for their removal. In contrast, on the Rh138Au63 alloy nanoparticle, the stability of reaction intermediates at corner sites is optimized, balancing the formation and removal processes. This indicates that the ML model has learned the scaling relationships from FP calculation data on single-crystal surfaces and effectively expresses the Sabatier principle. This study concluded that for a nanoparticle with a diameter of 2 nm and x = 0.31, the number of highly active corner sites increases, leading to maximum catalytic activity.
As MLFF has become capable of representing complex PESs of multielement systems, it has become possible to automatically explore reaction pathways and evaluate energy diagrams using the PES obtained from MLFF, without assuming the lattice gas model. Figure (d) illustrates an application of this approach to methanol (CH3OH) synthesis from a mixed gas of CO, CO2, and H2 on a Cu/Zn alloy surface. The MLFF used in this study was the first-generation NNP developed by Behler and Parrinello. Training data were collected through an active learning approach using reaction pathway exploration simulations based on the stochastic surface walking (SSW) method, which enables unbiased PES searching based on the bias-potential-driven dynamics and Metropolis MC method. Initially, training data were gathered using FP calculations with the SSW method, and an MLFF was generated. Next, additional SSW simulations were performed using the MLFF, during which structures appearing in the simulations were randomly selected for FP calculations. These new data were added to the training data, and the MLFF was updated. By iterating this process, FP data for a total of 63,539 structures were collected. The final MLFF achieved root-mean-square errors (RMSEs) of 2.99 meV atom–1 and 0.08 eV Å–1 for the entire training data set and 2.1 meV atom–1 for the energy of the test data set. Using MLFF-based SSW simulations, the stable elemental distribution of the alloy single-crystal surface was determined. Subsequently, reaction pathway exploration was performed on this surface using the SSW method. Figure (d) presents the results for the Cu(211) single-crystal surface. A total of 1,220,000 local minima on the energy surface were identified, along with 14,958 elementary steps and the corresponding structures of reactants, products, and transition states. The figure shows a frequency distribution visualized by compressing these data into two-dimensional collective variables (CV1, CV2). The red-colored regions indicate structures with high occurrence frequencies. Within this distribution, the lowest-energy reaction pathway from CO2 to CH3OH, as evaluated by MLFF, is highlighted. This pathway was also confirmed by FP calculations to be a low-energy pathway. Similar calculations were performed for Cu–Zn alloy single-crystal surfaces. It was concluded that step sites where Zn atoms avoid direct adjacency serve as the active sites for CH3OH synthesis and that CO2 is the primary source of CH3OH. Additionally, it was revealed that during the reaction, –Zn–OH–Zn– chains appear at the edges, deactivating the edge sites. CO plays a crucial role in reducing these chains and restoring the active sites.
The three cases (b), (c), and (d) all focus on gas-phase catalytic reactions. Although, to the best of our knowledge, comprehensive exploration and analysis of reaction pathways such as those presented here have not yet been conducted for electrochemical reactions at liquid–solid interfaces, the applicability of MLFFs to interfacial electrochemistry has already been demonstrated. ,,,,,− In principle, the method proposed by Shi and co-workers is also applicable to liquid-phase reactions and holds promise as an effective approach for analyzing complex electrochemical reactions.
Materials Explorations
One of the ultimate goals of computational materials science is to explore unknown materials through computational methods and identify those with useful functionalities. Electrochemical systems can be considered one of the fields where FP calculations have been most effectively utilized. A pioneering example of this approach is the calculation of redox potentials for Li insertion reactions in cathode materials for Li-ion batteries using eq . Ceder and co-workers applied this method to propose Li x Al y Co1–y O2 and experimentally demonstrated that, in agreement with FP-based predictions, increasing the Al content y leads to an increase in the open-circuit voltage (OCV). Since then, FP calculations have become an established standard for predicting the OCV of cathode materials. The same equation has also played a crucial role in the search for electrochemical catalyst materials. As described in the previous section, the use of scaling relations and the BEP relationship has enabled dimensionality reduction of key factors, allowing the activities of a wide range of electrochemical catalytic reactions to be systematically represented as volcano plots based on a few key descriptors, as shown in Figure . ,− These descriptors were used as indicators to search for alloy catalysts, leading to the proposal of advanced materials. ,,
The exploration of materials using FP calculations has also been extended to solid electrolytes by incorporating simplified interatomic interaction models. For example, Kahle and co-workers proposed a Li-ion solid electrolyte through a multistep screening process, as illustrated in Figure (a). First, band gaps were evaluated through crystal structure optimization using FP calculations, narrowing down the initial set of 1400 candidate materials to 900 insulating materials. Next, for the filtered materials, the self-diffusion coefficient of Li ions was calculated using MD simulations with an interaction model known as the pinball model, reducing the candidates to 130 materials. In the pinball model, Li+ ions are treated as independent particles that move within the field generated by a fixed crystalline host. The interaction between the host and the ions is computed using the electronic density and wave functions of the host, obtained from FP calculations performed solely on the host. Since the energy surface governing the motion of Li ions as independent particles is determined only from the structural optimization of the host, the computational efficiency is significantly improved. Finally, further evaluation was conducted using FPMD simulations, leading to the selection of five candidate materials. Jang and co-workers also conducted a multistep screening process and proposed a novel Na-ion solid electrolyte. Figure (b) outlines this process. Initially, electrostatically stable crystal structures were selected from an initial pool of 523 million candidates based on energy estimates derived from the Ewald summation, narrowing the set to 850 structures. Next, FP-based structural optimization was performed to assess the stability, further reducing the candidates to 170 materials. These were subjected to FPMD simulations with a large time step of 10 fs, refining the selection to 44 materials. Finally, for the remaining candidates, FPMD simulations were conducted with a 1 fs time step at temperatures ranging from 500 to 900 K over 100 ps. Using the Nernst–Einstein equation, ionic conductivities were calculated, and electrolytes with conductivities on the order of 10 mS cm–1 were identified. A key common feature of both studies is the use of physics-based models, such as the pinball model and the Ewald summation, to efficiently screen large sets of candidate materials.
11.
Four examples of FP-based materials exploration: (a) high-throughput computational screening of solid-state Li-ion conductors (adapted with permission from ref Copyright 2020 Royal Society of Chemistry), (b) high-throughput exploration of Na-ion sulfide solid electrolytes (reproduced from ref under the Creative Commons Attribution-NonCommercial License), (c) large-scale exploration of stable crystal structures (reproduced from ref under CC-BY 4.0), and (d) large-scale screening of stable and ion-conductive solid electrolytes (adapted with permission from ref Copyright 2024 American Chemical Society).
In recent years, with the advancements of MLFFs, general-purpose MLFFs have been developed to represent interactions in arbitrary materials composed of any element from the periodic table. These models are now being used for large-scale materials screening. One of the pioneering studies is the stability exploration conducted by Cheng and Ong. In this study, a general-purpose MLFF called M3GNet, based on CNN as described previously, was employed. M3GNet was trained on the FP calculation database from the Materials Project. The exploration began by substituting elements in crystal structures from the ICSD database. Through this process, 31 million virtual crystal structures were generated. Their energies were evaluated using the MLFF, and the structures with the lowest energy on the convex hull were selected. At this stage, the number of candidates was reduced to 2000 structures, which were further evaluated using FP calculations. Finally, 1578 materials were identified as stable. By employing the MLFF that better reproduces FP calculations in the screening process, this approach enables a more accurate assessment of material stability compared to previous methods.
Merchant et al. conducted a large-scale stability exploration using a multistep process incorporating ML models. Figure (c) presents a diagram outlining the overall scheme. The exploration was carried out through two routes: the structural pipeline and the compositional pipeline. As the first step, crystal structures from the Materials Project and the OQMD database were used as the source. By applying a method called Symmetry Aware Partial Substitution, which efficiently generates new crystal structures by substituting elements while considering symmetry, a total of 109 candidate crystal structures were generated. In the structural pipeline, the screening was performed using the ML model GNoME. While the details of GNoME remain unclear, it is an MPNN that predicts energy from graph representations of crystal structures. The model was pretrained on FP data from 69,000 crystal structures in the Materials Project. Structures identified as stable by GNoME were clustered, and their stability was further evaluated using FP calculations. The newly obtained data were added to the database, and the GNoME model was updated. By iterating this process, the accuracy of GNoME improved to 11 meV atom–1, leading to the proposal of 2.2 million crystal structures that were more stable than those reported in previous studies. The compositional pipeline employed a crystal structure prediction (CSP) method. An ML model predicting stability based on elemental composition was used to identify stable compositions, from which 100 candidate crystal structures were generated. For these structures, a CSP known as Ab initio Random Structure Searching (AIRSS) was applied to determine the most stable crystal structures. Among the structures obtained through these processes, 381,000 structures were found to be the lowest-energy structures on the convex hull, and 736 of them were reported to have been experimentally synthesized. Furthermore, the vast amount of FP data collected in this study was used to train a general-purpose MLFF based on NequIP, a type of MPNN employing equivariant descriptors and network, as described previously. This MLFF, referred to as the GNoME potential, was utilized in MD simulations, leading to the reported discovery of 623 new superionic conductors.
More recently, Chen et al. explored solid electrolytes using MLFF. Figure (d) outlines the overall process. In this study, approximately 30 million candidate crystal structures were generated by substituting elements in crystal structures from the ICSD database. Structural optimization was performed using M3GNet, and approximately 600,000 crystal structures within a 50 meV atom–1 energy range from the convex hull were selected. These structures were sequentially evaluated using ML models for band gap and redox potential predictions, FP calculations for stability assessment, and M3GNet-based MD simulations for ion diffusion, ultimately narrowing down the candidates to 147 structures. Further screening incorporating factors such as cost and density reduced the number of final candidate materials to 18. Among these, Na x Li3–x YCl6 was selected and synthesized. They reported that doping Li into the parent phase Na3YCl6 led to a two-order-of-magnitude increase in ionic conductivity, from 10–8 to 10–6 S cm–1, demonstrating an interesting effect. The authors have developed and provided a cloud-based system that integrates the functionalities of ML models and FP calculations.
As evident from Figure , these two recent studies propose a systematic and large-scale materials exploration by utilizing ML models for high-speed screening of extensive candidate pools. Before breakthroughs in data-driven science, the initial stages of this screening process relied heavily on the intuition of experienced scientists–an art-like process that was often difficult for others to logically explain. Since the early days of materials informatics, AI and ML have been recognized as technologies capable of systematizing this process. Today, this transformation is gradually becoming a reality, with MLFF playing a crucial role as one of its driving forces.
Conclusions
We presented an overview of technological trends and future prospects for AI and ML applications in electrochemical systems. In particular, we focused on MLFFs which can represent interatomic interactions across diverse materials within a unified framework. MLFFs offer general applicability comparable to FP calculations while being several orders of magnitude faster. This advantage has driven the rapid growth of MLFF applications in ionics and electrochemical reactions. MLFFs enable finite-temperature MD simulations of ionic conductivity, as well as reaction thermodynamics and kinetics, with sufficient statistical sampling–capabilities that were previously out of reach for conventional classical force fields or FP calculations. In addition, systematic coarse-graining techniques have been developed to extract interparticle interactions and transport coefficients from MD data, allowing large-scale simulations over extended length and time scales while preserving statistical mechanical rigor. Furthermore, with advancements in algorithms, computing power, and FP databases, training on FP data for 105 to 106 crystal structures has become possible, leading to the development of general-purpose MLFFs, which are now being used for large-scale materials discovery. As discussed in Introduction section, the development and application of MLFF are built upon the foundations of classical mechanics, quantum mechanics, and statistical mechanics, which have been established through theoretical and computational sciences. This represents a prime example of how data-driven science is merging with traditional scientific disciplines to drive further advancements.
While MLFF is an extremely promising technology, many challenges remain for further applications. Finally, based on the overview of MLFF development and applications presented in this paper, we summarize below the key research challenges for MLFF. As the first topic, we outline the challenges related to the advancement of MLFF itself.
Accuracy and transferability: For both inactive and active learning algorithms, the current use of MLFF consists of two main processes: (1) data preparation and training, and (2) simulations for property evaluations. One common issue in MLFF applications arises when a trained model is used for long MD simulations on large systems. During simulations, atomic configurations may emerge that lie outside the interpolation range of the training data, potentially causing the simulation to fail. Here, failure does not simply refer to minor errors, such as inaccuracies typical of classical force fields. Instead, it refers to catastrophic breakdowns, such as atoms overlapping unnaturally or the simulation cell expanding (or collapsing) uncontrollably. In such cases, the model must be retrained, and the simulation restarted from the beginning. This results in repeated trial-and-error cycles that substantially increase both human effort and computational cost. To mitigate this issue, it is essential to improve the transferrability of MLFF models, enabling them to achieve high accuracy with a minimal amount of training data.
Error detections and model validations: Even if more training data become available and the training efficiency of MLFF improves, it remains difficult to guarantee a priori that MLFF will provide accurate results for all materials under all conditions. There are already countless MLFFs, and the number of MLFF models and their applications will continue to grow. In this situation, it is essential to continuously demonstrate the validity of the presented computational results. The validity of MLFF predictions should not simply be judged by whether the predicted properties match experimental results. This is because the FP calculations that MLFF aims to reproduce may themselves produce results that deviate from experimental findings. FP calculations are not necessarily expected to match experimental outcomes, and if MLFF yields results that coincide with experiments due to erroneous regression, it becomes difficult to reproduce the process, and the results lose scientific validity. The validity of MLFF predictions should first be evaluated by confirming that the regression results statistically remain within the interpolation range of the training data in the descriptor space. For linear regression or kernel regression, this can likely be assessed through statistical analyses, such as Bayesian error estimation and projection error evaluation. For neural networks, in practice, QBC may be the only feasible approach for such validation. Establishing uncertainty evaluation methods and carefully examining the relationship between estimated uncertainty and actual errors for various properties are crucial. It is necessary for the scientific community to systematically accumulate knowledge on the conditions under which MLFF provides reliable predictions, as well as on effective methods for assessing their validity.
Physical interpretation of MLFF: MLFF has been advancing rapidly, continuously improving in both accuracy and applicability. At the same time, however, the representation of interatomic interactions has become increasingly complex. The development of MLFF began with the representation of many-body interactions using nonlinear functions of two-body and three-body invariant descriptors, following the body-order expansion concept illustrated in Figure . Subsequently, fully many-body invariant descriptors were introduced, leading to the emergence of linear regression models that directly realize the body-order expansion framework. However, state-of-the-art methods today, which are considered superior in terms of both data efficiency and accuracy, rely on complex MPNN models incorporating many-body equivariant descriptors. While it is undeniable that this technology has driven a major breakthrough, why is this formulation necessary? What is the optimal representation for efficiently capturing interatomic interactions? Answering these fundamental questions is crucial for the continued development of MLFF.
Robust and user-friendly MLFF software packages: Dozens to hundreds of MLFF methods and training strategies are proposed every six months, and many software packages are publicly available. However, the differences between these methods are often difficult to interpret–even for specialists in MLFFs–making it challenging for users to identify the approach best suited to their application. This situation contrasts sharply with the well-established landscape of traditional MD and FP simulation packages. In addition to advancing state-of-the-art MLFF algorithms, the development of robust and user-friendly software packages is crucial for promoting the broader adoption of MLFFs.
The challenges related to the applications of MLFFs in electrochemistry are as follows.
All-atom simulations of electric double layers: In FP calculations, it is challenging to represent the entire electric double layer using an explicit atom model. As a result, model charges and implicit solvation models are often introduced to evaluate electrochemical properties. However, as discussed in this article, implicit solvation models rely on empirical parameters. Similar to the challenges of error detection and model validation in MLFFs, agreement between the results of a model that combines empirical models with FP calculations and experimental data does not, by itself, demonstrate the soundness of the model, as FP calculations themselves can deviate from experimental results. To validate these models, the finite-temperature behavior of the electric double layer must be evaluated at the FP level. The MLFF-based approach, following eq , represents one attempt to address this issue. However, it still relies on a first-order approximation based on eq . A more advanced approach–ideally one without approximations–is needed to explicitly represent the polarization effects of the electric double layer and to explicitly evaluate the potential dependence of electrochemical properties with higher accuracy. Ultimately, the most rigorous method would be to conduct all-atom MD simulations in a sufficiently large system where electrochemical reactions do not alter the chemical potentials of electrons and ions in the subsystem. MLFF is expected to make this possible. Furthermore, in such efforts, incorporating long-range interactions and potential response behavior into MLFF is expected to enhance training efficiency and applicability, leading to improved computational accuracy and efficiency.
Unified multiscale simulation framework and software: All materials follow quantum mechanics. Therefore, FP calculation methods and their software can be applied to a wide range of phenomena. This makes it easier to generalize methods and standardize computational programs. However, when the system size increases from the atomic scale to the mesoscale, complex heterogeneous structures emerge, and the number of involved elementary processes becomes enormous, necessitating approximations. As a result, different approximations are developed for different phenomena, leading to fragmentation of methods and a loss of universality. This is also why it is difficult to develop generalized coarse-grained mesoscale models, as well as universal kMC or mean-field models for complex reaction networks. The case studies introduced in this article suggest that MLFF may provide a unified framework for systematically handling diverse elementary processes. For example, MLFFs can be used to represent interactions between coarse-grained particles, free energy landscapes, and local diffusion coefficients, as well as to automatically extract reaction pathways–all within a single ML framework. The development of a unified mesoscale algorithm and software framework based on MLFF represents an important and promising research challenge.
ML-aided materials explorations: The large-scale acceleration of materials discovery using MLFF has become a mainstream approach in materials exploration. A key challenge in this approach is to extract new physically interpretable principles for material design from the vast amount of generated data. MLFF is an information science-based approach, yet it has been developed in accordance with fundamental physical principles. One major advantage of simulations using MLFF is that they provide complete atomic coordinate information and, if necessary, insights into electronic structure. In materials development, it is not only important to identify promising candidate materials through high-throughput screening but also to physically elucidate the mechanisms underlying their high performance and establish new guiding principles for material design. For discovering such principles, materials informatics approaches that uncover hidden patterns within large data sets can be effective. Ultimately, it is essential to formulate mechanisms and guiding principles based on physical laws that can be understood by humans.
MLFF is a core technology that enables accurate property evaluation, multiscale simulations, and large-scale materials exploration. Its applications are not limited to ionics and electrochemical reactions, as discussed in this article, but will undoubtedly continue to expand across a wide range of condensed matter physics and materials science.
Vocabulary
Machine learning force field (MLFF): a supervised regression model that represents the potential energy surface of a material based on interatomic interactions.
Δ-machine learning (Δ-ML): a supervised regression model that represents the difference between potential energy surfaces calculated by two different first-principles methods.
Body-order expansion: A systematic approach to modeling interactions in a system by decomposing the total potential energy into contributions from one-body (individual atoms), two-body (pairs of atoms), three-body (triplets), and higher-order terms. This expansion represents increasingly complex many-body effects and is commonly used in the development of interatomic potentials.
Equilibrium Molecular Dynamics (EMD): A simulation technique where a molecular system is allowed to evolve at thermodynamic equilibrium, i.e., no net external forces or gradients. EMD is used to calculate properties like diffusion coefficients, radial distribution functions, and thermodynamic averages by sampling equilibrium trajectories.
Non-Equilibrium Molecular Dynamics (NEMD): A simulation method where external perturbations, such as temperature gradients, shear, or electric fields, are applied to drive the system out of equilibrium. NEMD is commonly used to study transport properties like thermal conductivity, viscosity, and response to external fields.
Thermodynamic integration (TI): A method for computing free energy differences by integrating the ensemble average of a derivative (usually of the potential energy) along a continuous path between two thermodynamic states.
Thermodynamic perturbation theory (TPT): An approach for estimating the free energy difference between two systems by treating one as a small perturbation of the other, using statistical averages obtained from simulations of the reference system.
The authors declare no competing financial interest.
References
- Newell, R. ; Raimi, D. ; Villanueva, S. ; Prest, B. . Global Energy Outlook 2021: Pathways from Paris 2020. https://www.rff.org/publications/reports/global-energy-outlook-2021-pathways-from-paris/. (accessed June, 5, 2025).
- Bp Statistical Review of World Energy 2020; 2020; BP p.l.c., London. https://www.bp.com/content/dam/bp/business-sites/en/global/corporate/pdfs/energy-economics/statistical-review/bp-stats-review-2020-full-report.pdf. (accessed June, 5, 2025).
- Lewis N. S., Nocera D. G.. Powering the Planet: Chemical Challenges in Solar Energy Utilization. Proc. Natl. Acad. Sci. U.S.A. 2006;103:15729–15735. doi: 10.1073/pnas.0603395103. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Mizushima K., Jones P., Wiseman P., Goodenough J.. Li x CoO2 (0<x< – 1): A New cathode Material for Batteries of High Energy Density. Mater. Res. Bull. 1980;15:783–789. doi: 10.1016/0025-5408(80)90012-4. [DOI] [Google Scholar]
- Wilson M. S., Gottesfeld S.. Thin-film Catalyst Layers for Polymer Electrolyte Fuel Cell Electrodes. J. Appl. Electrochem. 1992;22:1–7. doi: 10.1007/BF01093004. [DOI] [Google Scholar]
- Barbir F.. PEM Electrolysis for Production of Hydrogen from Renewable Energy Sources. Sol. Energy. 2005;78:661–669. doi: 10.1016/j.solener.2004.09.003. [DOI] [Google Scholar]
- Yoshino A.. The Birth of the Lithium-Ion Battery. Angew. Chem., Int. Ed. 2012;51:5798–5800. doi: 10.1002/anie.201105006. [DOI] [PubMed] [Google Scholar]
- Goodenough J. B., Park K.-S.. The Li-Ion Rechargeable Battery: A Perspective. J. Am. Chem. Soc. 2013;135:1167–1176. doi: 10.1021/ja3091438. [DOI] [PubMed] [Google Scholar]
- Carmo M., Fritz D. L., Mergel J., Stolten D.. A Comprehensive Review on PEM Water Electrolysis. Int. J. Hydrogen Energy. 2013;38:4901–4934. doi: 10.1016/j.ijhydene.2013.01.151. [DOI] [Google Scholar]
- Yoshida T., Kojima K.. Toyota MIRAI Fuel Cell Vehicle and Progress Toward a Future Hydrogen Society. Interface Mag. 2015;24:45. doi: 10.1149/2.F03152if. [DOI] [Google Scholar]
- Miller H. A., Bouzek K., Hnat J., Loos S., Bernäcker C. I., Weißgärber T., Röntzsch L., Meier-Haack J.. Green Hydrogen from Anion Exchange Membrane Water Electrolysis: A Review of Recent Developments in Critical Materials and Operating Conditions. Sustainable Energy Fuels. 2020;4:2114–2133. doi: 10.1039/C9SE01240K. [DOI] [Google Scholar]
- Cullen D. A., Neyerlin K. C., Ahluwalia R. K., Mukundan R., More K. L., Borup R. L., Weber A. Z., Myers D. J., Kusoglu A.. New Roads and Challenges for Fuel Cells in Heavy-Duty Transportation. Nat. Energy. 2021;6:462–474. doi: 10.1038/s41560-021-00775-z. [DOI] [Google Scholar]
- Kodama K., Nagai T., Kuwaki A., Jinnouchi R., Morimoto Y.. Challenges in Applying Highly Active Pt-Based Nanostructured Catalysts for Oxygen Reduction Reactions to Fuel Cell Vehicles. Nat. Nanotechnol. 2021;16:140–147. doi: 10.1038/s41565-020-00824-w. [DOI] [PubMed] [Google Scholar]
- Du N., Roy C., Peach R., Turnbull M., Thiele S., Bock C.. Anion-Exchange Membrane Water Electrolyzers. Chem. Rev. 2022;122:11830–11895. doi: 10.1021/acs.chemrev.1c00854. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Zhu Z., Jiang T., Ali M., Meng Y., Jin Y., Cui Y., Chen W.. Rechargeable Batteries for Grid Scale Energy Storage. Chem. Rev. 2022;122:16610–16751. doi: 10.1021/acs.chemrev.2c00289. [DOI] [PubMed] [Google Scholar]
- Helmholtz H.. Ueber Einige Gesetze der Vertheilung Elektrischer Ströme in Körperlichen Leitern mit Anwendung auf die Thierisch-Elektrischen Versuche. Ann. Phys. 1853;165:211–233. doi: 10.1002/andp.18531650603. [DOI] [Google Scholar]
- Gouy M.. Sur la Constitution de la Charge Électrique á la Surface dún Électrolyte. J. Phys. Theor. Appl. 1910;9:457–468. doi: 10.1051/jphystap:019100090045700. [DOI] [Google Scholar]
- Chapman D. L.. A Contribution to the Theory of Electrocapillarity. London, Edinburgh Dublin Philos. Mag. J. Sci. 1913;25:475–481. doi: 10.1080/14786440408634187. [DOI] [Google Scholar]
- Debye P., Hückel E.. Zur Theorie der Elektrolyte. I. Gefrierpunktserniedrigung und verwandte Erscheinungen. Phys. Z. 1923;24:185–206. [Google Scholar]
- Stern O.. Zur Theorie der Elektrolytischen Doppelschicht. Z. Elektrochem. Angew. Phys. Chem. 1924;30:508–516. doi: 10.1002/bbpc.192400182. [DOI] [Google Scholar]
- Martin, R. M. Electronic Structure: Basic Theory and Practical Methods; Cambridge University Press: Cambridge, 2004. [Google Scholar]
- Nørskov J. K., Bligaard T., Rossmeisl J., Christensen C. H.. Towards the Computational Design of Solid Catalysts. Nat. Chem. 2009;1:37–46. doi: 10.1038/nchem.121. [DOI] [PubMed] [Google Scholar]
- Cheng J., Liu X., VandeVondele J., Sulpizi M., Sprik M.. Redox Potentials and Acidity Constants from Density Functional Theory Based Molecular Dynamics. Acc. Chem. Res. 2014;47:3522–3529. doi: 10.1021/ar500268y. [DOI] [PubMed] [Google Scholar]
- Urban A., Seo D.-H., Ceder G.. Computational Understanding of Li-ion Batteries. npj Comput. Mater. 2016;2:16002. doi: 10.1038/npjcompumats.2016.2. [DOI] [Google Scholar]
- Kulkarni A., Siahrostami S., Patel A., Nørskov J. K.. Understanding Catalytic Activity Trends in the Oxygen Reduction Reaction. Chem. Rev. 2018;118:2302–2312. doi: 10.1021/acs.chemrev.7b00488. [DOI] [PubMed] [Google Scholar]
- Nitopi S., Bertheussen E., Scott S. B., Liu X., Engstfeld A. K., Horch S., Seger B., Stephens I. E. L., Chan K., Hahn C., Nørskov J. K., Jaramillo T. F., Chorkendorff I.. Progress and Perspectives of Electrochemical CO2 Reduction on Copper in Aqueous Electrolyte. Chem. Rev. 2019;119:7610–7672. doi: 10.1021/acs.chemrev.8b00705. [DOI] [PubMed] [Google Scholar]
- Xiao Y., Wang Y., Bo S.-H., Kim J. C., Miara L. J., Ceder G.. Understanding Interface Stability in Solid-State Batteries. Nat. Rev. Mater. 2020;5:105–126. doi: 10.1038/s41578-019-0157-5. [DOI] [Google Scholar]
- Qing G., Ghazfar R., Jackowski S. T., Habibzadeh F., Ashtiani M. M., Chen C.-P., Smith M. R. III, Hamann T. W.. Recent Advances and Challenges of Electrocatalytic N2 Reduction to Ammonia. Chem. Rev. 2020;120:5437–5516. doi: 10.1021/acs.chemrev.9b00659. [DOI] [PubMed] [Google Scholar]
- Zhong M., Tran K., Min Y.. et al. Accelerated Discovery of CO2 Electrocatalysts Using Active Machine Learning. Nature. 2020;581:178–183. doi: 10.1038/s41586-020-2242-8. [DOI] [PubMed] [Google Scholar]
- Ringe S., Hörmann N. G., Oberhofer H., Reuter K.. Implicit Solvation Methods for Catalysis at Electrified Interfaces. Chem. Rev. 2022;122:10777–10820. doi: 10.1021/acs.chemrev.1c00675. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Zhang C., Cheng J., Chen Y.. et al. 2023 Roadmap on Molecular Modelling of Electrochemical Energy Materials. J. Phys.: Energy. 2023;5:041501. doi: 10.1088/2515-7655/acfe9b. [DOI] [Google Scholar]
- Daw M. S., Baskes M. I.. Embedded-Atom Method: Derivation and Application to Impurities, Surfaces, and Other Defects in Metals. Phys. Rev. B. 1984;29:6443. doi: 10.1103/PhysRevB.29.6443. [DOI] [Google Scholar]
- Rappe A. K., Casewit C. J., Colwell K. S., Goddard W. A. III, Skiff W. M.. UFF, a Full Periodic Table Force Field for Molecular Mechanics and Molecular Dynamics Simulations. J. Am. Chem. Soc. 1992;114:10024–10035. doi: 10.1021/ja00051a040. [DOI] [Google Scholar]
- Cornell W. D., Cieplak P., Bayly C. I., Gould I. R., Merz K. M., Ferguson D. M., Spellmeyer D. C., Fox T., Caldwell J. W., Kollman P. A.. A Second Generation Force Field for the Simulation of Proteins, Nucleic Acids, and Organic Molecules. J. Am. Chem. Soc. 1995;117:5179–5197. doi: 10.1021/ja00124a002. [DOI] [Google Scholar]
- Bowman, G. R. ; Pande, V. S. ; Noé, F. . An Introduction to Markov State Models and Their Application to Long Timescale Molecular Simulation; Springer Netherlands: Dordrecht, 2014. [Google Scholar]
- Senftle T. P., Hong S., Islam M. M., Kylasa S. B., Zheng Y., Shin Y. K., Junkermeier C., Engel-Herbert R., Janik M. J., Aktulga H. M., Verstraelen T., Grama A., van Duin A. C. T.. The ReaxFF Reactive Force-Field: Development, Applications and Future Directions. npj Comput. Mater. 2016;2:15011. doi: 10.1038/npjcompumats.2015.11. [DOI] [Google Scholar]
- Kmiecik S., Gront D., Kolinski M., Wieteska L., Dawid A. E., Kolinski A.. Coarse-Grained Protein Models and Their Applications. Chem. Rev. 2016;116:7898–7936. doi: 10.1021/acs.chemrev.6b00163. [DOI] [PubMed] [Google Scholar]
- Bruix A., Margraf J. T., Andersen M., Reuter K.. First-Principles-Based Multiscale Modelling of Heterogeneous Catalysis. Nat. Catal. 2019;2:659–670. doi: 10.1038/s41929-019-0298-3. [DOI] [Google Scholar]
- Motagamwala A. H., Dumesic J. A.. Microkinetic Modeling: A Tool for Rational Catalyst Design. Chem. Rev. 2021;121:1049–1076. doi: 10.1021/acs.chemrev.0c00394. [DOI] [PubMed] [Google Scholar]
- Jeanmairet G., Rotenberg B., Salanne M.. Microscopic Simulations of Electrochemical Double-Layer Capacitors. Chem. Rev. 2022;122:10860–10898. doi: 10.1021/acs.chemrev.1c00925. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Jones T. E., Teschner D., Piccinin S.. Toward Realistic Models of the Electrocatalytic Oxygen Evolution Reaction. Chem. Rev. 2024;124:9136–9223. doi: 10.1021/acs.chemrev.4c00171. [DOI] [PubMed] [Google Scholar]
- Allen, M. P. ; Tildesley, D. J. . Computer Simulation of Liquids; Oxford University Press, 2017. [Google Scholar]
- Frenkel, D. ; Smit, B. . Understanding Molecular Simulation, 2nd ed.; Academic Press, Inc.: USA, 2001. [Google Scholar]
- Noid W. G.. Perspective: Coarse-Grained Models for Biomolecular Systems. J. Chem. Phys. 2013;139:090901. doi: 10.1063/1.4818908. [DOI] [PubMed] [Google Scholar]
- Saunders M. G., Voth G. A.. Coarse-Graining Methods for Computational Biology. Annu. Rev. Biophys. 2013;42:73–93. doi: 10.1146/annurev-biophys-083012-130348. [DOI] [PubMed] [Google Scholar]
- Joshi S. Y., Deshmukh S. A.. A Review of Advancements in Coarse-Grained Molecular Dynamics Simulations. Mol. Simul. 2021;47:786–803. doi: 10.1080/08927022.2020.1828583. [DOI] [Google Scholar]
- Jin J., Pak A. J., Durumeric A. E. P., Loose T. D., Voth G. A.. Bottom-up Coarse-Graining: Principles and Perspectives. J. Chem. Theory Comput. 2022;18:5759–5791. doi: 10.1021/acs.jctc.2c00643. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Sun R., Han Y., Swanson J. M. J., Tan J. S., Rose J. P., Voth G. A.. Molecular Transport through Membranes: Accurate Permeability Coefficients from Multidimensional Potentials of Mean Force and Local Diffusion Constants. J. Chem. Phys. 2018;149:072310. doi: 10.1063/1.5027004. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Nagai T., Tsurumaki S., Urano R., Fujimoto K., Shinoda W., Okazaki S.. Position-Dependent Diffusion Constant of Molecules in Heterogeneous Systems as Evaluated by the Local Mean Squared Displacement. J. Chem. Theory Comput. 2020;16:7239–7254. doi: 10.1021/acs.jctc.0c00448. [DOI] [PubMed] [Google Scholar]
- Medford A. J., Shi C., Hoffmann M. J., Lausche A. C., Fitzgibbon S. R., Bligaard T., Nørskov J. K.. CatMAP: A Software Package for Descriptor-Based Microkinetic Mapping of Catalytic Trends. Catal. Lett. 2015;145:794–807. doi: 10.1007/s10562-015-1495-6. [DOI] [Google Scholar]
- Reuter K.. Ab Initio Thermodynamics and First-Principles Microkinetics for Surface Catalysis. Catal. Lett. 2016;146:541–563. doi: 10.1007/s10562-015-1684-3. [DOI] [Google Scholar]
- The Royal Swedish Academy of Sciences, Scientific Background on the Nobel Prize in Physics 2024 2024. https://www.nobelprize.org/prizes/physics/2024/advanced-information. (accessed January 15, 2024).
- The Royal Swedish Academy of Sciences, Scientific Background on the Nobel Prize in Chemistry 2024 2024. https://www.nobelprize.org/prizes/chemistry/2024/advanced-information. (accessed January,15, 2024).
- Butler K. T., Davies D. W., Cartwright H., Isayev O., Walsh A.. Machine learning for molecular and materials science. Nature. 2018;559:547–555. doi: 10.1038/s41586-018-0337-2. [DOI] [PubMed] [Google Scholar]
- Kitchin J. R.. Machine Learning in Catalysis. Nat. Catal. 2018;1:230–232. doi: 10.1038/s41929-018-0056-y. [DOI] [Google Scholar]
- Tabor D. P., Roch L. M., Saikin S. K., Kreisbeck C., Sheberla D., Montoya J. H., Dwaraknath S., Aykol M., Ortiz C., Tribukait H., Amador-Bedolla C., Brabec C. J., Maruyama B., Persson K. A., Aspuru-Guzik A.. Accelerating the Discovery of Materials for Clean Energy in the Era of Smart Automation. Nat. Rev. Mater. 2018;3:5–20. doi: 10.1038/s41578-018-0005-z. [DOI] [Google Scholar]
- Lamoureux P. S., Winther K. T., Garrido Torres J. A., Streibel V., Zhao M., Bajdich M., Abild-Pedersen F., Bligaard T.. Machine Learning for Computational Heterogeneous Catalysis. ChemCatChem. 2019;11:3581–3601. doi: 10.1002/cctc.201900595. [DOI] [Google Scholar]
- Wei J., Chu X., Sun X.-Y., Xu K., Deng H.-X., Chen J., Wei Z., Lei M.. Machine Learning in Materials Science. InfoMat. 2019;1:338–358. doi: 10.1002/inf2.12028. [DOI] [Google Scholar]
- Freeze J. G., Kelly H. R., Batista V. S.. Search for Catalysts by Inverse Design: Artificial Intelligence, Mountain Climbers, and Alchemists. Chem. Rev. 2019;119:6595–6612. doi: 10.1021/acs.chemrev.8b00759. [DOI] [PubMed] [Google Scholar]
- Chen C., Zuo Y., Ye W., Li X., Deng Z., Ong S. P.. A Critical Review of Machine Learning of Energy Materials. Adv. Energy Mater. 2020;10:1903242. doi: 10.1002/aenm.201903242. [DOI] [Google Scholar]
- Toyao T., Maeno Z., Takakusagi S., Kamachi T., Takigawa I., Shimizu K.-i.. Machine Learning for Catalysis Informatics: Recent Applications and Prospects. ACS Catal. 2020;10:2260–2297. doi: 10.1021/acscatal.9b04186. [DOI] [Google Scholar]
- Ma S., Liu Z.-P.. Machine Learning for Atomic Simulation and Activity Prediction in Heterogeneous Catalysis: Current Status and Future. ACS Catal. 2020;10:13213–13226. doi: 10.1021/acscatal.0c03472. [DOI] [Google Scholar]
- Keith J. A., Vassilev-Galindo V., Cheng B., Chmiela S., Gastegger M., Müller K.-R., Tkatchenko A.. Combining Machine Learning and Computational Chemistry for Predictive Insights Into Chemical Systems. Chem. Rev. 2021;121:9816–9872. doi: 10.1021/acs.chemrev.1c00107. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Lv C., Zhou X., Zhong L., Yan C., Srinivasan M., Seh Z. W., Liu C., Pan H., Li S., Wen Y., Yan Q.. Machine Learning: An Advanced Platform for Materials Development and State Prediction in Lithium-Ion Batteries. Adv. Mater. 2022;34:2101474. doi: 10.1002/adma.202101474. [DOI] [PubMed] [Google Scholar]
- Lombardo T., Duquesnoy M., El-Bouysidy H.. et al. Artificial Intelligence Applied to Battery Research: Hype or Reality? Chem. Rev. 2022;122:10899–10969. doi: 10.1021/acs.chemrev.1c00108. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Mai H., Le T. C., Chen D., Winkler D. A., Caruso R. A.. Machine Learning for Electrocatalyst and Photocatalyst Design and Discovery. Chem. Rev. 2022;122:13478–13515. doi: 10.1021/acs.chemrev.2c00061. [DOI] [PubMed] [Google Scholar]
- Esterhuizen J. A., Goldsmith B. R., Linic S.. Interpretable Machine Learning for Knowledge Generation in Heterogeneous Catalysis. Nat. Catal. 2022;5:175–184. doi: 10.1038/s41929-022-00744-z. [DOI] [Google Scholar]
- Szymanski N. J., Rendy B., Fei Y.. et al. An Autonomous Laboratory for the Accelerated Synthesis of Novel Materials. Nature. 2023;624:86–91. doi: 10.1038/s41586-023-06734-w. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Yao Z., Lum Y., Johnston A., Mejia-Mendoza L. M., Zhou X., Wen Y., Aspuru-Guzik A., Sargent E. H., Seh Z. W.. Machine Learning for a Sustainable Energy Future. Nat. Rev. Mater. 2023;8:202–215. doi: 10.1038/s41578-022-00490-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Mou T., Pillai H. S., Wang S., Wan M., Han X., Schweitzer N. M., Che F., Xin H.. Bridging the Complexity Gap in Computational Heterogeneous Catalysis with Machine Learning. Nat. Catal. 2023;6:122–136. doi: 10.1038/s41929-023-00911-w. [DOI] [Google Scholar]
- Tom G., Schmid S. P., Baird S. G.. et al. Self-Driving Laboratories for Chemistry and Materials Science. Chem. Rev. 2024;124:9633–9732. doi: 10.1021/acs.chemrev.4c00055. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Behler J.. Perspective: Machine Learning Potentials for Atomistic Simulations. J. Chem. Phys. 2016;145:170901. doi: 10.1063/1.4966192. [DOI] [PubMed] [Google Scholar]
- Bartók A. P., De S., Poelking C., Bernstein N., Kermode J. R., Csányi G., Ceriotti M.. Machine Learning Unifies the Modeling of Materials and Molecules. Sci. Adv. 2017;3:e1701816. doi: 10.1126/sciadv.1701816. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Deringer V. L., Caro M. A., Csányi G.. Machine Learning Interatomic Potentials as Emerging Tools for Materials Science. Adv. Mater. 2019;31:1902765. doi: 10.1002/adma.201902765. [DOI] [PubMed] [Google Scholar]
- Zuo Y., Chen C., Li X., Deng Z., Chen Y., Behler J., Csányi G., Shapeev A. V., Thompson A. P., Wood M. A., Ong S. P.. Performance and Cost Assessment of Machine Learning Interatomic Potentials. J. Phys. Chem. A. 2020;124:731–745. doi: 10.1021/acs.jpca.9b08723. [DOI] [PubMed] [Google Scholar]
- Jinnouchi R., Miwa K., Karsai F., Kresse G., Asahi R.. On-the-Fly Active Learning of Interatomic Potentials for Large-Scale Atomistic Simulations. J. Phys. Chem. Lett. 2020;11:6946–6955. doi: 10.1021/acs.jpclett.0c01061. [DOI] [PubMed] [Google Scholar]
- Mueller T., Hernandez A., Wang C.. Machine Learning for Interatomic Potential Models. J. Chem. Phys. 2020;152:050902. doi: 10.1063/1.5126336. [DOI] [PubMed] [Google Scholar]
- Behler J.. Four Generations of High-Dimensional Neural Network Potentials. Chem. Rev. 2021;121:10037–10072. doi: 10.1021/acs.chemrev.0c00868. [DOI] [PubMed] [Google Scholar]
- Unke O. T., Chmiela S., Sauceda H. E., Gastegger M., Poltavsky I., Schütt K. T., Tkatchenko A., Müller K.-R.. Machine Learning Force Fields. Chem. Rev. 2021;121:10142–10186. doi: 10.1021/acs.chemrev.0c01111. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Musil F., Grisafi A., Bartók A. P., Ortner C., Csányi G., Ceriotti M.. Physics-Inspired Structural Representations for Molecules and Materials. Chem. Rev. 2021;121:9759–9815. doi: 10.1021/acs.chemrev.1c00021. [DOI] [PubMed] [Google Scholar]
- Deringer V. L., Bartók A. P., Bernstein N., Wilkins D. M., Ceriotti M., Csányi G.. Gaussian Process Regression for Materials and Molecules. Chem. Rev. 2021;121:10073–10141. doi: 10.1021/acs.chemrev.1c00022. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Anstine D. M., Isayev O.. Machine Learning Interatomic Potentials and Long-Range Physics. J. Phys. Chem. A. 2023;127:2417–2431. doi: 10.1021/acs.jpca.2c06778. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Batatia I., Batzner S., Kovács D. P., Musaelian A., Simm G. N. C., Drautz R., Ortner C., Kozinsky B., Csányi G.. The Design Space of E(3)-Equivariant Atom-Centred Interatomic Potentials. Nat. Mach. Intell. 2025;7:56–67. doi: 10.1038/s42256-024-00956-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Curtarolo S., Setyawan W., Hart G. L., Jahnatek M., Chepulskii R. V., Taylor R. H., Wang S., Xue J., Yang K., Levy O., Mehl M. J., Stokes H. T., Demchenko D. O., Morgan D.. AFLOW: An Automatic Framework for High-Throughput Materials Discovery. Comput. Mater. Sci. 2012;58:218–226. doi: 10.1016/j.commatsci.2012.02.005. [DOI] [Google Scholar]
- Jain A., Ong S. P., Hautier G., Chen W., Richards W. D., Dacek S., Cholia S., Gunter D., Skinner D., Ceder G., Persson K. A.. The Materials Project: A Materials Genome Approach to Accelerating Materials Innovation. APL Mater. 2013;1:011002. doi: 10.1063/1.4812323. [DOI] [Google Scholar]
- Ramakrishnan R., Dral P. O., Rupp M., von Lilienfeld O. A.. Quantum Chemistry Structures and Properties of 134 Kilo Molecules. Sci. Data. 2014;1:140022. doi: 10.1038/sdata.2014.22. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Choudhary K., Garrity K. F., Reid A. C. E.. et al. The Joint Automated Repository for Various Integrated Simulations (JARVIS) for Data-Driven Materials Design. npj Comput. Mater. 2020;6:173. doi: 10.1038/s41524-020-00440-1. [DOI] [Google Scholar]
- Draxl C., Scheffler M.. The NOMAD Laboratory: From Data Sharing to Artificial Intelligence. J. Phys.: Mater. 2019;2:036001. doi: 10.1088/2515-7639/ab13bb. [DOI] [Google Scholar]
- Talirz L., Kumbhar S., Passaro E.. et al. Materials Cloud, a Platform for Open Computational Science. Sci. Data. 2020;7:299. doi: 10.1038/s41597-020-00637-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Chanussot L., Das A., Goyal S.. et al. Open Catalyst 2020 (OC20) Dataset and Community Challenges. ACS Catal. 2021;11:6059–6072. doi: 10.1021/acscatal.0c04525. [DOI] [Google Scholar]
- Szabo, A. ; Ostlund, N. S. . Modern Quantum Chemistry: Introduction to Advanced Electronic Structure Theory; Dover Publications: New York, 1989. [Google Scholar]
- Parr, R. G. ; Yang, W. . Density-Functional Theory of Atoms and Molecules; International Series of Monographs on Chemistry; Oxford University Press: Oxford, 1989; Vol. 16. [Google Scholar]
- Dirac P. A. M.. The Quantum Theory of the Electron. Proc. R. Soc. London, Ser. A. 1928;117:610–624. [Google Scholar]
- Koelling D. D., Harmon B. N.. A Technique for Relativistic Spin-Polarised Calculations. J. Phys. C: Solid State Phys. 1977;10:3107. doi: 10.1088/0022-3719/10/16/019. [DOI] [Google Scholar]
- Hamann D. R., Schlüter M., Chiang C.. Norm-Conserving Pseudopotentials. Phys. Rev. Lett. 1979;43:1494. doi: 10.1103/PhysRevLett.43.1494. [DOI] [Google Scholar]
- Vanderbilt D.. Soft Self-Consistent Pseudopotentials in a Generalized Eigenvalue Formalism. Phys. Rev. B. 1990;41:7892. doi: 10.1103/PhysRevB.41.7892. [DOI] [PubMed] [Google Scholar]
- Troullier N., Martins J. L.. Efficient Pseudopotentials for Plane-Wave Calculations. Phys. Rev. B. 1991;43:1993. doi: 10.1103/PhysRevB.43.1993. [DOI] [PubMed] [Google Scholar]
- Blöchl P. E.. Projector Augmented-Wave Method. Phys. Rev. B. 1994;50:17953. doi: 10.1103/PhysRevB.50.17953. [DOI] [PubMed] [Google Scholar]
- Kresse G., Joubert D.. From Ultrasoft Pseudopotentials to the Projector Augmented-Wave Method. Phys. Rev. B. 1999;59:1758. doi: 10.1103/PhysRevB.59.1758. [DOI] [Google Scholar]
- Kresse G., Furthmüller J.. Efficient Iterative Schemes for Ab Initio Total-Energy Calculations Using a Plane-Wave Basis Set. Phys. Rev. B. 1996;54:11169. doi: 10.1103/PhysRevB.54.11169. [DOI] [PubMed] [Google Scholar]
- Kresse G., Furthmüller J.. Efficiency of Ab-Initio Total Energy Calculations for Metals and Semiconductors using a Plane-Wave Basis Set. Comput. Mater. Sci. 1996;6:15–50. doi: 10.1016/0927-0256(96)00008-0. [DOI] [PubMed] [Google Scholar]
- VandeVondele J., Krack M., Mohamed F., Parrinello M., Chassaing T., Hutter J.. Quickstep: Fast and Accurate Density Functional Calculations Using a Mixed Gaussian and Plane Waves Approach. Comput. Phys. Commun. 2005;167:103–128. doi: 10.1016/j.cpc.2004.12.014. [DOI] [Google Scholar]
- Giannozzi P., Baroni S., Bonini N.. et al. QUANTUM ESPRESSO: A Modular and Open-Source Software Project for Quantum Simulations of Materials. J. Phys.: Condens. Matter. 2009;21:395502. doi: 10.1088/0953-8984/21/39/395502. [DOI] [PubMed] [Google Scholar]
- Born M., Oppenheimer R.. Zur Quantentheorie der Molekeln. Ann. Phys. 1927;389:457–484. doi: 10.1002/andp.19273892002. [DOI] [Google Scholar]
- Feynman, R. P. ; Hibbs, A. R. . Quantum Mechanics and Path Integrals; Dover Publications: Mineola, NY, 2010. [Google Scholar]
- Cao J., Voth G. A.. The Formulation of Quantum Statistical Mechanics Based on the Feynman Path Centroid Density. I. Equilibrium Properties. J. Chem. Phys. 1994;100:5093–5105. doi: 10.1063/1.467175. [DOI] [Google Scholar]
- Craig I. R., Manolopoulos D. E.. Quantum Statistics and Classical Mechanics: Real Time Correlation Functions from Ring Polymer Molecular Dynamics. J. Chem. Phys. 2004;121:3368–3373. doi: 10.1063/1.1777575. [DOI] [PubMed] [Google Scholar]
- Hohenberg P., Kohn W.. Inhomogeneous Electron Gas. Phys. Rev. 1964;136:B864. doi: 10.1103/PhysRev.136.B864. [DOI] [Google Scholar]
- Levy M.. Universal Variational Functionals of Electron Densities, First-Order Density Matrices, and Natural Spin-Orbitals and Solution of the v-Representability Problem. Proc. Natl. Acad. Sci. U.S.A. 1979;76:6062–6065. doi: 10.1073/pnas.76.12.6062. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Thomas L. H.. The Calculation of Atomic Fields. Math. Proc. Cambridge Philos. Soc. 1927;23:542–548. doi: 10.1017/S0305004100011683. [DOI] [Google Scholar]
- Fermi E.. Un Metodo Statistico per la Determinazione di Alcune Proprietá dell’Atomo. Rend. Accad. Naz. Lincei. 1927;6:602–607. [Google Scholar]
- Dirac P. A. M.. Note on Exchange Phenomena in the Thomas Atom. Math. Proc. Cambridge Philos. Soc. 1930;26:376–385. doi: 10.1017/S0305004100016108. [DOI] [Google Scholar]
- Weizsäcker C. F. v.. Zur Theorie der Kernmassen. Z. Phys. 1935;96:431–458. doi: 10.1007/BF01337700. [DOI] [Google Scholar]
- Kohn W., Sham L. J.. Self-Consistent Equations Including Exchange and Correlation Effects. Phys. Rev. 1965;140:A1133. doi: 10.1103/PhysRev.140.A1133. [DOI] [Google Scholar]
- Ceperley D. M., Alder B. J.. Ground State of the Electron Gas by a Stochastic Method. Phys. Rev. Lett. 1980;45:566. doi: 10.1103/PhysRevLett.45.566. [DOI] [Google Scholar]
- Perdew J. P., Zunger A.. Self-Interaction Correction to Density-Functional Approximations for Many-Electron Systems. Phys. Rev. B. 1981;23:5048. doi: 10.1103/PhysRevB.23.5048. [DOI] [Google Scholar]
- Perdew J. P., Parr R. G., Levy M., Balduz J. L.. Density-Functional Theory for Fractional Particle Number: Derivative Discontinuities of the Energy. Phys. Rev. Lett. 1982;49:1691. doi: 10.1103/PhysRevLett.49.1691. [DOI] [Google Scholar]
- Perdew J. P., Levy M.. Physical Content of the Exact Kohn-Sham Orbital Energies: Band Gaps and Derivative Discontinuities. Phys. Rev. Lett. 1983;51:1884. doi: 10.1103/PhysRevLett.51.1884. [DOI] [Google Scholar]
- Langreth D. C., Perdew J. P.. Exchange-Correlation Energy of a Metallic Surface: Wave-Vector Analysis. Phys. Rev. B. 1977;15:2884. doi: 10.1103/PhysRevB.15.2884. [DOI] [Google Scholar]
- Andersson Y., Langreth D. C., Lundqvist B. I.. van der Waals Interactions in Density-Functional Theory. Phys. Rev. Lett. 1996;76:102. doi: 10.1103/PhysRevLett.76.102. [DOI] [PubMed] [Google Scholar]
- Perdew J. P., Burke K., Ernzerhof M.. Generalized Gradient Approximation Made Simple. Phys. Rev. Lett. 1996;77:3865. doi: 10.1103/PhysRevLett.77.3865. [DOI] [PubMed] [Google Scholar]
- Seidl A., Görling A., Vogl P., Majewski J. A., Levy M.. Generalized Kohn-Sham Schemes and the Band-Gap Problem. Phys. Rev. B. 1996;53:3764. doi: 10.1103/PhysRevB.53.3764. [DOI] [PubMed] [Google Scholar]
- Perdew J. P., Ernzerhof M., Burke K.. Rationale for Mixing Exact Exchange with Density Functional Approximations. J. Chem. Phys. 1996;105:9982–9985. doi: 10.1063/1.472933. [DOI] [Google Scholar]
- Perdew, J. P. ; Schmidt, K. In Jacob’s Ladder of Density Functional Approximations for the Exchange-Correlation Energy, AIP Conference Proceedings; American Institute of Physics, 2001; pp 1–20. [Google Scholar]
- Sun J., Ruzsinszky A., Perdew J. P.. Strongly Constrained and Appropriately Normed Semilocal Density Functional. Phys. Rev. Lett. 2015;115:036402. doi: 10.1103/PhysRevLett.115.036402. [DOI] [PubMed] [Google Scholar]
- Medvedev M. G., Bushmarinov I. S., Sun J., Perdew J. P., Lyssenko K. A.. Density Functional Theory is Straying from the Path Toward the Exact Functional. Science. 2017;355:49–52. doi: 10.1126/science.aah5975. [DOI] [PubMed] [Google Scholar]
- Mermin N. D.. Thermal Properties of the Inhomogeneous Electron Gas. Phys. Rev. 1965;137:A1441. doi: 10.1103/PhysRev.137.A1441. [DOI] [Google Scholar]
- Methfessel M., Paxton A. T.. High-Precision Sampling for Brillouin-Zone Integration in Metals. Phys. Rev. B. 1989;40:3616. doi: 10.1103/PhysRevB.40.3616. [DOI] [PubMed] [Google Scholar]
- Anderson A. B., Albu T. V.. Ab Initio Determination of Reversible Potentials and Activation Energies for Outer-Sphere Oxygen Reduction to Water and the Reverse Oxidation Reaction. J. Am. Chem. Soc. 1999;121:11855–11863. doi: 10.1021/ja992735d. [DOI] [Google Scholar]
- Lozovoi A. Y., Alavi A., Kohanoff J., Lynden-Bell R. M.. Ab Initio Simulation of Charged Slabs at Constant Chemical Potential. J. Chem. Phys. 2001;115:1661–1669. doi: 10.1063/1.1379327. [DOI] [Google Scholar]
- Tavernelli I., Vuilleumier R., Sprik M.. Ab Initio Molecular Dynamics for Molecules with Variable Numbers of Electrons. Phys. Rev. Lett. 2002;88:213002. doi: 10.1103/PhysRevLett.88.213002. [DOI] [PubMed] [Google Scholar]
- Nørskov J. K., Rossmeisl J., Logadottir A., Lindqvist L., Kitchin J. R., Bligaard T., Jónsson H.. Origin of the Overpotential for Oxygen Reduction at a Fuel-Cell Cathode. J. Phys. Chem. B. 2004;108:17886–17892. doi: 10.1021/jp047349j. [DOI] [PubMed] [Google Scholar]
- Taylor C. D., Wasileski S. A., Filhol J.-S., Neurock M.. First Principles Reaction Modeling of the Electrochemical Interface: Consideration and Calculation of a Tunable Surface Potential from Atomic and Electronic Structure. Phys. Rev. B. 2006;73:165402. doi: 10.1103/PhysRevB.73.165402. [DOI] [Google Scholar]
- Otani M., Sugino O.. First-Principles Calculations of Charged Surfaces and Interfaces: A Plane-Wave Nonrepeated Slab Approach. Phys. Rev. B. 2006;73:115407. doi: 10.1103/PhysRevB.73.115407. [DOI] [Google Scholar]
- Skúlason E., Karlberg G. S., Rossmeisl J., Bligaard T., Greeley J., Jónsson H., Nørskov J. K.. Density Functional Theory Calculations for the Hydrogen Evolution Reaction in an Electrochemical Double Layer on the Pt(111) Electrode. Phys. Chem. Chem. Phys. 2007;9:3241–3250. doi: 10.1039/B700099E. [DOI] [PubMed] [Google Scholar]
- Jinnouchi R., Anderson A. B.. Electronic Structure Calculations of Liquid-Solid Interfaces: Combination of Density Functional Theory and Modified Poisson-Boltzmann Theory. Phys. Rev. B. 2008;77:245417. doi: 10.1103/PhysRevB.77.245417. [DOI] [Google Scholar]
- Rossmeisl J., Skúlason E., Björketun M. E., Tripkovic V., Nørskov J. K.. Modeling the Electrified Solid-Liquid Interface. Chem. Phys. Lett. 2008;466:68–71. doi: 10.1016/j.cplett.2008.10.024. [DOI] [Google Scholar]
- Cheng J., Sulpizi M., Sprik M.. Redox Potentials and pKa for Benzoquinone from Density Functional Theory Based Molecular Dynamics. J. Chem. Phys. 2009;131:154504. doi: 10.1063/1.3250438. [DOI] [PubMed] [Google Scholar]
- Bonnet N., Morishita T., Sugino O., Otani M.. First-Principles Molecular Dynamics at a Constant Electrode Potential. Phys. Rev. Lett. 2012;109:266101. doi: 10.1103/PhysRevLett.109.266101. [DOI] [PubMed] [Google Scholar]
- Mathew K., Sundararaman R., Letchworth-Weaver K., Arias T. A., Hennig R. G.. Implicit Solvation Model for Density-Functional Study of Nanocrystal Surfaces and Reaction Pathways. J. Chem. Phys. 2014;140:084106. doi: 10.1063/1.4865107. [DOI] [PubMed] [Google Scholar]
- Held A., Walter M.. Simplified Continuum Solvent Model with a Smooth Cavity Based on Volumetric Data. J. Chem. Phys. 2014;141:174108. doi: 10.1063/1.4900838. [DOI] [PubMed] [Google Scholar]
- Chan K., Nørskov J. K.. Electrochemical Barriers Made Simple. J. Phys. Chem. Lett. 2015;6:2663–2668. doi: 10.1021/acs.jpclett.5b01043. [DOI] [PubMed] [Google Scholar]
- Ringe S., Oberhofer H., Hille C., Matera S., Reuter K.. Function-Space-Based Solution Scheme for the Size-Modified Poisson-Boltzmann Equation in Full-Potential DFT. J. Chem. Theory Comput. 2016;12:4052–4066. doi: 10.1021/acs.jctc.6b00435. [DOI] [PubMed] [Google Scholar]
- Sundararaman R., Goddard W. A. III, Arias T. A.. Grand Canonical Electronic Density-Functional Theory: Algorithms and Applications to Electrochemistry. J. Chem. Phys. 2017;146:114104. doi: 10.1063/1.4978411. [DOI] [PubMed] [Google Scholar]
- Nishihara S., Otani M.. Hybrid Solvation Models for Bulk, Interface, and Membrane: Reference Interaction Site Methods Coupled with Density Functional Theory. Phys. Rev. B. 2017;96:115429. doi: 10.1103/PhysRevB.96.115429. [DOI] [Google Scholar]
- Jinnouchi R., Kodama K., Morimoto Y.. Electronic Structure Calculations on Electrolyte-Electrode Interfaces: Successes and Limitations. Curr. Opin. Electrochem. 2018;8:103–109. doi: 10.1016/j.coelec.2018.03.024. [DOI] [Google Scholar]
- Melander M. M., Kuisma M. J., Christensen T. E. K., Honkala K.. Grand-Canonical Approach to Density Functional Theory of Electrocatalytic Systems: Thermodynamics of Solid-Liquid Interfaces at Constant Ion and Electrode Potentials. J. Chem. Phys. 2019;150:041706. doi: 10.1063/1.5047829. [DOI] [PubMed] [Google Scholar]
- Mathew K., Kolluru V. S. C., Mula S., Steinmann S. N., Hennig R. G.. Implicit Self-Consistent Electrolyte Model in Plane-Wave Density-Functional Theory. J. Chem. Phys. 2019;151:234101. doi: 10.1063/1.5132354. [DOI] [PubMed] [Google Scholar]
- Islam S. M. R., Khezeli F., Ringe S., Plaisance C.. An Implicit Electrolyte Model for Plane Wave Density Functional Theory Exhibiting Nonlinear Response and a Nonlocal Cavity Definition. J. Chem. Phys. 2023;159:234117. doi: 10.1063/5.0176308. [DOI] [PubMed] [Google Scholar]
- Jinnouchi R.. Grand-Canonical First Principles-Based Calculations of Electrochemical Reactions. J. Electrochem. Soc. 2024;171:096502. doi: 10.1149/1945-7111/ad7326. [DOI] [Google Scholar]
- Stillinger F. H., Weber T. A.. Computer Simulation of Local Order in Condensed Phases of Silicon. Phys. Rev. B. 1985;31:5262. doi: 10.1103/PhysRevB.31.5262. [DOI] [PubMed] [Google Scholar]
- MacKerell A. D. Jr., Bashford D., Bellott M.. et al. All-Atom Empirical Potential for Molecular Modeling and Dynamics Studies of Proteins. J. Phys. Chem. B. 1998;102:3586–3616. doi: 10.1021/jp973084f. [DOI] [PubMed] [Google Scholar]
- Bishop, C. M. Pattern Recognition and Machine Learning; Information Science and Statistics; Springer: New York, 2006. [Google Scholar]
- Thompson A., Swiler L., Trott C., Foiles S., Tucker G.. Spectral Neighbor Analysis Method for Automated Generation of Quantum-Accurate Interatomic Potentials. J. Comput. Phys. 2015;285:316–330. doi: 10.1016/j.jcp.2014.12.018. [DOI] [Google Scholar]
- Wood M. A., Thompson A. P.. Extending the accuracy of the SNAP interatomic potential form. J. Chem. Phys. 2018;148:241721. doi: 10.1063/1.5017641. [DOI] [PubMed] [Google Scholar]
- Shapeev A. V.. Moment Tensor Potentials: A Class of Systematically Improvable Interatomic Potentials. Multiscale Model. Simul. 2016;14:1153–1173. doi: 10.1137/15M1054183. [DOI] [Google Scholar]
- Drautz R.. Atomic Cluster Expansion for Accurate and Transferable Interatomic Potentials. Phys. Rev. B. 2019;99:014104. doi: 10.1103/PhysRevB.99.014104. [DOI] [Google Scholar]
- Cheng B.. Cartesian Atomic Cluster Expansion for Machine Learning Interatomic Potentials. npj Comput. Mater. 2024;10:157. doi: 10.1038/s41524-024-01332-4. [DOI] [Google Scholar]
- Bartók A. P., Payne M. C., Kondor R., Csányi G.. Gaussian Approximation Potentials: The Accuracy of Quantum Mechanics, without the Electrons. Phys. Rev. Lett. 2010;104:136403. doi: 10.1103/PhysRevLett.104.136403. [DOI] [PubMed] [Google Scholar]
- Jinnouchi R., Karsai F., Kresse G.. On-the-Fly Machine Learning Force Field Generation: Application to Melting Points. Phys. Rev. B. 2019;100:014105. doi: 10.1103/PhysRevB.100.014105. [DOI] [PubMed] [Google Scholar]
- Vandermause J., Torrisi S. B., Batzner S., Xie Y., Sun L., Kolpak A. M., Kozinsky B.. On-the-Fly Active Learning of Interpretable Bayesian Force Fields for Atomistic Rare Events. npj Comput. Mater. 2020;6:20. doi: 10.1038/s41524-020-0283-z. [DOI] [Google Scholar]
- Behler J., Parrinello M.. Generalized Neural-Network Representation of High-Dimensional Potential-Energy Surfaces. Phys. Rev. Lett. 2007;98:146401. doi: 10.1103/PhysRevLett.98.146401. [DOI] [PubMed] [Google Scholar]
- Artrith N., Morawietz T., Behler J.. High-Dimensional Neural-Network Potentials for Multicomponent Systems: Applications to Zinc Oxide. Phys. Rev. B. 2011;83:153101. doi: 10.1103/PhysRevB.83.153101. [DOI] [Google Scholar]
- Ko T. W., Finkler J. A., Goedecker S., Behler J.. A Fourth-Generation High-Dimensional Neural Network Potential with Accurate Electrostatics Including Non-Local Charge Transfer. Nat. Commun. 2021;12:398. doi: 10.1038/s41467-020-20427-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Batzner S., Musaelian A., Sun L., Geiger M., Mailoa J. P., Kornbluth M., Molinari N., Smidt T. E., Kozinsky B.. E(3)-Equivariant Graph Neural Networks for Data-Efficient and Accurate Interatomic Potentials. Nat. Commun. 2022;13:2453. doi: 10.1038/s41467-022-29939-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Musaelian A., Batzner S., Johansson A., Sun L., Owen C. J., Kornbluth M., Kozinsky B.. Learning Local Equivariant Representations for Large-Scale Atomistic Dynamics. Nat. Commun. 2023;14:579. doi: 10.1038/s41467-023-36329-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Batatia I., Kovács D. P., Simm G., Ortner C., Csányi G.. MACE: Higher Order Equivariant Message Passing Neural Networks for Fast and Accurate Force Fields. Adv. Neural Inf. Process. Syst. 2022;35:11423–11436. [Google Scholar]
- Chen C., Ong S. P.. A Universal Graph Deep Learning Interatomic Potential for the Periodic Table. Nat. Comput. Sci. 2022;2:718–728. doi: 10.1038/s43588-022-00349-3. [DOI] [PubMed] [Google Scholar]
- Deng B., Zhong P., Jun K., Riebesell J., Han K., Bartel C. J., Ceder G.. CHGNet as a Pretrained Universal Neural Network Potential for Charge-Informed Atomistic Modelling. Nat. Mach. Intell. 2023;5:1031–1041. doi: 10.1038/s42256-023-00716-3. [DOI] [Google Scholar]
- Gai L., Shin Y. K., Raju M., van Duin A. C. T., Raman S.. Atomistic Adsorption of Oxygen and Hydrogen on Platinum Catalysts by Hybrid Grand Canonical Monte Carlo/Reactive Molecular Dynamics. J. Phys. Chem. C. 2016;120:9780–9793. doi: 10.1021/acs.jpcc.6b01064. [DOI] [Google Scholar]
- Khaliullin R. Z., Eshet H., Kühne T. D., Behler J., Parrinello M.. Graphite-Diamond Phase Coexistence Study Employing a Neural-Network Mapping of the Ab Initio Potential Energy Surface. Phys. Rev. B. 2010;81:100103. doi: 10.1103/PhysRevB.81.100103. [DOI] [Google Scholar]
- Eshet H., Khaliullin R. Z., Kühne T. D., Behler J., Parrinello M.. Ab Initio Quality Neural-Network Potential for Sodium. Phys. Rev. B. 2010;81:184107. doi: 10.1103/PhysRevB.81.184107. [DOI] [Google Scholar]
- Khaliullin R. Z., Eshet H., Kühne T. D., Behler J., Parrinello M.. Nucleation Mechanism for the Direct Graphite-to-Diamond Phase Transition. Nat. Mater. 2011;10:693–697. doi: 10.1038/nmat3078. [DOI] [PubMed] [Google Scholar]
- Artrith N., Behler J.. High-Dimensional Neural Network Potentials for Metal Surfaces: A Prototype Study for Copper. Phys. Rev. B. 2012;85:045439. doi: 10.1103/PhysRevB.85.045439. [DOI] [Google Scholar]
- Sosso G. C., Miceli G., Caravati S., Behler J., Bernasconi M.. Neural Network Interatomic Potential for the Phase Change Material GeTe. Phys. Rev. B. 2012;85:174103. doi: 10.1103/PhysRevB.85.174103. [DOI] [Google Scholar]
- Natarajan S. K., Behler J.. Neural Network Molecular Dynamics Simulations of Solid-Liquid Interfaces: Water at Low-Index Copper Surfaces. Phys. Chem. Chem. Phys. 2016;18:28704–28725. doi: 10.1039/C6CP05711J. [DOI] [PubMed] [Google Scholar]
- Morawietz T., Singraber A., Dellago C., Behler J.. How van der Waals Interactions Determine the Unique Properties of Water. Proc. Natl. Acad. Sci. U.S.A. 2016;113:8368–8373. doi: 10.1073/pnas.1602375113. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Hellström M., Behler J.. Concentration-Dependent Proton Transfer Mechanisms in Aqueous NaOH Solutions: From Acceptor-Driven to Donor-Driven and Back. J. Phys. Chem. Lett. 2016;7:3302–3306. doi: 10.1021/acs.jpclett.6b01448. [DOI] [PubMed] [Google Scholar]
- Miwa K., Ohno H.. Molecular Dynamics Study on β-Phase Vanadium Monohydride with Machine Learning Potential. Phys. Rev. B. 2016;94:184109. doi: 10.1103/PhysRevB.94.184109. [DOI] [Google Scholar]
- Quaranta V., Hellström M., Behler J.. Proton-Transfer Mechanisms at the Water-ZnO Interface: The Role of Presolvation. J. Phys. Chem. Lett. 2017;8:1476–1483. doi: 10.1021/acs.jpclett.7b00358. [DOI] [PubMed] [Google Scholar]
- Miwa K., Ohno H.. Interatomic Potential Construction with Self-Learning and Adaptive Database. Phys. Rev. Mater. 2017;1:053801. doi: 10.1103/PhysRevMaterials.1.053801. [DOI] [Google Scholar]
- Bartók A. P., Kermode J., Bernstein N., Csányi G.. Machine Learning a General-Purpose Interatomic Potential for Silicon. Phys. Rev. X. 2018;8:041048. doi: 10.1103/PhysRevX.8.041048. [DOI] [Google Scholar]
- Miwa K., Asahi R.. Molecular Dynamics Simulations with Machine Learning Potential for Nb-Doped Lithium Garnet-Type Oxide Li7–x La3(Zr2–x Nb x )O12 . Phys. Rev. Mater. 2018;2:105404. doi: 10.1103/PhysRevMaterials.2.105404. [DOI] [Google Scholar]
- Gerrits N., Shakouri K., Behler J., Kroes G.-J.. Accurate Probabilities for Highly Activated Reaction of Polyatomic Molecules on Surfaces Using a High-Dimensional Neural Network Potential: CHD3 + Cu(111) J. Phys. Chem. Lett. 2019;10:1763–1768. doi: 10.1021/acs.jpclett.9b00560. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Eckhoff M., Behler J.. From Molecular Fragments to the Bulk: Development of a Neural Network Potential for MOF-5. J. Chem. Theory Comput. 2019;15:3793–3809. doi: 10.1021/acs.jctc.8b01288. [DOI] [PubMed] [Google Scholar]
- Miwa K., Asahi R.. Path integral study on C15-type Laves TiCr2 hydride. Int. J. Hydrogen Energy. 2019;44:23708–23715. doi: 10.1016/j.ijhydene.2019.07.086. [DOI] [Google Scholar]
- Jinnouchi R., Lahnsteiner J., Karsai F., Kresse G., Bokdam M.. Phase Transitions of Hybrid Perovskites Simulated by Machine-Learning Force Fields Trained on the Fly with Bayesian Inference. Phys. Rev. Lett. 2019;122:225701. doi: 10.1103/PhysRevLett.122.225701. [DOI] [PubMed] [Google Scholar]
- Sivaraman G., Krishnamoorthy A. N., Baur M., Holm C., Stan M., Csányi G., Benmore C., Vázquez-Mayagoitia Á.. Machine-learned interatomic potentials by active learning: amorphous and liquid hafnium dioxide. npj Comput. Mater. 2020;6:104. doi: 10.1038/s41524-020-00367-7. [DOI] [Google Scholar]
- Paleico M. L., Behler J.. Global Optimization of Copper Clusters at the ZnO(101 0) Surface Using a DFT-Based Neural Network Potential and Genetic Algorithms. J. Chem. Phys. 2020;153:054704. doi: 10.1063/5.0014876. [DOI] [PubMed] [Google Scholar]
- Weinreich J., Römer A., Paleico M. L., Behler J.. Properties of α-Brass Nanoparticles. 1. Neural Network Potential Energy Surface. J. Phys. Chem. C. 2020;124:12682–12695. doi: 10.1021/acs.jpcc.0c00559. [DOI] [Google Scholar]
- Eckhoff M., Schönewald F., Risch M., Volkert C. A., Blöchl P. E., Behler J.. Closing the Gap between Theory and Experiment for Lithium Manganese Oxide Spinels Using a High-Dimensional Neural Network Potential. Phys. Rev. B. 2020;102:174102. doi: 10.1103/PhysRevB.102.174102. [DOI] [Google Scholar]
- Liu Y.-B., Yang J.-Y., Xin G.-M., Liu L.-H., Csányi G., Cao B.-Y.. Machine Learning Interatomic Potential Developed for Molecular Simulations on Thermal Properties of β-Ga2O3 . J. Chem. Phys. 2020;153:144501. doi: 10.1063/5.0027643. [DOI] [PubMed] [Google Scholar]
- Jinnouchi R., Karsai F., Kresse G.. Making Free-Energy Calculations Routine: Combining First Principles with Machine Learning. Phys. Rev. B. 2020;101:060201. doi: 10.1103/PhysRevB.101.060201. [DOI] [Google Scholar]
- Shapeev A. V., Podryabinkin E. V., Gubaev K., Tasnádi F., Abrikosov I. A.. Elinvar Effect in β-Ti Simulated by on-the-Fly Trained Moment Tensor Potential. New J. Phys. 2020;22:113005. doi: 10.1088/1367-2630/abc392. [DOI] [Google Scholar]
- Korotaev P., Shapeev A.. Lattice Dynamics of Yb x Co4Sb12 Skutterudite by Machine-Learning Interatomic Potentials: Effect of Filler Concentration and Disorder. Phys. Rev. B. 2020;102:184305. doi: 10.1103/PhysRevB.102.184305. [DOI] [Google Scholar]
- Verdi C., Karsai F., Liu P., Jinnouchi R., Kresse G.. Thermal Transport and Phase Transitions of Zirconia by on-the-Fly Machine-Learned Interatomic Potentials. npj Comput. Mater. 2021;7:156. doi: 10.1038/s41524-021-00630-5. [DOI] [Google Scholar]
- Rosenbrock C. W., Gubaev K., Shapeev A. V., Pártay L. B., Bernstein N., Csányi G., Hart G. L. W.. Machine-Learned Interatomic Potentials for Alloys and Alloy Phase Diagrams. npj Comput. Mater. 2021;7:24. doi: 10.1038/s41524-020-00477-2. [DOI] [Google Scholar]
- Miwa K.. Multibaric Sampling for Machine Learning Potential Construction. Phys. Rev. B. 2021;103:144106. doi: 10.1103/PhysRevB.103.144106. [DOI] [Google Scholar]
- Jinnouchi R., Karsai F., Verdi C., Kresse G.. First-Principles Hydration Free Energies of Oxygenated Species at Water-Platinum Interfaces. J. Chem. Phys. 2021;154:094107. doi: 10.1063/5.0036097. [DOI] [PubMed] [Google Scholar]
- Liu P., Verdi C., Karsai F., Kresse G.. Phase Transitions of Zirconia: Machine-Learned Force Fields beyond Density Functional Theory. Phys. Rev. B. 2022;105:L060102. doi: 10.1103/PhysRevB.105.L060102. [DOI] [Google Scholar]
- Vandermause J., Xie Y., Lim J. S., Owen C. J., Kozinsky B.. Active Learning of Reactive Bayesian Force Fields Applied to Heterogeneous Catalysis Dynamics of H/Pt. Nat. Commun. 2022;13:5183. doi: 10.1038/s41467-022-32294-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Jinnouchi R.. Molecular Dynamics Simulations of Proton Conducting Media Containing Phosphoric Acid. Phys. Chem. Chem. Phys. 2022;24:15522–15531. doi: 10.1039/D2CP00484D. [DOI] [PubMed] [Google Scholar]
- Musaelian, A. ; Batzner, S. ; Johansson, A. ; Kozinsky, B. In Scaling the Leading Accuracy of Deep Equivariant Models to Biomolecular Simulations of Realistic Size, Proceedings of the International Conference for High Performance Computing, Networking, Storage and Analysis; ACM, 2023; pp 1–12. [Google Scholar]
- Chikuma H., Takasao G., Wada T., Chammingkwan P., Behler J., Taniike T.. Accelerating Non-Empirical Structure Determination of Ziegler-Natta Catalysts with a High-Dimensional Neural Network Potential. J. Phys. Chem. C. 2023;127:11683–11691. doi: 10.1021/acs.jpcc.3c01511. [DOI] [Google Scholar]
- Witt W. C., van der Oord C., Gelz̀inytė E., Järvinen T., Ross A., Darby J. P., Ho C. H., Baldwin W. J., Sachs M., Kermode J., Bernstein N., Csányi G., Ortner C.. ACEpotentials.jl: A Julia Implementation of the Atomic Cluster Expansion. J. Chem. Phys. 2023;159:164101. doi: 10.1063/5.0158783. [DOI] [PubMed] [Google Scholar]
- Xie Y., Vandermause J., Ramakers S., Protik N. H., Johansson A., Kozinsky B.. Uncertainty-Aware Molecular Dynamics from Bayesian Active Learning for Phase Transformations and Thermal Transport in SiC. npj Comput. Mater. 2023;9:36. doi: 10.1038/s41524-023-00988-8. [DOI] [Google Scholar]
- Liu P., Wang J., Avargues N., Verdi C., Singraber A., Karsai F., Chen X.-Q., Kresse G.. Combining Machine Learning and Many-Body Calculations: Coverage-Dependent Adsorption of CO on Rh(111) Phys. Rev. Lett. 2023;130:078001. doi: 10.1103/PhysRevLett.130.078001. [DOI] [PubMed] [Google Scholar]
- Miwa K.. Restricted Multicanonical Sampling for Machine Learning Potential Construction. Phys. Rev. B. 2023;107:054313. doi: 10.1103/PhysRevB.107.054313. [DOI] [Google Scholar]
- Jinnouchi R., Minami S., Karsai F., Verdi C., Kresse G.. Proton Transport in Perfluorinated Ionomer Simulated by Machine-Learned Interatomic Potential. J. Phys. Chem. Lett. 2023;14:3581–3588. doi: 10.1021/acs.jpclett.3c00293. [DOI] [PubMed] [Google Scholar]
- Minami S., Jinnouchi R.. Accelerating Anhydrous Proton Conduction via Anion Rotation and Hydrogen Bond Recombination: a Machine-Learning Molecular Dynamics. J. Mater. Chem. A. 2023;11:16104–16114. doi: 10.1039/D3TA03164K. [DOI] [Google Scholar]
- Zheng H., Fey L. T. W., Li X.-G., Hu Y.-J., Qi L., Chen C., Xu S., Beyerlein I. J., Ong S. P.. Multi-Scale Investigation of Short-Range Order and Dislocation Glide in MoNbTi and TaNbTi Multi-Principal Element Alloys. npj Comput. Mater. 2023;9:89. doi: 10.1038/s41524-023-01046-z. [DOI] [Google Scholar]
- Gubaev K., Zaverkin V., Srinivasan P., Duff A. I., Kästner J., Grabowski B.. Performance of Two Complementary Machine-Learned Potentials in Modelling Chemically Complex Systems. npj Comput. Mater. 2023;9:129. doi: 10.1038/s41524-023-01073-w. [DOI] [Google Scholar]
- Kovács D. P., Batatia I., Arany E. S., Csányi G.. Evaluation of the MACE force field architecture: From medicinal chemistry to materials science. J. Chem. Phys. 2023;159:044118. doi: 10.1063/5.0155322. [DOI] [PubMed] [Google Scholar]
- Merchant A., Batzner S., Schoenholz S. S., Aykol M., Cheon G., Cubuk E. D.. Scaling Deep Learning for Materials Discovery. Nature. 2023;624:80–85. doi: 10.1038/s41586-023-06735-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Owen C. J., Xie Y., Johansson A., Sun L., Kozinsky B.. Low-Index Mesoscopic Surface Reconstructions of Au Surfaces Using Bayesian Force Fields. Nat. Commun. 2024;15:3790. doi: 10.1038/s41467-024-48192-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
- de Hijes P. M., Dellago C., Jinnouchi R., Schmiedmayer B., Kresse G.. Comparing Machine Learning Potentials for Water: Kernel-Based Regression and Behler-Parrinello Neural Networks. J. Chem. Phys. 2024;160:114107. doi: 10.1063/5.0197105. [DOI] [PubMed] [Google Scholar]
- Liebetrau M., Dorenkamp Y., Bünermann O., Behler J.. Hydrogen Atom Scattering at the Al2O3(0001) Surface: a Combined Experimental and Theoretical Study. Phys. Chem. Chem. Phys. 2024;26:1696–1708. doi: 10.1039/D3CP04729F. [DOI] [PubMed] [Google Scholar]
- Zare M., Sahsah D., Saleheen M., Behler J., Heyden A.. Hybrid Quantum Mechanical, Molecular Mechanical, and Machine Learning Potential for Computing Aqueous-Phase Adsorption Free Energies on Metal Surfaces. J. Chem. Theory Comput. 2024;20:8247–8260. doi: 10.1021/acs.jctc.4c00869. [DOI] [PubMed] [Google Scholar]
- Romano S., de Hijes P. M., Meier M., Kresse G., Franchini C., Dellago C.. Structure and Dynamics of the Magnetite(001)/Water Interface from Molecular Dynamics Simulations Based on a Neural Network Potential. J. Chem. Theory Comput. 2025;21:1951–1960. doi: 10.1021/acs.jctc.4c01507. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Minami S., Kin M., Takahashi K., Sato T., Jinnouchi R.. Impact of Local Water Uptake on Proton Conduction in Covalent Organic Framework Revealed by Machine-Learning Potentials. Chem. Mater. 2024;36:9535–9546. doi: 10.1021/acs.chemmater.4c01351. [DOI] [Google Scholar]
- Jinnouchi R.. Machine-Learning Surrogate Models for Particle Insertions and Element Substitutions. J. Chem. Phys. 2024;161:194110. doi: 10.1063/5.0240275. [DOI] [PubMed] [Google Scholar]
- Li Z., Hu Z., Wang Y., Ouyang R.. Molecular Dynamics Simulation of Co-Fe-Based Perovskite Oxide/Water Interfaces. J. Phys. Chem. C. 2024;128:10035–10041. doi: 10.1021/acs.jpcc.4c03394. [DOI] [Google Scholar]
- Zotov N., Gubaev K., Wörner J., Grabowski B.. Moment Tensor Potential for Static and Dynamic Investigations of Screw Dislocations in Bcc Nb. Modell. Simul. Mater. Sci. Eng. 2024;32:035032. doi: 10.1088/1361-651X/ad2d68. [DOI] [Google Scholar]
- Rybin N., Shapeev A.. A Moment Tensor Potential for Lattice Thermal Conductivity Calculations of α and β phases of Ga2O3 . J. Appl. Phys. 2024;135:205108. doi: 10.1063/5.0206215. [DOI] [Google Scholar]
- Pan J., Cheng H., Yan G., Zhang L., Yu W., Shen S.. Atomic Cluster Expansion Interatomic Potential for Defects and Thermodynamics of Cu-W System. J. Appl. Phys. 2024;136:155108. doi: 10.1063/5.0231779. [DOI] [Google Scholar]
- Liu M., Wang J., Hu J., Liu P., Niu H., Yan X., Li J., Yan H., Yang B., Sun Y., Chen C., Kresse G., Zuo L., Chen X.-Q.. Layer-by-Layer Phase Transformation in Ti3O5 Revealed by Machine-Learning Molecular Dynamics Simulations. Nat. Commun. 2024;15:3079. doi: 10.1038/s41467-024-47422-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Maxson T., Szilvási T.. Transferable Water Potentials Using Equivariant Neural Networks. J. Phys. Chem. Lett. 2024;15:3740–3747. doi: 10.1021/acs.jpclett.4c00605. [DOI] [PubMed] [Google Scholar]
- Allam O., Maghsoodi M., Jang S. S., Snow S. D.. Unveiling Competitive Adsorption in TiO2 Photocatalysis through Machine-Learning-Accelerated Molecular Dynamics, DFT, and Experimental Methods. ACS Appl. Mater. Interfaces. 2024;16:36215–36223. doi: 10.1021/acsami.4c02334. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Yang G., Liu Y.-B., Yang L., Cao B.-Y.. Machine-Learned Atomic Cluster Expansion Potentials for Fast and Quantum-Accurate Thermal Simulations of Wurtzite AlN. J. Appl. Phys. 2024;135:085105. doi: 10.1063/5.0188905. [DOI] [Google Scholar]
- Omranpour A., Behler J.. A High-Dimensional Neural Network Potential for Co3O4 . J. Phys.: Condens. Matter. 2025;37:095701. doi: 10.1088/1361-648X/ad9f09. [DOI] [PubMed] [Google Scholar]
- Jinnouchi R., Karsai F., Kresse G.. Absolute Standard Hydrogen Electrode Potential and Redox Potentials of Atoms and Molecules: Machine Learning Aided First Principles Calculations. Chem. Sci. 2025;16:2335–2343. doi: 10.1039/D4SC03378G. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Jinnouchi R., Minami S.. The Melamine-Driven Solvation Effect Promotes Oxygen Reduction on a Platinum Catalyst: Machine Learning-Aided Free Energy Calculations. J. Phys. Chem. Lett. 2025;16:265–273. doi: 10.1021/acs.jpclett.4c03437. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Blank T. B., Brown S. D., Calhoun A. W., Doren D. J.. Neural Network Models of Potential Energy Surfaces. J. Chem. Phys. 1995;103:4129–4137. doi: 10.1063/1.469597. [DOI] [Google Scholar]
- Schütt K. T., Sauceda H. E., Kindermans P.-J., Tkatchenko A., Müller K.-R.. SchNet - A Deep Learning Architecture for Molecules and Materials. J. Chem. Phys. 2018;148:241722. doi: 10.1063/1.5019779. [DOI] [PubMed] [Google Scholar]
- Xie T., Grossman J. C.. Crystal Graph Convolutional Neural Networks for an Accurate and Interpretable Prediction of Material Properties. Phys. Rev. Lett. 2018;120:145301. doi: 10.1103/PhysRevLett.120.145301. [DOI] [PubMed] [Google Scholar]
- Li Z., Kermode J. R., De Vita A.. Molecular Dynamics with on-the-Fly Machine Learning of Quantum-Mechanical Forces. Phys. Rev. Lett. 2015;114:096405. doi: 10.1103/PhysRevLett.114.096405. [DOI] [PubMed] [Google Scholar]
- Podryabinkin E. V., Shapeev A. V.. Active Learning of Linearly Parametrized Interatomic Potentials. Comput. Mater. Sci. 2017;140:171–180. doi: 10.1016/j.commatsci.2017.08.031. [DOI] [Google Scholar]
- Marzari N., Pizzi G., Cepellotti A., Togo A., Goedecker S., Jain A., Curtarolo S., Kozinsky B., Persson K. A.. Materials Cloud: A Platform for Open Materials Science. Nat. Rev. Mater. 2021;6:12–25. [Google Scholar]
- Batatia, I. ; Benner, P. ; Chiang, Y. . et al. A Foundation Model for Atomistic Materials Chemistry. 2024, arXiv:2401.00096v2. arXiv.org e-Print archive. 10.48550/arXiv.2401.00096. (accessed June 5, 2025). [DOI]
- Kahle L., Marcolongo A., Marzari N.. High-Throughput Computational Screening for Solid-State Li-Ion Conductors. Energy Environ. Sci. 2020;13:928–948. doi: 10.1039/C9EE02457C. [DOI] [Google Scholar]
- Jang S.-H., Tateyama Y., Jalem R.. High-Throughput Data-Driven Prediction of Stable High-Performance Na-Ion Sulfide Solid Electrolytes. Adv. Funct. Mater. 2022;32:2206036. doi: 10.1002/adfm.202206036. [DOI] [Google Scholar]
- Chen C., Nguyen D. T., Lee S. J., Baker N. A., Karakoti A. S., Lauw L., Owen C., Mueller K. T., Bilodeau B. A., Murugesan V., Troyer M.. Accelerating Computational Materials Discovery with Machine Learning and Cloud High-Performance Computing: from Large-Scale Screening to Experimental Validation. J. Am. Chem. Soc. 2024;146:20009–20018. doi: 10.1021/jacs.4c03849. [DOI] [PubMed] [Google Scholar]
- Klawohn S., Darby J. P., Kermode J. R., Csányi G., Caro M. A., Bartók A. P.. Gaussian Approximation Potentials: Theory, Software Implementation and Application Examples. J. Chem. Phys. 2023;159:174108. doi: 10.1063/5.0160898. [DOI] [PubMed] [Google Scholar]
- Podryabinkin E., Garifullin K., Shapeev A., Novikov I.. MLIP-3: Active Learning on Atomic Environments with Moment Tensor Potentials. J. Chem. Phys. 2023;159:084112. doi: 10.1063/5.0155887. [DOI] [PubMed] [Google Scholar]
- Singraber A., Behler J., Dellago C.. Library-Based LAMMPS Implementation of High-Dimensional Neural Network Potentials. J. Chem. Theory Comput. 2019;15:1827–1840. doi: 10.1021/acs.jctc.8b00770. [DOI] [PubMed] [Google Scholar]
- Schütt K., Kindermans P.-J., Felix H. E. S., Chmiela S., Tkatchenko A., Müller K.-R.. SchNet: AContinuous-Filter Convolutional Neural Network for Modeling Quantum Interactions. Adv. Neural Inf. Process. Syst. 2017;30:992–1002. [Google Scholar]
- Zhang L., Han J., Wang H., Car R., E W.. Deep Potential Molecular Dynamics: A Scalable Model with the Accuracy of Quantum Mechanics. Phys. Rev. Lett. 2018;120:143001. doi: 10.1103/PhysRevLett.120.143001. [DOI] [PubMed] [Google Scholar]
- Zeng J., Zhang D., Lu D.. et al. DeePMD-kit v2: A Software Package for Deep Potential Models. J. Chem. Phys. 2023;159:054801. doi: 10.1063/5.0155600. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Lysogorskiy Y., van der Oord C., Bochkarev A., Menon S., Rinaldi M., Hammerschmidt T., Mrovec M., Thompson A., Csányi G., Ortner C., Drautz R.. Performant Implementation of the Atomic Cluster Expansion (PACE) and Application to Copper and Silicon. npj Comput. Mater. 2021;7:97. doi: 10.1038/s41524-021-00559-9. [DOI] [Google Scholar]
- Kovács, D. Machine Learning Force Fields for Molecular Chemistry. Ph.D. Thesis; University of Cambridge: Cambridge, UK, 2024. [Google Scholar]
- Paszke A.. PyTorch: An Imperative Style, High-Performance Deep Learning Library. Adv. Neural Inf. Process. Syst. 2019;32:8026–8037. [Google Scholar]
- Abadi, M. ; Barham, P. ; Chen, J. . et al. In TensorFlow: A System for Large-Scale Machine Learning, 12th USENIX Symposium on Operating Systems Design and Implementation; OSDI: Savannah, GA, 2016; pp 265–283. [Google Scholar]
- Plimpton S.. Fast Parallel Algorithms for Short-Range Molecular Dynamics. J. Comput. Phys. 1995;117:1–19. doi: 10.1006/jcph.1995.1039. [DOI] [Google Scholar]
- Clark S. J., Segall M. D., Pickard C. J., Hasnip P. J., Probert M. I. J., Refson K., Payne M. C.. First Principles Methods Using CASTEP. Z. Kristallogr. - Cryst. Mater. 2005;220:567–570. doi: 10.1524/zkri.220.5.567.65075. [DOI] [Google Scholar]
- Larsen A. H., Mortensen J. J., Blomqvist J.. et al. The Atomic Simulation Environmenta Python Library for Working with Atoms. J. Phys.: Condens. Matter. 2017;29:273002. doi: 10.1088/1361-648X/aa680e. [DOI] [PubMed] [Google Scholar]
- MedeA version 3.10. Materials Design, Inc., San Diego, USA, 2025.
- Materials Studio. Dassault Systémes, BIOVIA, San Diego, USA, 2024
- Glielmo A., Zeni C., De Vita A.. Efficient Nonparametric n-Body Force Fields from Machine Learning. Phys. Rev. B. 2018;97:184307. doi: 10.1103/PhysRevB.97.184307. [DOI] [Google Scholar]
- Mayer J. E.. The Statistical Mechanics of Condensing Systems. I. J. Chem. Phys. 1937;5:67–73. doi: 10.1063/1.1749933. [DOI] [Google Scholar]
- Smith J. S., Isayev O., Roitberg A. E.. ANI-1: An Extensible Neural Network Potential with DFT Accuracy at Force Field Computational Cost. Chem. Sci. 2017;8:3192–3203. doi: 10.1039/C6SC05720A. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Seko A., Togo A., Tanaka I.. Group-Theoretical High-Order Rotational Invariants for Structural Representations: Application to Linearized Machine Learning Interatomic Potential. Phys. Rev. B. 2019;99:214108. doi: 10.1103/PhysRevB.99.214108. [DOI] [Google Scholar]
- Unke O. T., Meuwly M.. PhysNet: A Neural Network for Predicting Energies, Forces, Dipole Moments, and Partial Charges. J. Chem. Theory Comput. 2019;15:3678–3693. doi: 10.1021/acs.jctc.9b00181. [DOI] [PubMed] [Google Scholar]
- Faraji S., Ghasemi S. A., Rostami S., Rasoulkhani R., Schaefer B., Goedecker S., Amsler M.. High Accuracy and Transferability of a Neural Network Potential through Charge Equilibration for Calcium Fluoride. Phys. Rev. B. 2017;95:104105. doi: 10.1103/PhysRevB.95.104105. [DOI] [Google Scholar]
- Grisafi A., Ceriotti M.. Incorporating Long-Range Physics in Atomic-Scale Machine Learning. J. Chem. Phys. 2019;151:204105. doi: 10.1063/1.5128375. [DOI] [PubMed] [Google Scholar]
- Faller C., Kaltak M., Kresse G.. Density-Based Long-Range Electrostatic Descriptors for Machine Learning Force Fields. J. Chem. Phys. 2024;161:214701. doi: 10.1063/5.0245615. [DOI] [PubMed] [Google Scholar]
- Gilmer, J. ; Schoenholz, S. S. ; Riley, P. F. ; Vinyals, O. ; Dahl, G. E. In Neural Message Passing for Quantum Chemistry, Proceedings of the 34th International Conference on Machine Learning; PMLR, 2017; pp 1263–1272. [Google Scholar]
- Chmiela S., Sauceda H. E., Müller K.-R., Tkatchenko A.. Towards Exact Molecular Dynamics Simulations with Machine-Learned Force Fields. Nat. Commun. 2018;9:3887. doi: 10.1038/s41467-018-06169-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Chen C., Ye W., Zuo Y., Zheng C., Ong S. P.. Graph Networks as a Universal Machine Learning Framework for Molecules and Crystals. Chem. Mater. 2019;31:3564–3572. doi: 10.1021/acs.chemmater.9b01294. [DOI] [Google Scholar]
- Gasteiger J., Groß J., Günnemann S.. Directional Message Passing on Molecular Graphs via Synthetic Coordinates. Adv. Neural Inf. Process. Syst. 2021;34:15421–15433. [Google Scholar]
- Choudhary K., DeCost B.. Atomistic Line Graph Neural Network for improved materials property predictions. npj Comput. Mater. 2021;7:185. doi: 10.1038/s41524-021-00650-1. [DOI] [Google Scholar]
- Faber F. A., Christensen A. S., Huang B., von Lilienfeld O. A.. Alchemical and Structural Distribution Based Representation for Universal Quantum Machine Learning. J. Chem. Phys. 2018;148:241717. doi: 10.1063/1.5020710. [DOI] [PubMed] [Google Scholar]
- Schölkopf B., Smola A., Müller K.-R.. Nonlinear Component Analysis as a Kernel Eigenvalue Problem. Neural Comput. 1998;10:1299–1319. doi: 10.1162/089976698300017467. [DOI] [Google Scholar]
- Rappe A. K., Goddard W. A. III. Charge Equilibration for Molecular Dynamics Simulations. J. Phys. Chem. A. 1991;95:3358–3363. doi: 10.1021/j100161a070. [DOI] [Google Scholar]
- Ghasemi S. A., Hofstetter A., Saha S., Goedecker S.. Interatomic Potentials for Ionic Systems with Density Functional Accuracy Based on Charge Densities Obtained by a Neural Network. Phys. Rev. B. 2015;92:045131. doi: 10.1103/PhysRevB.92.045131. [DOI] [Google Scholar]
- Hirshfeld F. L.. Bonded-Atom Fragments for Describing Molecular Charge Densities. Theor. Chim. Acta. 1977;44:129–138. doi: 10.1007/BF00549096. [DOI] [Google Scholar]
- Siepmann J. I., Sprik M.. Influence of Surface Topology and Electrostatic Potential on Water/Electrode Systems. J. Chem. Phys. 1995;102:511–524. doi: 10.1063/1.469429. [DOI] [Google Scholar]
- Coretti A., Scalfi L., Bacon C., Rotenberg B., Vuilleumier R., Ciccotti G., Salanne M., Bonella S.. Mass-Zero Constrained Molecular Dynamics for Electrode Charges in Simulations of Electrochemical Systems. J. Chem. Phys. 2020;152:194701. doi: 10.1063/5.0007192. [DOI] [PubMed] [Google Scholar]
- Takahashi K., Nakano H., Sato H.. A Polarizable Molecular Dynamics Method for Electrode-Electrolyte Interfacial Electron Transfer under the Constant Chemical-Potential-Difference Condition on the Electrode Electrons. J. Chem. Phys. 2020;153:054126. doi: 10.1063/5.0020619. [DOI] [PubMed] [Google Scholar]
- Coretti A., Bacon C., Berthin R., Serva A., Scalfi L., Chubak I., Goloviznina K., Haefele M., Marin-Laflèche A., Rotenberg B., Bonella S., Salanne M.. MetalWalls: Simulating electrochemical interfaces between polarizable electrolytes and metallic electrodes. J. Chem. Phys. 2022;157:184801. doi: 10.1063/5.0101777. [DOI] [PubMed] [Google Scholar]
- Mulliken R. S.. Electronic Population Analysis on LCAO-MO Molecular Wave Functions. I. J. Chem. Phys. 1955;23:1833–1840. doi: 10.1063/1.1740588. [DOI] [Google Scholar]
- Breneman C. M., Wiberg K. B.. Determining Atom-Centered Monopoles from Molecular Electrostatic Potentials. The Need for High Sampling Density in Formamide Conformational Analysis. J. Comput. Chem. 1990;11:361–373. doi: 10.1002/jcc.540110311. [DOI] [Google Scholar]
- Bader, R. F. Atoms in Molecules: A Quantum Theory; Oxford University Press: Oxford, 1990. [Google Scholar]
- Gonze X., Lee C.. Dynamical Matrices, Born Effective Charges, Dielectric Permittivity Tensors, and Interatomic Force Constants from Density-Functional Perturbation Theory. Phys. Rev. B. 1997;55:10355. doi: 10.1103/PhysRevB.55.10355. [DOI] [Google Scholar]
- Bultinck P., Van Alsenoy C., Ayers P. W., Carbó-Dorca R.. Critical Analysis and Extension of the Hirshfeld Atoms in Molecules. J. Chem. Phys. 2007;126:144111. doi: 10.1063/1.2715563. [DOI] [PubMed] [Google Scholar]
- Grisafi A., Bussy A., Salanne M., Vuilleumier R.. Predicting the Charge Density Response in Metal Electrodes. Phys. Rev. Mater. 2023;7:125403. doi: 10.1103/PhysRevMaterials.7.125403. [DOI] [Google Scholar]
- Nigam J., Pozdnyakov S., Fraux G., Ceriotti M.. Unified Theory of Atom-Centered Representations and Message-Passing Machine-Learning Schemes. J. Chem. Phys. 2022;156:204115. doi: 10.1063/5.0087042. [DOI] [PubMed] [Google Scholar]
- Seung, H. S. ; Opper, M. ; Sompolinsky, H. In Query by Committee, Proceedings of the Fifth Annual Workshop on Computational Learning Theory; ACM: New York, NY, USA, 1992; pp 287–294. [Google Scholar]
- Smith J. S., Nebgen B., Lubbers N., Isayev O., Roitberg A. E.. Less Is More: Sampling Chemical Space With Active Learning. J. Chem. Phys. 2018;148:241733. doi: 10.1063/1.5023802. [DOI] [PubMed] [Google Scholar]
- Zhang L., Lin D.-Y., Wang H., Car R., E W.. Active Learning of Uniformly Accurate Interatomic Potentials for Materials Simulation. Phys. Rev. Mater. 2019;3:023804. doi: 10.1103/PhysRevMaterials.3.023804. [DOI] [Google Scholar]
- Tokita A. M., Behler J.. How to Train a Neural Network Potential. J. Chem. Phys. 2023;159:121501. doi: 10.1063/5.0160326. [DOI] [PubMed] [Google Scholar]
- Bartók A. P., Kondor R., Csányi G.. On Representing Chemical Environments. Phys. Rev. B. 2013;87:184115. doi: 10.1103/PhysRevB.87.184115. [DOI] [Google Scholar]
- Willatt M. J., Musil F., Ceriotti M.. Atom-Density Representations for Machine Learning. J. Chem. Phys. 2019;150:154110. doi: 10.1063/1.5090481. [DOI] [PubMed] [Google Scholar]
- Grisafi A., Wilkins D. M., Csányi G., Ceriotti M.. Symmetry-Adapted Machine Learning for Tensorial Properties of Atomistic Systems. Phys. Rev. Lett. 2018;120:036002. doi: 10.1103/PhysRevLett.120.036002. [DOI] [PubMed] [Google Scholar]
- Thomas, N. ; Smidt, T. ; Kearnes, S. ; Yang, L. ; Li, L. ; Kohlhoff, K. ; Riley, P. . Tensor Field Networks: Rotation- and Translation-Equivariant Neural Networks for 3D Point Clouds. 2018, arXiv:1802.08219v3. arXiv.org e-Print archive. 10.48550/arXiv.1802.08219. (accessed June 5, 2025). [DOI]
- Anderson, B. ; Hy, T.-S. ; Kondor, R. In Cormorant: Covariant Molecular Neural Networks, Proceedings of the 33rd International Conference on Neural Information Processing Systems; Curran Associates Inc.: Red Hook, NY, USA, 2019; pp 14537–14546. [Google Scholar]
- Grisafi A., Fabrizio A., Meyer B., Wilkins D. M., Corminboeuf C., Ceriotti M.. Transferable Machine-Learning Model of the Electron Density. ACS Cent. Sci. 2019;5:57–64. doi: 10.1021/acscentsci.8b00551. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Artrith N., Urban A., Ceder G.. Efficient and Accurate Machine-Learning Interpolation of Atomic Energies in Compositions with Many Species. Phys. Rev. B. 2017;96:014112. doi: 10.1103/PhysRevB.96.014112. [DOI] [Google Scholar]
- Gastegger M., Schwiedrzik L., Bittermann M., Berzsenyi F., Marquetand P.. wACSF-Weighted Atom-Centered Symmetry Functions as Descriptors in Machine Learning Potentials. J. Chem. Phys. 2018;148:241709. doi: 10.1063/1.5019667. [DOI] [PubMed] [Google Scholar]
- Willatt M. J., Musil F., Ceriotti M.. Feature Optimization for Atomistic Machine Learning Yields a Data-Driven Construction of the Periodic Table of the Elements. Phys. Chem. Chem. Phys. 2018;20:29661–29668. doi: 10.1039/C8CP05921G. [DOI] [PubMed] [Google Scholar]
- Kostiuchenko T., Körmann F., Neugebauer J., Shapeev A.. Impact of Lattice Relaxations on Phase Transitions in a High-Entropy Alloy Studied by Machine-Learning Potentials. npj Comput. Mater. 2019;5:55. doi: 10.1038/s41524-019-0195-y. [DOI] [Google Scholar]
- Nigam J., Pozdnyakov S., Ceriotti M.. Recursive Evaluation and Iterative Contraction of N-Body Equivariant Features. J. Chem. Phys. 2020;153:121101. doi: 10.1063/5.0021116. [DOI] [PubMed] [Google Scholar]
- Jinnouchi R., Karsai F., Verdi C., Asahi R., Kresse G.. Descriptors Representing Two- and Three-Body Atomic Distributions and Their Effects on the Accuracy of Machine-Learned Inter-Atomic Potentials. J. Chem. Phys. 2020;152:234102. doi: 10.1063/5.0009491. [DOI] [PubMed] [Google Scholar]
- Goscinski A., Musil F., Pozdnyakov S., Nigam J., Ceriotti M.. Optimal Radial Basis for Density-Based Atomic Representations. J. Chem. Phys. 2021;155:104106. doi: 10.1063/5.0057229. [DOI] [PubMed] [Google Scholar]
- Zeni C., Rossi K., Glielmo A., de Gironcoli S.. Compact Atomic Descriptors Enable Accurate Predictions via Linear Models. J. Chem. Phys. 2021;154:224112. doi: 10.1063/5.0052961. [DOI] [PubMed] [Google Scholar]
- Darby J. P., Kermode J. R., Csányi G.. Compressing Local Atomic Neighbourhood Descriptors. npj Comput. Mater. 2022;8:166. doi: 10.1038/s41524-022-00847-y. [DOI] [Google Scholar]
- Pozdnyakov S. N., Willatt M. J., Bartók A. P., Ortner C., Csányi G., Ceriotti M.. Incompleteness of Atomic Structure Representations. Phys. Rev. Lett. 2020;125:166001. doi: 10.1103/PhysRevLett.125.166001. [DOI] [PubMed] [Google Scholar]
- Lim J. S., Vandermause J., van Spronsen M. A., Musaelian A., Xie Y., Sun L., O’Connor C. R., Egle T., Molinari N., Florian J., Duanmu K., Madix R. J., Sautet P., Friend C. M., Kozinsky B.. Evolution of Metastable Structures at Bimetallic Surfaces from Microscopy and Machine-Learning Molecular Dynamics. J. Am. Chem. Soc. 2020;142:15907–15916. doi: 10.1021/jacs.0c06401. [DOI] [PubMed] [Google Scholar]
- Ochi S., Kamishima O., Mizusaki J., Kawamura J.. Investigation of Proton Diffusion in Nafion 117 Membrane by Electrical Conductivity and NMR. Solid State Ionics. 2009;180:580–584. doi: 10.1016/j.ssi.2008.12.035. [DOI] [Google Scholar]
- Sasaki R., Gao B., Hitosugi T., Tateyama Y.. Nonequilibrium Molecular Dynamics for Accelerated Computation of Ion-Ion Correlated Conductivity beyond Nernst-Einstein Limitation. npj Comput. Mater. 2023;9:48. doi: 10.1038/s41524-023-00996-8. [DOI] [Google Scholar]
- Shimizu K., Otsuka R., Hara M., Minamitani E., Watanabe S.. Prediction of Born Effective Charges Using Neural Network to Study Ion Migration under Electric Fields: Applications to Crystalline and Amorphous Li3PO4 . Sci. Technol. Adv. Mater.:Methods. 2023;3:2253135. doi: 10.1080/27660400.2023.2253135. [DOI] [Google Scholar]
- Green M. S.. Markoff Random Processes and the Statistical Mechanics of Time-Dependent Phenomena. II. Irreversible Processes in Fluids. J. Chem. Phys. 1954;22:398–413. doi: 10.1063/1.1740082. [DOI] [Google Scholar]
- Kubo R.. Statistical-Mechanical Theory of Irreversible Processes. I. General Theory and Simple Applications to Magnetic and Conduction Problems. J. Phys. Soc. Jpn. 1957;12:570–586. doi: 10.1143/JPSJ.12.570. [DOI] [Google Scholar]
- Yang J., Tse J. S.. Li Ion Diffusion Mechanisms in LiFePO4: An ab Initio Molecular Dynamics Study. J. Phys. Chem. A. 2011;115:13045–13049. doi: 10.1021/jp205057d. [DOI] [PubMed] [Google Scholar]
- Mo Y., Ong S. P., Ceder G.. First Principles Study of the Li10GeP2S12 Lithium Super Ionic Conductor Material. Chem. Mater. 2012;24:15–17. doi: 10.1021/cm203303y. [DOI] [Google Scholar]
- Ong S. P., Mo Y., Richards W. D., Miara L., Lee H. S., Ceder G.. Phase Stability, Electrochemical Stability and Ionic Conductivity of the Li10 ± 1MP2X12 (M = Ge, Si, Sn, Al or P, and X = O, S or Se) Family of Superionic Conductors. Energy Environ. Sci. 2013;6:148–156. doi: 10.1039/C2EE23355J. [DOI] [Google Scholar]
- Wang Y., Richards W. D., Ong S. P., Miara L. J., Kim J. C., Mo Y., Ceder G.. Design Principles for Solid-State Lithium Superionic Conductors. Nat. Mater. 2015;14:1026–1031. doi: 10.1038/nmat4369. [DOI] [PubMed] [Google Scholar]
- Mills G., Jónsson H., Schenter G. K.. Reversible Work Transition State Theory: Application to Dissociative Adsorption of Hydrogen. Surf. Sci. 1995;324:305–337. doi: 10.1016/0039-6028(94)00731-4. [DOI] [Google Scholar]
- Henkelman G., Jónsson H.. Improved Tangent Estimate in the Nudged Elastic Band Method for Finding Minimum Energy Paths and Saddle Points. J. Chem. Phys. 2000;113:9978–9985. doi: 10.1063/1.1323224. [DOI] [Google Scholar]
- Van der Ven A., Ceder G.. Lithium Diffusion in Layered Li x CoO2 . Electrochem. Solid-State Lett. 2000;3:301 [Google Scholar]
- Van der Ven A., Ceder G., Asta M., Tepesch P. D.. First-Principles Theory of Ionic Diffusion with Nondilute Carriers. Phys. Rev. B. 2001;64:184307. doi: 10.1103/PhysRevB.64.184307. [DOI] [Google Scholar]
- Kang K., Ceder G.. Factors That Affect Li Mobility in Layered Lithium Transition Metal Oxides. Phys. Rev. B. 2006;74:094105. doi: 10.1103/PhysRevB.74.094105. [DOI] [Google Scholar]
- Choe Y.-K., Tsuchida E., Ikeshoji T., Yamakawa S., Hyodo S.-a.. Nature of Proton Dynamics in a Polymer Electrolyte Membrane, Nafion: a First-Principles Molecular Dynamics Study. Phys. Chem. Chem. Phys. 2009;11:3892–3899. doi: 10.1039/b819535h. [DOI] [PubMed] [Google Scholar]
- Devanathan R., Idupulapati N., Baer M. D., Mundy C. J., Dupuis M.. Ab Initio Molecular Dynamics Simulation of Proton Hopping in a Model Polymer Membrane. J. Phys. Chem. B. 2013;117:16522–16529. doi: 10.1021/jp410229u. [DOI] [PubMed] [Google Scholar]
- Nernst W.. Zur Kinetik der in Lösung Befindlichen Körper. Z. Phys. Chem. 1889;4:129–181. [Google Scholar]
- Einstein A.. Über die von der Molekularkinetischen Theorie der Wärme Geforderte Bewegung von in Ruhenden Flüssigkeiten Suspendierten Teilchen. Ann. Phys. 1905;322:549–560. doi: 10.1002/andp.19053220806. [DOI] [Google Scholar]
- Agmon N.. The Grotthuss Mechanism. Chem. Phys. Lett. 1995;244:456–462. doi: 10.1016/0009-2614(95)00905-J. [DOI] [Google Scholar]
- Kreuer K.-D.. Proton Conductivity: Materials and Applications. Chem. Mater. 1996;8:610–641. doi: 10.1021/cm950192a. [DOI] [Google Scholar]
- Kreuer K.-D., Portale G.. A Critical Revision of the Nano-Morphology of Proton Conducting Ionomers and Polyelectrolytes for Fuel Cell Applications. Adv. Funct. Mater. 2013;23:5390–5397. doi: 10.1002/adfm.201300376. [DOI] [Google Scholar]
- Agmon N., Bakker H. J., Campen R. K., Henchman R. H., Pohl P., Roke S., Thämer M., Hassanali A.. Protons and Hydroxide Ions in Aqueous Systems. Chem. Rev. 2016;116:7642–7672. doi: 10.1021/acs.chemrev.5b00736. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kusoglu A., Weber A. Z.. New Insights into Perfluorinated Sulfonic-Acid Ionomers. Chem. Rev. 2017;117:987–1104. doi: 10.1021/acs.chemrev.6b00159. [DOI] [PubMed] [Google Scholar]
- Eigen M.. Proton Transfer, Acid-Base Catalysis, and Enzymatic Hydrolysis. Part I: Elementary Processes. Angew. Chem., Int. Ed. Engl. 1964;3:1–19. doi: 10.1002/anie.196400011. [DOI] [Google Scholar]
- Zundel G., Metzger H.. Energiebänder der Tunnelnden Überschuß-Protonen in Flüssigen Säuren. Eine IR-Spektroskopische Untersuchung der Natur der Gruppierungen H5O2 + . Z. Phys. Chem. 1968;58:225–245. doi: 10.1524/zpch.1968.58.5_6.225. [DOI] [Google Scholar]
- de Grotthuss C. J. T.. Mémoire sur la Décomposition de l’Eau et des Corps Qu’elle Tient. Ann. Chim. 1806;58:54–73. [Google Scholar]
- Tuckerman M., Laasonen K., Sprik M., Parrinello M.. Ab Initio Molecular Dynamics Simulation of the Solvation and Transport of Hydronium and Hydroxyl Ions in Water. J. Chem. Phys. 1995;103:150–161. doi: 10.1063/1.469654. [DOI] [Google Scholar]
- Robinson, R. A. ; Stokes, R. H. . Electrolyte Solutions, 2nd ed.; Butterworths: London, 1959. [Google Scholar]
- Meiboom S.. Nuclear Magnetic Resonance Study of the Proton Transfer in Water. J. Chem. Phys. 1961;34:375–388. doi: 10.1063/1.1700960. [DOI] [Google Scholar]
- Zawodzinski T. A., Derouin C., Radzinski S., Sherman R. J., Smith V. T., Springer T. E., Gottesfeld S.. Water Uptake by and Transport Through Nafion 117 Membranes. J. Electrochem. Soc. 1993;140:1041. doi: 10.1149/1.2056194. [DOI] [Google Scholar]
- Zawodzinskijr T., Springer T. E., Uribe F., Gottesfeld S.. Characterization of Polymer Electrolytes for Fuel Cell Applications. Solid State Ionics. 1993;60:199–211. doi: 10.1016/0167-2738(93)90295-E. [DOI] [Google Scholar]
- Schmitt U. W., Voth G. A.. Multistate Empirical Valence Bond Model for Proton Transport in Water. J. Phys. Chem. B. 1998;102:5547–5551. doi: 10.1021/jp9818131. [DOI] [Google Scholar]
- Aeberhard P. C., Williams S. R., Evans D. J., Refson K., David W. I. F.. Ab initio Nonequilibrium Molecular Dynamics in the Solid Superionic Conductor LiBH4 . Phys. Rev. Lett. 2012;108:095901. doi: 10.1103/PhysRevLett.108.095901. [DOI] [PubMed] [Google Scholar]
- Falletta S., Cepellotti A., Johansson A., Tan C. W., Descoteaux M. L., Musaelian A., Owen C. J., Kozinsky B.. Unified Differentiable Learning of Electric Response. Nat. Commun. 2025;16:4031. doi: 10.1038/s41467-025-59304-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kutana A., Shimizu K., Watanabe S., Asahi R.. Representing Born Effective Charges with Equivariant Graph Convolutional Neural Networks. Sci. Rep. 2025;15:16719. doi: 10.1038/s41598-025-01250-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Steinfeld, J. I. ; Francisco, J. S. ; Hase, W. L. . Chemical Kinetics and Dynamics, 2nd ed.; Prentice Hall: Upper Saddle River, NJ, 1998. [Google Scholar]
- Liu X., Cheng J., Sprik M.. Aqueous Transition-Metal Cations as Impurities in a Wide Gap Oxide: The Cu2+/Cu+ and Ag2+/Ag+ Redox Couples Revisited. J. Phys. Chem. B. 2015;119:1152–1163. doi: 10.1021/jp506691h. [DOI] [PubMed] [Google Scholar]
- Wang F., Cheng J.. Automated Workflow for Computation of Redox Potentials, Acidity Constants, and Solvation Free Energies Accelerated by Machine Learning. J. Chem. Phys. 2022;157:024103. doi: 10.1063/5.0098330. [DOI] [PubMed] [Google Scholar]
- Jinnouchi R., Karsai F., Kresse G.. Machine Learning-Aided First-Principles Calculations of Redox Potentials. npj Comput. Mater. 2024;10:107. doi: 10.1038/s41524-024-01295-6. [DOI] [Google Scholar]
- Wang F., Ma Z., Cheng J.. Accelerating Computation of Acidity Constants and Redox Potentials for Aqueous Organic Redox Flow Batteries by Machine Learning Potential-Based Molecular Dynamics. J. Am. Chem. Soc. 2024;146:14566–14575. doi: 10.1021/jacs.4c01221. [DOI] [PubMed] [Google Scholar]
- Aydinol M. K., Kohan A. F., Ceder G., Cho K., Joannopoulos J.. Ab Initio Study of Lithium Intercalation in Metal Oxides and Metal Dichalcogenides. Phys. Rev. B. 1997;56:1354. doi: 10.1103/PhysRevB.56.1354. [DOI] [Google Scholar]
- Kirkwood J. G.. Statistical Mechanics of Fluid Mixtures. J. Chem. Phys. 1935;3:300–313. doi: 10.1063/1.1749657. [DOI] [Google Scholar]
- Zwanzig R. W.. High-Temperature Equation of State by a Perturbation Method. I. Nonpolar Gases. J. Chem. Phys. 1954;22:1420–1426. doi: 10.1063/1.1740409. [DOI] [Google Scholar]
- Perdew J. P.. Accurate Density Functional for the Energy: Real-Space Cutoff of the Gradient Expansion for the Exchange Hole. Phys. Rev. Lett. 1985;55:1665. doi: 10.1103/PhysRevLett.55.1665. [DOI] [PubMed] [Google Scholar]
- Perdew J. P., Chevary J. A., Vosko S. H., Jackson K. A., Pederson M. R., Singh D. J., Fiolhais C.. Atoms, Molecules, Solids, and Surfaces: Applications of the Generalized Gradient Approximation for Exchange and Correlation. Phys. Rev. B. 1992;46:6671. doi: 10.1103/PhysRevB.46.6671. [DOI] [PubMed] [Google Scholar]
- Pham T. A., Zhang C., Schwegler E., Galli G.. Probing the Electronic Structure of Liquid Water with Many-Body Perturbation Theory. Phys. Rev. B. 2014;89:060202. doi: 10.1103/PhysRevB.89.060202. [DOI] [Google Scholar]
- Kühne T. D., Iannuzzi M., Del Ben M.. et al. CP2K: An Electronic Structure and Molecular Dynamics Software Package - Quickstep: Efficient and Accurate Electronic Structure Calculations. J. Chem. Phys. 2020;152:194103. doi: 10.1063/5.0007045. [DOI] [PubMed] [Google Scholar]
- Zhan C.-G., Dixon D. A.. Absolute Hydration Free Energy of the Proton from First-Principles Electronic Structure Calculations. J. Phys. Chem. A. 2001;105:11534–11540. doi: 10.1021/jp012536s. [DOI] [Google Scholar]
- Baik M.-H., Friesner R. A.. Computing Redox Potentials in Solution: Density Functional Theory as A Tool for Rational Design of Redox Agents. J. Phys. Chem. A. 2002;106:7407–7412. doi: 10.1021/jp025853n. [DOI] [Google Scholar]
- Kelly C. P., Cramer C. J., Truhlar D. G.. Aqueous Solvation Free Energies of Ions and Ion-Water Clusters Based on an Accurate Value for the Absolute Aqueous Solvation Free Energy of the Proton. J. Phys. Chem. B. 2006;110:16066–16081. doi: 10.1021/jp063552y. [DOI] [PubMed] [Google Scholar]
- Jaque P., Marenich A. V., Cramer C. J., Truhlar D. G.. Computational Electrochemistry: The Aqueous Ru3+| Ru2+ Reduction Potential. J. Phys. Chem. C. 2007;111:5783–5799. doi: 10.1021/jp066765w. [DOI] [Google Scholar]
- Jinnouchi R., Anderson A. B.. Aqueous and Surface Redox Potentials from Self-Consistently Determined Gibbs Energies. J. Phys. Chem. C. 2008;112:8747–8750. doi: 10.1021/jp802627s. [DOI] [Google Scholar]
- Jinnouchi R., Hatanaka T., Morimoto Y., Osawa M.. First Principles Study of Sulfuric Acid Anion Adsorption on a Pt(111) Electrode. Phys. Chem. Chem. Phys. 2012;14:3208–3218. doi: 10.1039/c2cp23172g. [DOI] [PubMed] [Google Scholar]
- Neugebauer H., Bohle F., Bursch M., Hansen A., Grimme S.. Benchmark Study of Electrochemical Redox Potentials Calculated with Semiempirical and DFT Methods. J. Phys. Chem. A. 2020;124:7166–7176. doi: 10.1021/acs.jpca.0c05052. [DOI] [PubMed] [Google Scholar]
- Lybrand T. P., Ghosh I., McCammon J. A.. Hydration of Chloride and Bromide Anions: Determination of Relative Free Energy by Computer Simulation. J. Am. Chem. Soc. 1985;107:7793–7794. doi: 10.1021/ja00311a112. [DOI] [Google Scholar]
- Lybrand T. P., McCammon J. A., Wipff G.. Theoretical Calculation of Relative Binding Affinity in Host-Guest Systems. Proc. Natl. Acad. Sci. U.S.A. 1986;83:833–835. doi: 10.1073/pnas.83.4.833. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Straatsma T. P., Berendsen H. J. C.. Free Energy of Ionic Hydration: Analysis of a Thermodynamic Integration Technique to Evaluate Free Energy Differences by Molecular Dynamics Simulations. J. Chem. Phys. 1988;89:5876–5886. doi: 10.1063/1.455539. [DOI] [Google Scholar]
- Hammer B., Hansen L. B., Nørskov J. K.. Improved Adsorption Energetics within Density-Functional Theory Using Revised Perdew-Burke-Ernzerhof Functionals. Phys. Rev. B. 1999;59:7413. doi: 10.1103/PhysRevB.59.7413. [DOI] [Google Scholar]
- Grimme S., Antony J., Ehrlich S., Krieg H.. A Consistent and Accurate Ab Initio Parametrization of Density Functional Dispersion Correction (DFT-D) for the 94 Elements H-Pu. J. Chem. Phys. 2010;132:154104. doi: 10.1063/1.3382344. [DOI] [PubMed] [Google Scholar]
- Grimme S.. Density Functional Theory with London Dispersion Corrections. WIREs Comput. Mol. Sci. 2011;1:211–228. doi: 10.1002/wcms.30. [DOI] [Google Scholar]
- Adamo C., Barone V.. Toward reliable density functional methods without adjustable parameters: The PBE0 model. J. Chem. Phys. 1999;110:6158–6170. doi: 10.1063/1.478522. [DOI] [Google Scholar]
- Heyd J., Scuseria G. E., Ernzerhof M.. Hybrid Functionals Based on a Screened Coulomb Potential. J. Chem. Phys. 2003;118:8207–8215. doi: 10.1063/1.1564060. [DOI] [Google Scholar]
- Heyd J., Scuseria G. E., Ernzerhof M.. Erratum: Hybrid Functionals Based on a Screened Coulomb Potential [J. Chem. Phys. 118, 8207 (2003)] J. Chem. Phys. 2006;124:219906. doi: 10.1063/1.2204597. [DOI] [Google Scholar]
- Becke A. D.. Density-Functional Thermochemistry. III. The Role of Exact Exchange. J. Chem. Phys. 1993;98:5648–5652. doi: 10.1063/1.464913. [DOI] [Google Scholar]
- Gauthier J. A., Dickens C. F., Ringe S., Chan K.. Practical Considerations for Continuum Models Applied to Surface Electrochemistry. ChemPhysChem. 2019;20:3074–3080. doi: 10.1002/cphc.201900536. [DOI] [PubMed] [Google Scholar]
- Domínguez-Flores F., Melander M. M.. Approximating Constant Potential DFT with Canonical DFT and Electrostatic Corrections. J. Chem. Phys. 2023;158:144701. doi: 10.1063/5.0138197. [DOI] [PubMed] [Google Scholar]
- Janak J. F.. Proof That ∂E/∂ni = ε in Density-Functional Theory. Phys. Rev. B. 1978;18:7165. doi: 10.1103/PhysRevB.18.7165. [DOI] [Google Scholar]
- Ceder G., Chiang Y.-M., Sadoway D. R., Aydinol M. K., Jang Y.-I., Huang B.. Identification of Cathode Materials for Lithium Batteries Guided by First-Principles Calculations. Nature. 1998;392:694–696. doi: 10.1038/33647. [DOI] [Google Scholar]
- Nørskov J. K., Bligaard T., Logadottir A., Kitchin J. R., Chen J. G., Pandelov S., Stimming U.. Trends in the Exchange Current for Hydrogen Evolution. J. Electrochem. Soc. 2005;152:J23. doi: 10.1149/1.1856988. [DOI] [Google Scholar]
- Man I. C., Su H.-Y., Calle-Vallejo F., Hansen H. A., Martínez J. I., Inoglu N. G., Kitchin J., Jaramillo T. F., Nørskov J. K., Rossmeisl J.. Universality in Oxygen Evolution Electrocatalysis on Oxide Surfaces. ChemCatChem. 2011;3:1159–1165. doi: 10.1002/cctc.201000397. [DOI] [Google Scholar]
- Liu X., Xiao J., Peng H., Hong X., Chan K., Nørskov J. K.. Understanding Trends in Electrochemical Carbon Dioxide Reduction Rates. Nat. Commun. 2017;8:15438. doi: 10.1038/ncomms15438. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Skúlason E., Bligaard T., Gudmundsdóttir S., Studt F., Rossmeisl J., Abild-Pedersen F., Vegge T., Jónsson H., Nørskov J. K.. A Theoretical Evaluation of Possible Transition Metal Electro-Catalysts for N2 Reduction. Phys. Chem. Chem. Phys. 2012;14:1235–1245. doi: 10.1039/C1CP22271F. [DOI] [PubMed] [Google Scholar]
- Kuhl K. P., Hatsukade T., Cave E. R., Abram D. N., Kibsgaard J., Jaramillo T. F.. Electrocatalytic Conversion of Carbon Dioxide to Methane and Methanol on Transition Metal Surfaces. J. Am. Chem. Soc. 2014;136:14107–14113. doi: 10.1021/ja505791r. [DOI] [PubMed] [Google Scholar]
- Hoshi N., Hori Y.. Electrochemical Reduction of Carbon Dioxide at a Series of Platinum Single Crystal Electrodes. Electrochim. Acta. 2000;45:4263–4270. doi: 10.1016/S0013-4686(00)00559-4. [DOI] [Google Scholar]
- Hatsukade T., Kuhl K. P., Cave E. R., Abram D. N., Jaramillo T. F.. Insights into the Electrocatalytic Reduction of CO2 on Metallic Silver Surfaces. Phys. Chem. Chem. Phys. 2014;16:13814–13819. doi: 10.1039/C4CP00692E. [DOI] [PubMed] [Google Scholar]
- Trasatti S.. The Absolute Electrode Potential: An Explanatory Note (Recommendations 1986) Pure Appl. Chem. 1986;58:955–966. doi: 10.1351/pac198658070955. [DOI] [Google Scholar]
- Hummelshøj J. S., Blomqvist J., Datta S., Vegge T., Rossmeisl J., Thygesen K. S., Luntz A. C., Jacobsen K. W., Nørskov J. K.. Communications: Elementary oxygen electrode reactions in the aprotic Li-air battery. J. Chem. Phys. 2010;132:071101. doi: 10.1063/1.3298994. [DOI] [PubMed] [Google Scholar]
- Greeley J., Jaramillo T. F., Bonde J., Chorkendorff I., Nørskov J. K.. Computational High-Throughput Screening of Electrocatalytic Materials for Hydrogen Evolution. Nat. Mater. 2006;5:909–913. doi: 10.1038/nmat1752. [DOI] [PubMed] [Google Scholar]
- Markovic N., Gasteiger H., Ross P. N.. Kinetics of Oxygen Reduction on Pt(hkl) Electrodes: Implications for the Crystallite Size Effect with Supported Pt Electrocatalysts. J. Electrochem. Soc. 1997;144:1591. doi: 10.1149/1.1837646. [DOI] [Google Scholar]
- Blizanac B. B., Lucas C. A., Gallagher M. E., Arenz M., Ross P. N., Marković N. M.. Anion Adsorption, CO Oxidation, and Oxygen Reduction Reaction on a Au(100) Surface: The pH Effect. J. Phys. Chem. B. 2004;108:625–634. doi: 10.1021/jp036483l. [DOI] [Google Scholar]
- Blizanac B. B., Ross P. N., Marković N. M.. Oxygen Reduction on Silver Low-Index Single-Crystal Surfaces in Alkaline Solution: Rotating Ring DiskAg(hkl ) Studies. J. Phys. Chem. B. 2006;110:4735–4741. doi: 10.1021/jp056050d. [DOI] [PubMed] [Google Scholar]
- Stamenkovic V., Mun B. S., Mayrhofer K. J., Ross P. N., Markovic N. M., Rossmeisl J., Greeley J., Nørskov J. K.. Changing the Activity of Electrocatalysts for Oxygen Reduction by Tuning the Surface Electronic Structure. Angew. Chem., Int. Ed. 2006;45:2897–2901. doi: 10.1002/anie.200504386. [DOI] [PubMed] [Google Scholar]
- Shao M. H., Huang T., Liu P., Zhang J., Sasaki K., Vukmirovic M. B., Adzic R. R.. Palladium Monolayer and Palladium Alloy Electrocatalysts for Oxygen Reduction. Langmuir. 2006;22:10409–10415. doi: 10.1021/la0610553. [DOI] [PubMed] [Google Scholar]
- Stamenkovic V. R., Fowler B., Mun B. S., Wang G., Ross P. N., Lucas C. A., Marković N. M.. Improved Oxygen Reduction Activity on Pt3Ni(111) via Increased Surface Site Availability. Science. 2007;315:493–497. doi: 10.1126/science.1135941. [DOI] [PubMed] [Google Scholar]
- Greeley J., Stephens I. E. L., Bondarenko A. S., Johansson T. P., Hansen H. A., Jaramillo T. F., Rossmeisl J., Chorkendorff I., Nørskov J. K.. Alloys of Platinum and Early Transition Metals as Oxygen Reduction Electrocatalysts. Nat. Chem. 2009;1:552–556. doi: 10.1038/nchem.367. [DOI] [PubMed] [Google Scholar]
- Stephens I. E. L., Bondarenko A. S., Perez-Alonso F. J., Calle-Vallejo F., Bech L., Johansson T. P., Jepsen A. K., Frydendal R., Knudsen B. P., Rossmeisl J., Chorkendorff I.. Tuning the Activity of Pt(111) for Oxygen Electroreduction by Subsurface Alloying. J. Am. Chem. Soc. 2011;133:5485–5491. doi: 10.1021/ja111690g. [DOI] [PubMed] [Google Scholar]
- Hansen H. A., Viswanathan V., Nørskov J. K.. Unifying Kinetic and Thermodynamic Analysis of 2 e– and 4 e– Reduction of Oxygen on Metal Surfaces. J. Phys. Chem. C. 2014;118:6706–6718. doi: 10.1021/jp4100608. [DOI] [Google Scholar]
- Wada N., Nakamura M., Hoshi N.. Structural Effects on the Oxygen Reduction Reaction on Pt Single-Crystal Electrodes Modified with Melamine. Electrocatalysis. 2020;11:275–281. doi: 10.1007/s12678-020-00584-0. [DOI] [Google Scholar]
- Asahi M., Yamazaki S., Taguchi N., Ioroi T.. Facile Approach to Enhance Oxygen Reduction Activity by Modification of Platinum Nanoparticles by Melamine-Formaldehyde Polymer. J. Electrochem. Soc. 2019;166:F498. doi: 10.1149/2.0641908jes. [DOI] [Google Scholar]
- Daimon H., Yamazaki S., Asahi M., Ioroi T., Inaba M.. A Strategy for Drastic Improvement in the Durability of Pt/C and PtCo/C Alloy Catalysts for the Oxygen Reduction Reaction by Melamine Surface Modification. ACS Catal. 2022;12:8976–8985. doi: 10.1021/acscatal.2c01942. [DOI] [Google Scholar]
- Ikeshoji T., Otani M.. Toward Full Simulation of the Electrochemical Oxygen Reduction Reaction on Pt Using First-Principles and Kinetic Calculations. Phys. Chem. Chem. Phys. 2017;19:4447–4453. doi: 10.1039/C6CP08466D. [DOI] [PubMed] [Google Scholar]
- Kristoffersen H. H., Chan K., Vegge T., Hansen H. A.. Energy-Entropy Competition in Cation-Hydroxyl Interactions at the Liquid Water-Pt(111) Interface. Chem. Commun. 2020;56:427–430. doi: 10.1039/C9CC07769C. [DOI] [PubMed] [Google Scholar]
- Limmer D. T., Willard A. P., Madden P., Chandler D.. Hydration of Metal Surfaces Can Be Dynamically Heterogeneous and Hydrophobic. Proc. Natl. Acad. Sci. U.S.A. 2013;110:4200–4205. doi: 10.1073/pnas.1301596110. [DOI] [Google Scholar]
- Rossmeisl J., Karlberg G. S., Jaramillo T., Nørskov J. K.. Steady State Oxygen Reduction and Cyclic Voltammetry. Faraday Discuss. 2009;140:337–346. doi: 10.1039/B802129E. [DOI] [PubMed] [Google Scholar]
- Strasser P., Koh S., Anniyev T., Greeley J., More K., Yu C., Liu Z., Kaya S., Nordlund D., Ogasawara H., Toney M. F., Nilsson A.. Lattice-Strain Control of the Activity in Dealloyed Core-Shell Fuel Cell Catalysts. Nat. Chem. 2010;2:454–460. doi: 10.1038/nchem.623. [DOI] [PubMed] [Google Scholar]
- Shao Y., Andersson L., Knijff L., Zhang C.. Finite-Field Coupling via Learning the Charge Response Kernel. Electron. Struct. 2022;4:014012. doi: 10.1088/2516-1075/ac59ca. [DOI] [Google Scholar]
- Grisafi A., Salanne. Accelerating QM/MM Simulations of Electrochemical Interfaces through Machine Learning of Electronic Charge Densities. J. Chem. Phys. 2024;161:024109. doi: 10.1063/5.0218379. [DOI] [PubMed] [Google Scholar]
- Bird, R. B. ; Stewart, W. E. ; Lightfoot, E. N. . Transport Phenomena, 2nd ed.; John Wiley & Sons: New York, 2002. [Google Scholar]
- Liwo A., Ołdziej S., Pincus M. R., Wawak R. J., Rackovsky S., Scheraga H. A.. A United-Residue Force Field for Off-Lattice Protein-Structure Simulations. I. Functional Forms and Parameters of Long-Range Side-Chain Interaction Potentials from Protein Crystal Data. J. Comput. Chem. 1997;18:849–873. doi: 10.1002/(SICI)1096-987X(199705)18:7<849::AID-JCC1>3.0.CO;2-R. [DOI] [Google Scholar]
- Izvekov S., Voth G. A.. A Multiscale Coarse-Graining Method for Biomolecular Systems. J. Phys. Chem. B. 2005;109:2469–2473. doi: 10.1021/jp044629q. [DOI] [PubMed] [Google Scholar]
- Maupetit J., Tuffery P., Derreumaux P.. A Coarse-Grained Protein Force Field for Folding and Structure Prediction. Proteins:Struct., Funct., Bioinf. 2007;69:394–408. doi: 10.1002/prot.21505. [DOI] [PubMed] [Google Scholar]
- Marrink S. J., Risselada H. J., Yefimov S., Tieleman D. P., de Vries A. H.. The MARTINI Force Field: Coarse Grained Model for Biomolecular Simulations. J. Phys. Chem. B. 2007;111:7812–7824. doi: 10.1021/jp071097f. [DOI] [PubMed] [Google Scholar]
- Husic B. E., Charron N. E., Lemm D., Wang J., Pérez A., Majewski M., Krämer A., Chen Y., Olsson S., de Fabritiis G., Noé F., Clementi C.. Coarse Graining Molecular Dynamics with Graph Neural Networks. J. Chem. Phys. 2020;153:194101. doi: 10.1063/5.0026133. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Loose T. D., Sahrmann P. G., Qu T. S., Voth G. A.. Coarse-Graining with Equivariant Neural Networks: A Path Toward Accurate and Data-Efficient Models. J. Phys. Chem. B. 2023;127:10564–10572. doi: 10.1021/acs.jpcb.3c05928. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Duschatko, B. R. ; Fu, X. ; Owen, C. ; Xie, Y. ; Musaelian, A. ; Jaakkola, T. ; Kozinsky, B. . Thermodynamically Informed Multimodal Learning of High-Dimensional Free Energy Models in Molecular Coarse Graining. 2024, arXiv:2405.19386v1. arXiv.org e-Print archive. https://arxiv.org/abs/2405.19386. (accessed June 5, 2025).
- Nagai T., Kikkawa N., Jinnouchi R., Kimura M., Okazaki S.. Machine Learning Model to Predict Free-Energy Landscape and Position-Dependent Diffusion Constant to Extend the Scale of Dynamic Monte Carlo Simulations. J. Chem. Theory Comput. 2025;21:2598–2611. doi: 10.1021/acs.jctc.4c01552. [DOI] [PubMed] [Google Scholar]
- Nagai T., Fujimoto K., Okazaki S.. Three-Dimensional Free-Energy Landscape of Hydrogen and Oxygen Molecules in Polymer Electrolyte Membranes: Insight into Diffusion Paths. J. Chem. Phys. 2022;156:044507. doi: 10.1063/5.0075969. [DOI] [PubMed] [Google Scholar]
- Nørskov J., Bligaard T., Logadottir A., Bahn S., Hansen L., Bollinger M., Bengaard H., Hammer B., Sljivancanin Z., Mavrikakis M., Xu Y., Dahl S., Jacobsen C.. Universality in Heterogeneous Catalysis. J. Catal. 2002;209:275–278. doi: 10.1006/jcat.2002.3615. [DOI] [Google Scholar]
- Ulissi Z. W., Medford A. J., Bligaard T., Nørskov J. K.. To Address Surface Reaction Network Complexity Using Scaling Relations Machine Learning and DFT Calculations. Nat. Commun. 2017;8:14621. doi: 10.1038/ncomms14621. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Jinnouchi R., Asahi R.. Predicting Catalytic Activity of Nanoparticles by a DFT-Aided Machine-Learning Algorithm. J. Phys. Chem. Lett. 2017;8:4279–4283. doi: 10.1021/acs.jpclett.7b02010. [DOI] [PubMed] [Google Scholar]
- Shi Y.-F., Kang P.-L., Shang C., Liu Z.-P.. Methanol Synthesis from CO2/CO Mixture on Cu-Zn Catalysts from Microkinetics-Guided Machine Learning Pathway Search. J. Am. Chem. Soc. 2022;144:13401–13414. doi: 10.1021/jacs.2c06044. [DOI] [PubMed] [Google Scholar]
- Otani M., Hamada I., Sugino O., Morikawa Y., Okamoto Y., Ikeshoji T.. Electrode Dynamics from First Principles. J. Phys. Soc. Jpn. 2008;77:024802. doi: 10.1143/JPSJ.77.024802. [DOI] [PubMed] [Google Scholar]
- Kronberg R., Laasonen K.. Reconciling the Experimental and Computational Hydrogen Evolution Activities of Pt(111) through DFT-Based Constrained MD Simulations. ACS Catal. 2021;11:8062–8078. doi: 10.1021/acscatal.1c00538. [DOI] [Google Scholar]
- Rosca V., Duca M., de Groot M. T., Koper M. T. M.. Nitrogen Cycle Electrocatalysis. Chem. Rev. 2009;109:2209–2244. doi: 10.1021/cr8003696. [DOI] [PubMed] [Google Scholar]
- Taylor H. S.. A theory of the catalytic surface. Proc. R. Soc. London, Ser. A. 1925;108:105–111. [Google Scholar]
- Boudart, M. Catalysis by Supported Metals; Boudart, M. , Ed.; Academic Press, 1969; Vol. 20, pp 153–166. [Google Scholar]
- Zambelli T., Wintterlin J., Trost J., Ertl G.. Identification of the “Active Sites” of a Surface-Catalyzed Reaction. Science. 1996;273:1688–1690. doi: 10.1126/science.273.5282.1688. [DOI] [Google Scholar]
- Nørskov J. K., Bligaard T., Hvolbæk B., Abild-Pedersen F., Chorkendorff I., Christensen C. H.. The Nature of the Active Site in Heterogeneous Metal Catalysis. Chem. Soc. Rev. 2008;37:2163–2171. doi: 10.1039/b800260f. [DOI] [PubMed] [Google Scholar]
- Marković N., Ross P.. Surface Science Studies of Model Fuel Cell Electrocatalysts. Surf. Sci. Rep. 2002;45:117–229. doi: 10.1016/S0167-5729(01)00022-X. [DOI] [Google Scholar]
- Henkelman G., Jónsson H.. A Dimer Method for Finding Saddle Points on High Dimensional Potential Surfaces Using Only First Derivatives. J. Chem. Phys. 1999;111:7010–7022. doi: 10.1063/1.480097. [DOI] [Google Scholar]
- Sprik M., Ciccotti G.. Free Energy from Constrained Molecular Dynamics. J. Chem. Phys. 1998;109:7737–7744. doi: 10.1063/1.477419. [DOI] [Google Scholar]
- Winther K. T., Hoffmann M. J., Boes J. R., Mamun O., Bajdich M., Bligaard T.. Catalysis-Hub.org, an Open Electronic Structure Database for Surface Reactions. Sci. Data. 2019;6:75. doi: 10.1038/s41597-019-0081-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Abild-Pedersen F., Greeley J., Studt F., Rossmeisl J., Munter T. R., Moses P. G., Skúlason E., Bligaard T., Nørskov J. K.. Scaling Properties of Adsorption Energies for Hydrogen-Containing Molecules on Transition-Metal Surfaces. Phys. Rev. Lett. 2007;99:016105. doi: 10.1103/PhysRevLett.99.016105. [DOI] [PubMed] [Google Scholar]
- Bligaard T., Nørskov J., Dahl S., Matthiesen J., Christensen C., Sehested J.. The Brønsted-Evans-Polanyi Relation and the Volcano Curve in Heterogeneous Catalysis. J. Catal. 2004;224:206–217. doi: 10.1016/j.jcat.2004.02.034. [DOI] [Google Scholar]
- Nørskov J. K., Abild-Pedersen F., Studt F., Bligaard T.. Density Functional Theory in Surface Chemistry and Catalysis. Proc. Natl. Acad. Sci. U.S.A. 2011;108:937–943. doi: 10.1073/pnas.1006652108. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Medford A. J., Vojvodic A., Hummelshøj J. S., Voss J., Abild-Pedersen F., Studt F., Bligaard T., Nilsson A., Nørskov J. K.. From the Sabatier Principle to a Predictive Theory of Transition-Metal Heterogeneous Catalysis. J. Catal. 2015;328:36–42. doi: 10.1016/j.jcat.2014.12.033. [DOI] [Google Scholar]
- Weininger D.. SMILES, a Chemical Language and Information System. 1. Introduction to Methodology and Encoding Rules. J. Chem. Inf. Comput. Sci. 1988;28:31–36. doi: 10.1021/ci00057a005. [DOI] [Google Scholar]
- Jinnouchi R., Hirata H., Asahi R.. Extrapolating Energetics on Clusters and Single-Crystal Surfaces to Nanoparticles by Machine-Learning Scheme. J. Phys. Chem. C. 2017;121:26397–26405. doi: 10.1021/acs.jpcc.7b08686. [DOI] [Google Scholar]
- Shang C., Liu Z.-P.. Stochastic Surface Walking Method for Structure Prediction and Pathway Searching. J. Chem. Theory Comput. 2013;9:1838–1845. doi: 10.1021/ct301010b. [DOI] [PubMed] [Google Scholar]
- Artrith N., Kolpak A. M.. Understanding the Composition and Activity of Electrocatalytic Nanoalloys in Aqueous Solvents: A Combination of DFT and Accurate Neural Network Potentials. Nano Lett. 2014;14:2670–2676. doi: 10.1021/nl5005674. [DOI] [PubMed] [Google Scholar]
- Quaranta V., Behler J., Hellström M.. Structure and Dynamics of the Liquid-Water/Zinc-Oxide Interface from Machine Learning Potential Simulations. J. Phys. Chem. C. 2019;123:1293–1304. doi: 10.1021/acs.jpcc.8b10781. [DOI] [Google Scholar]
- Andrade M. F. C., Ko H.-Y., Zhang L., Car R., Selloni A.. Free Energy of Proton Transfer at the Water-TiO2 Interface from Ab Initio Deep Potential Molecular Dynamics. Chem. Sci. 2020;11:2335–2341. doi: 10.1039/C9SC05116C. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Mikkelsen A. E. G., Kristoffersen H. H., Schiøtz J., Vegge T., Hansen H. A., Jacobsen K. W.. Structure and Energetics of Liquid Water-Hydroxyl Layers on Pt(111) Phys. Chem. Chem. Phys. 2022;24:9885–9890. doi: 10.1039/D2CP00190J. [DOI] [PubMed] [Google Scholar]
- Kahle L., Marcolongo A., Marzari N.. Modeling Lithium-Ion Solid-State Electrolytes with a Pinball Model. Phys. Rev. Mater. 2018;2:065405. doi: 10.1103/PhysRevMaterials.2.065405. [DOI] [Google Scholar]
- Hellenbrandt M.. The Inorganic Crystal Structure Database (ICSD)-Present and Future. Crystallogr. Rev. 2004;10:17–22. doi: 10.1080/08893110410001664882. [DOI] [Google Scholar]
- Saal J. E., Kirklin S., Aykol M., Meredig B., Wolverton C.. Materials Design and Discovery with High-Throughput Density Functional Theory: The Open Quantum Materials Database (OQMD) JOM. 2013;65:1501–1509. doi: 10.1007/s11837-013-0755-4. [DOI] [Google Scholar]
- Pickard C. J., Needs R. J.. Ab Initio Random Structure Searching. J. Phys.: Condens. Matter. 2011;23:053201. doi: 10.1088/0953-8984/23/5/053201. [DOI] [PubMed] [Google Scholar]











