Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2025 Dec 17.
Published in final edited form as: J Chem Theory Comput. 2025 Dec 10;21(24):12709–12724. doi: 10.1021/acs.jctc.5c01400

A universal augmentation framework for long-range electrostatics in machine learning interatomic potentials

Dongjin Kim 1, Xiaoyu Wang 1, Santiago Vargas 2, Peichen Zhong 3, Daniel S King 3, Theo Jaffrelot Inizan 3, Bingqing Cheng 1,2,3,4,5,*
PMCID: PMC12707604  NIHMSID: NIHMS2128398  PMID: 41368735

Abstract

Most current machine learning interatomic potentials (MLIPs) rely on short-range approximations, without explicit treatment of long-range electrostatics. To address this, we recently developed the Latent Ewald Summation (LES) method, which infers electrostatic interactions, polarization, and Born effective charges (BECs), just by learning from energy and force training data. Here, we present LES as a standalone library, compatible with any short-range MLIP, and demonstrate its integration with methods such as MACE, NequIP, Allegro, CACE, CHGNet, and UMA. We benchmark LES-enhanced models on distinct systems, including bulk water, polar dipeptides, and gold dimer adsorption on defective substrates, and show that LES not only captures correct electrostatics but also improves accuracy. Additionally, we scale LES to large and chemically diverse data by training MACELES-OFF on the SPICE set containing molecules and clusters, making a universal MLIP with electrostatics for organic systems, including biomolecules. MACELES-OFF is more accurate than its short-range counterpart (MACE-OFF) trained on the same dataset, predicts dipoles and BECs reliably, and has better descriptions of bulk liquids. By enabling efficient long-range electrostatics without directly training on electrical properties, LES paves the way for electrostatic foundation MLIPs.

I. INTRODUCTION

The short-range approximation underlies most established machine learning interatomic potentials (MLIPs) [1, 2], as it enables the decomposition of total energy into individual contributions from local atomic environments, allowing efficient learning and inference. However, it is increasingly recognized that explicitly modeling long-range interactions is essential for systems with significant electrostatics and dielectric response, such as electrochemical interfaces [3], charged molecules [4, 5], and ionic [6] and polar materials [7].

While various approaches have been developed to incorporate long-range interactions into MLIPs [4, 5, 718], many require specialized training labels beyond energy and forces: the fourth-generation high-dimensional neural network potentials (4G-HDNNPs) [9], the charge constraint ACE model [19], and BAMBOO [12] are trained to reproduce atomic partial charges from reference quantum mechanical calculations; the self-consistent field neural network (SCFNN) [10] and the deep potential long-range (DPLR) model [6] learn from the positions of maximally localized Wannier centers (MLWCs) for insulating systems; and PhysNet [8], AIMNET2 [20], and SO3LR [21] utilize the dipole moments of gas-phase molecules. These additional data requirements limit the applicability of such methods to standard datasets containing only atomic positions, energies, forces, and sometimes stresses.

By contrast, a few methods can learn long-range effects directly from standard datasets [4, 5, 7, 14, 15, 22]. These include message-passing schemes such as Ewald message passing [15] and RANGE [18], and Coulombinspired long-range message-passing model FeNNix [23], and descriptor-based approaches like LODE [4, 5, 24] and the density-based long-range descriptor [22]. However, in these methods the long-range contributions are not explicitly related to electrostatics due to charged atoms, making it difficult to extract electrical response properties.

The Latent Ewald Summation (LES) framework [2527] overcomes these limitations by inferring atomic charges and the resulting long-range electrostatics directly from total energy and force data, without training on explicit charge labels. LES can thus be trained on any dataset used for short-ranged MLIPs. The inference of atomic charges not only enhances physical interpretability, but also enables the extraction of polarization (P) and Born effective charges (BECs) [27]. These quantities allow computation of electrical response properties, including dielectric constants, ionic conductivities, ferroelectricity, and infrared (IR) spectra, and also enable MLIP-based molecular dynamics under applied electric fields.

Here, we introduce LES as a universal electrostatics augmentation framework for short-ranged MLIPs. Unlike prior methods that are tightly integrated with specific architectures, LES is implemented as a standalone library that can be combined with a wide range of MLIP models [2841]. We provide open-source patches for several MLIP packages (MACE [33], NequIP [32], CACE [34], MatGL [36], Allegro [42], and UMA [43]), enabling seamless integration. We benchmark and compare LES-augmented models across diverse systems with the baseline short-ranged MLIPs, demonstrating improved accuracy and correct long-range electrostatics through the inference of BECs.

Additionally, to demonstrate scalability to large, chemically diverse datasets, we train MACELES-OFF, an MLIP with explicit long-range electrostatics for organic molecules, using the SPICE dataset [44]. We find that MACELES-OFF outperforms its short-range counterpart (MACE-OFF [45]) and accurately predicts BECs, despite never being trained on charge or polarization data.

II. ARCHITECTURE

The integration of the LES library with a host MLIP is illustrated in Fig. 1. The black box outlines the standard workflow of a short-range MLIP: Given a molecular or periodic system, the base MLIP computes invariant and/or equivariant features for each atom i, based on either local atomic environment descriptors [28, 31, 46] or message-passing architectures [47]. The local invariant descriptors Bi are mapped via a neural network to atomic energies Ei, which are then summed to obtain the total short-range energy Esr=i=1NEi.

Figure 1.

Figure 1.

Schematic illustration of the LES integration with a short-ranged MLIP. The black box shows a standard MLIP workflow, where invariant or equivariant atomic features are computed, the invariant features Bi are mapped to atomic energies Ei and summed to yield the short-range energy Esr. The red box shows the LES module, which either predicts latent charges qiles from atomic features Bi or receives them from the host MLIP. Based on atomic positions ri and simulation cell size L, LES then computes the long-range energy Elr via Ewald summation or pairwise interactions. Optionally, LES also enables the computation of Born effective charges (BECs) as shown in the shaded red region. The blue arrows indicate auto-differentiation operations that are used to compute forces Fi or BECs.

The red box in Fig. 1 shows the LES module that is written in PyTorch, and the black arrows crossing the two boxes indicate communication between LES and the host MLIP. The LES module can either take the Bi features to predict the latent charge of i of each atom, qiles, via a neural network, or this charge prediction part can happen inside the MLIP and be passed to LES. These qles charges are then used to compute the long-range energy contribution Elr. For periodic systems, the Ewald summation is used:

Elr=12ε0V0<k<kc1k2e-σ2k2/2|Sk|2, (1)

with

Sk=i=1Nqileseikri, (2)

where ε0 is the vacuum permittivity, k is the reciprocal wave vector determined by the periodic cell of the system L, V is the cell volume, and σ is a smearing factor (typically chosen to be 1Å). For finite systems, the long-range energy is simply computed using the pairwise direct sum:

Elr=1214πε0i=1Nj=1N1-φrijqiqjrij, (3)

where the complementary error function φ(r)=erfcr2σ.

The Elr computed in LES is returned to the host MLIP package to be added to the total potential energy, E=Esr+Elr. Inside the host MLIP, forces and stresses are obtained through automatic differentiation of the total energy with respect to atomic positions and cell strain, respectively. The training procedure remains unchanged from the original MLIP package, using conventional loss functions for energy, forces, and stresses with user-defined weights.

Optionally, as indicated by the red shaded area in Fig. 1, LES can predict the Born effective charge tensors Zαβ*, where α and β indicate Cartesian directions. The theory is described in Ref. [27]. In brief, for a finite system such as a gas-phase molecule, the BEC for atom i is

Ziαβ*=Pαriβ, (4)

where P=i=1Nqiri is the polarization or the dipole moment. For a homogeneous bulk system under periodic boundary conditions (PBCs),

Ziαβ*=limk0Rexp-ikriαPα(k)riβ, (5)

with Pα(k)=i=1Nεqilesikexpikriα. In practice, we set k to be the smallest reciprocal lattice vector, which is justified by the convergence tests using replicated cells presented in Methods. The high-frequency (electronic) relative permittivity ε is an extra parameter for the system that can be easily obtained from experimental measurements, such as refractive index, or from density-functional perturbation theory (DFPT) calculations [48] with frozen nuclei.

Overall, the computation inside LES is light, and the communication between LES and the host MLIP is minimized. This design allows the LES augmentation to be implemented as a drop-in module that preserves the architecture, training, and inference workflows of the host MLIP package. As long as the host MLIP is implemented in PyTorch and produces per-atom features used for local energy prediction, LES can be patched in with minimal effort.

Thus far, we have integrated LES into several representative MLIPs with distinct architectures, which we briefly describe here. NequIP [32, 49] is a message passing neural network (MPNN) with E(3)-equivariant convolutions, and it typically uses 3–5 message passing layers with a perceptive field that can be as large as about 20Å. Allegro [42, 49] is a scalable, strictly local many-body E(3)-equivariant deep neural network interatomic potential that circumvents the need for atom-centered message passing. The architecture learns scalar and tensorial features associated with directed edges in the atomic graph, enabling prediction of pairwise energies and subsequent reconstruction of per-atom energies. CACE [34] is based on atomic cluster expansion (ACE) [31]. With one layer (nl=1), CACE is a local atomic descriptor-based model employing polynomial expansions of atomic clusters with body order ν in Cartesian coordinates. When ν=2, three-body interactions are captured and (C)ACE has the same expressive power as earlier MLIP methods, including HDNNPs [28], Gaussian Approximation Potentials [29], and DeepMD [6]. MACE [33] is a MPNN that combines equivariant representations with the physically grounded structure of atomic cluster expansions. Thanks to the high-body-order messages, typically only one message passing layer on top of the base ACE layer is needed. MatGL is a graph deep learning library for materials modeling [36]. Instead of using equivariant message passing, MatGL utilizes node and line graph representations to incorporate three-body interactions, such as M3GNet [50] and CHGNet [51]. UMA is an equivariant MPNN that was designed for large datasets such as OMol25[52] and OMat24[53]. UMA extends several ideas from the preceding eSEN [37] architecture, including the eSCN [54] convolution, added envelope functions for smoothness, and non-discrete node representations. UMA furthur includes Mixture of Linear Experts (MoLE) [43] and global embedding of charge, spin, and task inputs. These models span a wide range of architectural paradigms–descriptor-based versus message-passing, invariant versus equivariant representations–and demonstrate that LES can be universally incorporated without requiring architecture-specific modifications.

III. BENCHMARKS

Here, we benchmark the performance of the aforementioned MLIP architectures (MACE, NequIP, CACE, CHGNet, and Allegro) with and without LES augmentation across three distinct systems, exemplifying the key challenges and practical importance of long-range electrostatic interactions. We also tested UMA and UMA-LES, but we found the baseline UMA model to perform substantially worse compared to the other baseline models despite exhaustive hyperparameter tuning, presumably because UMA is designed for large models fitted to large datasets and is not tuned for small datasets.. In the bulk water system, while short-range interactions play an important role in the hydrogen-bond network and structural properties [55], long-range electrostatics are crucial for its dielectric response and interfacial behavior [56]. The polar dipeptides represent gas-phase systems with strong multipole interactions [44]. The Au2-MgO(001) system involves the adsorption of metallic species on oxide substrates, where long-range interactions critically affect adsorption energy and charge redistribution, particularly upon substrate doping [9, 26]. In addition to standard metrics such as energy and force root mean square errors (RMSEs), we also evaluate the prediction of physical observables (BECs and adsorption energies) not in the training set.

A. Water

We trained baseline and LES-augmented MLIPs using the energy and forces from the RPBE-D3 bulk water dataset [57], which contains 604 train and 50 test configurations, each of 64 water molecules. The configurations were generated from MD trajectories at temperatures ranging from 270 K to 420 K at water density under room temperature (997kg/m3), using an on-the-fly learning scheme. For testing LES, we also used the BEC values for an additional 100 snapshots of bulk water at experimental density and room temperature [57], computed also from the RPBE-D3 DFT. We assumed the experimental high-frequency permittivity (ϵ=1.78) of water in the prediction of BECs based on Eqn. (5).

Fig. 2c shows force RMSEs for each MLIP architecture, as the energy RMSE values remain uniformly low (0.1–0.3 meV/atom) except for CHGNet (0.8–0.9 meV/atom) and are thus omitted here. For the baseline MLIPs as indicated by the hollow bars in Fig. 2c, we observe the usual trend typically expected from short-ranged models: Without message-passing layers (nl=1), increasing the cutoffs (r) improves the accuracy (e.g., comparing CACE ν=2r=4.5Å and r=5.5Å models), and so does boosting the body order (CACE r=4.5Åν=2 and ν=3). More layers (nl) effectively increase the perceptive field of the short-ranged MLIPs and reduce the errors, for both equivariant MPNNs (MACE and NequIP) and the invariant CHGNet. Increasing the rotation order () in E(3) equivariant representations of messages also helps the accuracy, as seen in MACE, NequIP and Allegro models.

Figure 2.

Figure 2.

Benchmark short-ranged and LES-augmented MLIPs for bulk water. a: A representative snapshot of bulk water. b: A parity plot comparing Born effective charge (BEC) tensors Z* for 100 bulk water from LES-augmented NequIP (r=4.5Å model, nl=3,=2) against RPBE-D3 DFT (Ref. [57]). The main panel compares the diagonal elements of BEC (Zαα*), and the inset shows the off-diagonal elements Zαβ* with αβ. c: The force root mean square errors (RMSEs) for baseline MLIPs (hollow bars), and with LES augmentation (solid bars). d: The RMSEs for BEC predictions using LES and corresponding R2 coefficients. The cutoff r, the number of layers nl, the order of irreducible representations (rotation order) in message passing, and body order ν are indicated for each MLIP, as applicable.

The solid bars in Fig. 2c show the force RMSEs for the LES-augmented models. Comparing the solid bars with the hollow ones, LES reduces the errors for all models, while the extent of improvement is somewhat dependent on the architecture of the short-ranged baseline. For this bulk water dataset, LES benefits most of the MLIPs with a smaller effective perceptive field. That is, the MACE and the CACE models without message-passing (nl=1) see the largest force error reductions, and the NequIP and CHGNet with several message-passing layers have modest reductions. Similarly, the non-message-passing architecture Allegro also shows a meaningfully large reduction, consistent with this trend, while its number of layers (nl) corresponds to the body order (e.g., nl=3 represents five-body tensor features) rather than message-passing depth. The other hyperparameters of the base MLIPs, ν and , do not seem to have a strong influence on the efficacy of LES.

Moreover, we ask the question: Is LES able to capture the correct electrostatic interactions and predict physical atomic charges? This is answered via the BEC benchmark shown in Fig. 2d, which presents the RMSEs of the BECs compared to the DFT calculated values. Importantly, for all MLIPs, regardless of model architecture, hyperparameters, or effective cutoffs, LES can accurately predict the BECs. Fig. 2b shows the parity plot for one of the models, which illustrates the quality of the agreement with the DFT ground truth. We do not find any obvious correlation between the accuracy of the BEC prediction and the baseline MLIP architecture, nor with the force RMSE errors. However, there is a weak inverse correlation with the reduction in force RMSE errors (see Fig. 11 in Methods). This suggests that LES is a robust approach to infer BECs, regardless of the underlying model performance. The BECs are not only sensitive indicators of the correct electrostatics, but are also useful in atomistic simulations for electrical response properties. For instance, in Ref. [27], we showed that such predictions of BECs can enable accurate prediction of water IR spectra under zero or finite external electric fields.

Figure 11.

Figure 11.

Correlation analysis between force and Born effective charge (BEC) RMSEs for LES-augmented MLIPs. a, b: Correlation between force RMSE and BEC RMSE for water (a) and dipeptide (b) systems. c, d: Correlation between force RMSE reduction after LES augmentation and BEC RMSE for water (c) and dipeptide (d) systems.

B. Dipeptides

We trained baseline and LES-augmented MLIPs using energy and forces from a dataset of polar dipeptide conformers selected from SPICE [26, 44], which contains 550 configurations (90% train/10% test split) For testing LES, we computed BEC values at the ωB97-mD3(BJ)/Def2SVP level of theory. As these systems are in a vacuum, the high-frequency permittivity is ϵ=1, and one can use Eqn. (4) directly when predicting BECs using LES.

Figs. 3c and d show the energy and force RMSE values for each MLIP architecture. For the baseline MLIPs indicated by the hollow bars, we again observe the usual trends that increasing body order (CACE r=4.5Åν=2 and ν=3), and the number of layers (CHGNet r=4.5Ånl=3 and nl=4) lead to improved accuracy. Additionally, increasing improves MACE, NequIP and Allegro models.

Figure 3.

Figure 3.

Benchmark short-ranged and LES-augmented MLIPs for dipeptide. a: A representative snapshot of a dipeptide conformer composed of arginine and aspartic acid from the SPICE dataset [44]. b: A parity plot comparing Born effective charge (BEC) tensors for 55 dipeptide configurations [44] from LES-augmented NequIP (r=4.5Å,nl=3,=2) model against DFT. The main panel compares the diagonal elements of BEC (Zαα*), and the inset shows the off-diagonal elements (Zαβ* with αβ) c, d: The energy (c) and force (d) root mean square errors (RMSEs) for baseline MLIPs (hollow bars), and with LES augmentation (solid bars). e: The RMSEs for BEC predictions using LES and corresponding R2 coefficients. The cutoff r, the number of layers nl, the order of irreducible representations (rotation order) of messages, and body order ν are indicated for each MLIP, as applicable.

The solid bars in Figs. 3c and d represent the energy and force RMSEs for the LES-augmented models. Comparing the solid and hollow bars, we find again that LES augmentation consistently reduces both energy and force prediction errors across all baseline MLIPs, regardless of their specifics: hyperparameters such as r and ν do not appear to strongly influence the effectiveness of LES, nor do the perceptive fields of baseline models. This implies that LES provides an improved molecular representation on top of message passing.

As is seen in Fig. 3e, LES-augmented models all show excellent agreement with DFT reference values for BEC predictions (up to R2=0.97). As an illustration, Fig. 3b shows a parity plot. We also investigated whether the quality of the BEC predictions is correlated with force RMSE values or RMSE reduction, as shown in Figs. 11b and d. As is seen, there are no clear-cut correlations between BEC and force or energy RMSE improvement, showing that LES remains a robust way to compute charge regardless of the relative performance of the baseline model. As shown in our previous study, the ability of LES to capture electrostatic interactions also extends to dipole moments, quadrupole moments, and IR spectra [26]. Electrostatics strongly affect peptide backbone conformations in dipeptides [58], which suggests LES augmentation can be helpful for protein structure, folding, binding, and other biological functions.

C. Au2 on MgO(001)

We trained baseline and LES-augmented MLIPs using energy and forces from an Au2-MgO(001) dataset from Ko et al. [9], consisting of 5000 configurations (90% train/10% test split). The configurations consist of a gold dimer adsorbed on an MgO(001) surface with two adsorption geometries: an upright non-wetting and a parallel wetting configuration. In some configurations, three Al dopant atoms were introduced in the fifth subsurface layer (Fig. 4a). Despite their distance of over 10Å away from the gold dimer, the dopant atoms significantly influence the electronic structure and the energetics between the wetting and non-wetting configurations.

Figure 4.

Figure 4.

Benchmark short-ranged and LES-augmented MLIPs for Au2 on MgO(001). a: Representative configurations for wetting (top) or non-wetting (bottom) Au2 adsorbed on doped (top) or undoped (bottom) MgO(001) surface. b, c: The energy (b) and force (c) root mean square errors (RMSEs) for baseline MLIPs (hollow bars), and with LES augmentation (solid bars). d: The energy difference (Ewetting-Enon–wetting) in eV for doped (top) and undoped (bottom) substrates. Baseline (hollow bars) and LES-augmented models (solid bars) compared with reference DFT results (gray bars and horizontal dashed lines). The cutoff r, the number of layers nl, the order of irreducible representations (rotation order) of messages, and body order ν are indicated for each MLIP, as applicable.

Figs. 4b and c present the energy and force RMSEs for each MLIP architecture. For all baseline MLIP models indicated by the hollow bars, increasing the perceptive field by either raising the cutoff radius or increasing the number of layers leads to significant improvements in both energy and force accuracy, for both the equivariant MPNNs (MACE, NequIP) and the invariant CHGNet. When the perceptive field is smaller than the distance between the dopants and the adsorbates, the MLIPs perform poorly.

The solid bars in Figs. 4b and c demonstrate that the LES augmentation rather consistently improves energy and force predictions across all models, with particularly large gains for models with a smaller perceptive field. For example, the force RMSE reductions for MACE r=5.5Ånl=1, NequIP r=4Ånl=2=1, CHGNet r=4.5Ånl=1, CACE r=5.5Ånl=1ν=3, CACE r=5.5Ånl=1ν=2, and Allegro r=5.5Ånl=2=1 are about 80%.

Fig. 4d shows the prediction of a physical observable, the energy difference (Ewetting-Enon–wetting) between wetting and non-wetting configurations for doped and undoped substrates. This is computed by performing geometry optimizations of the positions of the gold atoms, with the substrate fixed, for both doped and undoped surfaces. For the undoped MgO substrate, the non-wetting configuration is more stable, while the Al-doping stabilizes the wetting configuration. As shown, the baseline MLIPs with small perceptive fields of less than about 9Å struggle to distinguish between undoped and Aldoped structures. Either increasing the perceptive field or adding LES significantly enhances the accuracy of predicting this energy difference, closely matching DFT reference values of 934.8 meV for undoped and −66.9 meV for doped surfaces. Thus, although LES may not further improve the prediction for already highly accurate models, it assures that long-range effects associated with doping and adsorption are captured in all cases.

Notably, the CACE model with three-body interactions (r=5.5Ånl=1ν=2) has the same body order and expressiveness as the HDNNPs [28]. Ko et al. [9] showed that the SR HDNNP fails to distinguish the energetics of the doped and undoped systems, while the 4G-HDNNP, which was trained on DFT partial charges and uses a charge equilibration scheme, can reproduce the physical behavior. Our observation is consistent with theirs, except that no training on DFT charges nor charge equilibration is needed.

IV. A LONG-RANGE TRANSFERABLE MLIP FOR ORGANICS

In the above, we benchmarked LES on specific systems and relatively small datasets. Here, we explore whether the long-range correction scheme works at scale, for large and chemically diverse datasets. This is not only relevant for validating the scaling capability of the method, but is practically important for constructing emerging “foundation models” [59] that work across the periodic table.

The MACE-OFF23 (small (S), medium (M), and large (L)) [45] is a recent set of short-range transferable machine learning force fields for organic molecules, which are based on the MACE architecture and trained on finite organic systems: small molecules of up to 50 atoms from about 85% of the SPICE dataset version 1 [44] with a neutral formal charge and containing ten chemical elements H, C, N, O, F, P, S, Cl, Br, and I, as well as 50–90 atom molecules randomly selected from the QMugs dataset [60]. The dataset contains the energies and forces computed at the ωB97M-D3(BJ)/def2-TZVPPD DFT level of accuracy. We trained the MACELES-OFF model based on the same training/validation split of MACE-OFF23.

We note that our main purpose here is to showcase the effect of the LES augmentation compared to the baseline model, rather than fitting the most accurate model possible. The MACELES-OFF model we trained used hyperparameters of cutoff radius r=4.5Å, chemical channels k=192, the order of the equivariant features =1, and float32 precision. The comparison of these hyperparameters with the original MACE-OFF23 models is presented in Table II. Compared to the MACE-OFF23(S), the MACELES-OFF has the same r but higher k and . Thanks to its lower precision as well as little overhead from LES, the MACELES-OFF is about as fast as the MACE-OFF23(S), and much faster than the MACE-OFF23(M) and MACE-OFF23(L) models, as shown in the timing benchmark in Fig. 13 of Methods.

Table II.

Hyperparameters of the MACE-OFF and MACELES-OFF models. All MACE-OFF models (23(S), 23(M), 23(L), and 24(M)) are taken from Ref. [45]. 24(M) was trained on the updated SPICE version 2 [77].

23(S) 23(M) 23(M)b 23(L) 24(M) LES
Cutoff r(Å) 4.5 5.0 6.0 5.0 6.0 4.5
Channels k 96 128 128 192 128 192
Irrep order 0 1 1 2 1 1
SPICE version 1 1 1 1 2 1
Precision float64 float64 float64 float64 float64 float32

Figure 13.

Figure 13.

Computational performance benchmarks of molecular dynamics (MD) simulations using the MACELES-OFF and the MACE-OFF23 [45] models. The timing of MD simulations for bulk liquid water with varying system sizes was performed on an NVIDIA L40S GPU. Benchmarks were performed with both single-precision (float32) and double-precision (float64) models using ASE and OpenMM implementations.

In Fig. 5, we compare the RMSE values for energy and force for the three original MACE-OFF23 (S, M, L) and MACELES-OFF, on different subsets of test configurations (representative snapshots in Fig. 6a). Overall, the MACELES-OFF shows RMSEs comparable to the MACE-OFF23(L) that uses a larger cutoff r=5Å and higher =2 (see Table II). Compared to the MACE-OFF23(S) model that shares the same cutoff, the errors are about halved. Amongst all the subsets, the MACELES-OFF shows particularly better accuracy for water clusters, solvated amino acids, and dipeptides, which may indicate that the model is well-suited for simulating biological systems.

Figure 5.

Figure 5.

Comparison of test set root mean square errors (RMSEs) for energy and forces of the MACE-OFF23 (S, M, L) and MACELES-OFF models for organic molecules, with the underlying DFT reference values [45]. Exact values are provided in Table III.

Figure 6.

Figure 6.

Assessment of dipole moments and Born effective charges (BECs) predicted by the MACELES-OFF for subsets of the MACE-OFF test set. a: Representative configurations from each subset of the MACE-OFF test set. b: Parity plots comparing predicted dipole moment components (μ) against reference DFT values. c: Parity plots comparing predicted BEC tensors against DFT reference. Data points are colored by atomic number, with diagonal BEC components (Zαα*) shown in the main plots and off-diagonal Zαβ* components shown in inset plots.

To test whether LES is able to extract the correct electrostatics, we investigate if the model is able to infer the dipole moments μ (same as the polarization of finite systems) and BECs. For about 50 configurations in each of the test subsets, we calculated reference BECs and dipole moments at the ωB97m-D3(BJ)/Def2SVP level of theory (calculation details in Methods). Fig. 6b shows good agreement between the reference and the predicted dipole moments μ. Fig. 6c compares the DFT and the BECs computed by LES via taking the derivative of the predicted μ as in Eqn. (4), showing broad agreement in both the on- and off-diagonal components over the broad range of elements and chemical species contained in the test set. These agreements thus demonstrate that accurate inference of dipole moments and BECs is achievable in foundation-like models trained only on energy and forces.

a. Water

A key requirement for organic force fields is the accurate description of bulk water. Note that the training set only contains small, non-periodic water clusters; the application to the simulations of bulk liquid is in itself a generalization task.

We performed equilibrium NVT molecular dynamics simulations of bulk water at 300 K and experimental density, with further details provided in the Methods section. Fig. 7a demonstrates the capability of the MACELES-OFF model to accurately capture the radial distribution function (RDF) of water, and shows a result comparable to those from the previous MACE-OFF23 models (S, M) and experimental measurements [61].

Figure 7.

Figure 7.

Bulk water properties predicted by the MACELES-OFF compared to the MACE-OFF23 and experimental data. a: Oxygen-Oxygen radial distribution functions in bulk water from the MACE-OFF23 models (S, M) and the MACELES-OFF model compared with experimental value (Ref. [61]). b: The infrared (IR) absorption spectra of bulk liquid water computed using the MACELES-OFF (red). For comparison, we also show the experimental IR spectrum [62] (gray shading), the classical MD result from MACE-OFF23(S) (solid blue), and the result with quantum nuclear corrections to MACE-OFF23(S) (dashed blue), as reported in Ref. [45]. c: Density isobar of liquid water at 1 atm. Both the MACELES-OFF (red circles) and the MACE-OFF23(M) (green squares) models show similar isobaric density characteristics for liquid water at 1 atm, with experimental data [65] included for reference.

We computed an IR spectrum from the MD trajectory using predicted BECs, as illustrated in Fig. 7b, with additional details given in the Methods section. For comparison, we also included the IR spectrum predicted using MACE-OFF23(S), computed from the autocorrelation function of the time derivative of the total dipole moment predicted using a separate MACE-OFF23-μ model along classical MD trajectories [45], as well as experimental results [62]. The intramolecular bending mode (1640cm-1), the intermolecular low-frequency libration band 650cm-1, and the hydrogen-bond translational stretching mode (200cm-1) are all well-captured by the MACELES-OFF, whereas the MACE-OFF23(S) classical MD result reproduces the band shapes but not the intensities. The predicted intramolecular vibrational mode (OH stretching band 3400cm-1) is blue-shifted relative to the experiment, which is due to our use of classical MD without nuclear quantum effects (NQEs) [63]. To illustrate the impact of NQEs, we additionally show the IR spectrum obtained from the MACE-OFF23(S) with path-integral MD (PIMD) simulations within the PIGS approximation [64], as reported in Ref. [45]. Incorporating NQEs brings the MACE-OFF23(S) frequencies into close agreement with experiment, especially by correcting the blue shift in the OH stretching region.

We simulated the temperature dependence of liquid water density based on NPT simulations conducted at 1 atm, as shown in Fig. 7c. Comparative analysis demonstrates that MACELES-OFF achieves superior density prediction, exhibiting about 10% overestimation of liquid water density compared to the original 18% overestimation reported by the MACE-OFF23(M) model [45]. This improvement highlights the importance of long-range dipole-dipole interactions in predicting bulk properties such as density. We also note that the MACE-OFF model can predict water density closer to experimental values by further increasing the cutoff to 6Å, which corresponds to a perceptive field of 12Å.

b. Molecular liquids

We investigated the performance of MACELES-OFF in predicting condensed-phase properties such as densities (ρliq) and enthalpy of vaporization (ΔvapH) under ambient conditions (298 K and 1 atm), as presented in Fig. 8. Reproducing such properties was observed to present challenges for short-ranged MLIPs [45, 67]. We selected a comprehensive test set of 39 molecular liquids with boiling points greater than 320 K, referencing the MACEOFF benchmark [45]. This set encompasses diverse chemical classes highly relevant to chemistry and biology, including alcohols, amines, amides, ethers, aromatics/hydrocarbons, ketones/aldehydes, thiols/sulfides, and other functionalized compounds.

Figure 8.

Figure 8.

Predicted densities and heats of vaporization for diverse molecular liquids under ambient conditions (298 K and 1 atm) using the MACELES-OFF. a: Density (ρliq) predictions from the MACELES-OFF (red circles) and the MACE-OFF23(M) (green squares), compared with experimental values [66]. b: Heats of vaporization (ΔvapH) predictions from the MACELES-OFF (red circles) and the MACE-OFF23(M) (green squares), compared with experimental values [66].

As shown in Fig. 8, the MACELES-OFF model demonstrates superior accuracy in predicting liquid density and enthalpy of vaporization compared to the MACE-OFF23(M). For liquid density, the MACELES-OFF achieves an MAE of 0.04g/cm3 and RMSE of 0.05g/cm3, which represents approximately half the error values of the MACE-OFF23(M) (MAE: 0.09g/cm3, RMSE: 0.15g/cm3). These low error values, coupled with a stronger correlation with experimental data (R=0.96) compared to the MACE-OFF23(M) model (R=0.89), indicate that the MACELES-OFF accurately learned both short- and long-range intermolecular interactions.

Similarly, regarding the calculations of ΔvapH, the MACELES-OFF model yields a MAE of 1.32 kcal/mol and RMSE of 1.62 kcal/mol with a correlation of R=0.84 and thus outperforms the MACE-OFF23(M) model that has an MAE (2.18 kcal/mol) and RMSE (2.53 kcal/mol) with R=0.87. Moreover, while the MACE-OFF23(M) model exhibits a systematic ΔvapH offset of about 2 kcal/mol, which is largely eliminated in the MACELES-OFF prediction.

c. Simulations of biological molecules

To probe the influence of the long-range interactions in MACELES-OFF on protein-protein and protein-water interactions, we studied two biological systems: solvated alanine dipeptide and the 1FSV protein [68].

For the alanine dipeptide in water, we computed the free energy surface (FES) using well-tempered metadynamics [69] simulations in explicit solvent (see Methods Section), as shown in Fig. 9. The AMBER ff99SB-ILDN force field exhibits an FES with four local minima, consistent with previous studies [23, 70]: antiparallel β-sheet (aβ), polyproline II (PPII), right-handed α-helix (α-H(R)), and left-handed α-helix (α-H(L)). In comparison, MACELES-OFF displays the deepest minimum at α-H(R). The free energy difference between α-H(R) and the second minimum (aβ) is approximately 1.5 kcal/mol, while the difference with PPII is around 3.0 kcal/mol, consistent with the relative ordering of minima and similar to the relative free energies reported for MP2/ccpVTZ level with implicit solvent [71]. Compared to the recent FENNIX-Bio1 model [23] that includes long-range interactions, nuclear quantum effects, and is trained on the high-level SPICE2(+)-ccECP dataset, MACELES-OFF yields a very similar overall shape of the FES. Interestingly, the right-handed α-helix region (αD in Fig. 9b), does not appear as a clear minimum but rather as a shallow valley on the FES by both AMBER ff99SB-ILDN or AMBER99 [23], likely due to insufficient treatment of water-protein interactions, an issue well tackled by MACELES-OFF.

Figure 9.

Figure 9.

Torsional free energy surfaces (FES) of solvated alanine dipeptide computed at 310 K using different potentials. a: FES based on AMBER ff99SB/TIP3P. b: FES based on MACELES-OFF. Four local minima are distinguished in the landscape and they correspond to distinct conformations defined by the central ϕ and ψ backbone angles: antiparallel β-sheet (aβ), polyproline II (PPII), right-handed α-helix (α-H(R)), and left-handed α-helix structures (α-H(L))). In addition, compared to AMBER ff99SB/TIP3P, MACELES-OFF identifies another local minimum that corresponds to the right-handed α-helix region (αD).

For the 1FSV protein, we performed 2 ns MD simulations in vacuum for all three models, starting from the crystallographic PDB structure. The 1FSV protein poses a significant challenge due to its 10 positively and 5 negatively charged groups. As shown in Fig. 10, both MACELES-OFF and MACE-OFF24(M) began folding the protein into a more compact conformation, characteristic of gas-phase behavior. However, their trajectories diverged over time. MACELES-OFF exhibited more stable dynamics, remaining close to AMBER ff99SB-ILDN [72], with a root-mean-square deviation (RMSD) of 5.14Å with respect to the 1FSV PDB structure, just 0.085Å lower than AMBER’s 5.22Å. In contrast, MACE-OFF24(M) showed continued RMSD drift and over-compaction, reaching 4.43Å.

Figure 10.

Figure 10.

The root mean square deviation (RMSD) of the 1FSV protein during a 2 ns gas-phase NVT MD simulation at 310 K. Time evolution of RMSD and protein structures compared to the initial PDB structure for AMBER ff99SB (blue), MACEOFF-24(M) (orange), and MACELES-OFF (red).

Further structural analysis confirmed these trends through the calculation of salt bridges and hydrogen bond networks. MACE-OFF24(M) exhibited the highest number of salt bridges (159, +35% vs. PDB) and hydrogen bonds (286, +55%), indicating an over-collapsed structure. By comparison, MACELES-OFF formed 149 salt bridges and 263 hydrogen bonds, values more consistent with AMBER (143 and 258, respectively).

Overall, the results above suggest that MACELES-OFF can provide a good description of the FES of a small peptide solvated in water. Comparing MACELES-OFF and MACEOFF for the folding of 1FSV in vacuum, long-range electrostatics play a critical role in governing the dynamics and structural stability of the protein. These findings underscore the importance of including long-range interactions in simulations of biomolecular systems.

V. DISCUSSION

We present LES as a “universal” augmentation framework for adding long-range electrostatics to short-range MLIPs. It is universal in three ways: First, LES can work together with many short-ranged MLIPs in an architecture-agnostic way as long as there are local invariant features for atoms. We patched LES to MACE, NequIP, CACE, Allegro, UMA and MatGL, and it is our ongoing commitment to include LES in other MLIPs, making it more widely accessible. Interestingly, LES works well with the CACE models with three-body interactions ν=2, which are mathematically equivalent to the earlier generation of MLIPs (e.g., HDNNPs [28], Gaussian Approximation Potentials [29], and DeepMD [6]). This suggests that LES will also complement well those lower-body-order MLIP methods. The LES-augmented HDNNPs would indeed be quite similar to the third-generation HDNNPs (3G-HDNNPs) [9], except that not explicitly trained on partial DFT charges. The fact that 3G-HDNNPs have poor performance in many systems [9] may be attributed to the explicit charge training, rather than inferring from energy and forces.

Second, the LES augmentation is useful across a wide range of systems. Across the three benchmarks (bulk water, dipeptides, and Au2 on MgO(001)), the inclusion of LES yields consistent, architecture-agnostic improvements. Even when the reduction of energy and force RMSEs is modest, LES significantly improves physical observables. For example, in the water and dipeptide set, LES achieves accurate BEC predictions despite training exclusively on energy and forces. BECs are not only well-defined physical quantities that serve as sensitive probes of electrostatics, but are also critical in calculating a range of electrical response properties such as dielectric constants, ionic conductivities, and dipole correlation functions [27].

Third, LES can be scaled to large datasets with chemical diversity and can be used to train universal MLIPs. As a demonstration, the MACELES-OFF model shows better accuracy compared to the short-ranged baseline models, encodes the correct electrostatics, and exhibits better transferability. For instance, the MACE-OFF23 models show an amazingly wide range of accurate predictions across small molecules, biological systems, and molecular crystals, but they are less accurate in predicting the densities and heats of vaporization of molecular liquids. Such shortcomings are attributed to the lack of long-range interactions [45]. Indeed, incorporating LES effectively solved such issues and provides much improved predictions for the liquids. In addition, electrical response properties such as IR spectra are readily available.

Additionally, using LES only requires standard energy and force labels. This is an advantage for scaling up the training of universal MLIPs with large-scale datasets, as most such datasets [52, 53, 73, 74], especially of extended systems, do not contain explicit polarization or charge information. However, the dipole moments of gas-phase molecules are more readily available, and in principle, one can include these in the training of partial charges under the LES framework, by adding a dipole loss function as in AIMNET2 [20], SO3LR [21], and MPNICE [75]. Towards true foundation models with full electrostatic physics, one also needs to address the issue of the field-dependence of atomic charges, which LES currently does not handle, but we have ongoing work along this direction.

To conclude, by incorporating long-range electrostatics and without training on labels other than energy, forces (and stresses), the LES method addresses a core limitation of current MLIPs. The LES library works as a drop-in patch for many short-ranged MLIPs, and demonstrates reliable and consistent performance across a wide range of architectures, chemical system types, and data sizes.

VI. METHODS

A. Implementation

LES is implemented in PyTorch [76] and is easily compatible with MLIP packages also implemented in Py-Torch. In CACE [34], we implemented a LESwrapper, and we note that CACE has its own native LES implementation [25] which yields consistent results. In MACE [33], we added a MACELES model which can be used in the same way as the original MACE model, and all training and evaluation procedures stay the same. For NequIP [32] and Allegro [42], we developed an extension package, NequIP-LES, leveraging the modular structure of the redesigned NequIP framework [49]. One can specify the baseline model through the base_model key and enable BEC inference using the compute_bec flag. The framework automatically augments the baseline architecture by adding charge prediction and Ewald summation modules, while keeping all training and evaluation procedures unchanged. After training, BECs can be computed using either the callback or the model modifier options. In MatGL [36], the inference and training are treated in different modules. We extended the Potential module with calc_BEC = True to include LES and autograd to compute the BECs for the inference. We extended the PotentialLightningModule with include_long_range = True to call Potential with long-range interactions for the training. For UMA [43], we implemented LES wrappers to handle translation between FAIRChem’s data structures and the LES code. Inference, in particular BECs, is handled via an ad hoc module for computing and saving BEC values. These wrappers are used in custom energy-force output heads in the UMA implementation. UMA-LES can be used by simply changing the output head and specifying LES parameters to the model. Training and inference without BECs are entirely unchanged from standard FAIRChem procedures.

B. Details for benchmarks

The training scripts with hyperparameters as well as the trained models for the water, dipeptides, and Au2 on MgO(001) datasets are included in the SI repository. For the LES augmentation part, we always use the default parameters: σ=1Å,dl=2Å (which corresponds to a k-point cutoff of kc=π in the Ewald summation).

Fig. 11 shows the relation between force RMSE and BEC RMSE for LES-augmented MLIPs (panels a and b), and the relation between the reduction in force RMSE after LES augmentation and the corresponding BEC RMSE (panels c and d), for water and dipeptide systems.

Table I summarizes the energy differences (see Fig. 4) between wetting and non-wetting configurations for doped and undoped substrates in the Au2-MgO(001) system.

Table I.

Energy difference (Ewetting-Enon–wetting) in meV between the wetting and non-wetting configurations for doped and undoped substrates.

MLIP Hyperparameters Undoped Doped
MACE r=5.5Å,nl=1 419.0 419.0
MACE LES r=5.5Å,nl=1 919.3 −56.5
NequIP r=4.0Å,nl=2,=1 390.2 390.3
NequIP LES r=4.0Ånl=2,=1 973.3 −45.1
CHGNet r=4.5Å,nl=1 411.9 411.9
CHGNet LES r=4.5Å,nl=1 925.5 −103.5
CACE r=5.5Å,nl=1,v=2 421.5 421.4
CACE LES r=5.5Å,nl=1,v=2 931.7 −72.2
CACE r=5.5Å,nl=1,v=3 430.9 430.9
CACE LES r=5.5Å,nl=1,v=3 930.9 −71.1
Allegro r=5.5Å,nl=2,=1 429.3 429.6
Allegro LES r=5.5Å,nl=2,=1 934.7 −69.4
DFT reference 934.8 −66.9

Fig. 12a compares the DFT tensors from DFT and those predicted by the LES-augmented CACE model (r=4.5Ånl=1,ν=3) for the same 100 water configurations used in Fig. 2b. Fig. 12b repeats the same comparison for the corresponding 2 × 2 × 2 replicated supercells to conduct a convergence test for k in Eqn. 5. The inferred BECs are nearly identical, with the supercell showing slightly improved agreement with DFT (R2 coefficients increasing from 0.976 to 0.985 over all tensor components). These results confirm that LES yields convergent and reliable BEC predictions under periodic boundary conditions.

Figure 12.

Figure 12.

Parity plots comparing Born effective charge (BEC) tensors Z* from the LES-augmented CACE (r=4.5Å,nl=1,ν=3) with RPBE-D3 DFT (Ref. [57]) for 100 bulk water configurations. Main panels compare the diagonal elements of BEC Zαα*, and the inset shows the off-diagonal elements (Zαβ* with αβ). a: Primitive cell; b: 2 × 2 × 2 replicated supercell.

C. Details on training the MACELES-OFF model

The training scripts as well as the trained models for the MACELES-OFF are included in the SI repository. The hyperparameters are summarized and compared with the original MACE-OFF models in Table II.

Table III summarizes the test errors for each dataset. For the PubChem dataset, 13 outlier configurations were removed from a total of 34,093 test structures in accordance with the MACE-OFF paper [45].

Table III.

Test set root mean square errors (RMSEs) for energy and forces of the MACE-OFF23 (S, M, L) and MACELES-OFF models for organic molecules compared to the underlying DFT reference data [45].

PubChem DES370K Monomers DES370K Dimers Dipeptides Solvated Amino Acids Water QMugs
MAE E (meV/at) 23S 1.41 1.04 0.98 0.84 1.60 1.67 1.03
23M 0.91 0.63 0.58 0.52 1.21 0.76 0.69
23L 0.88 0.59 0.54 0.42 0.98 0.83 0.45
LES 0.71 0.54 0.47 0.52 0.82 0.69 0.76
F (meV/Å) 23S 35.68 17.63 16.31 25.07 38.56 28.53 41.45
23M 20.57 9.36 9.02 14.27 23.26 15.27 23.58
23L 14.75 6.58 6.62 10.19 19.43 13.57 16.93
LES 17.24 7.65 7.48 11.46 18.37 12.54 20.11
RMSE E (meV/at) 23S 2.74 1.47 1.51 1.26 1.98 2.07 1.28
23M 2.02 0.90 0.91 0.85 1.55 0.99 0.89
23L 2.48 0.84 0.87 0.70 1.32 0.99 0.58
LES 1.45 0.72 0.72 0.69 1.01 0.90 0.94
F (meV/Å) 23S 61.83 26.15 28.48 36.97 53.55 39.33 62.46
23M 40.51 14.31 16.81 22.25 32.19 21.40 36.73
23L 33.34 10.27 12.81 16.19 26.91 18.78 27.17
LES 35.26 11.70 14.42 17.62 25.31 17.08 32.57

Reference Born effective charges (BECs) and dipole moments for the MACELES-OFF were computed in PySCF [78] using the ωB97m-D3(BJ) functional in the Def2SVP basis set. Code for calculating BECs can be found in the infrared module of the properties extension of PySCF https://github.com/pyscf/properties.git.

D. Details on MD simulations using the MACELES-OFF model

a. Water

We performed equilibrium NVT simulations of bulk water using the MACE-OFF23 (S, M) and MACELES-OFF models in ASE [79] at a density of 0.997g/cm3 and 300 K for a system of 64 water molecules, employing the Nosé-Hoover thermostat for Fig. 7a and b. Each MD simulation was run for 200,000 steps with a timestep of 0.25 fs, corresponding to 50 ps.

The time-dependent current of polarization of the system is given by J(t)=i=1NZi*(t)vi(t), where Zi*(t) is the BEC tensor and vi(t) is the velocity of an atom i. The IR spectra are obtained via the Fourier transform of the current-current autocorrelation function,

Iω0TdtJ0Jte-iωt. (6)

The IR spectrum for Fig. 7b was calculated from the MACELES-OFF MD trajectory, and the resulting raw intensity was smoothed by convolving the spectrum with a Gaussian kernel.

We simulated the density isobar of liquid water at 1 atm by investigating four temperatures: 275 K, 295 K, 315 K, and 336 K. These NPT simulations employed a system of 64 water molecules, with a 1.0 fs timestep over 300 ps of simulation time, utilizing the Nosé-Hoover thermostat and Berendsen barostat implemented in ASE.

b. Molecular liquids

For the preparation of MD simulations in the liquid phase, SMILES strings for the organic molecules were first obtained from the PubChem database [80] and subsequently converted to PDB format using the Python package RDKit [81]. The initial molecular configurations were generated at 60% of experimental density using the GROMACS command-line interface [82], followed by structural pre-optimization with the General AMBER Force Field (GAFF) [83]. A two-stage optimization protocol was then implemented: geometry optimization using the MACELES-OFF model with the BFGS algorithm in ASE, followed by NVT simulations for 5,000 steps using a Nosé-Hoover thermostat with a timestep of 1.0 fs to achieve structural relaxation. Production NPT simulations were performed for 300 ps using a 1.0 fs timestep, using the same Nosé-Hoover thermostat coupled with the Martyna-Tobias-Klein (MTK) barostat [84]. Final density and potential energy values were computed by averaging over the last 150 ps to ensure proper equilibration. To compute the enthalpy of vaporization, potential energies for the gas phase were obtained based on simulations of isolated molecules in a non-periodic box. The ideal gas approximation was employed to account for pressure effects in the gas phase.

It should also be noted the 1,1,2,2-tetrachloroethane system vaporized under NPT simulation using MACELES-OFF, although it remained liquid using MACE-OFF23(M). This might be due to the difficulty of preparing the initial condition of this system. Despite this exception, the MACELES-OFF demonstrated successful simulation of another chlorinated hydrocarbon, 1,4-dichlorobutane, a structurally analogous compound that was an outlier in the MACE-OFF23(M) predictions of ΔvapH. Furthermore, the MACELES-OFF accurately predicted both the density and enthalpy of vaporization for chlorinated aniline, thereby further proving its capacity to model intra- and intermolecular interactions involving chlorine atoms.

Simulated density, enthalpy of vaporization, and optimized initial configurations for both liquid and gas phases are provided in the SI repository.

c. Simulations of biological molecules

The PDB structure of alanine dipeptide was solvated in a cubic box of 28Å using the addSolvent method in OpenMM, resulting in a system of 2,050 atoms. The system was modeled using the AMBER ff99SB-ILDN force field [72] and the TIP3P water model [85]. The structure was first geometry-optimized with a tolerance of 10 kJ/mol/nm, followed by 50 ps of NPT equilibration at 310 K and 1.013 bar using a Monte Carlo barostat and Langevin thermostat, with a timestep of 1.0 fs and a friction coefficient of 1.0ps1. Production NVT simulations with well-tempered metadynamics were performed for 2 ns for MACELES-OFF and 5 ns for AMBER ff99SB-ILDN/TIP3P along the central ϕ and ψ backbone angles using ASE patched with the PLUMED code (version 2.9) [86, 87].

For both collective variables (ϕ and ψ), we employed a Gaussian height of 2.5 kJ/mol and a bias factor of 30. Gaussian potentials were deposited on the energy landscape every 500 steps, and their width was adapted on the fly based on the local diffusivity [88]. The FES was computed from reweighting.

For the 1FSV protein, the PDB structure [68] was placed in a cubic simulation box at a density of 0.6g/cm3 to avoid interatomic overlaps and to maintain periodic boundary conditions as required by the LES framework. Geometry optimization was performed in three successive stages using the FIRE optimizer [89], with gradually decreasing convergence criteria down to a maximum force of 0.2eV/Å. We then ran 2 ns of MD production in the NVT ensemble at 310 K using a 1 fs timestep and a friction coefficient of 0.1ps-1, starting from the PDB crystallographic structure.

Salt bridge analysis was performed by counting the number of interactions within a 4.0Å cutoff between the side-chain oxygen atoms of acidic residues (e.g., ASP) and the nitrogen atoms of basic residues (e.g., ARG). Similarly, hydrogen bonds were assessed by counting interactions within a 3.5Å cutoff between donor and acceptor atoms.

d. Timing

Fig. 13 compares the inference speed of the MACE-OFF23 (S, M, L) models and the MACELES-OFF model for MD simulations of liquid water using both the ASE and OpenMM implementations on a single NVIDIA L40S GPU with 48 GB of memory. All MACE-OFF23 models were also converted to single precision (float32) to ensure a fair comparison with MACELES-OFF, which was trained and evaluated in float32 precision. As shown, the inclusion of long-range interaction via LES introduces only minimal computational overhead. The MACELES-OFF model achieves a comparable inference speed to the MACE-OFF23(M) at the same precision (float32), which shares the same irreducible representation order (=1), and also to the MACE-OFF23(S) at double-precision. Moreover, the MACELES-OFF model is significantly faster than the MACE-OFF23(L), which exhibits similar RMSE values to MACELES-OFF, as summarized in Table III. No significant speed differences were observed between the ASE and OpenMM implementations for any of the models, and all models demonstrate favorable scaling with system size. In practice, the MACELES-OFF model supports simulations with up to approximately 6,000 atoms on a single GPU, whereas the MACE-OFF23(S) float32 model using ASE can handle up to around 35,000 atoms.

Figure 14.

Figure 14.

For Table of Contents Only

Acknowledgements

Research reported in this publication was supported by the National Institute Of General Medical Sciences of the National Institutes of Health under Award Number R35GM159986. The content is solely the responsibility of the authors and does not necessarily represent the official views of the National Institutes of Health. D.K. and B.C. acknowledge funding from Toyota Research Institute Synthesis Advanced Research Challenge. T.J.I, D.S.K. and P.Z. acknowledge funding from BIDMaP Postdoctoral Fellowship. T.J.I used resources of the National Energy Research Scientific Computing Center (NERSC), a Department of Energy Office of Science User Facility using NERSC award DOEERCAP0031751 ‘GenAI@NERSC’. The authors thank Bowen Deng for valuable discussions on MatGL implementation, and thank Gabor Csanyi for stimulating discussions.

Footnotes

Competing Interests B.C. has an equity stake in AIMATX Inc. The University of California, Berkeley has filed a provisional patent for the Latent Ewald Summation algorithm.

Data availability

The training sets, training scripts, and trained potentials are available at https://github.com/ChengUCB/les_fit.

Code availability

The LES library is publicly available at https://github.com/ChengUCB/les. The CACE package with the LES implementation is available at https://github.com/BingqingCheng/cace. The MACE package with the LES implementation is available at https://github.com/ACEsuit/mace The NequIP and Allegro LES extension package is available at https://github.com/ChengUCB/NequIP–LES. The MatGL package with the LES implementation is available at https://github.com/ChengUCB/matgl. The UMA package with the LES implementation is available at https://github.com/santi921/fairchem/tree/les_branch.

References

  • [1].Keith John A, Vassilev-Galindo Valentin, Cheng Bingqing, Chmiela Stefan, Gastegger Michael, Müller Klaus-Robert, and Tkatchenko Alexandre, “Combining machine learning and computational chemistry for predictive insights into chemical systems,” Chemical reviews 121, 9816–9872 (2021). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [2].Unke Oliver T, Chmiela Stefan, Sauceda Huziel E, Gastegger Michael, Poltavsky Igor, Schütt Kristof T, Tkatchenko Alexandre, and Müller Klaus-Robert, “Machine learning force fields,” Chemical Reviews 121, 10142–10186 (2021). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [3].Niblett Samuel P, Galib Mirza, and Limmer David T, “Learning intermolecular forces at liquid-vapor interfaces,” The Journal of chemical physics 155 (2021). [Google Scholar]
  • [4].Grisafi Andrea and Ceriotti Michele, “Incorporating long-range physics in atomic-scale machine learning,” The Journal of chemical physics 151 (2019). [Google Scholar]
  • [5].Huguenin-Dumittan Kevin K, Loche Philip, Haoran Ni, and Ceriotti Michele, “Physics-inspired equivariant descriptors of nonbonded interactions,” The Journal of Physical Chemistry Letters 14, 9612–9618 (2023). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [6].Zhang Linfeng, Wang Han, Muniz Maria Carolina, Panagiotopoulos Athanassios Z., Car Roberto, and Weinan E, “A deep potential model with long-range electrostatic interactions,” The Journal of Chemical Physics 156 (2022). [Google Scholar]
  • [7].Monacelli Lorenzo and Marzari Nicola, “Electrostatic interactions in atomistic and machine-learned potentials for polar materials,” arXiv preprint arXiv:2412.01642 (2024). [Google Scholar]
  • [8].Unke Oliver T and Meuwly Markus, “Physnet: A neural network for predicting energies, forces, dipole moments, and partial charges,” Journal of chemical theory and computation 15, 3678–3693 (2019). [DOI] [PubMed] [Google Scholar]
  • [9].Ko Tsz Wai, Finkler Jonas A, Goedecker Stefan, and Behler Jörg, “A fourth-generation high-dimensional neural network potential with accurate electrostatics including non-local charge transfer,” Nature communications 12, 398 (2021). [Google Scholar]
  • [10].Gao Ang and Remsing Richard C, “Self-consistent determination of long-range electrostatics in neural network potentials,” Nature communications 13, 1572 (2022). [Google Scholar]
  • [11].Sifain Andrew E, Lubbers Nicholas, Nebgen Benjamin T, Smith Justin S, Lokhov Andrey Y, Isayev Olexandr, Roitberg Adrian E, Barros Kipton, and Tretiak Sergei, “Discovering a transferable charge assignment model using machine learning,” The journal of physical chemistry letters 9, 4495–4501 (2018). [DOI] [PubMed] [Google Scholar]
  • [12].Gong Sheng, Zhang Yumin, Mu Zhenliang, Pu Zhichen, Wang Hongyi, Han Xu, Yu Zhiao, Chen Mengyi, Zheng Tianze, Wang Zhi, Chen Lifei, Yang Zhenze, Wu Xiaojie, Shi Shaochen, Gao Weihao, Yan Wen, and Xiang Liang, “A predictive machine learning force-field framework for liquid electrolyte development,” Nature Machine Intelligence 7, 543–552 (2025) [Google Scholar]
  • [13].Shaidu Yusuf, Pellegrini Franco, Küçükbenli Emine, Lot Ruggero, and de Gironcoli Stefano, “Incorporating long-range electrostatics in neural network potentials via variational charge equilibration from shortsighted ingredients,” npj Computational Materials 10, 47 (2024). [Google Scholar]
  • [14].Yu Hongyu, Hong Liangliang, Chen Shiyou, Gong Xingao, and Xiang Hongjun, “Capturing long-range interaction with reciprocal space neural network,” arXiv preprint arXiv:2211.16684 (2022). [Google Scholar]
  • [15].Kosmala Arthur, Gasteiger Johannes, Gao Nicholas, and Günnemann Stephan, “Ewald-based long-range message passing for molecular graphs,” in International Conference on Machine Learning (PMLR, 2023) pp. 17544–17563. [Google Scholar]
  • [16].Plé Thomas, Lagardère Louis, and Piquemal Jean-Philip, “Force-field-enhanced neural network interactions: from local equivariant embedding to atom-in-molecule properties and long-range effects,” Chemical Science 14, 12554–12569 (2023) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [17].Inizan Théo Jaffrelot, Plé Thomas, Adjoua Olivier, Ren Pengyu, Gökcan Hatice, Isayev Olexandr, Lagardère Louis, and Piquemal Jean-Philip, “Scalable hybrid deep neural networks/polarizable potentials biomolecular simulations including long-range effects,” Chemical Science 14, 5438–5452 (2023) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [18].Caruso Alessandro, Venturin Jacopo, Giambagli Lorenzo, Rolando Edoardo, Noé Frank, and Clementi Cecilia, “Extending the range of graph neural networks: Relaying attention nodes for global encoding,” arXiv preprint arXiv:2502.13797 (2025). [Google Scholar]
  • [19].Rinaldi Matteo, Bochkarev Anton, Lysogorskiy Yury, and Drautz Ralf, “Charge-constrained atomic cluster expansion,” Physical Review Materials 9, 033802 (2025) [Google Scholar]
  • [20].Anstine Dylan M., Zubatyuk Roman, and Isayev Olexandr, “Aimnet2: a neural network potential to meet your neutral, charged, organic, and elemental-organic needs,” Chemical Science 16, 10228–10244 (2025). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [21].Adil Kabylda J Frank Thorben, Dou ergio Suarez, Khabibrakhmanov Almaz, Sandonas Leonardo Medrano, Unke Oliver T., Chmiela Stefan, Müller Klaus-Robert, and Tkatchenko Alexandre, “Molecular simulations with a pretrained neural network and universal pairwise force fields,” ChemRxiv (2025) [Google Scholar]
  • [22].Faller Carolin, Kaltak Merzuk, and Kresse Georg, “Density-based long-range electrostatic descriptors for machine learning force fields,” The Journal of Chemical Physics 161, 214701 (2024) [DOI] [PubMed] [Google Scholar]
  • [23].Plé Thomas, Adjoua Olivier, Benali Anouar, Posenitskiy Evgeny, Villot Corentin, Lagardère Louis, and Piquemal Jean-Philip, “A foundation model for accurate atomistic simulations in drug design,” ChemRxiv (2025). [Google Scholar]
  • [24].Loche Philip, Huguenin-Dumittan Kevin K., Honarmand Melika, Xu Qianjun, Rumiantsev Egor, How Wei Bin, Langer Marcel F., and Ceriotti Michele, “Fast and flexible long-range models for atomistic machine learning,” The Journal of Chemical Physics 162, 142501 (2025). [DOI] [PubMed] [Google Scholar]
  • [25].Cheng Bingqing, “Latent ewald summation for machine learning of long-range interactions,” npj Computational Materials 11, 80 (2025). [Google Scholar]
  • [26].King Daniel S., Kim Dongjin, Zhong Peichen, and Cheng Bingqing, “Machine learning of charges and long-range interactions from energies and forces,” Nature Communications 16, 8763 (2025). [Google Scholar]
  • [27].Zhong Peichen, Kim Dongjin, King Daniel S, and Cheng Bingqing, “Machine learning interatomic potential can infer electrical response,” arXiv preprint arXiv:2504.05169 (2025). [Google Scholar]
  • [28].Behler Jörg and Parrinello Michele, “Generalized neural-network representation of high-dimensional potential-energy surfaces,” Physical review letters 98, 146401 (2007). [DOI] [PubMed] [Google Scholar]
  • [29].Bartók Albert P, Payne Mike C, Kondor Risi, and Csányi Gábor, “Gaussian approximation potentials: The accuracy of quantum mechanics, without the electrons,” Physical review letters 104, 136403 (2010). [DOI] [PubMed] [Google Scholar]
  • [30].Shapeev Alexander V, “Moment tensor potentials: A class of systematically improvable interatomic potentials,” Multiscale Modeling & Simulation 14, 1153–1173 (2016). [Google Scholar]
  • [31].Drautz Ralf, “Atomic cluster expansion for accurate and transferable interatomic potentials,” Physical Review B 99, 014104 (2019). [Google Scholar]
  • [32].Batzner Simon, Musaelian Albert, Sun Lixin, Geiger Mario, Mailoa Jonathan P, Kornbluth Mordechai, Molinari Nicola, Smidt Tess E, and Kozinsky Boris, “E (3)-equivariant graph neural networks for data-efficient and accurate interatomic potentials,” Nature communications 13, 2453 (2022). [Google Scholar]
  • [33].Batatia Ilyes, Kovacs David P, Simm Gregor, Ortner Christoph, and Csányi Gábor, “Mace: Higher order equivariant message passing neural networks for fast and accurate force fields,” Advances in Neural Information Processing Systems 35, 11423–11436 (2022). [Google Scholar]
  • [34].Cheng Bingqing, “Cartesian atomic cluster expansion for machine learning interatomic potentials,” npj Computational Materials 10, 157 (2024). [Google Scholar]
  • [35].Qu Eric and Krishnapriyan Aditi S., “The importance of being scalable: Improving the speed and accuracy of neural network interatomic potentials across chemical domains,” in The Thirty-eighth Annual Conference on Neural Information Processing Systems (2024). [Google Scholar]
  • [36].Ko Tsz Wai, Deng Bowen, Nassar Marcel, Barroso-Luque Luis, Liu Runze, Qi Ji, Thakur Atul C., Mishra Adesh Rohan, Liu Elliott, Ceder Gerbrand, Miret Santiago, and Ong Shyue Ping, “Materials graph library (matgl), an open-source graph deep learning library for materials science and chemistry,” npj Computational Materials 11, 253 (2025). [Google Scholar]
  • [37].Fu Xiang, Wood Brandon M., Barroso-Luque Luis, Levine Daniel S., Gao Meng, Dzamba Misko, and Zitnick C. Lawrence, “Learning smooth and expressive interatomic potentials for physical property prediction,” arXiv preprint arXiv:2502.12147 (2025). [Google Scholar]
  • [38].Pozdnyakov Sergey and Ceriotti Michele, “Smooth, exact rotational symmetrization for deep learning on point clouds,” in Advances in Neural Information Processing Systems, Vol. 36 (Curran Associates, Inc., 2023) pp. 79469–79501. [Google Scholar]
  • [39].Pelaez Raul P., Simeon Guillem, Galvelis Raimondas, Mirarchi Antonio, Eastman Peter, Doerr Stefan, Thölke Philipp, Markland Thomas E., and De Fabritiis Gianni, “Torchmd-net 2.0: Fast neural network potentials for molecular simulations,” Journal of Chemical Theory and Computation 20, 4076–4087 (2024) [DOI] [PubMed] [Google Scholar]
  • [40].Haghighatlari Mojtaba, Li Jie, Guan Xingyi, Zhang Oufan, Das Akshaya, Stein Christopher J., Heidar-Zadeh Farnaz, Liu Meili, Head-Gordon Martin, Bertels Luke, Hao Hongxia, Leven Itai, and Head-Gordon Teresa, “Newtonnet: A newtonian message passing network for deep learning of interatomic potentials and forces,” Digital Discovery 1, 333–343 (2022) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [41].Liao Yi-Lun and Smidt Tess, “Equiformer: Equivariant graph attention transformer for 3d atomistic graphs,” in International Conference on Learning Representations (2023). [Google Scholar]
  • [42].Musaelian Albert, Batzner Simon, Johansson Anders, Sun Lixin, Owen Cameron J, Kornbluth Mordechai, and Kozinsky Boris, “Learning local equivariant representations for large-scale atomistic dynamics,” Nature Communications 14, 579 (2023). [Google Scholar]
  • [43].Wood Brandon M., Dzamba Misko, Fu Xiang, Gao Meng, Shuaibi Muhammed, Barroso-Luque Luis, Abdelmaqsoud Kareem, Gharakhanyan Vahe, Kitchin John R., Levine Daniel S., Michel Kyle, Sriram Anuroop, Cohen Taco, Das Abhishek, Rizvi Ammar, Sahoo Sushree Jagriti, Ulissi Zachary W., and Zitnick C. Lawrence, “Uma: A family of universal models for atoms,” arXiv preprint arXiv:2506.23971 (2025). [Google Scholar]
  • [44].Eastman Peter, Behara Pavan Kumar, Dotson David L., Galvelis Raimondas, Herr John E., Horton Josh T., Mao Yuezhi, Chodera John D., Pritchard Benjamin P., Wang Yuanqing, De Fabritiis Gianni, and Markland Thomas E, “Spice, a dataset of drug-like molecules and peptides for training machine learning potentials,” Scientific Data 10, 11 (2023). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [45].Kovács Dávid Péter, Moore J. Harry, Browning Nicholas J., Batatia Ilyes, Horton Joshua T., Pu Yixuan, Kapil Venkat, Witt William C., Magdău Ioan-Bogdan, Cole Daniel J., and Csányi Gábor, “Mace-off: Shortrange transferable machine learning force fields for organic molecules,” Journal of the American Chemical Society (2025). [Google Scholar]
  • [46].Wang Han, Zhang Linfeng, Han Jiequn, and Weinan E, “Deepmd-kit: A deep learning package for many-body potential energy representation and molecular dynamics,” Computer Physics Communications 228, 178–184 (2018). [Google Scholar]
  • [47].Schütt Kristof, Kindermans Pieter-Jan, Sauceda Felix Huziel Enoc, Chmiela Stefan, Tkatchenko Alexandre, and Müller Klaus-Robert, “Schnet: A continuous-filter convolutional neural network for modeling quantum interactions,” Advances in neural information processing systems 30 (2017). [Google Scholar]
  • [48].Gonze Xavier and Lee Changyol, “Dynamical matrices, born effective charges, dielectric permittivity tensors, and interatomic force constants from density-functional perturbation theory,” Physical Review B 55, 10355 (1997). [Google Scholar]
  • [49].Chuin Wei Tan, Descoteaux Marc L., Kotak Mit, de Miranda Nascimento Gabriel, Kavanagh Seán R., Zichi Laura, Wang Menghang, Saluja Aadit, Hu Yizhong R., Smidt Tess, Johansson Anders, Witt William C., Kozinsky Boris, and Musaelian Albert, “High-performance training and inference for deep equivariant interatomic potentials,” arXiv preprint arXiv:2504.16068 (2025). [Google Scholar]
  • [50].Chen Chi and Ong Shyue Ping, “A universal graph deep learning interatomic potential for the periodic table,” Nature Computational Science 2, 718–728 (2022). [DOI] [PubMed] [Google Scholar]
  • [51].Deng Bowen, Zhong Peichen, Jun KyuJung, Riebesell Janosh, Han Kevin, Bartel Christopher J, and Ceder Gerbrand, “Chgnet as a pretrained universal neural network potential for charge-informed atomistic modelling,” Nature Machine Intelligence 5, 1031–1041 (2023). [Google Scholar]
  • [52].Levine Daniel S., Shuaibi Muhammed, Spotte-Smith Evan Walter Clark, Taylor Michael G., Hasyim Muhammad R., Michel Kyle, Batatia Ilyes, Csányi Gábor, Dzamba Misko, Eastman Peter, Frey Nathan C., Fu Xiang, Gharakhanyan Vahe, Krishnapriyan Aditi S., Rackers Joshua A., Raja Sanjeev, Rizvi Ammar, Rosen Andrew S., Ulissi Zachary, Vargas Santiago, Zitnick C. Lawrence, Blau Samuel M., and Wood Brandon M., “The open molecules 2025 (omol25) dataset, evaluations, and models,” arXiv prerpint arXiv:2505.08762 (2025). [Google Scholar]
  • [53].Barroso-Luque Luis, Shuaibi Muhammed, Fu Xiang, Wood Brandon M., Dzamba Misko, Gao Meng, Rizvi Ammar, Zitnick C. Lawrence, and Ulissi Zachary W., “Open materials 2024 (omat24) inorganic materials dataset and models,” arXiv prerpint arXiv:2410.12771 (2024). [Google Scholar]
  • [54].Passaro Saro and Zitnick C. Lawrence, “Reducing so(3) convolutions to so(2) for efficient equivariant gnns,” arXiv preprint arXiv:2302.03655 (2023). [Google Scholar]
  • [55].Remsing Richard C., Rodgers Jocelyn M., and Weeks John D., “Deconstructing classical water models at interfaces and in bulk,” Journal of Statistical Physics 145, 313–334 (2011). [Google Scholar]
  • [56].Gao Ang, Remsing Richard C., and Weeks John D., “Short solvent model for ion correlations and hydrophobic association,” Proceedings of the National Academy of Sciences 117, 1293–1302 (2020). [Google Scholar]
  • [57].Schmiedmayer Bernhard and Kresse Georg, “Derivative learning of tensorial quantities—predicting finite temperature infrared spectra from first principles,” The Journal of Chemical Physics 161 (2024). [Google Scholar]
  • [58].Avbelj Franc, Grdadolnik Simona Golic, Grdadolnik Joze, and Baldwin Robert L., “Intrinsic backbone preferences are fully present in blocked amino acids,” Proceedings of the National Academy of Sciences 103, 1272–1277 (2006) [Google Scholar]
  • [59].Yuan Eric C.-Y., Liu Yunsheng, Chen Junmin, Zhong Peichen, Raja Sanjeev, Kreiman Tobias, Vargas Santiago, Xu Wenbin, Head-Gordon Martin, Yang Chao, Blau Samuel M., Cheng Bingqing, Krishnapriyan Aditi, and Head-Gordon Teresa, “Foundation models for atomistic simulation of chemistry and materials,” arXiv preprint arXiv:2503.10538 (2025). [Google Scholar]
  • [60].Isert Clemens, Atz Kenneth, Jiménez-Luna José, and Schneider Gisbert, “Qmugs, quantum mechanical properties of drug-like molecules,” Scientific Data 9, 273 (2022). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [61].Skinner Lawrie B, Benmore CJ, Neuefeind Joerg C, and Parise John B, “The structure of water around the compressibility minimum,” The Journal of chemical physics 141, 214507 (2014). [DOI] [PubMed] [Google Scholar]
  • [62].Bertie John E. and Lan Zhida, “Infrared intensities of liquids xx: The intensity of the oh stretching band of liquid water revisited, and the best current values of the optical constants of h2o(l) at 25°c between 15,000 and 1 cm-1,” Applied Spectroscopy 50, 1047–1057 (1996). [Google Scholar]
  • [63].Marsalek Ondrej and Markland Thomas E, “Quantum dynamics and spectroscopy of ab initio liquid water: The interplay of nuclear and electronic quantum effects,” The journal of physical chemistry letters 8, 1545–1551 (2017). [DOI] [PubMed] [Google Scholar]
  • [64].Musil Félix, Zaporozhets Iryna, Noé Frank, Clementi Cecilia, and Kapil Venkat, “Quantum dynamics using path integral coarse-graining,” The Journal of Chemical Physics 157, 181102 (2022) [DOI] [PubMed] [Google Scholar]
  • [65].Haynes William M, CRC handbook of chemistry and physics (CRC press, 2016). [Google Scholar]
  • [66].Linstorm P, “Nist chemistry webbook, nist standard reference database number 69,” J. Phys. Chem. Ref. Data, Monograph 9, 1–1951 (1998). [Google Scholar]
  • [67].Magdău Ioan-Bogdan, Arismendi-Arrieta Daniel J, Smith Holly E, Grey Clare P, Hermansson Kersti, and Csányi Gábor, “Machine learning force fields for molecular liquids: Ethylene carbonate/ethyl methyl carbonate binary solvent,” npj Computational Materials 9, 146 (2023). [Google Scholar]
  • [68].Dahiyat Bassil I. and Mayo Stephen L., “De Novo Protein Design: Fully Automated Sequence Selection,” Science 278, 82–87 (1997). [DOI] [PubMed] [Google Scholar]
  • [69].Barducci Alessandro, Bussi Giovanni, and Parrinello Michele, “Well-Tempered Metadynamics: A Smoothly Converging and Tunable Free-Energy Method,” Physical Review Letters 100, 020603 (2008) [DOI] [PubMed] [Google Scholar]
  • [70].Zhang Shuting, Schweitzer-Stenner Reinhard, and Urbanc Brigita, “Do Molecular Dynamics Force Fields Capture Conformational Dynamics of Alanine in Water?” Journal of Chemical Theory and Computation 16, 510–527 (2020). [DOI] [PubMed] [Google Scholar]
  • [71].Wang Zhi-Xiang and Duan Yong, “Solvation effects on alanine dipeptide: A MP2/cc-pVTZ//MP2/6–31G** study of (ϕ,ψ) energy maps and conformers in the gas phase, ether, and water,” Journal of Computational Chemistry 25, 1699–1716 (2004) [DOI] [PubMed] [Google Scholar]
  • [72].Lindorff-Larsen Kresten, Piana Stefano, Palmo Kim, Maragakis Paul, Klepeis John L., Dror Ron O., and Shaw David E., “Improved side-chain torsion potentials for the Amber ff99SB protein force field,” Proteins: Structure, Function, and Bioinformatics 78, 1950–1958 (2010) [Google Scholar]
  • [73].Jain Anubhav, Ong Shyue Ping, Hautier Geoffroy, Chen Wei, Richards William Davidson, Dacek Stephen, Cholia Shreyas, Gunter Dan, Skinner David, Ceder Gerbrand, and Persson Kristin A., “Commentary: The materials project: A materials genome approach to accelerating materials innovation,” APL Mater. 1, 11002 (2013). [Google Scholar]
  • [74].Tran Richard, Lan Janice, Shuaibi Muhammed, Wood Brandon M., Goyal Siddharth, Das Abhishek, Heras-Domingo Javier, Kolluru Adeesh, Rizvi Ammar, Shoghi Nima, Sriram Anuroop, Therrien Félix, Abed Jehad, Voznyy Oleksandr, Sargent Edward H., Ulissi Zachary, and Zitnick C. Lawrence, “The open catalyst 2022 (oc22) dataset and challenges for oxide electrocatalysts,” ACS Catalysis 13, 3066–3084 (2023) [Google Scholar]
  • [75].Weber John L., Guha Rishabh D., Agarwal Garvit, Wei Yujing, Fike Aidan A., Xie Xiaowei, Stevenson James, Leswing Karl, Halls Mathew D., Abel Robert, and Jacobson Leif D., “Efficient long-range machine learning force fields for liquid and materials properties,” arXiv preprint arXiv:2505.06462 (2025). [Google Scholar]
  • [76].Paszke Adam, Gross Sam, Massa Francisco, Lerer Adam, Bradbury James, Chanan Gregory, Killeen Trevor, Lin Zeming, Gimelshein Natalia, Antiga Luca, Desmaison Alban, Kopf Andreas, Yang Edward, DeVito Zachary, Raison Martin, Tejani Alykhan, Chilamkurthy Sasank, Steiner Benoit, Fang Lu, Bai Junjie, and Chintala Soumith, “Pytorch: An imperative style, high-performance deep learning library,” in Advances in Neural Information Processing Systems, Vol. 32 (Curran Associates, Inc., 2019). [Google Scholar]
  • [77].Eastman Peter, Behara Pavan Kumar, Dotson David, Galvelis Raimondas, Herr John, Horton Josh, Mao Yuezhi, Chodera John, Pritchard Benjamin, Wang Yuanqing, De Fabritiis Gianni, and Markland Thomas, “Spice 2.0.1,” Zenodo https://zenodo.org/records/10975225 (2024). [Google Scholar]
  • [78].Sun Qiming, Berkelbach Timothy C., Blunt Nick S., Booth George H., Guo Sheng, Li Zhendong, Liu Junzi, McClain James D., Sayfutyarova Elvira R., Sharma Sandeep, Wouters Sebastian, and Kin-Lic Chan Garnet, “Pyscf: the python-based simulations of chemistry framework,” Wiley Interdisciplinary Reviews: Computational Molecular Science 8, e1340 (2018). [Google Scholar]
  • [79].Larsen Ask Hjorth, Mortensen Jens Jørgen, Blomqvist Jakob, Castelli Ivano E, Christensen Rune, Dułak Marcin, Friis Jesper, Groves Michael N, Hammer Bjørk, Hargus Cory, Hermes Eric D, Jennings Paul C, Jensen Peter Bjerre, Kermode James, Kitchin John R, Kolsbjerg Esben Leonhard, Kubal Joseph, Kaasbjerg Kristen, Lysgaard Steen, Maronsson Jón Bergmann, Maxson Tristan, Olsen Thomas, Pastewka Lars, Peterson Andrew, Rostgaard Carsten, Schiøtz Jakob, Schütt Ole, Strange Mikkel, Thygesen Kristian S, Vegge Tejs, Vilhelmsen Lasse, Walter Michael, Zeng Zhenhua, and Jacobsen Karsten W, “The atomic simulation environment—a Python library for working with atoms,” Journal of Physics: Condensed Matter 29, 273002 (2017) [DOI] [PubMed] [Google Scholar]
  • [80].Kim Sunghwan, Chen Jie, Cheng Tiejun, Gindulyte Asta, He Jia, He Siqian, Li Qingliang, Shoemaker Benjamin A., Thiessen Paul A., Yu Bo, Zaslavsky Leonid, Zhang Jian, and Bolton Evan E., “Pubchem 2025 update,” Nucleic Acids Research 53, D1516–D1525 (2025). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [81].Landrum Greg, Tosco Paolo, Kelley Brian, Rodriguez Ricardo, Cosgrove David, Vianello Riccardo, sriniker, Gedeck Peter, Jones Gareth, Kawashima Eisuke, Schneider Nadine, Nealschneider Dan, Dalke Andrew, Swain Matt, Cole Brian, Turk Samo, Savelev Aleksandr, tadhurst-cdd, Vaucher Alain, Wójcikowski Maciej, Take Ichiru, Walker Rachel, Scalfani Vincent F., Faara Hussein, Ujihara Kazuya, Probst Daniel, Lehtivarjo Juuso, godin guillaume, Pahl Axel, and Maeder Niels, “rdkit/rdkit: 2025_03_2 (q1 2025) release,” Zenodo https://zenodo.org/records/15286010 (2025). [Google Scholar]
  • [82].Berendsen HJC, van der Spoel D, and van Drunen R, “Gromacs: A message-passing parallel molecular dynamics implementation,” Computer Physics Communications 91, 43–56 (1995) [Google Scholar]
  • [83].Wang Junmei, Wang Wei, Kollman Peter A, and Case David A, “Automatic atom type and bond type perception in molecular mechanical calculations,” Journal of molecular graphics and modelling 25, 247–260 (2006). [DOI] [PubMed] [Google Scholar]
  • [84].Yoshida Haruo, “Construction of higher order symplectic integrators,” Physics Letters A 150, 262–268 (1990) [Google Scholar]
  • [85].Jorgensen William L., Chandrasekhar Jayaraman, Madura Jeffry D., Impey Roger W., and Klein Michael L., “Comparison of simple potential functions for simulating liquid water,” The Journal of Chemical Physics 79, 926–935 (1983). [Google Scholar]
  • [86].Bonomi Massimiliano, Bussi Giovanni, Camilloni Carlo, Tribello Gareth A., Banáš Pavel, Barducci Alessandro, Bernetti Mattia, Bolhuis Peter G., Bottaro Sandro, Branduardi Davide, Capelli Riccardo, Carloni Paolo, Ceriotti Michele, Cesari Andrea, Chen Haochuan, Chen Wei, Colizzi Francesco, De Sandip, De La Pierre Marco, Donadio Davide, Drobot Viktor, Ensing Bernd, Ferguson Andrew L., Filizola Marta, Fraser James S., Fu Haohao, Gasparotto Piero, Gervasio Francesco Luigi, Giberti Federico, Gil-Ley Alejandro, Giorgino Toni, Heller Gabriella T., Hocky Glen M., Iannuzzi Marcella, Invernizzi Michele, Jelfs Kim E., Jussupow Alexander, Kirilin Evgeny, Laio Alessandro, Limongelli Vittorio, Lindorff-Larsen Kresten, Löhr Thomas, Marinelli Fabrizio, Martin-Samos Layla, Masetti Matteo, Meyer Ralf, Michaelides Angelos, Molteni Carla, Morishita Tetsuya, Nava Marco, Paissoni Cristina, Papaleo Elena, Parrinello Michele, Pfaendtner Jim, Piaggi Pablo, Piccini GiovanniMaria, Pietropaolo Adriana, Pietrucci Fabio, Pipolo Silvio, Provasi Davide, Quigley David, Raiteri Paolo, Raniolo Stefano, Rydzewski Jakub, Salvalaglio Matteo, Sosso Gabriele Cesare, Spiwok Vojtěch, Šponer Jiří, Swenson David W. H., Tiwary Pratyush, Valsson Omar, Vendruscolo Michele, Voth Gregory A., White Andrew, and The PLUMED consortium, “Promoting transparency and reproducibility in enhanced molecular simulations,” Nature Methods 16, 670–673 (2019) [DOI] [PubMed] [Google Scholar]
  • [87].Tribello Gareth A., Bonomi Massimiliano, Branduardi Davide, Camilloni Carlo, and Bussi Giovanni, “PLUMED 2: New feathers for an old bird,” Computer Physics Communications 185, 604–613 (2014). [Google Scholar]
  • [88].Branduardi Davide, Bussi Giovanni, and Parrinello Michele, “Metadynamics with adaptive gaussians,” Journal of chemical theory and computation 8, 2247–2254 (2012). [DOI] [PubMed] [Google Scholar]
  • [89].Bitzek Erik, Koskinen Pekka, Gähler Franz, Moseler Michael, and Gumbsch Peter, “Structural relaxation made simple,” Physical Review Letters 97, 1–4 (2006). [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

The training sets, training scripts, and trained potentials are available at https://github.com/ChengUCB/les_fit.

The LES library is publicly available at https://github.com/ChengUCB/les. The CACE package with the LES implementation is available at https://github.com/BingqingCheng/cace. The MACE package with the LES implementation is available at https://github.com/ACEsuit/mace The NequIP and Allegro LES extension package is available at https://github.com/ChengUCB/NequIP–LES. The MatGL package with the LES implementation is available at https://github.com/ChengUCB/matgl. The UMA package with the LES implementation is available at https://github.com/santi921/fairchem/tree/les_branch.

RESOURCES