Significance
The interplay between molecular simulation and artificial intelligence has spurred many insights into chemical discovery, yet the data requirement of machine learning approaches remains a bottleneck. We tackle this challenge by developing a rigorous framework to conserve the symmetries of quantum chemical systems within deep neural networks. Our framework greatly improves the prediction of numerous molecular electronic and energetic properties, even for systems far from the training data. Through tight integration of quantum theory, geometry, and machine learning, our study offers a solution to vastly accelerate the in silico modelling of chemical phenomena in many domains, such as material design and catalysis.
Keywords: quantum chemistry, machine learning, equivariance
Abstract
Predicting electronic energies, densities, and related chemical properties can facilitate the discovery of novel catalysts, medicines, and battery materials. However, existing machine learning techniques are challenged by the scarcity of training data when exploring unknown chemical spaces. We overcome this barrier by systematically incorporating knowledge of molecular electronic structure into deep learning. By developing a physics-inspired equivariant neural network, we introduce a method to learn molecular representations based on the electronic interactions among atomic orbitals. Our method, OrbNet-Equi, leverages efficient tight-binding simulations and learned mappings to recover high-fidelity physical quantities. OrbNet-Equi accurately models a wide spectrum of target properties while being several orders of magnitude faster than density functional theory. Despite only using training samples collected from readily available small-molecule libraries, OrbNet-Equi outperforms traditional semiempirical and machine learning–based methods on comprehensive downstream benchmarks that encompass diverse main-group chemical processes. Our method also describes interactions in challenging charge-transfer complexes and open-shell systems. We anticipate that the strategy presented here will help to expand opportunities for studies in chemistry and materials science, where the acquisition of experimental or reference training data is costly.
Discovering new molecules and materials is central to tackling contemporary challenges in energy storage and drug discovery (1, 2). As the experimentally uninvestigated chemical space for these applications is immense, large-scale computational design and screening for new molecule candidates have the potential to vastly reduce the burden of laborious experiments and to accelerate discovery (3–5). A crucial task is to model the quantum chemical properties of molecules by solving the many-body Schrödinger equation, which is commonly addressed by ab initio electronic structure methods (6, 7), such as density functional theory (DFT) (Fig. 1A). While very successful, ab initio methods are laden with punitive computational requirements that makes it difficult to achieve a throughput on a scale of the unexplored chemical space.
In contrast, machine learning (ML) approaches are highly flexible as function approximators and thus, are promising for modeling molecular properties at a drastically reduced computational cost. A large class of ML-based molecular property predictors includes methods that use atomic coordinate–based input features, which closely resemble molecular mechanics (MM) descriptors (8–18); these methods will be referred to as Atomistic ML methods in the current work (Fig. 1B). Atomistic ML methods have been employed to solve challenging problems in molecular sciences, such as RNA structure prediction (19) and anomalous phase transitions (20). However, there remains a key discrepancy between Atomistic ML and ab initio approaches regarding the modeling of quantum chemical properties, as Atomistic ML approaches typically neglect the electronic degrees of freedom that are central for the description of important phenomena, such as electronic excitations, charge transfer, and long-range interactions. Moreover, recent work shows that Atomistic ML can struggle with transferability on downstream tasks where the molecules may chemically deviate from the training samples (21, 22) as is expected to be common for underexplored chemical spaces.
Recent efforts to embody quantum mechanics (QM) into molecular representations based on electronic structure theory have made breakthroughs in improving both the chemical and electronic transferability of ML-based molecular modeling (23–29). Leveraging a physical feature space extracted from QM simulations, such QM-informed ML methods have attained data efficiency that significantly surpasses Atomistic ML methods, especially when extrapolated to systems with length scales or chemical compositions unseen during training. Nevertheless, QM-informed ML methods still fall short in terms of the flexibility of modeling diverse molecular properties unlike their atomistic counterparts, as they are typically implemented for a limited set of learning targets, such as the electronic energy or the exchange–correlation potential. A key bottleneck hampering the broader applicability of QM-informed approaches is the presence of unique many-body symmetries necessitated by an explicit treatment on electron–electron interactions. Heuristic schemes have been used to enforce invariance (24, 26, 30–33) at a potential loss of information in their input features or expressivity in their ML models. Two objectives remain elusive for QM-informed ML: 1) incorporate the underlying physical symmetries with maximal data efficiency and model flexibility and 2) accurately infer downstream molecular properties for large chemical spaces at a computational resource requirement on par with existing empirical and Atomistic ML methods.
Herein, we introduce an end-to-end ML method for QM-informed molecular representations, OrbNet-Equi, in fulfillment of these two objectives. OrbNet-Equi featurizes a mean-field electronic structure via the atomic orbital basis and learns molecular representations through an ML model that is equivariant with respect to isometric basis transformations (Fig. 1 C–E). By the virtue of equivariance, OrbNet-Equi respects essential physical constraints of symmetry conservation so that the target quantum chemistry properties are learned independent of a reference frame. Underpinning OrbNet-Equi is a neural network designed with insights from recent advances in geometric deep learning (34–40) but with key architectural innovations to achieve equivariance based on the tensor-space algebraic structures entailed in atomic orbital–based molecular representations.
We demonstrate the data efficiency of OrbNet-Equi on learning molecular properties using input features obtained from tight-binding QM simulations, which are efficient and scalable to systems with thousands of atoms (41). We find that OrbNet-Equi consistently achieves lower prediction errors than existing Atomistic ML methods and our previous QM-informed ML method (26) on diverse target properties, such as electronic energies, dipole moments, electron densities, and frontier orbital energies. Specifically, our study on learning frontier orbital energies illustrates an effective strategy to improve the prediction of electronic properties by incorporating molecular orbital space information.
To showcase its transferability to complex real-world chemical spaces, we trained an OrbNet-Equi model on single-point energies of ∼236,000 molecules curated from readily available small-molecule libraries. The resulting model, OrbNet-Equi/Selected Drug-Like and Biofragment Conformers (SDC21), achieves a performance competitive to state-of-the-art composite DFT methods when tested on a wide variety of main-group quantum chemistry benchmarks while being up to 1,000-fold faster at run time. As a particular case study, we found that OrbNet-Equi/SDC21 substantially improved the prediction accuracy of ionization potentials relative to semiempirical QM methods, even though no radical species were included for training. Thus, our method has the potential to accelerate simulations for challenging problems in organic synthesis (42), battery design (43), and molecular biology (44). Detailed data analysis pinpoints viable future directions to systematically improve its chemical space coverage, opening a plausible pathway toward a generic hybrid physics–ML paradigm for the acceleration of molecular modeling and discovery.
Results
The OrbNet-Equi Methodology.
OrbNet-Equi featurizes a molecular system through mean-field QM simulations. Semiempirical tight-binding models (41) are used through this study since they can be solved rapidly for both small-molecule and extended systems, which enables deploying OrbNet-Equi to large chemical spaces. In particular, we employ the recently reported GFN (geometries [vibrational], frequencies, and noncovalent interactions [NCIs])-xTB (extended tight-binding) (45) QM model in which the mean-field electronic structure is obtained through self-consistently solving a tight-binding model system (Fig. 1C). Built upon , the inputs to the neural network comprise a stack of matrices defined as single-electron operators represented in the atomic orbitals (Fig. 1D),
[1] |
where A and B are both atom indices; and indicate a basis function in the set of atomic orbitals { Φ } centered at each atom. Motivated by mean-field electronic energy expressions, the input atomic orbital features are selected as using the Fock F, density P, core Hamiltonian H, and overlap S matrices of the tight-binding QM model (Materials and Methods), unless otherwise specified.
OrbNet-Equi learns a map to approximate the target molecular property y of high-fidelity electronic structure simulations or experimental measurements,
[2] |
where denotes a cost functional between the reference and predicted targets over training data. The learning problem described by Eq. 2 requires careful treatment on isometric coordinate transformations imposed on the molecular system because the coefficients of are defined up to a given viewpoint (Fig. 1E). Precisely, the atomic orbitals undergo a unitary linear recombination subject to three-dimensional (3D) rotations: , where denotes the Wigner-D matrix of degree l for a rotation operation . As a consequence of the basis changing induced by is transformed block wise:
[3] |
where the dagger symbol denotes an Hermitian conjugate. To account for the rototranslation symmetries, the neural network must be made equivariant with respect to all such isometric basis rotations: that is,
[4] |
which is fulfilled through our delicate design of the neural network in OrbNet-Equi (Fig. 2). The neural network iteratively updates a set of representations defined at each atom through its neural network modules and reads out predictions using a pooling layer located at the end of the network. During its forward pass, diagonal blocks of the inputs are first transformed into components that are isomorphic to orbital–angular–momentum eigenstates, which are then cast to the initial representations . Each subsequent module exploits off-diagonal blocks of to propagate nonlocal information among atomic orbitals and refine the representations , which resemble a process of applying time-evolution operators on quantum states. We provide a technical introduction to the neural network architecture in Materials and Methods. We incorporate other constraints on the learning task, such as size consistency solely through programming the pooling layer (Materials and Methods), therefore achieving task-agnostic modeling for diverse chemical properties. Additional details and theoretical results are provided in SI Appendix, sections S1 and S2.
Performance on Benchmark Datasets.
We begin with benchmarking OrbNet-Equi on the QM9 (Quantum Machines 9) dataset (46), which has been widely adopted for assessing ML-based molecular property prediction methods. QM9 contains 133,885 small organic molecules at optimized geometries, with target properties computed by DFT. Following previous works (13, 14, 17, 18, 38, 47), we take 110,000 random samples as the training set and 10,831 samples as the test set. We present results for both the “direct-learning” training strategy, which corresponds to training the model directly on the target property, and whenever applicable, the “delta-learning” strategy (48), which corresponds to training on the residual between output of the tight-binding QM model and the target level of theory.
We first trained OrbNet-Equi on two representative targets, the total electronic energy U0 and the molecular dipole moment vector (Fig. 3 A and B), for which a plethora of task-specific ML models has previously been developed (26, 49–53). The total energy U0 is predicted through a sum over atom-wise energy contributions, and the dipole moment is predicted through a combination of atomic partial charges and dipoles (SI Appendix, section S1.E). For U0 (Fig. 3A), the direct-learning results of OrbNet-Equi match the state-of-the-art kernel-based ML method [Faber–Christensen–Huang–Lilienfeld (FCHL)18/GPR (Gaussian Process Regression) (49) in terms of the test mean absolute error (MAE) while being scalable to large data regimes (Fig. 3A, training data size >20,000), where no competitive result has been reported before. With delta learning, OrbNet-Equi outperforms our previous QM-informed ML approach OrbNet (26) by in the test MAE. Because OrbNet also uses the GFN-xTB QM model for featurization and the delta-learning strategy for training, this improvement underscores the strength of our neural network design, which seamlessly integrates the underlying physical symmetries. Moreover, for dipole moments (Fig. 3B), OrbNet-Equi exhibits steep learning curve slopes regardless of the training strategy, highlighting its capability of learning rotational-covariant quantities at no sacrifice of data efficiency.
We then targeted the learning task of frontier molecular orbital (FMO) properties, in particular energies of the highest occupied molecular orbital (HOMO), the lowest unoccupied molecular orbital (LUMO), and the HOMO–LUMO gaps, which are important in the prediction of chemical reactivity and optical properties (54, 55). Because the FMOs are inherently defined in the electron energy space and are often spatially localized, it is expected to be challenging to predict FMO properties based on molecular representations in which a notion of electronic energy levels is absent. OrbNet-Equi overcame this obstacle by breaking the orbital filling degeneracy of its input features to encode plausible electron excitations near the FMO energy levels: that is, adding energy-weighted density matrices of “hole-excitation” and that of “particle-excitation” :
[5] |
[6] |
where ϵi and ni are the orbital energy and occupation number of the ith molecular orbital from tight-binding QM and denotes the molecular orbital coefficients with μ and ν indexing the atomic orbital basis. Here, the effective temperature parameters β are chosen as (atomic units), and a global attention-based pooling is used to ensure size-intensive predictions (SI Appendix, section S1.E.4). Fig. 3C shows that the inclusion of energy-weighted density matrices () indeed greatly enhanced model generalization on FMO energies, as evident from the drastic test MAE reduction against the model with default ground-state features () as well as the best result from Atomistic ML methods. Remarkably, for models using default ground-state features (Fig. 3C, red lines), we noticed a rank reversal behavior between direct-learning and delta-learning models as more training samples became available, mirroring similar observations from a recent Atomistic ML study (56). The absence of this cross-over when () are provided (Fig. 3C, blue) suggests that the origin of such a learning slowdown is the incompleteness of spatially degenerate descriptors, and the gap between delta-learning and direct-learning curves can be restored by breaking the energy-space degeneracy. This analysis reaffirms the role of identifying the dominant physical degrees of freedom in the context of the ML-based prediction of quantum chemical properties and is expected to benefit the modeling of relevant electrochemical and optical properties, such as redox potentials.
Furthermore, OrbNet-Equi is benchmarked on 12 targets of QM9 using the 110,000 full training set (SI Appendix, Table S1), for which we programmed its pooling layer to reflect the symmetry constraint of each target property (SI Appendix, section S1.E). We observed top-ranked performance on all targets with average test MAE around twofold lower than atomistic deep learning methods. In addition, we tested OrbNet-Equi on fitting molecular potential energy surfaces by training on multiple configurations of a molecule (SI Appendix, section S3.B). Results (SI Appendix, Tables S2 and S3) showed that OrbNet-Equi obtained energy and force prediction errors that match state-of-the-art ML potential methods (57, 58) on the Molecular Dynamics 17 dataset (57, 59), suggesting that our method also efficiently generalizes over the conformation degrees of freedom apart from being transferable across the chemical space. These extensive benchmarking studies confirm that our strategy is consistently applicable to a wide range of molecular properties.
Accurate Modeling for Electron Densities.
We next focus on the task of predicting the electron density , which plays an essential role in both the formulation of DFT and in the interpretation of molecular interactions. It is also more challenging than predicting the energetic properties from an ML perspective due to the need of preserving its real-space continuity and rotational covariance. OrbNet-Equi learns to output a set of expansion coefficients to represent the predicted electron density through a density-fitting basis set (Materials and Methods and SI Appendix, section S1.E.6),
[7] |
where is the maximum angular momentum in the density-fitting basis set for atom type zA and denotes the cardinality of basis functions with angular momentum l. We train OrbNet-Equi to learn DFT electron densities on the QM9 dataset of small organic molecules and the BfDB-SSI (Biofragment Database-sidechain-sidechain interactions) (60) dataset of amino acid side-chain dimers (Fig. 4) using the direct-learning strategy. OrbNet-Equi results are substantially better than Atomistic ML baselines in terms of the average L1 density error (Materials and Methods); specifically, OrbNet-Equi achieves an average of 0.191 ± 0.003% on BfDB-SSI using 2,000 training samples compared with 0.29% of Symmetry-Adapted Gaussian Process Regression (61) and an average of 0.206 ± 0.001% on QM9 using 123,835 training samples as compared with 0.28 to 0.36% of DeepDFT (62). Fig. 4A confirms that OrbNet-Equi predicts densities at consistently low errors across the real space and maintains a robust asymptotic decay behavior within low-density [] regions that are far from the molecular system. Furthermore, analysis of electric multipole moments derived from the predicted charge densities (SI Appendix, S3.C and Figs. S3 and S4) suggests that the OrbNet-Equi–predicted density yields multipole errors comparable with the deviations between different DFT functionals.
To understand whether the model generalizes to cases where charge transfer is significant, as in donor–acceptor systems, we introduce a simple baseline predictor termed monomer density superposition (MDS). The MDS electron density of a dimeric system is taken as the sum of independently computed DFT electron densities of the two monomers. OrbNet-Equi yields accurate predictions in the presence of charge redistribution induced by noncovalent effects, as identified by dimeric examples from the BfDB-SSI test set for which the MDS density (Fig. 4B, x axis) largely deviates from the DFT reference density of the dimer due to intermolecular interactions. One representative example is a strongly interacting glutamic acid–lysine system (Fig. 4 C and D) whose salt-bridge formation is known to be essential for the helical stabilization in protein folding (63), for which OrbNet-Equi predicts with significantly lower than that of MDS (). The accurate modeling of offers an opportunity for constructing transferable DFT models for extended systems by learning on both energetics and densities while at a small fraction of expense relative to solving the Kohn–Sham equations from scratch.
Transferability on Downstream Tasks.
Beyond data efficiency on established datasets in train–test split settings, a crucial but highly challenging aspect is whether the model accurately infers downstream properties after being trained on data that are feasible to obtain. To comprehensively evaluate whether OrbNet-Equi can be transferred to unseen chemical spaces without any additional supervision, we have trained an OrbNet-Equi model on a dataset curated from readily available small-molecule databases (Materials and Methods). The training dataset contains 235,834 samples with chemical space coverage for drug-like molecules and biological motifs containing chemical elements C, O, N, F, S, Cl, Br, I, P, Si, B, Na, K, Li, Ca, and Mg and thermalized geometries. The resulting OrbNet-Equi/SDC21 potential energy model is solely trained on DFT single-point energies using the delta-learning strategy. Without any fine-tuning, we directly apply OrbNet-Equi/SDC21 to downstream benchmarks that are recognized for assessing the accuracy of physics-based molecular modeling methods.
The task of ranking conformer energies of drug-like molecules is benchmarked via the Hutchison dataset of conformers of ∼700 molecules (21) (Fig. 5A and SI Appendix, Table S5, rows 1 and 2). On this task, OrbNet-Equi/SDC21 achieves a median R2 score of 0.87 ± 0.02 and R2 distributions closely matching the reference DFT theory on both neutral and charged systems. On the other hand, we notice that the median R2 of OrbNet-Equi/SDC21 with respect to the reference DFT theory (ωB97X-D3/def2-TZVP) is 0.96 ± 0.01, suggesting that the current performance on this task is saturated by the accuracy of DFT and can be systematically improved by applying fine-tuning techniques on higher-fidelity labels (64, 65). Timing results on the Hutchison dataset (SI Appendix, Table S4) confirm that the neural network inference time of OrbNet-Equi/SDC21 is on par with the GFN-xTB QM featurizer, resulting in an overall computational speed that is 100- to 1000-fold faster relative to existing cost-efficient composite DFT methods (21, 66, 67). To understand the model’s ability to describe dihedral energetics, which are crucial for virtual screening tasks, we benchmark OrbNet-Equi on the prediction of intramolecular torsion energy profiles using the TorsionNet500 (68) dataset, the most diverse benchmark set available for this problem (SI Appendix, Table S5, row 3). Although no explicit torsion angle sampling was performed during training data generation, OrbNet-Equi/SDC21 exhibits a barrier MAE of 0.173 ± 0.003 kcal/mol, much lower than the 1-kcal/mol threshold commonly considered for chemical accuracy. On the other hand, we notice an MAE of 0.7 kcal/mol for the TorsionNet model (68), which was trained on ∼1 million torsion energy samples. As shown in Fig. 5B, OrbNet-Equi/SDC21 robustly captures the torsion sectors of potential energy surface on an example challenging for both semiempirical QM (45, 69) and cost-efficient composite DFT (66) methods, precisely resolving both the suboptimal energy minima location at an dihedral angle as well as the barrier energy between two local minimas within a 1-kcal/mol chemical accuracy. Next, the ability to characterize noncovalent interactions (NCIs) is assessed on the S66X10 dataset (70) of intermolecular dissociation curves (SI Appendix, Table S6), on which OrbNet-Equi achieves an equilibrium-distance binding energy MAE of 0.35 ±0.09 kcal/mol with respect to the reference DFT theory compared against 1.55 ±0.17 kcal/mol of the GFN-xTB baseline. As shown from a uracil–uracil base pair example (Fig. 5C) for which high-fidelity wave function–based reference calculations have been reported, the binding energy curve along the intermolecular axis predicted by OrbNet-Equi/SDC21 agrees well with both DFT and the high-level CCSD(T) (Coupled Cluster single-double and perturbative triple) results. To further understand the accuracy and smoothness of the energy surfaces and the applicability on dynamics tasks, we perform geometry optimizations on the ROT34 dataset of 12 small organic molecules and the MCONF (Melatonin Conformer Space) dataset of 52 conformers of melatonin (71, 72) (Fig. 5D and SI Appendix, Table S5, rows 4 and 5). Remarkably, OrbNet-Equi/SDC21 consistently exhibits the lowest average rmsd among all physics-based and ML-based approaches (SI Appendix, Table S5), including the popular cost-efficient DFT method B97-3c (66). Further details regarding the numerical experiments and error metrics are provided in Materials and Methods.
Remarkably, on the G21IP dataset (73) of adiabatic ionization potentials, we find that the OrbNet-Equi/SDC21 model achieves prediction errors substantially lower than semiempirical QM methods (Fig. 5E and SI Appendix, Table S7), even though samples of open-shell signatures are expected to be rare from the training set (Materials and Methods). Such an improvement cannot be solely attributed to structure-based corrections since there are no or negligible geometrical changes between the neutral and ionized species for both the single-atom systems and several polyatomic systems (e.g., IP_66, a Phosphanide anion) in the G21IP dataset. This reveals that our method has the potential to be transferred to unseen electronic states in a zero-shot manner, which represents early evidence that a hybrid physics–ML strategy may unravel under-investigated chemical processes, such as unknown electron-catalyzed reactions (74).
To comprehensively study the transferability of OrbNet-Equi on complex underexplored main-group chemical spaces, we evaluate OrbNet-Equi/SDC21 on the challenging, community-recognized benchmark collection of the General Thermochemistry, Kinetics, and Noncovalent Interactions (GMTKN55) (75) datasets (Fig. 6). Prediction error statistics on the GMTKN55 benchmark are reported with three filtration schemes. First, we evaluate the WTMAD (weighted mean absolute deviation) error metrics (Materials and Methods) on reactions that only consist of neutral and closed-shell molecules with chemical elements CHONFSCl (Fig. 6A), as is supported by an Atomistic ML–based potential method, ANI-2x (76), which is trained on large-scale DFT data. OrbNet-Equi/SDC21 predictions are found to be highly accurate on this subset, as seen from the WTMAD with respect to CCSD(T) being on par with the DFT methods on all five reaction classes and significantly outperforming ANI-2x and the GFN family of semiempirical QM methods (45, 69). It is worth noting that OrbNet-Equi/SDC21 uses a much smaller number of training samples than the ANI-2x training set, which signifies the effectiveness of combining physics-based and ML-based modeling.
The second filtration scheme includes reactions that consist of closed-shell (but can be charged) molecules with chemical elements that have appeared in the SDC21 training dataset (Fig. 6B). Although all chemical elements and electronic configurations in this subset are contained in the training dataset, we note that unseen types of physical interactions or bonding types are included, such as in alkali metal clusters from the ALK8 subset (75) and short strong hydrogen bonds in the AHB21 subset (77). Therefore, assessments of OrbNet-Equi with this filtration strategy reflect its performance on cases where examples of atom-level physics are provided, but the chemical compositions are largely unknown. Despite this fact, the median WTMADs of OrbNet-Equi/SDC21 are still competitive with DFT methods on the tasks of small-system properties, large-system properties, and intramolecular interactions. On reaction barriers and intermolecular NCIs, OrbNet-Equi/SDC21 results fall behind DFT, but they still show improvements against the GFN-xTB baseline and match the accuracy of GFN2-xTB, which is developed with physic-based schemes to improve the descriptions on NCI properties against its predecessor GFN-xTB.
The last scheme includes all reactions in the GMTKN55 benchmarks containing chemical elements and spin states never seen during training (Fig. 6C), which represents on the most stringent test and reflects the performance of OrbNet-Equi/SDC21 when being indiscriminately deployed as a quantum chemistry method. When evaluated on the collection of all GMTKN55 tasks (Fig. 6, total panel), OrbNet-Equi/SDC21 maintains the lowest median WTMAD among methods considered here that can be executed at the computational cost of semiempirical QM calculations. Moreover, we note that failure modes on a few highly extrapolative subsets can be identified to diagnose cases that are challenging for the QM model used for featurization (SI Appendix, Table S7). For example, the fact that predictions are inaccurate on the W4 (Weizmann-4)-11 subset of atomization energies (78) and the G21EA subset of electron affinities (73) parallels the absence of an explicit treatment of triplet or higher-spin species within the formulation of the GFN family of tight-binding models. On the population level, the distribution of prediction WTMADs across GMTKN55 tasks also differs from that of GFN2-xTB, which implies that further incorporating physics-based approximations into the QM featurizer can complement the ML model, and thus, the accuracy boundary of semiempirical methods can be pushed to a regime where no known physical approximation is feasible.
Discussion
We have introduced OrbNet-Equi, a QM-informed geometric deep learning framework for learning molecular or material properties using representations in the atomic orbital basis. OrbNet-Equi shows excellent data efficiency for learning related to both energy-space and real-space properties, expanding the diversity of molecular properties that can be modeled by QM-informed ML. Despite only using readily available small-molecule libraries as training data, OrbNet-Equi offers an accuracy alternative to DFT methods on comprehensive main-group quantum chemistry benchmarks at a computation speed on par with semiempirical methods, thus offering a possible replacement for conventional ab initio simulations for general-purpose downstream applications. For example, OrbNet-Equi could immediately facilitate applications, such as screening electrochemical properties of electrolytes for the design of flow batteries (43) and performing accurate direct or hybrid QM/MM simulations for reactions in transition-metal catalysis (42, 79). The method can also improve the modeling for complex reactive biochemical processes (80) using multiscale strategies that have been demonstrated in our previous study (44), while conventional ab initio reference calculations can be prohibitively expensive even on a minimal subsystem.
The demonstrated transferability of OrbNet-Equi to seemingly dissimilar chemical species identifies a promising future direction of improving the accuracy and chemical space coverage through adding simple model systems of the absent types of physical interactions to the training data, a strategy that is consistent with using synthetic data to improve ML models (81), which has been demonstrated for improving the accuracy of DFT functionals (29). Additionally, OrbNet-Equi may provide valuable perspectives for the development of physics-based QM models by relieving the burden of parameterizing Hamiltonian parameters against specific target systems, potentially expanding their design space to higher-energy scales without sacrificing model accuracy. Because the framework presented here can be readily extended to alternative quantum chemistry models for either molecular or material systems, we expect OrbNet-Equi to broadly benefit studies in chemistry, materials science, and biotechnology.
Materials and Methods
The Unitary N-Body Tensor Equivariant Network Neural Network.
This section introduces Unitary N-Body Tensor Equivariant Network (UNiTE), the neural network model developed for the OrbNet-Equi method to enable learning equivariant maps between the input atomic orbital features and the property predictions . Given the inputs T, UNiTE first generates initial representations through its diagonal reduction module (Materials and Methods). Then, UNiTE updates the representations with t1 stacks of block convolution (Materials and Methods), message-passing (Materials and Methods), and point-wise interaction (Materials and Methods) modules, followed by t2 stacks of point-wise interaction modules. A pooling layer (Materials and Methods) outputs predictions using the final representations at as inputs.
is a stack of atom-wise representations (i.e., for a molecular system containing d atoms, ). The representation for the Ath atom, , is a concatenation of neurons that are associated with irreducible representations of group . Each neuron in is identified by a channel index , a “degree” index , and a “parity” index . The neuron is a vector of length and transforms as the lth irreducible representation of group (i.e., , where ⊕ denotes a vector concatenation operation and ). We use Nlp to denote the number of neurons with degree l and parity p in and to denote the total number of neurons in .
For a molecular/material system with atomic coordinates , the following equivariance properties with respect to isometric Euclidean transformations are fulfilled for any input gauge-invariant and Hermitian operator ; for all allowed indices ,
●translation invariance:
[8] |
where is an arbitrary global shift vector;
●rotation equivariance:
[9] |
for , where denotes a rotation matrix corresponding to standard Euler angles α, β, γ; and
●parity inversion equivariance:
[10] |
The initial vector representations are generated by decomposing diagonal subtensors of the input T into a spherical-tensor representation without explicitly solving tensor factorization based on the tensor product property of group . The intuition behind this operation is that the diagonal subtensors of T can be viewed as isolated systems interacting with an effective external field whose rotational symmetries are described by the Wigner–Eckart Theorem (82), which links tensor operators to their spherical counterparts and applies here within a natural generalization. Each update step is composed of 1) block convolution, 2) message passing, and 3) point-wise interaction modules, which are all equivariant with respect to index permutations and basis transformations. In an update step , each off-diagonal block of T corresponding to a pair of atoms is contracted with . This block-wise contraction operation can be interpreted as performing local convolutions using the blocks of T as convolution kernels and therefore, is called the block convolution module. The output block-wise representations are then passed into a message-passing module, which is analogous to a message-passing operation on edges in graph neural networks (83). The message-passing outputs are then fed into a point-wise interaction module with the previous step representation to finish the update . The point-wise interaction modules are constructed as a stack of multilayer perceptrons (MLPs), Clebsch–Gordan product operations, and skip connections. Within those modules, a matching layer assigns the channel indices of to indices of the atomic orbital basis.
We also introduce a normalization layer termed equivariant normalization (EvNorm) (Materials and Methods) to improve training and generalization of the neural network. EvNorm normalizes scales of the representations h while recording the direction-like information to be recovered afterward. EvNorm is fused with a point-wise interaction module through first applying EvNorm to the module inputs, then using an MLP to transform the normalized frame-invariant scale information, and finally, multiplying the recorded direction vector to the MLP’s output. Using EvNorm within the point-wise interaction modules is found to stabilize training and eliminate the need for tuning weight initializations and learning rates across different tasks.
The explicit expressions for the neural network modules are provided for quantum operators being one-electron operators, and therefore, the input tensors T are a stack of matrices (i.e., order 2 tensors). Without loss of generality, we also assume that T contains only one feature matrix. Additional technical aspects regarding the case of multiple input features, the inclusion of geometric descriptors, and implementation details are discussed in SI Appendix, section S1. The proofs regarding equivariance and theoretical generalizations to order N tensors are provided in SI Appendix, section S2.
The diagonal reduction module.
We define the shorthand notations and to index atomic orbitals. The initialization scheme for is based on the following proposition; for each diagonal block of T, , defined for an on-site atom pair ,
[11] |
there exists a set of T-independent coefficients such that the following linear transformation ψ,
[12] |
is injective and yields that satisfies equivariance (Eqs. 8–10).
The existence of Q is discussed in SI Appendix, Corollary S3. For the sake of computational feasibility, a physically motivated scheme is employed to tabulate Q and produce order 1 equivariant embeddings using on-site three-index overlap integrals :
[13] |
where are the atomic orbital basis and are auxiliary Gaussian-type basis functions defined as (for conciseness, at )
[14] |
where is a normalization constant such that following standard conventions (84). For numerical experiments considered in this work, the scale parameters γ are chosen as (in atomic units)
adheres to equivariance constraints due to its relation to Clebsch–Gordan coefficients (82). Note that the auxiliary basis is independent of the atomic numbers; thus, the resulting are of equal length for all chemical elements. can be efficiently generated using electronic structure programs, here done with ref. 85. The resulting in explicit form are
are then projected by learnable linear weight matrices such that the number of channels for each (l, p) matches the model specifications. The outputs are regarded as the initial representations to be passed into other modules.
The block convolution module.
In an update step , subblocks of T are first contracted with a stack of linearly transformed order 1 representations :
[15] |
which can be viewed as a one-dimensional convolution between each block (as convolution kernels) and the (as the signal) in the ith channel where is the convolution channel index. The block convolution produces block-wise representations for each block index . ρi is called a matching layer at atom A and channel i, defined as
[16] |
are learnable linear weight matrices specific to each degree index l, where Ml is the maximum principle quantum number for shells of angular momentum l within the atomic orbital basis used for featurization. The operation maps the feature dimension to valid atomic orbitals by indexing using , the principle quantum numbers of atomic orbitals μ for atom type zA.
The message-passing module.
Block-wise representations are then aggregated into each atom index A by summing over the indices B, analogous to a “message passing” between nodes and edges in common realizations of graph neural networks (83),
[17] |
up to a nonessential symmetrization and inclusion of point-cloud geometrical terms (SI Appendix, Eq. S5). in Eq. 17 are scalar-valued weights parameterized as SE(3) (the Special Euclidean Group in 3 dimensions)-invariant multihead attentions:
[18] |
where denotes an element-wise (Hadamard) product, and
[19] |
where denotes a two-layer MLP, are learnable linear functions, and denotes an attention head (one value in ). is chosen as Morlet wavelet basis functions:
[20] |
[21] |
where are learnable linear functions and γk are learnable frequency coefficients initialized as , where . Similar to the scheme proposed in transformers (40), the attention mechanism Eq. 18 improves the network capacity without increasing memory costs as opposed to explicitly expanding T.
The aggregated message is combined with the representation of current step through a point-wise interaction module (Materials and Methods) to complete the update .
EvNorm.
We define , where and are given by
[22] |
where denotes taking a neuron-wise regularized L2 norm:
[23] |
and are mean and variance estimates of the invariant content that can be obtained from either batch or layer statistics as in normalization schemes developed for scalar neural networks (86, 87); βklp are positive, learnable scalars controlling the fraction of vector scale information from h to be retained in , and ϵ is a numerical stability factor. The EvNorm operation Eq. 22 decouples h to the normalized frame-invariant representation suitable for being transformed by an MLP and a “pure-direction” that is later multiplied to the MLP-transformed normalized invariant content to finish updating h. Note that in Eq. 22, is always a fixed point of the map , and the vector directions information h is always preserved.
The point-wise interaction module and representation updates.
A point-wise interaction module (Eqs. 24–26) nonlinearly updates the atom-wise representations through , which consist of coupling another -equivariant representation g with and performing normalizations. In Eqs. 24–26, are Clebsch–Gordan coefficients of group is a Kronecker delta function, and and denote MLPs acting on the feature (nlp) dimension. and correspond to learnable linear weight matrices specific to the update step t and each (l, p).
[24] |
[25] |
[26] |
For , the updates are performed by combining with the aggregated messages from step t:
[27] |
where is called a reverse matching layer defined as
[28] |
[29] |
The operation maps the atomic orbital dimension in to a feature dimension with fixed length Ml using as the indices and flattens the outputs into shape . are learnable linear weight matrices to project the outputs into the shape of .
For , the updates are based on local information:
[30] |
Pooling layers and training.
A programmed pooling layer reads out the target prediction after the representations are updated to the last step . Pooling operations employed for obtaining main numerical results are detailed in SI Appendix, section S1.E; hyperparameter, training, and loss function details are provided in SI Appendix, section S4. As a concrete example, the dipole moment vector is predicted as , where is the 3D coordinate of atom A and atomic charges qA and atomic dipoles are predicted using scalar (l = 0) and Cartesian-coordinate vector (l = 1) components of , respectively.
QM-Informed Featurization Details and Gradient Calculations.
The QM-informed representation employed in this work is motivated by a pair of our previous works (26, 88), but in this study, the features are directly evaluated in the atomic orbital basis without the need of heuristic postprocessing algorithms to enforce rotational invariance.
In particular, this work (as well as refs. 26 and 88) constructs features based on the GFN-xTB semiempirical QM method (45). As a member of the class of mean-field quantum chemical methods, GFN-xTB centers around the self-consistent solution of the Roothaan–Hall equations:
[31] |
All boldface symbols are matrices represented in the atomic orbital basis. For the particular case of GFN-xTB, the atomic orbital basis is similar to STO-6G and comprises a set of hydrogen-like orbitals. C is the molecular orbital coefficient, which defines , and is a diagonal eigenvalue matrix of the molecular orbital energies. S is the overlap matrix and is given by
[32] |
where μ and ν index the atomic orbital basis . F is the Fock matrix and is given by
[33] |
H is the one-electron integral including electron–nuclear attraction and electron kinetic energy. G is the two-electron integrals comprising the electron–electron repulsion. Approximation of G is the key task for self-consistent field methods, and GFN-xTB provides an accurate and efficient tight-binding approximation for G. Finally, P is the (one electron–reduced) density matrix and is given by
[34] |
is the number of electrons, and a closed-shell singlet ground state is assumed for simplicity. Eqs. 31 and 33 are solved for P. The electronic energy E is related to the Fock matrix by
[35] |
The particular form of the GFN-xTB electronic energy can be found in ref. 45.
UNiTE is trained to predict the quantum chemistry properties of interest based on the inputs , P, S, with possible extensions (e.g., the energy-weighted density matrices). For the example of learning the DFT electronic energy with the delta-learning training strategy,
[36] |
Note that F, P, S, and H all implicitly depend on the atomic coordinates x and charge/spin-state specifications.
In addition to predicting E, it is also common to compute its gradient with respect to atomic nuclear coordinates x to predict the forces used for geometry optimization and molecular dynamics simulations. We directly differentiate the energy Eq. 36 to obtain energy-conserving forces. The partial derivatives of the UNiTE energy with respect to F, P, S, and H are determined through automatic differentiation. The resulting forces are computed through an adjoint approach developed in appendix D of our previous work (88), with the simplification that the Symmetry-Adapted Atomic Orbital transformation matrix X is replaced by the identity.
Dataset and Computational Details.
Training datasets.
The molecule datasets used in Performance on Benchmark Datasets and Accurate Modeling for Electron Densities were all previously published. Following section 2.1 of ref. 61, the 2,291 BFDb-SSI samples for training and testing are selected as the side-chain–side-chain dimers in the original BFDb-SSI dataset that contain 25 atoms and no sulfur element to allow for comparisons among methods.
The SDC21 dataset used for training the OrbNet-Equi/SDC21 model described in Transferability on Downstream Tasks is collected from several publicly accessible sources. First, 11,827 neutral SMILES (simplified molecular-input line-entry system) strings were extracted from the ChEMBL database (89). For each SMILES string, up to four conformers were generated and optimized at the GFN-xTB level. Nonequilibrium geometries of the conformers were generated using either normal mode sampling (90) at 300 K or ab initio molecular dynamics for 200 fs at 500 K in a ratio of 50/50%, resulting in a total of 178,836 structures. An additional 2,549 SMILES strings were extracted from ChEMBL, and random protonation states for these were selected using Dimorphite-DL (91) as well as another 2,211 SMILES strings, which were augmented by adding randomly selected salts from the list of common salts in the ChEMBL Structure Pipeline (92). For these two collections of modified ChEMBL SMILES strings, nonequilibrium geometries were created using the same protocol described earlier, resulting in 21,141 and 27,005 additional structures for the two sets, respectively. To compensate for the bias toward large drug-like molecules, ∼45,000 SMILES strings were enumerated using common bonding patterns, from which 9,830 conformers were generated from a randomly sampled subset. Lastly, molecules in the BFDb-SSI and JSCH-2005 datasets were added to the training dataset (60, 93). In total, the dataset consists of 237,298 geometries spanning the elements C, O, N, F, S, Cl, Br, I, P, Si, B, Na, K, Li, Ca, and Mg. For each geometry, DFT single-point energies were calculated on the dataset at the ωB97X-D3/def2-TZVP level of theory in Entos Qcore version 0.8.17 (85, 94, 95). We additionally filtered the geometries for which DFT calculation failed to converge or broken bonds between the equilibirum and nonequilibrium geometries are detected, resulting in 235,834 geometries used for training the OrbNet-Equi/SDC21 model.
Electronic structure computational details.
The dipole moment labels for the QM9 dataset used in Performance on Benchmark Datasets were calculated at the B3LYP level of DFT theory with def2-TZVP AO basis set to match the level of theory used for published QM9 labels using Entos Qcore version 1.1.0 (85, 95, 96). The electron density labels for QM9 and BFDb-SSI were computed at the ωB97X-D3/def2-TZVP level of DFT theory using def2-TZVP-JKFIT (97) for Coulomb and Exchange fitting, also as the electron charge density expansion basis . The density expansion coefficients d are calculated as
[37] |
where are AO basis indices and are density-fitting basis indices. Note that γ stands for the combined index in Eq. 7. P is the DFT AO density matrix, is the density-fitting basis overlap matrix, and are three-index overlap integrals between the AO basis and the density-fitting basis .
Benchmarking details and summary statistics.
For the mean L1 electronic density error over the test sets reported in Accurate Modeling for Electron Densities, we used 291 dimers as the test set for the BFDb-SSI dataset and 10,000 molecules as the test set for the QM9 dataset following the literature (61, 62). for each molecule in the test sets is computed using a 3D cubic grid of voxel spacing Bohr for the BFDb-SSI test set and voxel spacing Bohr for the QM9 test set, both with cutoffs at . We note that two baseline methods used slightly different normalization conventions when computing the dataset-averaged L1 density errors : 1) computing for each molecule and normalizing over the number of molecules in the test set (62) or 2) normalizing over the total number of electrons in the test set (61). We found that the average computed using normalization 2 is higher than normalization 1 by around 5% for our results. We follow their individual definitions for average for the quantitative comparisons described in the text: that is, using scheme 1 for QM9 but scheme 2 for BfDB-SSI.
For downstream task statistics reported in Fig. 5 and SI Appendix, Table S5, the results on the Hutchison dataset in Fig. 5A are calculated as the R2 correlation coefficients comparing the conformer energies of multiple conformers from a given model with the energies from DLPNO-CCSD(T). The median R2 vales in SI Appendix, Table S5 with respect to both DLPNO (domain-based local pair natural orbital)-CCSD(T) and ωB97X-D3/def2-TZVP are calculated over the R2 values for every molecule, and error bars are estimated by bootstrapping the pool of molecules. The error bars for TorsionNet500 and s66 × 10 are computed as 95% CIs. Geometry optimization experiments are performed through relaxing the reference geometries until convergence. Geometry optimization accuracies in Fig. 5D and SI Appendix, Table S5 are reported as the symmetry-corrected rmsd of the minimized geometry vs. the reference level of theory (ω B97X-D3/def2-TZVP) calculated over molecules in the benchmark set. Additional computational details for this task are provided in SI Appendix, section S3.E.
For the GMTKN55 (general main group thermochemistry, kinetics and noncovalent interactions) benchmark dataset collection, the reported CCSD(T)/complete basis set results are used as reference values. The weighted absolute deviation (WTAD) scores for producing Fig. 6 are defined similar to the updated weighted mean absolute deviation (WTMAD-2) in ref. 75 but computed for each reaction in GMTKN55:
[38] |
for the jth reaction in the ith task subset. Note that the subset-wise WTMAD-2 metric in SI Appendix, Table S7 is given by
[39] |
The overall WTMAD-2 is reproduced by
[40] |
Supplementary Material
Acknowledgments
Z.Q. acknowledges graduate research funding from Caltech and partial support from the Amazon–Caltech AI4Science fellowship. A.A. and T.F.M. acknowledge partial support from the Caltech DeLogi fund, and A.A. acknowledges support from a Caltech Bren professorship. Z.Q. acknowledges Bo Li, Vignesh Bhethanabotla, Dani Kiyasseh, Hongkai Zheng, Sahin Lale, and Rafal Kocielnik for proofreading and helpful comments on the manuscript.
Footnotes
Competing interest statement: A patent application related to this work has been filed. A.S.C., M.W., F.R.M., and T.F.M. are employees of Entos, Inc. or its affiliates. The software used for computing input features and gradients is proprietary to Entos, Inc.
This article is a PNAS Direct Submission.
This article contains supporting information online at https://www.pnas.org/lookup/suppl/doi:10.1073/pnas.2205221119/-/DCSupplemental.
Data Availability
Source data for results described in the text and SI Appendix, the training dataset, code, and evaluation examples have been deposited in Zenodo (https://zenodo.org/record/6568518#.YrtTKHbMK38) (99).
References
- 1.Ling C., A review of the recent progress in battery informatics. npj Comput. Mater 8, 1–22 (2022). [Google Scholar]
- 2.Vamathevan J., et al., Applications of machine learning in drug discovery and development. Nat. Rev. Drug Discov. 18, 463–477 (2019). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Baik M. H., Friesner R. A., Computing redox potentials in solution: Density functional theory as a tool for rational design of redox agents. J. Phys. Chem. A 106, 7407–7412 (2002). [Google Scholar]
- 4.Shoichet B. K., Virtual screening of chemical libraries. Nature 432, 862–865 (2004). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.Cordova M., Wodrich M. D., Meyer B., Sawatlon B., Corminboeuf C., Data-driven advancement of homogeneous nickel catalyst activity for aryl ether cleavage. ACS Catal. 10, 7021–7031 (2020). [Google Scholar]
- 6.Szabo A., Ostlund N. S., Modern Quantum Chemistry (Mineola, Dover, United Kingdom, 1996), pp. 231–239. [Google Scholar]
- 7.Kohn W., Sham L. J., Self-consistent equations including exchange and correlation effects. Phys. Rev. 140, A1133 (1965). [Google Scholar]
- 8.Mackerell A. D. Jr., Empirical force fields for biological macromolecules: Overview and issues. J. Comput. Chem. 25, 1584–1604 (2004). [DOI] [PubMed] [Google Scholar]
- 9.Behler J., Parrinello M., Generalized neural-network representation of high-dimensional potential-energy surfaces. Phys. Rev. Lett. 98, 146401 (2007). [DOI] [PubMed] [Google Scholar]
- 10.Behler J., Perspective: Machine learning potentials for atomistic simulations. J. Chem. Phys. 145, 170901 (2016). [DOI] [PubMed] [Google Scholar]
- 11.Christensen A. S., Bratholm L. A., Faber F. A., O. Anatole von Lilienfeld, FCHL revisited: Faster and more accurate quantum machine learning. J. Chem. Phys. 152, 044107 (2020). [DOI] [PubMed] [Google Scholar]
- 12.Zhang L., Han J., Wang H., Car R., W. E, Deep potential molecular dynamics: A scalable model with the accuracy of quantum mechanics. Phys. Rev. Lett. 120, 143001 (2018). [DOI] [PubMed] [Google Scholar]
- 13.Schütt K., et al., Schnet: A continuous-filter convolutional neural network for modeling quantum interactions. Adv. Neural Inf. Process. Syst. 30, 991–1001 (2017). [Google Scholar]
- 14.Unke O. T., Meuwly M., Physnet: A neural network for predicting energies, forces, dipole moments, and partial charges. J. Chem. Theory Comput. 15, 3678–3693 (2019). [DOI] [PubMed] [Google Scholar]
- 15.Liu Z., et al., Transferable multi-level attention neural network for accurate prediction of quantum chemistry properties via multitask learning. J. Chem. Inf. Model. 61, 1066–1082 (2021). [DOI] [PubMed] [Google Scholar]
- 16.Gasteiger J., Groß J., Günnemann S., “Directional message passing for molecular graphs” in International Conference on Learning Representations (ICLR ) (Technical University of Munich, Munich, Germany, 2020). [Google Scholar]
- 17.Liu Y., et al., Spherical message passing for 3D graph networks. arXiv [Preprint] (2021). https://arxiv.org/abs/2102.05013v2 (Accessed 13 April 2021).
- 18.Schütt K. T., Unke O. T., Gastegger M., Equivariant message passing for the prediction of tensorial properties and molecular spectra. arXiv [Preprint] (2021). https://arxiv.org/abs/2102.03150 (Accessed 13 April 2021).
- 19.Townshend R. J. L., et al., Geometric deep learning of RNA structure. Science 373, 1047–1051 (2021). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Cheng B., Mazzola G., Pickard C. J., Ceriotti M., Evidence for supercritical behaviour of high-pressure liquid hydrogen. Nature 585, 217–220 (2020). [DOI] [PubMed] [Google Scholar]
- 21.Folmsbee D., Hutchison G., Assessing conformer energies using electronic structure and machine learning methods. Int. J. Quantum Chem. 121, e26381 (2020). [Google Scholar]
- 22.Rosenberger D., Smith J. S., Garcia A. E., Modeling of peptides with classical and novel machine learning force fields: A comparison. J. Phys. Chem. B 125, 3598–3612 (2021). [DOI] [PubMed] [Google Scholar]
- 23.Cheng L., Welborn M., Christensen A. S., Miller T. F. III, A universal density matrix functional from molecular orbital-based machine learning: Transferability across organic molecules. J. Chem. Phys. 150, 131103 (2019). [DOI] [PubMed] [Google Scholar]
- 24.Chen Y., Zhang L., Wang H., W. E, Ground state energy functional with hartree–fock efficiency and chemical accuracy. J. Phys. Chem. A 124, 7155–7165 (2020). [DOI] [PubMed] [Google Scholar]
- 25.Li H., Collins C., Tanha M., Gordon G. J., Yaron D. J., A density functional tight binding layer for deep learning of chemical Hamiltonians. J. Chem. Theory Comput. 14, 5764–5776 (2018). [DOI] [PubMed] [Google Scholar]
- 26.Qiao Z., Welborn M., Anandkumar A., Manby F. R., Miller T. F. III, OrbNet: Deep learning for quantum chemistry using symmetry-adapted atomic-orbital features. J. Chem. Phys. 153, 124111 (2020). [DOI] [PubMed] [Google Scholar]
- 27.Li L., et al., Kohn-sham equations as regularizer: Building prior knowledge into machine-learned physics. Phys. Rev. Lett. 126, 036401 (2021). [DOI] [PubMed] [Google Scholar]
- 28.Nagai R., Akashi R., Sugino O., Completing density functional theory by machine learning hidden messages from molecules. npj Comput. Mater 6, 1–8 (2020). [Google Scholar]
- 29.Kirkpatrick J., et al., Pushing the frontiers of density functionals by solving the fractional electron problem. Science 374, 1385–1389 (2021). [DOI] [PubMed] [Google Scholar]
- 30.Welborn M., Cheng L., Miller T. F. III, Transferability in machine learning for electronic structure via the molecular orbital basis. J. Chem. Theory Comput. 14, 4772–4779 (2018). [DOI] [PubMed] [Google Scholar]
- 31.Dick S., Fernandez-Serra M., Machine learning accurate exchange and correlation functionals of the electronic density. Nat. Commun. 11, 3509 (2020). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 32.Low K., Coote M. L., Izgorodina E. I., Inclusion of more physics leads to less data: Learning the interaction energy as a function of electron deformation density with limited training data. J. Chem. Theory Comput. 18, 1607–1618 (2022). [DOI] [PubMed] [Google Scholar]
- 33.Karandashev K., von Lilienfeld O. A., An orbital-based representation for accurate quantum machine learning. arXiv [Preprint] (2021). https://arxiv.org/abs/2112.12877v1 (Accessed 24 March 2022). [DOI] [PubMed]
- 34.Bronstein M. M., Bruna J., Cohen T., Veličković P., Geometric deep learning: Grids, groups, graphs, geodesics, and gauges. arXiv [Preprint] (2021). https://arxiv.org/abs/2104.13478 (Accessed 25 October 2021).
- 35.Cohen T., Welling M., “Group equivariant convolutional networks” in International Conference on Machine Learning(PMLR) (JMLR, New York, NY, 2016), pp. 2990–2999. [Google Scholar]
- 36.Weiler M., Geiger M., Welling M., Boomsma W., Cohen T., 3D steerable CNNs: Learning rotationally equivariant features in volumetric data. arXiv [Preprint] (2018). https://arxiv.org/abs/1807.02547 (Accessed 25 October 2021).
- 37.Kondor R., Lin Z., Trivedi S., Clebsch–Gordan nets: A fully Fourier space spherical convolutional neural network. Adv. Neural Inf. Process. Syst. 31, 10138–10147 (2018). [Google Scholar]
- 38.Anderson B., Hy T. S., Kondor R., “Cormorant: Covariant molecular neural networks” in Advances in Neural Information Processing Systems, Wallach H.et al., Eds. (Curran Associates, Inc., 2019), vol. 32, pp. 14537–14546. [Google Scholar]
- 39.Thomas N., et al., Tensor field networks: Rotation-and translation-equivariant neural networks for 3D point clouds. arXiv [Preprint] (2018). https://arxiv.org/abs/1802.08219 (Accessed 18 November 2020).
- 40.Fuchs F. B., Worrall D. E., Fischer V., Welling M., Se (3)-transformers: 3D roto-translation equivariant attention networks. arXiv [Preprint] (2020). https://arxiv.org/abs/2006.10503 (Accessed 9 December 2020).
- 41.Bannwarth C., et al., Extended tight-binding quantum chemistry methods. Wiley Interdiscip. Rev. Comput. Mol. Sci. 11, e1493 (2021). [Google Scholar]
- 42.Chen L. D., et al., Embedded mean-field theory for solution-phase transition-metal polyolefin catalysis. J. Chem. Theory Comput. 16, 4226–4237 (2020). [DOI] [PubMed] [Google Scholar]
- 43.Janet J. P., Ramesh S., Duan C., Kulik H. J., Accurate multiobjective design in a space of millions of transition metal complexes with neural-network-driven efficient global optimization. ACS Cent. Sci. 6, 513–524 (2020). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 44.Dommer A., et al., #COVIDisAirborne: AI-enabled multiscale computational microscopy of delta SARS-CoV-2 in a respiratory aerosol. bioRxiv [Preprint] (2021). https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8609898/ (Accessed 24 March 2022). [DOI] [PMC free article] [PubMed]
- 45.Grimme S., Bannwarth C., Shushkov P., A robust and accurate tight-binding quantum chemical method for structures, vibrational frequencies, and noncovalent interactions of large molecular systems parametrized for all spd-block elements (Z=1–86). J. Chem. Theory Comput. 13, 1989–2009 (2017). [DOI] [PubMed] [Google Scholar]
- 46.Ramakrishnan R., Dral P. O., Rupp M., von Lilienfeld O. A., Quantum chemistry structures and properties of 134 kilo molecules. Sci. Data 1, 140022 (2014). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 47.Klicpera J., Giri S., Margraf J. T., Günnemann S., Fast and uncertainty-aware directional message passing for non-equilibrium molecules. arXiv [Preprint] (2020). https://arxiv.org/abs/2011.14115v2 (Accessed 13 April 2021).
- 48.Ramakrishnan R., Dral P. O., Rupp M., von Lilienfeld O. A., Big data meets quantum chemistry approximations: The Δ-machine learning approach. J. Chem. Theory Comput. 11, 2087–2096 (2015). [DOI] [PubMed] [Google Scholar]
- 49.Faber F. A., Christensen A. S., Huang B., von Lilienfeld O. A., Alchemical and structural distribution based representation for universal quantum machine learning. J. Chem. Phys. 148, 241717 (2018). [DOI] [PubMed] [Google Scholar]
- 50.Bartók A. P., et al., Machine learning unifies the modeling of materials and molecules. Sci. Adv. 3, e1701816 (2017). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 51.Christensen A. S., Faber F. A., von Lilienfeld O. A., Operators in quantum machine learning: Response properties in chemical space. J. Chem. Phys. 150, 064105 (2019). [DOI] [PubMed] [Google Scholar]
- 52.Veit M., Wilkins D. M., Yang Y., DiStasio R. A. Jr., Ceriotti M., Predicting molecular dipole moments by combining atomic partial charges and atomic dipoles. J. Chem. Phys. 153, 024113 (2020). [DOI] [PubMed] [Google Scholar]
- 53.Huang B., von Lilienfeld O. A., Quantum machine learning using atom-in-molecule-based fragments selected on the fly. Nat. Chem. 12, 945–951 (2020). [DOI] [PubMed] [Google Scholar]
- 54.Yang W., Mortier W. J., The use of global and local molecular parameters for the analysis of the gas-phase basicity of amines. J. Am. Chem. Soc. 108, 5708–5711 (1986). [DOI] [PubMed] [Google Scholar]
- 55.Silva G. L., Ediz V., Yaron D., Armitage B. A., Experimental and computational investigation of unsymmetrical cyanine dyes: Understanding torsionally responsive fluorogenic dyes. J. Am. Chem. Soc. 129, 5710–5718 (2007). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 56.Atz K., Isert C., Böcker M. N. A., Jiménez-Luna J., Schneider G., δ-quantum machine learning for medicinal chemistry. Phys. Chem. Chem. Phys. 24, 10775–10783 (2022). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 57.Christensen A. S., von Lilienfeld O. A., On the role of gradients for machine learning of molecular energies and forces. Mach. Learn. Sci. Technol. 1, 045018 (2020). [Google Scholar]
- 58.Batzner S., et al., E(3)-equivariant graph neural networks for data-efficient and accurate interatomic potentials. Nat. Commun. 13, 2453 (2022). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 59.Chmiela S., et al., Machine learning of accurate energy-conserving molecular force fields. Sci. Adv. 3, e1603015 (2017). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 60.Burns L. A., et al., The biofragment database (BFDb): An open-data platform for computational chemistry analysis of noncovalent interactions. J. Chem. Phys. 147, 161727 (2017). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 61.Fabrizio A., Grisafi A., Meyer B., Ceriotti M., Corminboeuf C., Electron density learning of non-covalent systems. Chem. Sci. (Camb.) 10, 9424–9432 (2019). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 62.Jørgensen P. B., Bhowmik A., Graph neural networks for fast electron density estimation of molecules, liquids, and solids. arXiv [Preprint] (2021). https://arxiv.org/abs/2112.00652 (Accessed 29 March 2022).
- 63.Marqusee S., Baldwin R. L., Helix stabilization by Glu-…Lys+ salt bridges in short peptides of de novo design. Proc. Natl. Acad. Sci. U.S.A. 84, 8898–8902 (1987). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 64.Smith J. S., et al., Approaching coupled cluster accuracy with a general-purpose neural network potential through transfer learning. Nat. Commun. 10, 2903 (2019). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 65.Zheng P., Zubatyuk R., Wu W., Isayev O., Dral P. O., Artificial intelligence-enhanced quantum chemical method with broad applicability. Nat. Commun. 12, 7022 (2021). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 66.Brandenburg J. G., Bannwarth C., Hansen A., Grimme S., B97-3c: A revised low-cost variant of the B97-D density functional method. J. Chem. Phys. 148, 064104 (2018). [DOI] [PubMed] [Google Scholar]
- 67.Grimme S., Hansen A., Ehlert S., Mewes J. M., r2SCAN-3c: A “Swiss army knife” composite electronic-structure method. J. Chem. Phys. 154, 064103 (2021). [DOI] [PubMed] [Google Scholar]
- 68.Rai B. K., et al., Torsionnet: A deep neural network to rapidly predict small-molecule torsional energy profiles with the accuracy of quantum mechanics. J. Chem. Inf. Model. 62, 785–800 (2022). [DOI] [PubMed] [Google Scholar]
- 69.Bannwarth C., Ehlert S., Grimme S., GFN2-xTB — An accurate and broadly parametrized self-consistent tight-binding quantum chemical method with multipole electrostatics and density-dependent dispersion contributions. J. Chem. Theory Comput. 15, 1652–1671 (2019). [DOI] [PubMed] [Google Scholar]
- 70.Smith D. G., Burns L. A., Patkowski K., Sherrill C. D., Revised damping parameters for the d3 dispersion correction to density functional theory. J. Phys. Chem. Lett. 7, 2197–2203 (2016). [DOI] [PubMed] [Google Scholar]
- 71.Risthaus T., Steinmetz M., Grimme S., Implementation of nuclear gradients of range-separated hybrid density functionals and benchmarking on rotational constants for organic molecules. J. Comput. Chem. 35, 1509–1516 (2014). [DOI] [PubMed] [Google Scholar]
- 72.Fogueri U. R., Kozuch S., Karton A., Martin J. M., The melatonin conformer space: Benchmark and assessment of wave function and DFT methods for a paradigmatic biological and pharmacological molecule. J. Phys. Chem. A 117, 2269–2277 (2013). [DOI] [PubMed] [Google Scholar]
- 73.Curtiss L. A., Raghavachari K., Trucks G. W., Pople J. A., Gaussian-2 theory for molecular energies of first-and second-row compounds. J. Chem. Phys. 94, 7221–7230 (1991). [Google Scholar]
- 74.Studer A., Curran D. P., The electron is a catalyst. Nat. Chem. 6, 765–773 (2014). [DOI] [PubMed] [Google Scholar]
- 75.Goerigk L., et al., A look at the density functional theory zoo with the advanced GMTKN55 database for general main group thermochemistry, kinetics and noncovalent interactions. Phys. Chem. Chem. Phys. 19, 32184–32215 (2017). [DOI] [PubMed] [Google Scholar]
- 76.Devereux C., et al., Extending the applicability of the ANI deep learning molecular potential to sulfur and halogens. J. Chem. Theory Comput. 16, 4192–4202 (2020). [DOI] [PubMed] [Google Scholar]
- 77.Lao K. U., Schäffer R., Jansen G., Herbert J. M., Accurate description of intermolecular interactions involving ions using symmetry-adapted perturbation theory. J. Chem. Theory Comput. 11, 2473–2486 (2015). [DOI] [PubMed] [Google Scholar]
- 78.Karton A., Daon S., Martin J. M., W4-11: A high-confidence benchmark dataset for computational thermochemistry derived from first-principles w4 data. Chem. Phys. Lett. 510, 165–178 (2011). [Google Scholar]
- 79.Li B., Xu H., Dang Y., Houk K., Dispersion and steric effects on enantio-/diastereoselectivities in synergistic dual transition-metal catalysis. J. Am. Chem. Soc. 144, 1971–1985 (2022). [DOI] [PubMed] [Google Scholar]
- 80.Lampret O., et al., The roles of long-range proton-coupled electron transfer in the directionality and efficiency of [FeFe]-hydrogenases. Proc. Natl. Acad. Sci. U.S.A. 117, 20520–20529 (2020). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 81.Chen W., Yu Z., Wang Z., Anandkumar A., “Automated synthetic-to-real generalization” in International Conference on Machine Learning (PMLR, 2020), pp. 1746–1756.
- 82.Sakurai J. J., Commins E. D., Modern Quantum Mechanics (Addison-Wesley, Reading, MA, revised ed., 1995). [Google Scholar]
- 83.Gilmer J., Schoenholz S. S., Riley P. F., Vinyals O., Dahl G. E., “Neural message passing for quantum chemistry” in Proceedings of the 34th International Conference on Machine Learning-Volume 70 (PMLR, 2017), pp. 1263–1272. [Google Scholar]
- 84.Dunning T. H., Hay P. J., Gaussian Basis Sets for Molecular Calculations in Methods of Electronic Structure Theory (Springer, 1977), pp. 1–27. [Google Scholar]
- 85.Manby F., et al., entos: A quantum molecular simulation package. ChemRxiv [Preprint] (2019). https://chemrxiv.org/engage/chemrxiv/article-details/60c740e10f50db1f643959db (Accessed 25 October 2021).
- 86.Ioffe S., Szegedy C., “Batch normalization: Accelerating deep network training by reducing internal covariate shift” in International Conference on Machine Learning (JMLR, 2015), pp. 448–456.
- 87.Ba J. L., Kiros J. R., Hinton G. E., Layer normalization. arXiv [Preprint] (2016). https://arxiv.org/abs/1607.06450 (Accessed 25 October 2021).
- 88.Qiao Z., et al., Multi-task learning for electronic structure to predict and explore molecular potential energy surfaces. arXiv [Preprint] (2020). https://arxiv.org/abs/2011.02680 (Accessed 24 March 2022).
- 89.Gaulton A., et al., The ChEMBL database in 2017. Nucleic Acids Res. 45, D945–D954 (2017). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 90.Smith J. S., Isayev O., Roitberg A. E., ANI-1, A data set of 20 million calculated off-equilibrium conformations for organic molecules. Sci. Data 4, 170193 (2017). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 91.Ropp P. J., Kaminsky J. C., Yablonski S., Durrant J. D., Dimorphite-DL: An open-source program for enumerating the ionization states of drug-like small molecules. J. Cheminformatics 11, 14 (2019). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 92.Bento A. P., et al., An open source chemical structure curation pipeline using RDKit. J. Cheminformatics 12, 51 (2020). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 93.Jurečka P., Šponer J., Cerný J., Hobza P., Benchmark database of accurate (MP2 and CCSD(T) complete basis set limit) interaction energies of small model complexes, DNA base pairs, and amino acid pairs. Phys. Chem. Chem. Phys. 8, 1985–1993 (2006). [DOI] [PubMed] [Google Scholar]
- 94.Lin Y. S., Li G. D., Mao S. P., Chai J. D., Long-range corrected hybrid density functionals with improved dispersion corrections. J. Chem. Theory Comput. 9, 263–272 (2013). [DOI] [PubMed] [Google Scholar]
- 95.Weigend F., Ahlrichs R., Balanced basis sets of split valence, triple zeta valence and quadruple zeta valence quality for H to Rn: Design and assessment of accuracy. Phys. Chem. Chem. Phys. 7, 3297–3305 (2005). [DOI] [PubMed] [Google Scholar]
- 96.Lee C., Yang W., Parr R. G., Development of the Colle-Salvetti correlation-energy formula into a functional of the electron density. Phys. Rev. B Condens. Matter 37, 785–789 (1988). [DOI] [PubMed] [Google Scholar]
- 97.Weigend F., Hartree-Fock exchange fitting basis sets for H to Rn. J. Comput. Chem. 29, 167–175 (2008). [DOI] [PubMed] [Google Scholar]
- 98.Huang B., von Lilienfeld O. A., Quantum machine learning using atom-in-molecule-based fragments selected on the fly. Nat. Chem. 12, 945–951 (2020). [DOI] [PubMed] [Google Scholar]
- 99.Qiao Z., et al., OrbNet-Equi: Informing geometric deep learning with electronic interactions to accelerate quantum chemistry. Zenodo. 10.5281/zenodo.6568518. Deposited 20 May 2022. [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
Source data for results described in the text and SI Appendix, the training dataset, code, and evaluation examples have been deposited in Zenodo (https://zenodo.org/record/6568518#.YrtTKHbMK38) (99).