Skip to main content
Chemical Science logoLink to Chemical Science
. 2018 Jan 18;9(8):2261–2269. doi: 10.1039/c7sc04934j

The TensorMol-0.1 model chemistry: a neural network augmented with long-range physics

Kun Yao a, John E Herr a, David W Toth a, Ryker Mckintyre a, John Parkhill a,
PMCID: PMC5897848  PMID: 29719699

graphic file with name c7sc04934j-ga.jpgWe construct a robust chemistry consisting of a nearsighted neural network potential, TensorMol-0.1, with screened long-range electrostatic and van der Waals physics. It is offered in an open-source Python package and achieves millihartree accuracy and a scalability to tens-of-thousands of atoms on ordinary laptops.

Abstract

Traditional force fields cannot model chemical reactivity, and suffer from low generality without re-fitting. Neural network potentials promise to address these problems, offering energies and forces with near ab initio accuracy at low cost. However a data-driven approach is naturally inefficient for long-range interatomic forces that have simple physical formulas. In this manuscript we construct a hybrid model chemistry consisting of a nearsighted neural network potential with screened long-range electrostatic and van der Waals physics. This trained potential, simply dubbed “TensorMol-0.1”, is offered in an open-source Python package capable of many of the simulation types commonly used to study chemistry: geometry optimizations, harmonic spectra, open or periodic molecular dynamics, Monte Carlo, and nudged elastic band calculations. We describe the robustness and speed of the package, demonstrating its millihartree accuracy and scalability to tens-of-thousands of atoms on ordinary laptops. We demonstrate the performance of the model by reproducing vibrational spectra, and simulating the molecular dynamics of a protein. Our comparisons with electronic structure theory and experimental data demonstrate that neural network molecular dynamics is poised to become an important tool for molecular simulation, lowering the resource barrier to simulating chemistry.

1. Introduction

Statistical models from machine learning experienced growing popularity in many areas of chemistry, such as in reducing the cost of simulating chemical systems,113 improving the accuracy of quantum methods,1422 generating force field parameters,23,24 predicting molecular properties2532 and designing new materials.3338 Neural network model chemistries (NNMCs) are one of the most powerful methods among this class of models.3945 They are shown to be capable of generating high quality potential energy surfaces (PESs) with different schemes such as summing over atoms or bonds,4657 many-body expansions5860 and permutation invariant polynomials.6164 They are also used to predict the properties of materials,6573 and even to find new drugs.7479 In spite of their growing popularity, traditional force fields remain more popular than NNMCs, even for screening and reactive applications. This paper develops an open-source, transferable neural network model chemistry called TensorMol-0.1 (Fig. 1). This model hybridizes a NNMC with the physical contributions to molecular energies that are familiar from Molecular Mechanics and corrections to Density Functional Theory (DFT).80 This approach yields a predictable reproduction of physical long-range forces, and features a linear-scaling inductive charge model. The charges do not depend on a quadratic-scaling polarization equation like a Thole-model,81 instead they are not fixed and respond well to geometry changes as we will show with reproduction of IR spectra.

Fig. 1. A schematic graph of TensorMol-0.1. Each element has its own charge network and energy network. The charge network predicts the atomic charges that yield the ab initio dipole moment. An atom index matrix is used to reassemble the molecular energies/dipoles from atom energies/charges. The Behler–Parinello type energy network produces a short-range embedded atomic energy, which is summed with the electrostatic energy and van der Waals energy to predict the total atomization energy of molecules at and away from equilibrium.

Fig. 1

Our group is one of several who have been pursuing transferable and black-box neural network model chemistries.7,16,46,50,54,82,83 The state of the art in this field is progressing rapidly. Readers may not appreciate that a model can achieve chemical accuracy for energies but have uselessly noisy forces. Models that provide energies at equilibrium, and those treating a fixed molecule or stoichiometry, are nowadays reliably produced.50 We will show that TensorMol-0.1 yields usefully accurate predictions of forces out-of-equilibrium by showing the reproduction of infrared spectra that closely approximate our source model chemistry (ωB97X-D, 6-311G**),84 and molecular dynamics. We outline several tricks that are required to ensure the stability of long-time molecular dynamics.

This force model is implemented in an open-source package that uses the TensorFlow tensor algebra system to compute descriptors and forces. The methodology can be used to propagate the dynamics of large molecules (105 atoms) on simple laptop computers. No force field refinement, atom assignment, or other interventions are needed to apply the method to a molecule of interest, so long as the elements are supported. The package is also interfaced with the I-PI path integral package,85 to allow for quantum simulations and enhanced sampling.

2. Methods

The community of neural network model chemistry developers is rapidly improving the accuracy and generality of these reactive force fields.4,48,49,63,82,86 The model developed in this paper includes several components that were the subject of recent developments in other groups.46,49,52,82,87 We will describe the details here from the bottom up, citing prior studies. Our notational convention will be that i, j, k… are the indices of atoms, qi is the charge on atom i, x, y and z are atomic numbers, A, B and C are molecules, and α, β… are the indices of basis functions which are the products of radial and angular functions. If a function depends on all the atomic coordinates of a molecule it will be written as a vector, and those functions which depend on only a few coordinates will be given explicit indices. The energy of TensorMol-0.1 is expressed as a sum of a short-range embedded n-body potential,49 long-range electrostatic potential and van der Waals force:

2. 1

In the above expression Ezi is a Behler–Parinello type energy network for the element z for atom i. This n-body potential takes as its argument Sα, the modified symmetry functions of Isayev and coworkers:82

2. 2
2. 3

Modern machine learning frameworks provide automatic differentiation of tensor algebraic expressions, allowing a force field developer to obtain the gradient of a molecular potential Inline graphic in a single line of code, once the expression for E(R) has been written. An important feature of our code is that this symmetry function is coded within the TensorFlow system,88 so all the parameters of this descriptor can be optimized alongside the network weights to minimize error. Our implementation of the symmetry function employs a list of nearest-pairs and triples within radial cutoffs such that the scaling of the overall network is asymptotically linear. On an ordinary laptop equipped with only a CPU, a force/energy call on 20 000 atoms takes less than a minute.

The second term of our energy expression is the damped-shifted force (DSF) Coulomb energy of Gezelter and coworkers.89 The charges are obtained from a sub-network which reproduces molecular dipole moments. Previous studies90,91 included electrostatic energy by training the networks to learn Hirshfeld charges. Our charge model enforces conservation of total charge by evenly spreading any required neutralizing charge over the entire molecule or unit cell. The damped-shifted force ensures the long range continuity and differentiability of the effective Coulomb potential with smooth cutoffs. We modify the DSF kernel at short range with an “elu” type non-linearity,92 such that the forces within the radius of the Behler–Parinello symmetry function smoothly approach zero, avoiding singularities and interference with the Behler–Parinello many-body potential. The range separation concept has a long history in chemistry whenever two models of a physical force have complementary cost or accuracy range advantages.84 The energy of the DSF kernel is expressed as:

2. 4

where EDSF(original) is the energy of the DSF kernel89 and Rswitch is the short range cutoff for the “elu” kernel. αelu and βelu are chosen so that the value and the gradient of EDSF are continuous at Rswitch. The damped-shifted force is well-suited to being combined with neural network models because it requires no Fourier transformation to treat periodic systems with linear scaling, and maps well onto TensorFlow. The last term is the van der Waals energy, which is calculated by following Grimme’s C6 scheme.80

We employed a two step training approach. First, the charge networks are trained to learn the atom charges that predict the dipole moment. The loss function can be written as follows:

2. 5

After the charge training is converged, we train the energy network. During the energy network training, the weights in charge networks are kept frozen, but they are still evaluated to calculate the electrostatic energy that is added to construct the total energy. Our Behler–Parinello many-body potential also absorbs the shape of the transition between the many-body and electrostatic regions. The learning target for the energy network includes both the DFT energy and DFT force. The loss function for the energy network training is:

2. 6

where ENN is obtained according to eqn (1) and FNN is calculated by taking the gradient of ENN with respect to the coordinates of the atoms. Natom is the number of the atoms in the system and γ is a parameter that controls the portion of force loss. We employ γ = 0.05. We trained two neural networks based on two sets of data (see Table 1). One network (the “water network”) was trained on a database that includes ∼370 000 water clusters with 1 to 21 water molecules. The other network was trained on ∼3 000 000 different geometries of ∼15 000 different molecules that only contain C, H, O and N and up to 35 atoms. Since these 15 000 molecules were sampled randomly from the Chemspider database, we will refer to this network as the “Chemspider network” in the following text. The training geometries were sampled with metadynamics93 and their energies calculated using the QChem94 program and ωB97X-D84 exchange correlation functional and a 6-311G** basis set.

Table 1. Training details and test RMSE of each learning target. The unit of energy RMSE, gradient RMSE and dipole RMSE is kcal mol–1 per atom, kcal mol–1 Å–1 per atom and Debye per atom, respectively.

Water network Chemspider network
Number of training cases 370 844 2 979 162
Training time (days) a 3 10
Energy RMSE 0.054 0.24
Gradient RMSE 0.49 2.4
Dipole RMSE 0.0082 0.024

aTraining was done on a single Nvidia K40 GPU

Each charge network and energy network contains three fully-connected hidden layers with 500 hidden neurons in each layer. For the Chemspider network, a network with three hidden layers, with 2000 hidden neurons in each layer, is used for each charge network and energy network. L2 regularization is used for both networks and dropout95 on the last layer was used for the Chemspider network to prevent overfitting, with a dropout probability of 0.3. We chose a soft-plus function (ln(1.0 + eαx)/α) with α = 100 as the non-linear activation function and used the Adaptive moment solver (Adam)96 to fix the weights of the network. The test sets were separated from the training data by randomly choosing 20% of the molecules at the outset, which were kept independent throughout. Besides water, we will present calculations from molecules strictly absent from either the training or test set.

To obtain linear scaling, TensorMol uses atom neighbor lists within cutoffs. This allows double precision energy, charge and force calculations of up to 24 000 atoms to be executed in less than 90 seconds on a 2015 Intel i7 2.5 GHz MacBook Pro (Fig. 2). Periodic evaluations are achieved by tessellation of a unit cell with summation of the energies of atoms within the cell. Periodic calculations require about three times more wall time to execute. Speedups greater than a factor of two are obtained automatically when using computers with GPUs (Fig. S7) or single-precision calculations.

Fig. 2. Aperiodic timings of an energy, charge and force call for cubic water clusters at a density of 1 g cm–3. The largest ∼60 Angstrom cube is 4× larger than the electrostatic cutoff. The slope of a log–log version of this curve is near unity, indicating the wall-time scaling of TensorMol. Inset figure: the cubic water cluster used for timing containing 1728 water molecules.

Fig. 2

3. Results

The root mean square error (RMSE) on the independent test set of the energy is 0.054 kcal mol–1 atom–1 and the RMSE of the force is 0.49 kcal mol–1 Å–1. The top panel of Fig. 3 is a plot of the potential energy surface (PES) of a water trimer when one of the water molecules is pulled away from the other two. One can see our neural network PES is not only in good agreement with the PES of the target method but is also smooth. To achieve this we use a variation of the soft-plus neuron rather than the rectified linear units that are popular in computer science. The latter train more efficiently, but produce discontinuous forces.

Fig. 3. Top panel: the PES of a water trimer when one water is pulled away from the other two. Bottom panel: the percentage contribution of the Behler–Parrinello atom-wise energy, electrostatic energy and van der Waals energy to the binding energy between the water that is pulled away and the other two waters. The Behler–Parrinello atom-wise energy contributes most of the binding energy at short range and the electrostatic energy is the dominant contribution at long range.

Fig. 3

The bottom panel shows the fractional contribution of each of the three energy components in eqn (1) to the binding energy along the trimer dissociation coordinate. At short range, most of the binding energy is contributed by the n-body neural network potential. When the distance between the monomer and the dimer approaches the cutoff distance of the neural network, the contribution of the neural network potential starts to decrease and the contribution of the electrostatic potential increases. After 6 Å, where the neural network symmetry functions on the atoms in the monomer have no contribution from the dimer, the neural network force drops smoothly to zero and the electrostatic interaction dominates. The small difference in the energy at 7 Å is due to the difference between the Madelung energy given by the learned charges, and the genuine physical cohesive forces at this distance. The dimer and monomer are beyond the symmetry function sensory radius, and so the charges are constant in this region. Future iterations of the charge network will use local-field information to improve this region of the PES. The learned inductive charges are of high quality considering their linear scaling cost. Fig. 4 shows the PES and dipole change of a water dimer when the hydrogen bond is broken by rotating the OH bond. Both the PES and dipole change fit well with the DFT results.

Fig. 4. Top left panel: the PES of breaking a hydrogen bond between two waters by rotating one water around the O–H bond. Top right, bottom left and bottom right panels: change in the x, y and z dipole components during the rotation, respectively. DFT (ωB97X-D/6-311G**) results are shown in dashed orange line and the TensorMol force field results are plotted in solid blue line.

Fig. 4

Given the increased dimensions of the Hessian, it is naturally a more stringent test to reproduce forces and infrared spectra than it is to simply produce energies. The top panel and bottom panel of Fig. 5 show the optimized geometries and IR spectra of a 10 water cluster and a 20 water cluster, respectively, generated using our force field and DFT. Each method uses its own equilibrium geometry, so this also tests the ability of TensorMol-0.1 to reproduce non-covalent geometries. Our force field quantitatively reproduces the IR spectra generated using DFT, both in terms of frequencies and intensities, especially for the water bend modes and inter-monomer modes. The Mean Absolute Error (MAE) of the frequencies in those two regions is 33.2 cm–1 for the 10 water cluster and 16.2 cm–1 for the 20 water cluster. The error is slightly larger in the water OH stretching region with a MAE of 34.2 cm–1 and 13.1 cm–1 for the 10 and 20 water clusters, respectively. This accuracy is comparable to high quality polarizable water force fields.6

Fig. 5. The simulated harmonic IR spectra of a 10 water cluster (top panel) and a 20 water cluster (bottom panel) generated using ωB97X-D/6-311G** (dashed orange line) and the TensorMol force field (solid blue line).

Fig. 5

Compared with traditional force fields, one major advantage of TensorMol is its reactivity. TensorMol is able to simulate a concerted proton transfer in a water hexamer, finding a minimum energy transition path. The PES calculated using a nudged elastic band (NEB) method97 with the TensorMol force field and the PES calculated using DFT are shown in Fig. 6. The barrier height predicted by TensorMol is 29.7 kcal mol–1, which is 6.7 kcal mol–1 lower than the prediction from DFT, which is remarkable considering the dearth of transition structures in the training data. Our method of sampling molecular geometries uses a meta-dynamics procedure described elsewhere,93 so these proton transfers do occur in the training data although extremely infrequently.

Fig. 6. The reaction energy profile converged from a nudged elastic band method along the reaction coordinate of conservative proton transfer in a water hexamer cluster. The reaction coordination is defined as (ROHRini)/(RfinalRini).

Fig. 6

Encouraged by our results from studying water, we developed a force field with applicability across the chemical space spanned by C, N, O and H. The Chemspider dataset that we used to train our force field covers a vast area of chemical space containing 15 thousand different molecules and 3 million geometries. The geometries are generated using a meta-dynamics procedure,98 which ensures that each new geometry is a fresh part of chemical space; energies up to 400kbT are sampled in the data. We describe the details of this meta-dynamics sampling algorithm, which we have found vital to achieving robust and transferable force fields elsewhere.93 The diversity of structures makes learning the Chemspider dataset a much harder task for the TensorMol-0.1 network; the test set RMSE of energy is 0.24 kcal mol–1 atom–1 and the RMSE of force is 2.4 kcal mol–1 atom–1. More importantly, the model usefully reproduces several elements of molecular structures, at and away from equilibrium, for molecules outside its training set. It robustly optimizes the geometries of typical organic molecules to structures that match DFT well, and yields infrared frequencies and intensities that are in good agreement with ab initio calculations. It is a black-box method that does not rely on any specific atom type, connectivity, etc., which one would need to specify in a traditional classical force field. The few proteins we have examined remain stable and near their experimental structures when optimized or propagated at room temperature using the TensorMol-0.1 force field.

Morphine was not included in our training set. The top right panel of Fig. 7 shows the geometry of morphine that is optimized with our force field. The RMSE of the bond lengths predicted by our force field is 0.0067 Å and the RMSE of the angles is 1.04 degrees, compared with the source DFT model chemistry. The upper left panel plots the harmonic IR spectra generated using DFT and the TensorMol force field, at their respective optimized geometries. One can see the IR spectrum generated using our force field is in good agreement with the DFT-generated IR spectrum. The MAE in our force field frequencies is 13.7 cm–1 compared with the DFT frequencies, which is about half of the MAE in the prediction using MMFF9499 (Fig. S3). Fig. 8 shows comparisons of the IR spectra that are generated using these two methods for aspirin, typrosine, caffeine and cholesterol. All four of these molecules were not included in the training set. The MAE in the frequencies predicted by our field is less than 20 cm–1 for all four molecules, compared with the target DFT frequencies. As Fig. S4 shows, the MAE in the frequencies calculated using MMFF94 are 2 to 3 times larger than the MAE in the frequencies calculated using our force field for these four molecules. The intensities calculated using MMFF94 are also qualitatively different to the DFT intensities. The concept of a chemical bond and force constant are not enforced in any way, yet good agreement with DFT is obtained at a tiny fraction of the original cost.

Fig. 7. The geometry of morphine as optimized by TensorMol-0.1 (upper right panel) and its harmonic IR spectra simulated using ωB97X-D/6-311G** (dashed orange line) and the TensorMol force field (solid blue line) (upper left panel). The lower panels show the real-time IR spectra obtained using TensorMol (solid green line), and the DFT results (dashed orange line) (left), and the conservation of energy maintained by the smoothness of the energy (right).

Fig. 7

Fig. 8. Harmonic IR spectra of four different molecules simulated using ωB97X-D/6-311G** (dashed orange line) and TensorMol-0.1 (solid blue line). All these molecules were not included in the training set.

Fig. 8

Traditional harmonic vibrational spectra require quadratic computational effort, which works against the speed advantage of a NNMC. For large systems, one can use the molecular dynamics functionality of TensorMol to simulate infrared spectra, Fourier transforming the dipole–dipole correlation function of conservative Newtonian dynamics, whose cost grows linearly with the size of the system. The lower left panel of Fig. 7 shows the infrared spectrum produced by propagation in TensorMol-0.1, also showcasing the good energy conservation of TensorMol. Unlike when using a traditional force field, in this case it is non-trivial to obtain smoothly differentiable NNMCs. 64-bit precision needs to be used as the network cannot be made too flexible, and smooth versions of the typical rectified linear units need to be used. Our package can be used in this way to simulate IR spectra of large systems with linear cost.

TensorMol-0.1 uses a relatively simple treatment of the electrostatic and van der Waals forces, which we would like to augment in the future with a many-body dispersion scheme.100 However, a main advantage of the approach used by TensorMol-0.1 is that no self-consistent polarization equation is solved even though the charges are inductive, which results in linear scaling and ease of inexpensively calculating the electrostatic energies of even very large molecules. At shorter ranges, non-covalent interactions like hydrogen bonds are dealt with by the Behler–Parinello portion of the network. The Chemspider training data include some examples of dimers and intra-molecular hydrogen bonds. To our surprise, the treatment of the inter-molecular interactions that were not targets for TensorMol-0.1 are satisfactory. Fig. 9 shows the optimized geometries and binding energies of two DNA base pairs calculated using our force field. The target DFT method predicts a binding energy of 18.3 kcal mol–1 for the thymine–adenine (TA) pair and a binding energy of 32.4 kcal mol–1 for the guanine–cytosine (GC) pair. The prediction using our force field is 1.2 kcal mol–1 smaller for the TA pair and 2.0 kcal mol–1 larger for the GC pair relative to DFT.

Fig. 9. The binding energy between DNA base pairs at their optimized geometries, calculated using DFT (ωB97x-D) and TensorMol methods. The difference between the binding energies calculated using DFT and TensorMol is <2 kcal mol–1.

Fig. 9

One holy grail in the field of neural network model chemistries is to simulate biological chemistry without QM-MM or bespoke force fields. Protein simulation also demonstrates several important features of neural network model chemistry: reasonable inter-molecular forces, stability, scalability and generalization far from small-molecule training data. TensorMol-0.1 was not trained on any peptide polymers and includes no biological data of any sort. To our pleasant surprise, even this first iteration of neural network model chemistry is accurate enough to perform rudimentary studies on small proteins. Fig. 10 shows geometries sampled from a 1 picosecond, periodic, 300 K TensorMol dynamics NVT trajectory in explicit solvent. The initial structure (included in the supplementary information) was generated from the PDB structure ; 2MZX using OpenMM’s automatic solvation and hydrogenation scripts,101 but includes nothing but atom coordinates. This short alpha-helix is stable, both in optimizations and dynamics, and the structures sampled during the dynamics simulation superficially resemble the solution NMR structure. Traditional force fields will always be less expensive (by some prefactor) than NNMCs, yet the reactivity advantages of NNMCs and the ease of set up will probably lead to a rapid adoption of these methods in the biological community.

Fig. 10. The left panel shows samples from a 1 picosecond NVT (Nosé) trajectory of solvated 2MZX at 300 K, simulated by our TensorMol force field in explicit water. The right panel is the NMR structure of ; 2MZX from the PDB database.

Fig. 10

4. Discussion and conclusions

We have presented a transferable neural network model chemistry, TensorMol-0.1, with long-range coulombic physics and a short-range n-body potential. The model is integrated in an open-source Python package, which provides many of the types of simulation commonly used in chemistry. The method can be used to scan conformational and chemical space along the singlet neutral potential energy surface with high throughput using bare atomic coordinates.

TensorMol-0.1 is not the final iteration of a neural network model chemistry, although it shows that DFT-quality predictions can be made by models with five orders of magnitude lower cost. Inexpensive post-DFT corrections such as many-body dispersion100 will become even more powerful when integrated with these potentials, opening the door to quantitative treatments of large systems. NNMCs may compete strongly with DFT packages, and provide an interesting complement to QM-MM-type simulations in the near future.

There are several clear paths to extend this work:

• generalize the descriptors to encode other physical atom properties besides charge (spin or polarizability)

• develop accurate descriptors whose cost grows linearly with the number of elements treated

• extend the range of the n-body embedding

• explore the hierarchy of physical detail between force fields and semi-empirical electronic structures

These are the directions of continuing study in our group and others.

Conflicts of interest

There are no conflicts to declare.

Supplementary Material

Footnotes

†Electronic supplementary information (ESI) available. See DOI: 10.1039/c7sc04934j

References

  1. Lopez-Bezanilla A., von Lilienfeld O. A. Phys. Rev. B. 2014;89:235411. [Google Scholar]
  2. Pilania G., Wang C., Jiang X., Rajasekaran S., Ramprasad R. Sci. Rep. 2013;3:2810. doi: 10.1038/srep02810. [DOI] [PMC free article] [PubMed] [Google Scholar]
  3. Schütt K., Glawe H., Brockherde F., Sanna A., Müller K., Gross E. Phys. Rev. B. 2014;89:205118. [Google Scholar]
  4. Bartok A. P., Payne M. C., Kondor R., Csanyi G. Phys. Rev. Lett. 2010;104:136403. doi: 10.1103/PhysRevLett.104.136403. [DOI] [PubMed] [Google Scholar]
  5. Kruglov I., Sergeev O., Yanilkin A., Oganov A. R. Sci. Rep. 2017;7:8512. doi: 10.1038/s41598-017-08455-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
  6. Medders G. R., Götz A. W., Morales M. A., Bajaj P., Paesani F. J. Chem. Phys. 2015;143:104102. doi: 10.1063/1.4930194. [DOI] [PubMed] [Google Scholar]
  7. Medders G. R., Babin V., Paesani F. J. Chem. Theory Comput. 2013;9:1103–1114. doi: 10.1021/ct300913g. [DOI] [PubMed] [Google Scholar]
  8. Reddy S. K., Straight S. C., Bajaj P., Huy Pham C., Riera M., Moberg D. R., Morales M. A., Knight C., Götz A. W., Paesani F. J. Chem. Phys. 2016;145:194504. doi: 10.1063/1.4967719. [DOI] [PubMed] [Google Scholar]
  9. Riera M., Mardirossian N., Bajaj P., Götz A. W., Paesani F. J. Chem. Phys. 2017;147:161715. doi: 10.1063/1.4993213. [DOI] [PubMed] [Google Scholar]
  10. Moberg D. R., Straight S. C., Knight C., Paesani F. J. Phys. Chem. Lett. 2017;8(12):2579–2583. doi: 10.1021/acs.jpclett.7b01106. [DOI] [PubMed] [Google Scholar]
  11. John S. T., Csanyi G. J. Phys. Chem. B. 2017;121(48):10934–10949. doi: 10.1021/acs.jpcb.7b09636. [DOI] [PubMed] [Google Scholar]
  12. Chmiela S., Tkatchenko A., Sauceda H. E., Poltavsky I., Schütt K. T., Müller K.-R. Sci. Adv. 2017;3:e1603015. doi: 10.1126/sciadv.1603015. [DOI] [PMC free article] [PubMed] [Google Scholar]
  13. Mones L., Bernstein N., Csanyi G. J. Chem. Theory Comput. 2016;12:5100–5110. doi: 10.1021/acs.jctc.6b00553. [DOI] [PubMed] [Google Scholar]
  14. Yao K., Parkhill J. J. Chem. Theory Comput. 2016;12:1139–1147. doi: 10.1021/acs.jctc.5b01011. [DOI] [PubMed] [Google Scholar]
  15. Snyder J. C., Rupp M., Hansen K., Blooston L., Müller K.-R., Burke K. J. Chem. Phys. 2013;139:224104. doi: 10.1063/1.4834075. [DOI] [PubMed] [Google Scholar]
  16. Brockherde F., Vogt L., Li L., Tuckerman M. E., Burke K., Müller K.-R. Nat. Commun. 2017;8:872. doi: 10.1038/s41467-017-00839-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
  17. Snyder J. C., Rupp M., Hansen K., Müller K.-R., Burke K. Phys. Rev. Lett. 2012;108:253002. doi: 10.1103/PhysRevLett.108.253002. [DOI] [PubMed] [Google Scholar]
  18. Li L., Snyder J. C., Pelaschier I. M., Huang J., Niranjan U.-N., Duncan P., Rupp M., Müller K.-R., Burke K. Int. J. Quantum Chem. 2016;116:819–833. [Google Scholar]
  19. Li L., Baker T. E., White S. R., Burke K. Phys. Rev. B. 2016;94:245129. [Google Scholar]
  20. Vu K., Snyder J. C., Li L., Rupp M., Chen B. F., Khelif T., Müller K.-R., Burke K. Int. J. Quantum Chem. 2015;115:1115–1128. [Google Scholar]
  21. McGibbon R. T., Taube A. G., Donchev A. G., Siva K., Hernández F., Hargus C., Law K.-H., Klepeis J. L., Shaw D. E. J. Chem. Phys. 2017;147:161725. doi: 10.1063/1.4986081. [DOI] [PubMed] [Google Scholar]
  22. Wu J., Shen L., Yang W. J. Chem. Phys. 2017;147:161732. doi: 10.1063/1.5006882. [DOI] [PMC free article] [PubMed] [Google Scholar]
  23. Fracchia F., Del Frate G., Mancini G., Rocchia W., Barone V. J. Chem. Theory Comput. 2018;14(1):255–273. doi: 10.1021/acs.jctc.7b00779. [DOI] [PMC free article] [PubMed] [Google Scholar]
  24. Li Y., Li H., Pickard IV F. C., Narayanan B., Sen F. G., Chan M. K., Sankaranarayanan S. K., Brooks B. R., Roux B. J. Chem. Theory Comput. 2017;13:4492–4503. doi: 10.1021/acs.jctc.7b00521. [DOI] [PMC free article] [PubMed] [Google Scholar]
  25. Rupp M., Tkatchenko A., Müller K.-R., Von Lilienfeld O. A. Phys. Rev. Lett. 2012;108:058301. doi: 10.1103/PhysRevLett.108.058301. [DOI] [PubMed] [Google Scholar]
  26. Hansen K., Biegler F., Ramakrishnan R., Pronobis W., Von Lilienfeld O. A., Müller K.-R., Tkatchenko A. J. Phys. Chem. Lett. 2015;6:2326–2331. doi: 10.1021/acs.jpclett.5b00831. [DOI] [PMC free article] [PubMed] [Google Scholar]
  27. Sun Y. T., Bai H., Li M.-Z., Wang W. J. Phys. Chem. Lett. 2017;8:3434–3439. doi: 10.1021/acs.jpclett.7b01046. [DOI] [PubMed] [Google Scholar]
  28. Jinnouchi R., Asahi R. J. Phys. Chem. Lett. 2017;8:4279–4283. doi: 10.1021/acs.jpclett.7b02010. [DOI] [PubMed] [Google Scholar]
  29. Ghiringhelli L. M., Vybiral J., Ahmetcik E., Ouyang R., Levchenko S. V., Draxl C., Scheffler M. New J. Phys. 2017;19:023017. [Google Scholar]
  30. Ouyang R., Curtarolo S., Ahmetcik E., Scheffler M. and Ghiringhelli L. M., arXiv preprint arXiv:1710.03319, 2017.
  31. Isayev O., Oses C., Toher C., Gossett E., Curtarolo S., Tropsha A. Nat. Commun. 2017;8:15679. doi: 10.1038/ncomms15679. [DOI] [PMC free article] [PubMed] [Google Scholar]
  32. Grisafi A., Wilkins D. M., Csanyi G. and Ceriotti M., arXiv preprint arXiv:1709.06757, 2017.
  33. Kim E., Huang K., Tomala A., Matthews S., Strubell E., Saunders A., McCallum A., Olivetti E. Sci. Data. 2017;4:170127. doi: 10.1038/sdata.2017.127. [DOI] [PMC free article] [PubMed] [Google Scholar]
  34. Hachmann J., Olivares-Amaya R., Atahan-Evrenk S., Amador-Bedolla C., Sánchez-Carrera R. S., Gold-Parker A., Vogt L., Brockway A. M., Aspuru-Guzik A. J. Phys. Chem. Lett. 2011;2:2241–2251. [Google Scholar]
  35. Hachmann J., Olivares-Amaya R., Jinich A., Appleton A. L., Blood-Forsythe M. A., Seress L. R., Roman-Salgado C., Trepte K., Atahan-Evrenk S., Er S. Energy Environ. Sci. 2014;7:698–704. [Google Scholar]
  36. Isayev O., Fourches D., Muratov E. N., Oses C., Rasch K., Tropsha A., Curtarolo S. Chem. Mater. 2015;27:735–743. [Google Scholar]
  37. Olivares-Amaya R., Amador-Bedolla C., Hachmann J., Atahan-Evrenk S., Sánchez-Carrera R. S., Vogt L., Aspuru-Guzik A. Energy Environ. Sci. 2011;4:4849–4861. [Google Scholar]
  38. Ulissi Z. W., Tang M. T., Xiao J., Liu X., Torelli D. A., Karamad M., Cummins K., Hahn C., Lewis N. S., Jaramillo T. F. ACS Catal. 2017;7:6600–6608. [Google Scholar]
  39. Handley C. M., Popelier P. L. J. Phys. Chem. A. 2010;114:3371–3383. doi: 10.1021/jp9105585. [DOI] [PubMed] [Google Scholar]
  40. Piquemal J.-P., Jordan K. D. J. Chem. Phys. 2017;147:161401. doi: 10.1063/1.5008887. [DOI] [PubMed] [Google Scholar]
  41. Mills K., Spanner M., Tamblyn I. Phys. Rev.A. 2017;96:042113. [Google Scholar]
  42. Malshe M., Raff L., Hagan M., Bukkapatnam S., Komanduri R. J. Chem. Phys. 2010;132:204103. doi: 10.1063/1.3431624. [DOI] [PubMed] [Google Scholar]
  43. Peterson A. A. J. Chem. Phys. 2016;145:074106. doi: 10.1063/1.4960708. [DOI] [PubMed] [Google Scholar]
  44. Cubuk E. D., Malone B. D., Onat B., Waterland A., Kaxiras E. J. Chem. Phys. 2017;147:024104. doi: 10.1063/1.4990503. [DOI] [PubMed] [Google Scholar]
  45. Carpenter B. K., Ezra G. S., Farantos S. C., Kramer Z. C., Wiggins S. J. Phys. Chem. B. 2017 doi: 10.1021/acs.jpcb.7b08707. [DOI] [PubMed] [Google Scholar]
  46. Behler J., Parrinello M. Phys. Rev. Lett. 2007;98:146401. doi: 10.1103/PhysRevLett.98.146401. [DOI] [PubMed] [Google Scholar]
  47. Behler J. Phys. Chem. Chem. Phys. 2011;13:17930–17955. doi: 10.1039/c1cp21668f. [DOI] [PubMed] [Google Scholar]
  48. Shakouri K., Behler J., Meyer J., Kroes G.-J. J. Phys. Chem. Lett. 2017;8:2131–2136. doi: 10.1021/acs.jpclett.7b00784. [DOI] [PMC free article] [PubMed] [Google Scholar]
  49. Behler J. Angew. Chem., Int. Ed. 2017;56:12828–12840. doi: 10.1002/anie.201703114. [DOI] [PubMed] [Google Scholar]
  50. Han J., Zhang L., Car R., E W. Comm Comput Phys. 2018;23:629–639. [Google Scholar]
  51. Khaliullin R. Z., Eshet H., Kühne T. D., Behler J., Parrinello M. Nat. Mater. 2011;10:693–697. doi: 10.1038/nmat3078. [DOI] [PubMed] [Google Scholar]
  52. Gastegger M., Behler J., Marquetand P. Chem. Sci. 2017;8:6924–6935. doi: 10.1039/c7sc02267k. [DOI] [PMC free article] [PubMed] [Google Scholar]
  53. Kobayashi R., Giofré D., Junge T., Ceriotti M., Curtin W. A. Phys. Rev. B: Condens. Matter Mater. Phys. 2017;1:053604. [Google Scholar]
  54. Yao K., Herr J. E., Brown S. N., Parkhill J. J. Phys. Chem. Lett. 2017;8(12):2689–2694. doi: 10.1021/acs.jpclett.7b01072. [DOI] [PubMed] [Google Scholar]
  55. Kolb B., Lentz L. C., Kolpak A. M. Sci. Rep. 2017;7:1192. doi: 10.1038/s41598-017-01251-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
  56. Lubbers N., Smith J. S. and Barros K., arXiv preprint arXiv:1710.00017, 2017.
  57. Khorshidi A., Peterson A. A. Comput. Phys. Commun. 2016;207:310–324. [Google Scholar]
  58. Yao K., Herr J. E., Parkhill J. J. Chem. Phys. 2017;146:014106. doi: 10.1063/1.4973380. [DOI] [PubMed] [Google Scholar]
  59. Manzhos S., Dawes R., Carrington T. Int. J. Quantum Chem. 2015;115:1012–1020. [Google Scholar]
  60. Manzhos S., Yamashita K., Carrington Jr T. Comput. Phys. Commun. 2009;180:2002–2012. [Google Scholar]
  61. Shao K., Chen J., Zhao Z., Zhang D. H. J. Chem. Phys. 2016;145:071101. doi: 10.1063/1.4961454. [DOI] [PubMed] [Google Scholar]
  62. Zhang Z., Zhang D. H. J. Chem. Phys. 2014;141:144309. doi: 10.1063/1.4897308. [DOI] [PubMed] [Google Scholar]
  63. Li J., Chen J., Zhao Z., Xie D., Zhang D. H., Guo H. J. Chem. Phys. 2015;142:204302. doi: 10.1063/1.4921412. [DOI] [PubMed] [Google Scholar]
  64. Conte R., Qu C., Bowman J. M. J. Chem. Theory Comput. 2015;11:1631–1638. doi: 10.1021/acs.jctc.5b00091. [DOI] [PubMed] [Google Scholar]
  65. Ma X., Li Z., Achenie L. E., Xin H. J. Phys. Chem. Lett. 2015;6:3528–3533. doi: 10.1021/acs.jpclett.5b01660. [DOI] [PubMed] [Google Scholar]
  66. Janet J. P., Kulik H. J. Chem. Sci. 2017;8:5137–5152. doi: 10.1039/c7sc01247k. [DOI] [PMC free article] [PubMed] [Google Scholar]
  67. Janet J. P., Kulik H. J. J. Phys. Chem. A. 2017;121(46):8939–8954. doi: 10.1021/acs.jpca.7b08750. [DOI] [PubMed] [Google Scholar]
  68. Häse F., Kreisbeck C., Aspuru-Guzik A. Chem. Sci. 2017;8:8419–8426. doi: 10.1039/c7sc03542j. [DOI] [PMC free article] [PubMed] [Google Scholar]
  69. Bereau T., DiStasio Jr R. A., Tkatchenko A. and von Lilienfeld O. A., arXiv preprint arXiv:1710.05871, 2017.
  70. Faber F. A., Hutchison L., Huang B., Gilmer J., Schoenholz S. S., Dahl G. E., Vinyals O., Kearnes S., Riley P. F., von Lilienfeld O. A. J. Chem. Theory Comput. 2017;13(11):5255–5264. doi: 10.1021/acs.jctc.7b00577. [DOI] [PubMed] [Google Scholar]
  71. Schütt K. T., Arbabzadah F., Chmiela S., Müller K. R., Tkatchenko A. Nat. Commun. 2017;8:13890. doi: 10.1038/ncomms13890. [DOI] [PMC free article] [PubMed] [Google Scholar]
  72. Zhou X.-X., Zeng W.-F., Chi H., Luo C., Liu C., Zhan J., He S.-M., Zhang Z. Anal. Chem. 2017;89(23):12690–12697. doi: 10.1021/acs.analchem.7b02566. [DOI] [PubMed] [Google Scholar]
  73. Timoshenko J., Lu D., Lin Y., Frenkel A. I. J. Phys. Chem. Lett. 2017;8(20):5091–5098. doi: 10.1021/acs.jpclett.7b02364. [DOI] [PubMed] [Google Scholar]
  74. Li J., Cai D. and He X., arXiv preprint arXiv:1709.03741, 2017.
  75. Ramsundar B., Liu B., Wu Z., Verras A., Tudor M., Sheridan R. P., Pande V. J. Chem. Inf. Model. 2017;57:2068–2076. doi: 10.1021/acs.jcim.7b00146. [DOI] [PubMed] [Google Scholar]
  76. Segler M., Preuß M. and Waller M. P., arXiv preprint arXiv:1702.00020, 2017.
  77. Guimaraes G. L., Sanchez-Lengeling B., Farias P. L. C. and Aspuru-Guzik A., arXiv preprint arXiv:1705.10843, 2017.
  78. Wei J. N., Duvenaud D., Aspuru-Guzik A. ACS Cent. Sci. 2016;2:725–732. doi: 10.1021/acscentsci.6b00219. [DOI] [PMC free article] [PubMed] [Google Scholar]
  79. Gómez-Bombarelli R., Duvenaud D., Hernández-Lobato J. M., Aguilera-Iparraguirre J., Hirzel T. D., Adams R. P. and Aspuru-Guzik A., arXiv preprint arXiv:1610.02415, 2016.
  80. Grimme S. J. Comput. Chem. 2006;27:1787–1799. doi: 10.1002/jcc.20495. [DOI] [PubMed] [Google Scholar]
  81. Thole B. T. Chem. Phys. 1981;59:341–350. [Google Scholar]
  82. Smith J. S., Isayev O., Roitberg A. E. Chem. Sci. 2017;8:3192–3203. doi: 10.1039/c6sc05720a. [DOI] [PMC free article] [PubMed] [Google Scholar]
  83. Smith J. S., Isayev O., Roitberg A. E. Sci. Data. 2017;4:170193. doi: 10.1038/sdata.2017.193. [DOI] [PMC free article] [PubMed] [Google Scholar]
  84. Chai J.-D., Head-Gordon M. Phys. Chem. Chem. Phys. 2008;10:6615–6620. doi: 10.1039/b810189b. [DOI] [PubMed] [Google Scholar]
  85. Ceriotti M., More J., Manolopoulos D. E. Comput. Phys. Commun. 2014;185:1019–1026. [Google Scholar]
  86. Deringer V. L., Csanyi G. Phys. Rev. B. 2017;95:094203. [Google Scholar]
  87. Morawietz T., Behler J. J. Phys. Chem. A. 2013;117:7356–7366. doi: 10.1021/jp401225b. [DOI] [PubMed] [Google Scholar]
  88. Abadi M., Agarwal A., Barham P., Brevdo E., Chen Z., Citro C., Corrado G. S., Davis A., Dean J., Devin M., Ghemawat S., Goodfellow I., Harp A., Irving G., Isard M., Jia Y., Jozefowicz R., Kaiser L., Kudlur M., Levenberg J., Mané D., Monga R., Moore S., Murray D., Olah C., Schuster M., Shlens J., Steiner B., Sutskever I., Talwar K., Tucker P., Vanhoucke V., Vasudevan V., Viégas F., Vinyals O., Warden P., Wattenberg M., Wicke M., Yu Y. and Zheng X., TensorFlow: Large-Scale Machine Learning on Heterogeneous Systems, 2015, http://tensorflow.org/.
  89. Fennell C. J., Gezelter J. D. J. Chem. Phys. 2006;124:234104. doi: 10.1063/1.2206581. [DOI] [PubMed] [Google Scholar]
  90. Morawietz T., Sharma V., Behler J. J. Chem. Phys. 2012;136:064103. doi: 10.1063/1.3682557. [DOI] [PubMed] [Google Scholar]
  91. Artrith N., Morawietz T., Behler J. Phys. Rev. B. 2011;83:153101. [Google Scholar]
  92. Clevert D.-A., Unterthiner T. and Hochreiter S., arXiv preprint arXiv:1511.07289, 2015.
  93. Herr J. E., Yao K., McIntyre R., Toth D. and Parkhill J., arXiv preprint arXiv:1712.07240, 2017.
  94. Shao Y., Gan Z., Epifanovsky E., Gilbert A. T., Wormit M., Kussmann J., Lange A. W., Behn A., Deng J., Feng X. Mol. Phys. 2015;113:184–215. [Google Scholar]
  95. Srivastava N., Hinton G. E., Krizhevsky A., Sutskever I., Salakhutdinov R. J. Mach. Learn. Res. 2014;15:1929–1958. [Google Scholar]
  96. Kingma D. and Ba J., arXiv preprint arXiv:1412.6980, 2014.
  97. Henkelman G., Uberuaga B. P., Jónsson H. J. Chem. Phys. 2000;113:9901–9904. [Google Scholar]
  98. Barducci A., Bussi G., Parrinello M. Phys. Rev. Lett. 2008;100:020603. doi: 10.1103/PhysRevLett.100.020603. [DOI] [PubMed] [Google Scholar]
  99. Halgren T. A. J. Comput. Chem. 1996;17:553–586. [Google Scholar]
  100. Tkatchenko A., DiStasio Jr R. A., Car R., Scheffler M. Phys. Rev. Lett. 2012;108:236402. doi: 10.1103/PhysRevLett.108.236402. [DOI] [PubMed] [Google Scholar]
  101. Eastman P., Swails J., Chodera J. D., McGibbon R. T., Zhao Y., Beauchamp K. A., Wang L.-P., Simmonett A. C., Harrigan M. P., Stern C. D., Wiewiora R. P., Brooks B. R., Pande V. S. PLoS Comput. Biol. 2017;13:1–17. doi: 10.1371/journal.pcbi.1005659. [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials


Articles from Chemical Science are provided here courtesy of Royal Society of Chemistry

RESOURCES