Abstract
The accuracy of computational models of water is key to atomistic simulations of biomolecules. We propose a computationally efficient way to improve the accuracy of the prediction of hydration free energies (HFEs) of small molecules: the remaining errors of the physics-based models relative to the experiment are predicted and mitigated by machine learning (ML) as a postprocessing step. Specifically, the trained graph convolutional neural network attempts to identify the “blind spots” in the physics-based model predictions, where the complex physics of aqueous solvation is poorly accounted for, and partially corrects for them.
The strategy is explored for five classical solvent models representing various accuracy/speed trade-offs, from the fast analytical generalized Born (GB) to the popular TIP3P explicit solvent model; experimental HFEs of small neutral molecules from the FreeSolv set are used for the training and testing. For all of the models, the ML correction reduces the resulting root-mean-square error relative to the experiment for HFEs of small molecules, without significant overfitting and with negligible computational overhead. For example, on the test set, the relative accuracy improvement is 47% for the fast analytical GB, making it, after the ML correction, almost as accurate as uncorrected TIP3P. For the TIP3P model, the accuracy improvement is about 39%, bringing the ML-corrected model’s accuracy below the 1 kcal/mol threshold. In general, the relative benefit of the ML corrections is smaller for more accurate physics-based models, reaching the lower limit of about 20% relative accuracy gain compared with that of the physics-based treatment alone. The proposed strategy of using ML to learn the remaining error of physics-based models offers a distinct advantage over training ML alone directly on reference HFEs: it preserves the correct overall trend, even well outside of the training set.
Graphical Abstract

1. Introduction
Atomistic modeling and simulation methods enable a modern molecular approach to biological research,1–4 including structure-based drug design.5 The ability of these methods6 to address biologically relevant problems is largely determined by the accuracy and computational efficiency of the treatment of complex solvation and electrostatic effects in biomolecules surrounded by water. A large variety of water models have been developed, yet none of the current models commonly used in practice is perfect:7–10 various compromises between speed and accuracy had to be made, see, e.g., Ref.11 for a recent review.
Broadly speaking, there are two different approaches to modeling aqueous solvation that are widely used in classical biomolecular simulations at atomic resolution: explicit and implicit solvation.11 The former treats each water molecule individually, at the same resolution as the target biomolecule. One of the simplest, most efficient, and still widely used models of this class is TIP3P,12 with rigid geometry and fixed partial charges on its 3 atoms.
A computationally attractive alternative to the explicit solvation is the implicit solvation, in which all of the solvent is approximated as (infinite) continuum with dielectric and non-polar properties of water.13–23 The relative computational efficiency of the approach is considered to be its main advantage over the explicit solvation. The history of the development of the implicit solvation framework is longer than 100 years: Max Born introduced his model24 for the solvation (hydration) free energy of a single ion in 1920. The generalization of this very simple, yet powerful model – generalized Born (GB) approximation18,25–57 – is the implicit solvation model often used in atomistic molecular dynamics (MD) simulations,58 where it can provide up to several orders of magnitude speed-up of conformational sampling relative to the explicit solvation.59 Needless to say, the GB model is separated from reality by a multitude of approximations,60 each with its own speed-accuracy trade-offs.
The implicit solvation framework is particularly well suited to adding individual physical effects to the base model. For example, the linear response Poisson-Boltzmann (PB) or GB models are manifestly invariant under charge inversion, but real water is not – as a result, the so-called charge hydration asymmetry (CHA)61–68 is missing from these models, but can be re-introduced explicitly.69–76 Likewise, efforts to introduce the “non-polar” effects of real water into the implicit solvation framework have a long history, starting from an empirical, single parameter surface energy term to more sophisticated, multi-parameter models45,77 aimed at capturing the more nuanced physics of solute-solvent interactions, including the hydrophobic effect. Approaches based directly on variational principles78–85 attempt to account for multiple physical effects all at once. Generally, the accuracy gains of implicit solvent models come at a price of reduced computational efficiency, and, often, introduce additional parameters to account for more subtle and diverse physics of hydration. Yet, even multi-parameter models are not perfect, while the lofty goal of capturing it all in a single formula, remains elusive. Breaking the stubborn “accuracy barrier”, while keeping the models computationally tractable, is still a major problem in the field of atomistic modeling and simulations.
The development of the explicit solvent models has followed a similar path in several respects. Over the decades since TIP3P was introduced, multiple accuracy improvements to classical explicit solvent models generally required either better accounting86,87 for the same overall physics already present at the conceptual level of the older model, or an explicit addition of the missing physics into the model, such as the inclusion of electronic polarization effects,88–90 often (but not always91,92) at a substantial additional computational cost.
Still, none of these classical models, explicit or implicit, account for all of the complex physics of liquid water, including various quantum effects. Quantum modeling of tens of thousands of water molecules at room temperature remains out of reach for time-scales relevant for most biomolecular systems and processes. In practical simulations, stubborn errors remain.
It is then not surprising that recent explosive development of machine learning (ML) techniques,93 including deep neural networks (DNNs),94–96 is already making a noticeable impact on this field,97–110 including works aimed directly at improving the accuracy of description of complex solvation effects.111–117 It should be noted that the majority of these recent works combine QM-based methodology with ML, while our interest here is purely classical approaches. Among recent purely ML-based approaches, a featurization algorithm, based on functional class fingerprints, and implemented within the DeepChem ML framework,95 was used in Ref.117 to predict hydration free energies (HFEs) of a diverse set of 642 neutral small molecules available in FreeSolv118 – arguably the largest public database of experimentally measured HFEs. The overall accuracy achieved with this approach was on par with, or even better than, several existing, computationally efficient classical models of solvation. Still, despite several novel features, the resulting accuracy did not reach that of the explicit models such as TIP3P or even some of the GB-based models.56,73 Apparently, the pure neural network based approach struggles with the complex physics of hydration, at least when the training set is under 1000 data points, providing a motivation for combining ML with classical physics-based water models to improve the accuracy of computationally efficient HFE prediction. In fact, in the context of water models, it was recently demonstrated97 that DNN potentials face accuracy limitations in predicting properties not included in the training process.
Physics-guided machine learning (PGML)119–121 is an emerging area of research that aims to utilize physics knowledge in the design and training of ML models to achieve better generalization performance on samples outside of training data. One of the research directions in PGML that has received considerable attention is focused on incorporating various physics-based constraints directly in the process of training ML models to provide additional sources of supervision to the ML models beyond the empirical loss observed on labeled data. This direction has been explored in several scientific applications including lake modeling,122,123 quantum mechanics,124 and solving partial differential equations.125,126 For prediction of HFEs in conjunction with physics-based modeling, ML is often used to train the parameters of the physics-based model, e.g., the force-field.113
Here we propose, and explore in detail, an alternative way to benefit from the Physics while harnessing the power of DNNs: in our approach the physics and ML parts of the prediction pipeline are completely decoupled. Namely, a physics-based model is applied first, while a “pure” DNN (ML) learns and attempts to correct the remaining errors relative to the experiment as a separate, post-processing step. The motivation for this approach is several-fold. First, we expect that it may automatically preserve the correct trends set by the underlying physics model outside of the ML training set. Second, assuming that the physics model is already reasonably accurate, the errors remaining after application of the physics-based treatment should be relatively small, so that the power of ML is not “wasted” on learning what the existing classical physics can already handle well. Note that within this approach, the larger fraction of the computed quantity automatically satisfies various physical constrains imposed by the physics-based model. The relatively small inaccuracies that remain are complex, hard to decompose into distinct independent components, and are overall difficult or even impossible to describe accurately within computationally facile models of classical physics. In contrast, DNNs may be just the right tool to handle these small, but complex and convoluted errors, difficult to attribute to specific bits of the missing physics. Finally, a potential practical benefit of the proposed strategy is that it should be easy to adopt for any physics-based model that predicts the quantity of interest, with no additional, and often non-trivial, effort to integrate the physics-based and ML parts.
We test the ability of the proposed approach to improve the accuracy of prediction of HFEs of small molecules. Accurate HFE prediction is important in its own right:127 this single number incorporates many facets of the complex physics of hydration, being able to predict it correctly is key to many types of computations common to molecular biophysics and computational biology.128 For example, the accuracy of predicting receptor-ligand binding energetics129,130 depends critically on how well the hydration effects are described,131–133 and these are encapsulated in HFEs. It is also important that a reasonably large data set of experimental HFEs is available,118 which makes it possible to train the DNN directly against experiment, as opposed to yet another model.
We apply our approach to several explicit and implicit physics models, with varying computational complexity and accuracy, to determine how much accuracy can be gained using this approach, and how this accuracy improvement depends on the accuracy of the physics model. If the small molecule test shows promise, further exploration of the approach may be warranted in the future.
The extensive “Methods” section begins with an introduction of the overall approach, then describes the choice of physics-based solvent (water) models used here; their details and parameters are given in the Supporting Information. The “Methods” section proceeds to define and describe the neural network and the data sets. Details of the parameter optimization and neural network training follow. The performance of “Physics + ML”, and “ML alone” approaches are presented in “Results and Discussion”, including model performance on molecules well outside of the training set. The overall findings are summarized and discussed further in “Conclusions”, along with limitations of the approach.
2. Methods
Every practical physics-based model for prediction of HFE of molecules has some residual error relative to the ground truth, that is experiment. Here we use this remaining error – the difference between the experimental values and the predictions of a given physics-based model – as the target for ML (DNN) models. The resulting prediction of the ML model is added to the physics-based prediction to reduce the remaining error, Fig. 1.
Figure 1:

Schematic showing our overall approach of using ML to reduce the remaining error between the hydration energies predicted by physics-based models and experiment. The HFE remaining error between physics-based and experimental predictions is fit via ML; this ML correction augments the HFE of the physics-based models, reducing the remaining error to experiment. The physics and the ML parts of the overall workflow are completely separate and independent: the output of the physics-based model becomes the input of the ML (DNN) part (the latter is not a physics-motivated ML potential).
Specifically, our ML models use the L2 loss function to compare ML-predicted corrections to the target to evaluate its performance as it learns. The L2 loss function is defined as:
| (1) |
where N is the number of data points, are the reference data points, are the predicted values, and RMSE is the root-mean-square error relative to the reference. A final trained model predicts the difference between the experiment and physics model predictions to reduce the overall error.
2.1. Physics-based Water Models
The five classical physics-based water models – one explicit and four implicit – used here as baselines for ML improvement are TIP3P, CHA-GB, GBNSR6, IGB5 (GB-OBC), and AASC, listed here in the approximate order of expected accuracy. The last model, AASC,134 is a relatively new addition to the implicit solvation family: it is an approximation of the apparent surface charge at the solute dielectric boundary that provides a description of the electric field around the solute in aqueous solution, allowing for a straightforward estimate of the polar part of the HFE. More details about the models, and their parameters, are provided in the Supporting Information.
The main rationale for choosing this specific subset from the many available classical solvent models is that we aim to test our new approach on a wide spectrum of accuracy/speed trade-offs of solvation models currently used in practice, see Table 1 below. The implicit solvation models alone provide the wide range we seek, but we also need to make sure our conclusions are robust to the type of solvent model, this is why we add an explicit solvent model to the mix. Our choice of the implicit solvent models attempts to illustrate two aspects of the evolution of the framework: accuracy improvements made while keeping the same level of the underlying physics, as in the IGB5(GB-OBC) to GBNSR6 step, and improvements due to the addition of new physics, absent from the previous level of approximation, as in the GB to CHA-GB step. Here, CHA-GB73 is the generalized Born model modified to take into account charge hydration asymmetry (CHA) – non-invariance of the polar part of the solvation energy upon solute charge inversion. Also, while we use IGB5 (GB-OBC) and AASC in conjunction with an early, single-parameter model for the non-polar part of the free energy, a more sophisticated model for non-polar energy is used with GBNSR6 and CHA-GB, see details in the Supporting Information. The choice for the only representative of the explicit solvent class – TIP3P – is further justified below. First, based on available, albeit limited published data,86 we do not expect the widely used, reasonably fast (fixed-charge) explicit solvation models to provide nearly as wide a range of HFE accuracy as do the implicit solvent models, so selecting just one member of the explicit solvent class should be enough for our purposes. While water models other than TIP3P, including special-purpose models,135,136 can often predict water properties more accurately than TIP3P, see, e.g., Ref.11 for a review, it is not obvious137 whether these necessarily provide substantial accuracy gains with respect to our main accuracy metric, which is the RMSE of HFE predictions of small molecules to experiment. At the same time, more accurate water models, such as polarziable ones, would make computing HFEs for a large enough data set much more computationally expensive than is already the case with TIP3P, Table 1. Since our goal is to evaluate our general strategy, rather than evaluating specific water models, we believe that limiting the representative examples to the above five models is appropriate. Note that we are deliberately not considering the wide class of HFE-predictive models that are QM-based, see, e.g., Ref.138 for a comprehensive review of this class of models. The reason for the exclusion of QM-based approaches is two-fold: most importantly, unlike classical models of solvation, which necessarily miss some physics of the process, Quantum Mechanics is capable of predicting molecular properties essentially exactly in principle. Also, from a practical perspective, QM-based models generally represent a different computational complexity class, at least compared to fast classical implicit solvent models.
Table 1:
The accuracy-efficiency range offered by the physics-based water models considered in this work. Shown are the RMSE (kcal/mol) relative to the experiment and approximate simulation time (for TIP3P model see Ref.139) to obtain a converged estimate of the HFE of a small molecule using one CPU core for each of the physics-based water models. The numbers are averages over the entire FreeSolv database.
| Model | RMSE | Approximate Simulation Time |
|---|---|---|
| TIP3P | 1.54 | 2 day/molecule139 |
| AASC | 2.51 | 100 ms/molecule |
| CHA-GB | 1.72 | 30 ms/molecule |
| GBNSR6 | 1.67 | 30 ms/molecule |
| IGB5 | 2.84 | 5 ms/molecule |
2.2. The Neural Network
The GraphConvModel implemented in TensorFlow by DeepChem95 and based on the graph convolutions described by Duvenaud et al.140 is used to predict the difference between experimental values and predictions of the physics models. As input, the model takes graph representations of molecules that have been created from SMILES strings using the ConvMolFeaturizer supplied by DeepChem. This represents each molecule as a graph, with atoms as nodes and bonds as edges. The atoms have 75 features, all of which are one-hot encoding of either atomic number, atom node degree, implicit valence, aromaticity, or number of attached hydrogens.140
The model itself consists of two convolutional layers and an atom-level dense layer with ReLU activation used for each layer as seen in Fig. 2. The first convolutional layer identifies small-scale patterns and starts with a graph convolution that aggregates information about graph nodes and their neighboring nodes. Next, a fraction of randomly chosen output nodes of the graph convolution are ignored in the dropout stage (a regularization technique used to improve the robustness of the model while training141). Note this step is skipped for all layers when the model is used to make predictions for new molecules. Then, graph pooling is applied, which decreases the size of the graph representation while maintaining the most important information. The second convolutional layer captures more high-level structural patterns in the molecule. It takes the output from the first graph pool layer as input and follows the same procedure as the first convolutional layer. The output from the second graph pool layer is used as input to the atom-level dense layer which transforms and combines the information from the convolutional layers into a better representation for regression. The graph gather takes the output of the dense layer after dropout and combines the information from the graph to a fixed-size representation. This representation is then used as input for the regression step which linearly transforms the data to give its final prediction. A diagram of the architecture of the model can be seen in Fig. 2.
Figure 2:

Schematic of the architecture of the GraphConvModel implemented in this paper. The model takes a SMILES string representing the molecule, which has been converted into a graph representation as its input. The model is composed of two convolutional layers and one atom-level dense layer which each use ReLU activation. Each convolutional layer has three main steps: the graph convolution with optimized convolutional layer size, dropout of an optimized fraction of nodes, and graph pooling which transforms the output for the next layer. The dense layer occurs after both convolutional layers and starts with the atom-level dense layer with optimized dense layer size. Dropout of an optimized fraction of nodes occurs next before the graph gather transforms the data into a fixed size. After this, the output of the graph gather is linearly transformed to the final output in the regression step.
2.3. Data sets
The ML-corrected solvation models are trained and evaluated using version 0.52 of the FreeSolv database118 which is found at the following URL: https://github.com/MobleyLab/FreeSolv. This database, described in Table 2, is a collection of experimental HFEs for 642 small neutral molecules. The experimental uncertainties on the HFE values listed in the database range from 0.03 to 1.93 kcal/mol, with a mean of 0.57 kcal/mol; for the majority (459 out of 642) of the molecules listed uncertainty is 0.6 kcal/mol. Higher uncertainties are generally correlated with higher absolute HFEs, so that an average relative error over the entire data set is ~20%. At the extremes, some molecules have relative uncertainty of the measured HFE as high as ~50%. Molecules with |HFE| < 1 kcal/mol were excluded from the relative error estimates above.
Table 2:
FreeSolv database version 0.52: experimental hydration free energies for small neutral molecules.
| Number of molecules in the dataset | 642 |
| Average number of heavy (non-H) atoms | 8 ± 4.5 |
| Largest molecule number of atoms (non-H) | 24 |
| Average hydration free energy, μ ± σ | −3.80 ± 3.85 kcal/mol |
| Hydration free energy range | (−25.47, 3.43) kcal/mol |
| Average experimental uncertainty, μ ± σ | 0.57 ± 0.31 kcal/mol |
| Experimental uncertainty range | (0.03, 1.93) kcal/mol |
| Elements (most to least frequent) | H, C, O, Cl, N, F, S, Br, P, I |
Since the size of the FreeSolv set does not approach that of the typical ML tasks, extra care was taken to ensure that the training and test sets were well-balanced,142 representing the full range of HFEs present in FreeSolv. The dataset was separated into a test set of 80 molecules and the remaining 562 molecules, representing the training and validation sets, respectively. The latter was further divided into 20 subsets of 28 or 29 molecules each used for 20-fold cross-validation. These sets were selected using stratified sampling, with groups each representing a different range of HFEs, to ensure that each partition had a similar distribution of HFE as the whole dataset. For more details see the Supporting Information or the code available on GitHub (see the Supporting Information).
2.4. Hyperparameter optimization
We optimize the hyperparameters for the GraphConvModel to find the best combination of high accuracy and low overfitting in predicting the error relative to the experiment for HFEs. Since our goal is a direct comparison of the performance of our proposed approach across 5 physics-based models of solvation, it makes sense to use one and the same neural network architecture and hyperparameters for all of the ML models. Here we chose to optimize the hyperparameters using TIP3P model as the base to which the DNN correction is added. We chose TIP3P as the target for hyperparameter optimization for the following reasons. First, the TIP3P model is among the most accurate physics-based models of solvation we utilize in this work, Table 1, therefore its HFE predictions are likely to be consistent with the main premise of our approach in that the errors remaining after an application of the physics-based model should be relatively small. Second, the development of TIP3P was completely agnostic to HFEs in FreeSolv – just like many other popular, classical atomistic water models, TIP3P was originally parameterized to describe experimental properties of liquid water at ambient conditions. This is yet another reason for why hyperparameters optimized using TIP3P should generalize well to the other physics models. In Section 3.4 we confirm that our conclusions are robust to the choice of hyperparameters, which suggests that optimizing these for each model individually is not necessary, as it would likely result in insignificant change in performance.
Each set of parameters was evaluated using 20-fold cross-validation where the training set partitions (see previous section) take turns being the validation set while the rest of the molecules make up the train set. For each fold, a model was trained on the training set and then used to make predictions for each molecule in the training set and separately on every molecule in the validation set. These predicted corrections were added to the TIP3P predictions for each molecule, and the root mean square error (RMSE) between the machine learning corrected TIP3P values and experiment was calculated for the validation set and the training set for each fold. The average validation RMSE value for each fold over two rounds of 20-fold cross-validation was used as the primary metric to evaluate each set of parameters. Two rounds of cross-validation were used to reduce random variation due to the stochastic nature of the training. The difference between the average RMSE for the validation set and training set was used as a metric to assess overfitting.
Each generated ML model for correction of TIP3P hydration was optimized over 500 epochs with the ADAM optimizer,143 as it was observed in preliminary testing (results not shown) that this number of epochs led to a reasonably well-converged loss function, while also mitigating the possibility of overfitting. This choice is further justified by the convergence of models trained on final optimized hyperparameters seen in Fig. 3. The batch size used for each of these models was 100, which represents about 20% of the training data. Batch normalization was disabled because its improvements to convergence speed were negligible due to the relatively small training set compared to typical ML tasks.
Figure 3:

Convergence of the loss function demonstrated for 100 separate ML models learning the TIP3P residuals errors for 500 epochs. The models were trained with optimized hyperparameters using the full training set of 562 molecules used for cross-validation. The L2 loss function is evaluated on this full training set for each epoch during training.
Next, the value for dropout used for all three layers, and dense layer size, were optimized in conjunction using two rounds of 20-fold cross-validation. The goal was to reduce overfitting to below 0.3 kcal/mol while maximizing accuracy. The value for dropout is the fraction of nodes in each layer randomly chosen each batch to be ignored during training highlighted as orange in Fig. 2. The dense layer size is the width of channels for the atom-level dense layer of the graph convolutional neural network as seen in green in Fig. 2. While optimizing dense layer size and dropout, two convolutional layers of size 32 were used for the neural network. The size of 32 was chosen to be smaller than the default value of 64 to reduce runtime but still large enough to produce accurate results. As seen in Fig. 4a, dropout was tested from 0 to 0.9 in increments of 0.1 and the dense layer size was tested from 10 to 906 nodes, on a 10-value logarithmic scale. The final parameters chosen were the ones with the lowest average validation RMSE, ensuring that the train-validation RMSE difference was less than 0.3 kcal/mol. The final values for dropout and dense layer size were 0.4 and 27 respectively. The accuracy performance of these final parameters relative to the full hyperparameter space searched is shown by the white squares in Fig. 4a.
Figure 4:

Hyperparameter exploration using cross-validation for (a) dropout and dense layer size, and (b) convolutional layer sizes using optimized values for dropout and dense layer size. In each case, the top and middle graphs show how different combinations of the hyperparameters affect model performance on training and validation data respectively. As seen on vertical color bar to the right, more red colors indicate high RMSE values while more purple colors indicate low RMSE values. The bottom graph shows the difference in RMSE between the training and validation sets meaning that more purple colors indicate a low difference in RMSE and as a result, low overfitting. The white squares indicate the final selected values for those hyperparameters. The thick black lines indicate the overfitting boundary; squares to the left of the black line for (a) and to the lower right for (b) indicate overfitting greater than 0.3 kcal/mol.
After that, the sizes of convolutional layers, shown in purple in Fig. 2, were optimized using two rounds of 20-fold cross-validation. The convolutional layer sizes are the widths of channels for the convolution layers. For this step, the optimized values of dropout and dense layer size were used. As seen in Fig. 4b, the sizes of the two convolutional layers were varied independently, each tested from 2 to 107 nodes, with 10 values growing proportional to x1.8 with x going from 1 to 10. The final values chosen for the sizes of convolutional layers were 53 and 38 for the first and second convolutional layers respectively. These values maximized the accuracy of the ML predictions while ensuring that overfitting remained below 0.3 kcal/mol. Fig. 4b shows how model accuracy and overfitting change as the sizes of the convolutional layers are changed.
In summary, this hyperparameter optimization resulted in the final hyperparameters of a dropout of 0.4, a dense layer size of 27 nodes, and convolutional layers with sizes of 53 and 38, respectively. Once the hyperparameters had been optimized, we used the same hyperparameter values for training ML models to correct all of the physics-based models. Doing so keeps the dimension of the neural net consistent for all of the comparisons made in this work.
2.5. Final training and testing
After final hyperparameters were determined, ML models were trained to correct for errors of each of the physics-based models. For each physics-based model, the target values used in the training were the difference between the experiment and HFE prediction based on that specific physics model. We also trained ML alone to predict HFE directly (technically, this model is trained to predict the difference between the experiment and the Null model assuming all HFEs to be zero); in what follows this model is referred to as “ML alone”. In total, 100 ML models were individually trained to make corrections for each physics-based model individually. In each case, the training set was the 562 molecules used for cross-validation, and the models were evaluated on the 80 molecule test set.
The ML models were evaluated using the RMSE between the experimental values and the final ML corrected physics-based model predictions for the training and test sets. These results, see Table S2 in the SI, were analyzed to gain a better understanding of the variation in performance caused by the stochastic optimization used by the GraphConvModel. Convergence of the loss function over 500 epochs for the 100 trained models for TIP3P is shown in Fig. 3.
One final ML model for each physics-based model was chosen as the ML model with the RMSE of the training set closest to the mean training set RMSE of the 100 models. Results for the final chosen ML models can be seen in Table 3.
Table 3:
Performance of the physics-based hydration models without ML corrections and with final trained ML corrections on the TEST set of 80 molecules and on the full TRAINING set of 562 small neutral molecules. The method for choosing the final ML model for each physic-based model is described in Section 2.5, and the summary of the average performance of ML models can be seen in Table S2 in the SI. The performance metrics – RMSE relative to the experimental HFEs and mean deviation – are in kcal/mol, and the compute speed is relative to the speed of IGB5 model. The Null “physics model” sets all HFE values to zero; thus, the physics + ML results for Null model effectively represent using ML alone.
| TIP3P | AASC | CHA-GB | GBNSR6 | IGB5 | Null (ML alone) | ||
|---|---|---|---|---|---|---|---|
| Data set | relative compute speed | 107 | 20 | 6 | 6 | 1 | N/A |
| TEST | physics model (RMSE) | 1.44 | 2.65 | 1.32 | 1.85 | 3.11 | 5.45 |
| physics + ML (RMSE) | 0.88 | 1.36 | 0.90 | 1.50 | 1.65 | 1.53 | |
| physics model (mean deviation) | −0.19 | 0.16 | 0.03 | 0.09 | 1.45 | −3.90 | |
| physics + ML (mean deviation) | 0.33 | −0.15 | −0.01 | 0.40 | 0.26 | −0.52 | |
| physics model (5% outlier RMSE) | 3.79 | 6.79 | 3.57 | 4.88 | 9.04 | 14.34 | |
| physics + ML (5% outlier RMSE) | 2.74 | 3.71 | 2.77 | 5.23 | 5.60 | 5.23 | |
| TRATNTNG | physics model (RMSE) | 1.56 | 2.49 | 1.77 | 1.64 | 2.80 | 5.40 |
| physics + ML (RMSE) | 0.63 | 0.91 | 0.70 | 0.67 | 0.93 | 0.94 | |
| physics model (mean deviation) | −0.33 | 0.15 | 0.18 | −0.06 | 1.38 | −3.79 | |
| physics + ML (mean deviation) | 0.20 | −0.23 | −0.01 | 0.13 | 0.08 | −0.37 | |
| physics model (5% outlier RMSE) | 4.83 | 7.75 | 5.84 | 5.07 | 8.38 | 14.26 | |
| physics + ML (5% outlier RMSE) | 1.98 | 2.96 | 2.38 | 2.16 | 3.19 | 2.68 | |
Source code for the hyperparameter optimization, the final training and testing of ML models, in addition to all of the saved ML models and data sets used, is available on GitHub (see the SI).
3. Results and Discussion
3.1. Error reduction achieved by Physics + ML
For every physics-based model tested, ML decreases the RMSE to experiment for both training and testing data as seen in Fig. 5a. It is also clear that more accurate physics-based models consistently provide better overall predictive accuracy after ML corrections. In most cases, final RMSE values for both the training and test sets were better for the more accurate physics-based models and worse for the less accurate ones. This highlights the benefit of using more accurate physics-based models for further improvement with ML. However, the relative improvement for more accurate physics-based models was significantly less than that for less accurate models as seen in Fig. 5b. The two least accurate models, IGB5 and AASC, both had over 45% (almost two-fold) relative improvement for test set data while the other more accurate models each had only around 30% relative improvement.
Figure 5:

For every physics model tested, ML reduces the error relative to the experiment for both training and testing data, but the improvements due to ML correction tend to decrease as it is applied to more accurate physics models. Each set of blue (TRAINING) and green (TEST) points with an arrow between them represents a single physics model. (a) Physics + ML model RMSE values plotted against physics model alone. The y=x line (dash) represents no improvement; points below the line show improvement due to ML corrections. (b) Relative improvement due to ML corrections for each of the physics models.
Fig. 6 shows a more detailed view of how the ML-corrected physics models perform compared to the uncorrected models on the test set. As seen in the left column, the ML corrected predictions match up better to experiment and have fewer outliers than uncorrected predictions for each physics model – in particular, the negative trend line bias of IGB5 and AASC is mostly corrected after the ML correction is applied. The center column of Fig. 6 clearly shows that the distribution of error to experiment for each physics model is significantly narrower after ML corrections, although some outliers do persist. Finally, it can be seen in the right column of Fig. 6 that for most, but not all of the molecules in the test set, ML corrections improved the physics model predictions.
Figure 6:

(Previous page.) Error analysis of ML corrections to the five physics-based models of hydration, on the TEST set. (Left:) Scatter plots of predictions of five different physics models before (red) and after (green) ML corrections against experiment. Least squares regression lines are shown as dotted lines. (Center:) The distribution of errors relative to the experiment for five different physics models (unnormalized probability density) before (red) and after (green) their residual errors were predicted and corrected by ML. (Right:) The distribution of how much ML improved predictions (unnormalized probability density), with points to the left of zero indicating that the physics prediction was made worse.
Table 3 shows detailed results for the performance of each physics-based model without and with ML corrections on both the training and test sets. After ML corrections, the two least accurate physics models, IGB5 and AASC, were both within around 0.2 kcal/mol in accuracy (on the test set) from the uncorrected explicit TIP3P model. In both cases, the absolute improvement to the RMSE was over 1.2 kcal/mol and the final test set RMSE was just over half of that for the uncorrected models. A major reason for this significant improvement is that the ML corrections worked well to correct outlier physics model predictions with particularly large errors, greatly reducing the outlier RMSE for the two models seen on the bottom two rows of each section of Table 3. In contrast, the most accurate physics models saw less relative improvement, but the ML corrections were still able to bring TIP3P and CHA-GB RMSE values below 1 kcal/mol. Interestingly, ML-corrected GBNSR6 saw the least improvement and performed comparably to ML-corrected AASC and IGB5 on the test set after ML corrections, despite being a far more accurate physics model than these two. We see at least two possible explanations for the poorer then expected accuracy of GBNSR after the ML corrections. First, recall that the particular ML model chosen is a result of a stochastic optimization. Averaging over multiple optimization outcomes (see Table S2 in the SI) gives a more consistent relative performance ranking of the ML-corrected models, where, on average, GBNSR6 performs better than AASC and IGB5, as expected. In addition, we note that the ZAP9 radii set, used here with GBNSR6, was optimized exclusively using the FreeSolv database and, as a result, may be overfit to the database. Relative overfitting of ML corrections was highest for GBNSR6 at 55%, lowest for CHA-GB at 23%, and remained between 25% and 45% for the other physics models. IGB5 had the highest mean deviation to experiment, but after ML corrections, this systemic error was significantly reduced in the training and test sets as seen in Table 3.
3.2. Error reduction by ML Alone
Based on the accuracy metrics in Table 3, ML alone predictions of HFEs show performance similar to the ML-corrected AASC, GBNSR6, and IGB5 models on the FreeSolv database. The error distribution seen in Fig. 7b is also comparable to those ML-corrected physics models. However, as seen in Fig. 7a, the trend line for ML alone deviates significantly from experiment for molecules with high absolute HFE values. In contrast, the ML-corrected AASC, GBNSR6, and IGB5 models follow the correct trend across the entire range of HFEs in the FreeSolv database as seen in Fig. 6 (left) We expect that ML alone will likely perform significantly worse for molecules that are very different from those in the FreeSolv database; we will test that expectation in the next section.
Figure 7:

(a) Scatter plot of ML alone model predictions for the TEST set against experimental values. The model deviates from experiment near the extremes, as highlighted by the dotted green trend line – linear regression fit y = 0.72x − 0.58, which is significantly off from experiment. (A perfect match to experiment is indicated the y=x grey line). (b) Distribution of the errors of ML alone predictions on the TEST set.
3.3. Application to blocked amino acids
To test how generalizable the ML models trained on small neutral molecules may be, they were used to predict the HFE values for a set of 44 structures of blocked amino acids:73,144 two conformations for each of 20 blocked amino acids and for 2 additional neutralized ASP and GLU analogs. Unfortunately, there is no experimental data for this set, which includes charged species, therefore TIP3P calculated polar solvation values73,144 were used as a reasonable proxy. Note that our goal here is limited to identifying gross errors, if any. As seen in Fig. 8, the ML alone model performs very poorly on this amino acid dataset, predicting HFE values of around −10 kcal/mol for all amino acids, whereas the TIP3P calculated values for this set have a mean of −28.9 and a range of −12.0 to −82.7 kcal/mol. It is expected that the ML model performs poorly on the charged amino acids, with their high absolute HFEs – these were completely absent from the training set. However, it is somewhat surprising that the ML alone model performs as poorly as it does on uncharged amino acids, given that there exist several amino acid analogs in the FreeSolv dataset. A possible reason for this poor performance for the ML alone model is that the HFEs for the uncharged blocked amino acids (−30 to −12 kcal/mol) are not within the typical range of values used in the training SM set (−12 to 3 kcal/mol). Only 15 of the 642 molecules in the FreeSolv database have HFE values outside this typical range, meaning that all of the amino acids, even neutral ones, have HFEs that are outside the normal range of molecules used to train the ML models.
Figure 8:

ML alone fails to preserve the physically correct trend for HFEs of blocked amino acids, while physics with ML corrections follows the correct trend. Shown are (a) ML alone model and (b) CHA-GB before (red x’s) and after (green circles) ML corrections on the blocked amino acid data set. In both cases, the ML models are trained on small molecules in the FreeSolv database and evaluated here on a completely separate set of blocked amino acids, see “Methods”.
As seen in Figure 8a, the ML alone model performs very poorly when in predicting HFEs for the amino acids. This result shows that the ML alone model cannot predict HFEs with any useful accuracy for molecules with HFEs significantly outside the range of HFEs it was trained on.
In contrast, when the ML correction model is applied to predictions of each of the physics models (except TIP3P), the overall correct trend is preserved as seen in Figure 8b and Figure S1 in the SI, which is in stark contrast to the results using ML alone. We refrain from quantifying possible improvement of accuracy that the ML corrections may provide for physics-based HFEs of blocked amino acids. This is mainly because we would have to compare these against available TIP3P calculated values, which have their own error relative to the experiment. A minor technical point is that the reference TIP3P HFE values available to us for the blocked amino-acids are for the polar solvation part of the total hydration energy, rather than the total HFE values that the models were trained for. While the difference between the two is relatively small for these highly polar molecules, which justifies using the polar component to make the qualitative statements above, it is not completely negligible.
3.4. Impact of hyperparameter optimization
So far, the hyperparameters used for all the final ML models were optimized for TIP3P for reasons outlined in “Methods”. Distinct, but computationally inexpensive ML models using these hyperparameters were then trained and evaluated for each of the physics-based models. We find that despite the hyperparameters being optimized for TIP3P, each of the final physics + ML models works to significantly reduce the error to experiment for its respective physics-based model, Table 3. To test how the hyperparameters may impact performance for the different ML-corrected physics models, we re-trained the ML models using two additional hyperparameters sets and compared them to the results based on the previously optimized hyperparameters. For both sets of hyperparameters, the differences in test and training set accuracy compared to the optimized hyperparameters are relatively small as seen in Table S3 of the SI. In general, it appears that a reasonable set of hyperparameter values similar to what is used here works relatively well and does not affect any of the key conclusions made in this work. Based on these results, it does not appear necessary to optimize hyperparameters for each physics model separately. A more detailed analysis of these results can be found in Section 5 of the SI.
4. Conclusions
This work investigates use of ML (DNN) to improve the accuracy of estimating hydration free energies (HFEs) of small molecules by predicting and correcting the errors of physics-based atomistic water models. This strategy is explored for several popular implicit and explicit water models and evaluated mainly on experimental HFEs from the FreeSolv database, arguably the largest publicly available database of small neutral molecules. The graph convolutional deep neural network (DNN) implemented as the GraphConvModel in DeepChem is utilized to try to learn the remaining error that a physics model alone fails to properly account for. A distinct feature of the approach, which sets it apart it from many previous works that combined physics and ML to improve accuracy of HFE (or other molecular property) predictions is a complete separation of the physics and ML (DNN) parts of the overall workflow: the output of the physics-based model becomes the input of the ML part, the latter being a “pure” DNN, not a physics-motivated ML model.110 That is the proposed strategy aims to render unto clear and traceable physics all that is physics’.
We find that for every physics-based water model employed to test our approach, the subsequent ML treatment reduces the RMSE relative to the experiment, improving individual physics-based HFE predictions for most molecules in the test set. The ML treatment also reduces outlier errors for most physics-based models. With the ML correction, even relatively low-accuracy implicit solvation models achieve the final accuracy comparable to the uncorrected explicit solvent TIP3P model, which is orders of magnitude more computationally expensive. A noteworthy trend that emerges from this investigation is that the relative improvement of the physics-based HFE accuracy due to subsequent application of the ML tends to be lower for higher accuracy physics models. Moreover, the amount of relative accuracy improvement does not appears to be sensitive to the type of the physics model; for example, the relative benefit of the ML correction is essentially the same for the explicit solvent model TIP3P and the implicit solvent model CHA-GB, which have very comparable accuracy on the small molecule sets considered here. For all of the physics models, the ML correction carries negligible additional computational expense. A noteworthy feature of our approach is that the remaining error that the DNN has to correct is relatively small. Therefore, the final ML-corrected solution retains most of the “physicality” of the underlying physics-based model, and likely contains only a relatively small amount of “untraceable physics” often attributable to purely DNN solutions.
For practical application, the strategy we propose requires no additional effort to integrate the DNN post-processing correction with the physics-based model. With any of the five physics-based solvent models considered in this work, regardless of their specific implementation, the approach could be called “plug-and-play”, in that the HFE of any new small molecule can be predicted by simply running one of the pre-trained DNN models as a post-processing step, following a completely independent pre-processing by the chosen physics model. For a physics model not yet on the list, hyperparameter re-optimization is not required, which makes the task of using the approach still relatively straightforward. Given predictions for the FreeSolv database by the new physics model, the corresponding DNN correction model can be trained in roughly 30s on a CPU core. We see this computational facility as one of the advantages of the approach, which we hope will encourage further evaluation. The source code for the hyperparameter optimization, the final training and testing of ML models, and all the saved ML models and data sets are available on GitHub, see the Supporting Information.
We have also compared, in some detail, the “physics + ML” with the corresponding “ML alone” approach, the comparison showing several noteworthy trends. First, the computationally inexpensive ML alone predicts HFE with similar RMSE accuracy as several of the physics-based models after the ML correction; these include AASC, IGB5 (GB-OBC), and GBNSR6 (the latter using ZAP9 radii) implicit solvent models. However, the ML alone has significantly worse accuracy than the ML corrected TIP3P and CHA-GB models, the latter two representing the upper edge of the accuracy range of the physics-based solvent models explore here. Second, significant deviations from experimental HFE values become apparent for the “ML alone” solution at the edges of the FreeSolv database: the overall trend line predicted by the ML is clearly off the mark, unlike that of all of the physics-based, or “physics + ML” models tested.
When used on molecules outside of the range of HFEs in the FreeSolv data set, ML alone performs very poorly, while even the least accurate “physics + ML” models reproduce the correct overall trends, even well outside the training range. This comparison emphasizes a key advantage of the approach, in which ML is used only to correct the relatively small error remaining after the physics-based prediction has already been made, independently. The approach guarantees that the correct overall trend (if present in the physics-based model) persists in the final prediction: if the molecule’s HFE is close to the range of HFEs in the training set, the physics-based prediction is likely further finessed by the ML follow-up. Given the above considerations, about the only clear advantage to the ML alone, compared to the physics+ML approach, would be a significantly faster run time in the case of relatively slow physics models such as TIP3P. However, this advantage becomes less obvious in the case of the implicit solvent physics models that are already fast enough.
The proposed approach has several limitations. First, as mentioned above, the benefit of the ML correction for the accuracy of HFE predictions is expected to be minimal or non-existent for molecules that differ significantly from those in the FreeSolv database: for these molecules, the accuracy of the final prediction is close to that of the physics model alone. Without further investigation, this particular limitation is hard to avoid unless the training set is expanded significantly. In the near future, this limitation can be potentially be mitigated somewhat by applying a more sophisticated train/test partitioning of the existing data set of small molecule HFEs. Second, all the HFE values used for the training have experimental uncertaintties, but these uncertainties were not considered when training the ML models. The uncertainty sets a limit for the achievable accuracy, and also means that the ML models may learn systematic experimental error. The concern is mitigated by an observation that the errors of the final predicted HFEs are still appreciably higher than the average experimental uncertainty of the FreSolve set. In the future, one can consider accounting for the experimental uncertainties in the training procedure. Third, some inevitable over-fitting is still present in the final models. The degree of over-fitting varies between the models, for more accurate physics models it is close to, or even smaller than, the target of 30%, but becomes larger for less accurate models, including ML alone. In the future, one can explore setting a tighter over-fitting threshold at the hyper-parameter optimization stage. Forth, a limitation of the specific featurizer (part of DeepChem) employed here, is that it is agnostic to conformation of the molecule, sacrificing exact positional data for generalizability. As a consequence, changes in the HFE due to conformational transitions in the same molecule are only accounted for by the physics part of the procedure. We believe that this specific limitation may also be overcome in the future. Finally, given the trend of lower relative improvement for more accurate physics models, it remains to be seen whether the proposed approach would have a noticeable impact for physics models significantly more accurate than those (TIP3P or CHA-GB) at the upper end of the accuracy range of HFE predictive power explored in this work.
In summary, we have proposed a strategy for improving predictions made by physics-based water models by using deep neural networks to make corrections as an independent post-processing step. The strategy, tested on improving accuracy of hydration free energy predictions, shows promise, making it worthwhile to pursue further exploration of this approach, while also encouraging its adoption for immediate practical applications.
Supplementary Material
Acknowledgement
We thank Virginia Tech Advanced Research Computing for providing the computational resources. This study is supported by NIH R01 GM144596 to A.V.O. and NIH R16GM146633 to N.F.
Footnotes
Supporting Information Available
Details of the physics-based models of solvation; additional tables and figures.
References
- (1).Adcock S; McCammon J Molecular Dynamics: Survey of Methods for Simulating the Activity of Proteins. Chemical Reviews 2006, 106, 1589–1615. [DOI] [PMC free article] [PubMed] [Google Scholar]
- (2).Karplus M; Kuriyan J Molecular dynamics and protein function. Proc Natl Acad Sci U S A 2005, 102, 6679–6685. [DOI] [PMC free article] [PubMed] [Google Scholar]
- (3).Karplus M; McCammon JA Molecular dynamics simulations of biomolecules. Nat Struct Biol 2002, 9, 646–652. [DOI] [PubMed] [Google Scholar]
- (4).Raman EP; MacKerell AD Spatial Analysis and Quantification of the Thermodynamic Driving Forces in Protein-Ligand Binding: Binding Site Variability. J. Am. Chem. Soc 2015, 137, 2608–2621. [DOI] [PMC free article] [PubMed] [Google Scholar]
- (5).Jorgensen WL The Many Roles of Computation in Drug Discovery. Science 2004, 303, 1813–1818. [DOI] [PubMed] [Google Scholar]
- (6).Schlick T; Portillo-Ledesma S; Myers CG; Beljak L; Chen J; Dakhel S; Darling D; Ghosh S, et al. Biomolecular Modeling and Simulation: A Prospering Multidisciplinary Field. Annu Rev Biophys 2021, 50, 267–301. [DOI] [PMC free article] [PubMed] [Google Scholar]
- (7).Jorgensen WL; Tirado-Rives J Potential energy functions for atomic-level simulations of water and organic and biomolecular systems. Proc. Natl. Acad. Sci. U. S. A 2005, 102, 6665–6670. [DOI] [PMC free article] [PubMed] [Google Scholar]
- (8).Guillot B A reappraisal of what we have learnt during three decades of computer simulations on water. J. Mol. Liq 2002, 101, 219–260. [Google Scholar]
- (9).Mark P; Nilsson L Structure and Dynamics of the TIP3P, SPC, and SPC/E Water Models at 298 K. J Phys Chem A 2001, 105, 9954–9960. [Google Scholar]
- (10).Vega C; Abascal JLF Simulating water with rigid non-polarizable models: a general perspective. Phys. Chem. Chem. Phys 2011, 13, 19663–19688. [DOI] [PubMed] [Google Scholar]
- (11).Onufriev AV; Saeed I Water models for biomolecular simulations. Wiley Interdiscip. Rev.: Comput. Mol. Sci 2018, 8, e1347. [Google Scholar]
- (12).Jorgensen WL; Chandrasekhar J; Madura JD; Impey RW; Klein ML Comparison of simple potential functions for simulating liquid water. J Chem Phys 1983, 79, 926–935. [Google Scholar]
- (13).Cramer CJ; Truhlar DG Implicit Solvation Models: Equilibria, Structure, Spectra, and Dynamics. Chem. Rev 1999, 99, 2161–2200, [DOI] [PubMed] [Google Scholar]
- (14).Honig B; Nicholls A Classical Electrostatics in Biology and Chemistry. Science 1995, 268, 1144–1149. [DOI] [PubMed] [Google Scholar]
- (15).Beroza P; Case DA Calculation of Proton Binding Thermodynamics in Proteins. Methods Enzymol. 1998, 295, 170–189. [DOI] [PubMed] [Google Scholar]
- (16).Madura JD; Davis ME; Gilson MK; Wade RC; Luty BA; McCammon JA Biological Applications of Electrostatic Calculations and Brownian Dynamics. Rev. Comp. Chem 1994, 5, 229–267. [Google Scholar]
- (17).Gilson MK Theory of Electrostatic Interactions in Macromolecules. Curr. Opin. Struct. Biol 1995, 5, 216–223. [DOI] [PubMed] [Google Scholar]
- (18).Scarsi M; Apostolakis J; Caflisch A Continuum Electrostatic Energies of Macromolecules in Aqueous Solutions. J. Phys. Chem. A 1997, 101, 8098–8106. [Google Scholar]
- (19).Luo R; David L; Gilson MK Accelerated Poisson-Boltzmann calculations for static and dynamic systems. J. Comput. Chem 2002, 23, 1244–1253. [DOI] [PubMed] [Google Scholar]
- (20).Simonson T Electrostatics and Dynamics of Proteins. Rep. Prog. Phys 2003, 66, 737–787. [Google Scholar]
- (21).Baker N; Bashford D; Case D Implicit Solvent Electrostatics in Biomolecular Simulation. New Algorithms for Macromolecular Simulation. 2006; pp 263–295. [Google Scholar]
- (22).Bardhan JP Biomolecular electrostatics – I want your solvation (model). Computational Science & Discovery 2012, 5, 013001.+. [Google Scholar]
- (23).Li L; Li C; Zhang Z; Alexov E On the Dielectric “Constant” of Proteins: Smooth Dielectric Function for Macromolecular Modeling and Its Implementation in DelPhi. Journal of chemical theory and computation 2013, 9, 2126–2136. [DOI] [PMC free article] [PubMed] [Google Scholar]
- (24).Born M Volumes and heats of hydration of ions. Z. Phys 1920, 1, 45–48. [Google Scholar]
- (25).Hoijtink G; De Boer E; Van der Meij P; Weijland W Reduction Potentials of Various Aromatic Hydrocarbons and Their Univalent Anions. Recl. Trav. Chim. Pays-Bas 1956, 75, 487–503. [Google Scholar]
- (26).Tucker SC; Truhlar DG Generalized Born Fragment Charge Model for Solvation Effects as a Function of Reaction Coordinate. Chem. Phys. Lett 1989, 157, 164–170. [Google Scholar]
- (27).Qiu D; Shenkin P; Hollinger F; Still WC The GB/SA continuum model for solvation. A fast analytical method for the calculation of approximate Born radii. J. Phys. Chem. A 1997, 101, 3005–3014. [Google Scholar]
- (28).Jayaram B; Liu Y; Beveridge DJ A modification of the generalized Born theory for improved estimates of solvation energies and pK shifts. J. Chem. Phys 1998, 109, 1465–1470. [Google Scholar]
- (29).Still WC; Tempczyk A; Hawley RC; Hendrickson T Semianalytical Treatment of Solvation for Molecular Mechanics and Dynamics. J. Am. Chem. Soc 1990, 112, 6127–6129. [Google Scholar]
- (30).Onufriev A; Bashford D; Case D Modification of the Generalized Born Model Suitable for Macromolecules. J. Phys. Chem. B 2000, 104, 3712–3720. [Google Scholar]
- (31).Dominy BN; Brooks CL Development of a Generalized Born Model Parametrization for Proteins and Nucleic Acids. J. Phys. Chem. B 1999, 103, 3765–3773. [Google Scholar]
- (32).Bashford D; Case D Generalized Born models of macromolecular solvation effects. Annu. Rev. Phys. Chem 2000, 51, 129–152. [DOI] [PubMed] [Google Scholar]
- (33).Calimet N; Schaefer M; Simonson T Protein molecular dynamics with the generalized Born/ACE solvent model. Proteins: Structure, Function, and Genetics 2001, 45, 144–158. [DOI] [PubMed] [Google Scholar]
- (34).Hawkins GD; Cramer CJ; Truhlar DG Pairwise solute descreening of solute charges from a dielectric medium. Chem. Phys. Lett 1995, 246, 122–129. [Google Scholar]
- (35).Hawkins GD; Cramer CJ; Truhlar DG Parametrized models of aqueous free energies of solvation based on pairwise descreening of solute atomic charges from a dielectric medium. J. Phys. Chem 1996, 100, 19824–19836. [Google Scholar]
- (36).Schaefer M; Karplus M A Comprehensive Analytical Treatment of Continuum Electrostatics. J. Phys. Chem 1996, 100, 1578–1599. [Google Scholar]
- (37).Feig M; Im W; Brooks CL Implicit solvation based on generalized Born theory in different dielectric environments. J. Chem. Phys 2004, 120, 903–911. [DOI] [PubMed] [Google Scholar]
- (38).Onufriev A; Bashford D; Case DA Modification of the generalized Born model suitable for macromolecules. The Journal of Physical Chemistry B: Condensed matter, materials, surfaces, interfaces and biophysical chemistry 2000, 104, 3712–3720. [Google Scholar]
- (39).Lee MS; F. R. Salsbury J; Brooks CL III Novel generalized Born methods. J. Chem. Phys 2002, 116, 10606–10614. [Google Scholar]
- (40).Lee MS; Feig M; Salsbury FR; Brooks CL New analytic approximation to the standard molecular volume definition and its application to generalized Born calculations. Journal of Computational Chemistry 2003, 24, 1348–56. [DOI] [PubMed] [Google Scholar]
- (41).Onufriev A; Bashford D; Case DA Exploring protein native states and large-scale conformational changes with a modified generalized Born model. Proteins 2004, 55, 383–94. [DOI] [PubMed] [Google Scholar]
- (42).Srinivasan J; Trevathan M; Beroza P; Case D Application of a pairwise generalized Born model to proteins and nucleic acids: Inclusion of salt effects. Theor. Chem. Accts 1999, 101, 426–434. [Google Scholar]
- (43).Tsui V; Case D Molecular dynamics simulations of nucleic acids using a generalized Born solvation model. J. Am. Chem. Soc 2000, 122, 2489–2498. [Google Scholar]
- (44).Wang T; Wade R Implicit solvent models for flexible protein-protein docking by molecular dynamics simulation. Proteins 2003, 50, 158–169. [DOI] [PubMed] [Google Scholar]
- (45).Gallicchio E; Levy RM AGBNP: an analytic implicit solvent model suitable for molecular dynamics simulations and high-resolution modeling. J. Comput. Chem 2004, 25, 479–499. [DOI] [PubMed] [Google Scholar]
- (46).Nymeyer H; García AE Simulation of the folding equilibrium of alpha-helical peptides: a comparison of the generalized Born approximation with explicit solvent. Proc Natl Acad Sci U S A 2003, 100, 13934–13939. [DOI] [PMC free article] [PubMed] [Google Scholar]
- (47).Ghosh A; Rapp CS; Friesner RA Generalized Born Model Based on a Surface Integral Formulation. J. Phys. Chem. B 1998, 102, 10983–10990. [Google Scholar]
- (48).Im W; Lee MS; Brooks CL Generalized born model with a simple smoothing function. Journal of Computational Chemistry 2003, 24, 1691–1702. [DOI] [PubMed] [Google Scholar]
- (49).Haberthür U; Caflisch A FACTS: Fast analytical continuum treatment of solvation. J Comput Chem 2007, 29, 701–715. [DOI] [PubMed] [Google Scholar]
- (50).Grant JA; Pickup BT; Sykes MJ; Kitchen CA; Nicholls A The Gaussian Generalized Born model: application to small molecules. Phys. Chem. Chem. Phys 2007, 9, 4913–4922. [DOI] [PubMed] [Google Scholar]
- (51).Tjong H; Zhou HX GBr6: A Parameterization-Free, Accurate, Analytical Generalized Born Method. J. Phys. Chem. B 2007, 111, 3055–3061. [DOI] [PubMed] [Google Scholar]
- (52).Labute P The generalized Born/volume integral implicit solvent model: Estimation of the free energy of hydration using London dispersion instead of atomic surface area. Journal of Computational Chemistry 2008, 29, 1693–1698. [DOI] [PubMed] [Google Scholar]
- (53).Onufriev A; Case DA; Bashford D Effective Born radii in the generalized Born approximation: the importance of being perfect. Journal of Computational Chemistry 2002, 23, 1297–304. [DOI] [PubMed] [Google Scholar]
- (54).Onufriev AV; Sigalov G A strategy for reducing gross errors in the generalized Born models of implicit solvation. The Journal of Chemical Physics 2011, 134, 164104.+, [DOI] [PMC free article] [PubMed] [Google Scholar]
- (55).Lange AW; Herbert JM Improving Generalized Born Models by Exploiting Connections to Polarizable Continuum Models. I. An Improved Effective Coulomb Operator. J. Chem. Theory Comput 2012, 8, 1999–2011. [DOI] [PubMed] [Google Scholar]
- (56).Aguilar B; Shadrach R; Onufriev A Reducing the secondary structure bias in the generalized Born model via R6 effective radii. J. Chem. Theory Comput 2010, 6, 3613–3630. [Google Scholar]
- (57).Forouzesh N; Izadi S; Onufriev AV Grid-based surface generalized Born model for calculation of electrostatic binding free energies. Journal of chemical information and modeling 2017, 57, 2505–2513. [DOI] [PMC free article] [PubMed] [Google Scholar]
- (58).Nguyen H; Maier J; Huang H; Perrone V; Simmerling C Folding Simulations for Proteins with Diverse Topologies Are Accessible in Days with a Physics-Based Force Field and Implicit Solvent. J. Am. Chem. Soc 2014, 136, 13959–13962. [DOI] [PMC free article] [PubMed] [Google Scholar]
- (59).Anandakrishnan R; Drozdetski A; Walker RC; Onufriev AV Speed of Conformational Change: Comparing Explicit and Implicit Solvent Molecular Dynamics Simulations. Biophysical journal 2015, 108, 1153–1164, [DOI] [PMC free article] [PubMed] [Google Scholar]
- (60).Onufriev A In Modeling Solvent Environments; Feig M, Ed.; Wiley: USA, 2010; pp 127–165, Chapter: public policy not applicable. [Google Scholar]
- (61).Buckingham AD A theory of ion-solvent interaction. Discussions of the Faraday Society 1957, 24, 151–157. [Google Scholar]
- (62).Rashin AA; Honig B Reevaluation of the Born model of ion hydration. J Phys Chem 1985, 89, 5588–5593. [Google Scholar]
- (63).Hirata F; Redfern P; Levy RM Viewing the Born model for ion hydration through a microscope. Int J Quantum Chem 1988, 34, 179–190. [Google Scholar]
- (64).Roux B; Yu HA; Karplus M Molecular basis for the Born model of ion solvation. J Phys Chem 1990, 94, 4683–4688. [Google Scholar]
- (65).Hummer G; Pratt LR; Garcia AE Free Energy of Ionic Hydration. J Phys Chem 1996, 100, 1206–1215. [Google Scholar]
- (66).Bell RML; Rasaiah JC From hydrophobic to hydrophilic behaviour: A simulation study of solvation entropy and free energy of simple solutes. J Chem Phys 1997, 107, 1981–1991. [Google Scholar]
- (67).Grossfield A Dependence of ion hydration on the sign of the ion’s charge. J Chem Phys 2005, 122, 024506. [DOI] [PubMed] [Google Scholar]
- (68).Mobley DL; Barber AE; Fennell CJ; Dill KA Charge asymmetries in hydration of polar solutes. J Phys Chem B 2008, 112, 2405–2414. [DOI] [PMC free article] [PubMed] [Google Scholar]
- (69).Purisima EO; Sulea T Restoring Charge Asymmetry in Continuum Electrostatics Calculations of Hydration Free Energies. J Phys Chem B 2009, 113, 8206–8209. [DOI] [PubMed] [Google Scholar]
- (70).Corbeil CR; Sulea T; Purisima EO Rapid Prediction of Solvation Free Energy. 2. The First-Shell Hydration (FiSH) Continuum Model. J Chem Theory Comput 2010, 6, 1622–1637. [DOI] [PubMed] [Google Scholar]
- (71).Mukhopadhyay A; Fenley AT; Tolokh IS; Onufriev AV Charge hydration asymmetry: the basic principle and how to use it to test and improve water models. J. Phys. Chem. B 2012, 116, 9776–9783. [DOI] [PMC free article] [PubMed] [Google Scholar]
- (72).Bardhan JP; Jungwirth P; Makowski L Affine-response model of molecular solvation of ions: Accurate predictions of asymmetric charging free energies. J. Chem. Phys 2012, 137. [DOI] [PMC free article] [PubMed] [Google Scholar]
- (73).Mukhopadhyay A; Aguilar BH; Tolokh IS; Onufriev AV Introducing Charge Hydration Asymmetry into the Generalized Born Model. J. Chem. Theory Comput 2014, 10, 1788–1794. [DOI] [PMC free article] [PubMed] [Google Scholar]
- (74).Bardhan JP; Knepley MG Communication: Modeling charge-sign asymmetric solvation free energies with nonlinear boundary conditions. The Journal of chemical physics 2014, 141, 131103. [DOI] [PMC free article] [PubMed] [Google Scholar]
- (75).Mukhopadhyay A; Tolokh IS; Onufriev AV Accurate Evaluation of Charge Asymmetry in Aqueous Solvation. The Journal of Physical Chemistry B 2015, 119, 6092–6100, [DOI] [PubMed] [Google Scholar]
- (76).Cox SJ; Mandadapu KK; Geissler PL Quadrupole-mediated dielectric response and the charge-asymmetric solvation of ions in water. The Journal of Chemical Physics 2021, 154, 244502. [DOI] [PubMed] [Google Scholar]
- (77).Gallicchio E; Paris K; Levy RM The AGBNP2 Implicit Solvation Model. J. Chem. Theory Comput 2009, 5, 2544–2564. [DOI] [PMC free article] [PubMed] [Google Scholar]
- (78).Fogolari F; Briggs JM On the variational approach to Poisson-Boltzmann free energies. Chemical Physics Letters 1997, 281, 135–139. [Google Scholar]
- (79).Dzubiella J; Swanson JM; McCammon JA Coupling nonpolar and polar solvation free energies in implicit solvent models. J Chem Phys 2006, 124, 084905. [DOI] [PubMed] [Google Scholar]
- (80).Abrashkin A; Andelman D; Orland H Dipolar Poisson-Boltzmann Equation: Ions and Dipoles Close to Charge Interfaces. Physical Review Letters 2007, 99, 077801.+. [DOI] [PubMed] [Google Scholar]
- (81).Chen Z; Baker NA; Wei GW Differential geometry based solvation model I: Eulerian formulation. Journal of Computational Physics 2010, 229, 8231–8258. [DOI] [PMC free article] [PubMed] [Google Scholar]
- (82).Chen J Effective Approximation of Molecular Volume Using Atom-Centered Dielectric Functions in Generalized Born Models. Journal of Chemical Theory and Computation 2010, 6, 2790–2803. [DOI] [PubMed] [Google Scholar]
- (83).Thomas DG; Chun J; Chen Z; Wei G; Baker NA Parameterization of a geometric flow implicit solvation model. J. Comput. Chem 2013, 34, 687–695. [DOI] [PMC free article] [PubMed] [Google Scholar]
- (84).Xia K; Feng X; Chen Z; Tong Y; Wei G-W Multiscale geometric modeling of macromolecules I: Cartesian representation. Journal of Computational Physics 2014, 257, 912–936. [DOI] [PMC free article] [PubMed] [Google Scholar]
- (85).Wei G-W; Zhou Y In Variational Methods in Molecular Modeling; Wu J, Ed.; Molecular Modeling and Simulation; Springer; Singapore, 2017; pp 181–221. [Google Scholar]
- (86).Izadi S; Anandakrishnan R; Onufriev AV Building Water Models: A Different Approach. J. Phys. Chem. Lett 2014, 5, 3863–3871. [DOI] [PMC free article] [PubMed] [Google Scholar]
- (87).Wang LP; Martinez TJ; Pande VS Building Force Fields: An Automatic, Systematic, and Reproducible Approach. J. Phys. Chem. Lett 2014, 5, 1885–1891. [DOI] [PMC free article] [PubMed] [Google Scholar]
- (88).Jing Z; Liu C; Cheng SY; Qi R; Walker BD; Piquemal J-P; Ren P Polarizable Force Fields for Biomolecular Simulations: Recent Advances and Applications. Annu. Rev. Biophys 2019, 48, 371–394. [DOI] [PMC free article] [PubMed] [Google Scholar]
- (89).Huang J; Lemkul JA; Eastman PK; MacKerell AD Jr. Molecular dynamics simulations using the drude polarizable force field on GPUs with OpenMM: Implementation, validation, and benchmarks. Journal of Computational Chemistry 2018, 39, 1682–1689. [DOI] [PMC free article] [PubMed] [Google Scholar]
- (90).Lambros E; Paesani F How good are polarizable and flexible models for water: Insights from a many-body perspective. The Journal of Chemical Physics 2020, 153, 060901. [DOI] [PubMed] [Google Scholar]
- (91).Leontyev IV; Stuchebrukhov AA Polarizable Mean-Field Model of Water for Biological Simulations with AMBER and CHARMM Force Fields. J. Chem. Theory Comput 2012, 8, 3207–3216. [DOI] [PMC free article] [PubMed] [Google Scholar]
- (92).Xiong Y; Izadi S; Onufriev AV Fast Polarizable Water Model for Atomistic Simulations. Journal of Chemical Theory and Computation 2022, 18, 6324–6333. [DOI] [PMC free article] [PubMed] [Google Scholar]
- (93).Qiu J; Wu Q; Ding G; Xu Y; Feng S A survey of machine learning for big data processing. EURASIP Journal on Advances in Signal Processing 2016, 2016, 1–16. [Google Scholar]
- (94).LeCun Y; Bengio Y; Hinton G Deep learning. nature 2015, 521, 436–444. [DOI] [PubMed] [Google Scholar]
- (95).Ramsundar B; Eastman P; Walters P; Pande V; Leswing K; Wu Z Deep Learning for the Life Sciences; O’Reilly Media, 2019; https://www.amazon.com/Deep-Learning-Life-Sciences-Microscopy/dp/1492039837. [Google Scholar]
- (96).Goodfellow I; Bengio Y; Courville A Deep learning; MIT press, 2016. [Google Scholar]
- (97).Zhai Y; Caruso A; Bore SL; Luo Z; Paesani FA “short blanket” dilemma for a state-of-the-art neural network potential for water: Reproducing experimental properties or the physics of the underlying many-body interactions? The Journal of Chemical Physics 2023, 158, 084111. [DOI] [PubMed] [Google Scholar]
- (98).Cain S; Risheh A; Forouzesh N A Physics-Guided Neural Network for Predicting Protein–Ligand Binding Free Energy: From Host–Guest Systems to the PDBbind Database. Biomolecules 2022, 12, 919. [DOI] [PMC free article] [PubMed] [Google Scholar]
- (99).Smith JS; Isayev O; Roitberg AE ANI-1: an extensible neural network potential with DFT accuracy at force field computational cost. Chemical science 2017, 8, 3192–3203. [DOI] [PMC free article] [PubMed] [Google Scholar]
- (100).Dou B; Zhu Z; Merkurjev E; Ke L; Chen L; Jiang J; Zhu Y; Liu J; Zhang B; Wei G-W Machine Learning Methods for Small Data Challenges in Molecular Science. Chemical Reviews 2023, [DOI] [PMC free article] [PubMed] [Google Scholar]
- (101).Fedik N; Zubatyuk R; Kulichenko M; Lubbers N; Smith JS; Nebgen B; Messerly R; Li YW, et al. Extending machine learning beyond interatomic potentials for predicting molecular properties. Nature Reviews Chemistry 2022, 6, 653–672. [DOI] [PubMed] [Google Scholar]
- (102).Butler KT; Davies DW; Cartwright H; Isayev O; Walsh A Machine learning for molecular and materials science. Nature 2018, 559, 547–555. [DOI] [PubMed] [Google Scholar]
- (103).Wei H; Zhao Z; Luo R Machine-Learned Molecular Surface and Its Application to Implicit Solvent Simulations. Journal of chemical theory and computation 2021, 17, 6214–6224. [DOI] [PMC free article] [PubMed] [Google Scholar]
- (104).McNutt AT; Francoeur P; Aggarwal R; Masuda T; Meli R; Ragoza M; Sunseri J; Koes DR GNINA 1.0: molecular docking with deep learning. Journal of cheminformatics 2021, 13, 1–20. [DOI] [PMC free article] [PubMed] [Google Scholar]
- (105).Ragoza M; Masuda T; Koes DR Generating 3D molecules conditional on receptor binding sites with deep generative models. Chemical science 2022, 13, 2701–2713. [DOI] [PMC free article] [PubMed] [Google Scholar]
- (106).Veltri D; Kamath U; Shehu A Deep learning improves antimicrobial peptide recognition. Bioinformatics 2018, 34, 2740–2747. [DOI] [PMC free article] [PubMed] [Google Scholar]
- (107).Ricci-Lopez J; Aguila SA; Gilson MK; Brizuela CA Improving structure-based virtual screening with ensemble docking and machine learning. Journal of Chemical Information and Modeling 2021, 61, 5362–5376. [DOI] [PMC free article] [PubMed] [Google Scholar]
- (108).Zeng J; Giese TJ; Ekesan S; York DM Development of Range-Corrected Deep Learning Potentials for Fast, Accurate Quantum Mechanical/Molecular Mechanical Simulations of Chemical Reactions in Solution. J Chem Theory Comput 2021, 17, 6993–7009, J Chem Theory Comput. 2021 Nov 9;17(11):6993–7009. doi: 10.1021/acs.jctc.1c00201. Epub 2021 Oct 13. [DOI] [PMC free article] [PubMed] [Google Scholar]
- (109).Pinheiro M; Ge F; Ferré N; Dral PO; Barbatti M Choosing the right molecular machine learning potential. Chem. Sci 2021, 12, 14396–14413. [DOI] [PMC free article] [PubMed] [Google Scholar]
- (110).Ramakrishnan R; Dral PO; Rupp M; von Lilienfeld OA Big Data Meets Quantum Chemistry Approximations: The Δ-Machine Learning Approach. Journal of Chemical Theory and Computation 2015, 11, 2087–2096. [DOI] [PubMed] [Google Scholar]
- (111).Alibakhshi A; Hartke B Improved prediction of solvation free energies by machine-learning polarizable continuum solvation model. Nature Communications 2021, 12, 3584. [DOI] [PMC free article] [PubMed] [Google Scholar]
- (112).Vermeire FH; Green WH Transfer learning for solvation free energies: From quantum chemistry to experiments. Chemical Engineering Journal 2021, 418, 129307. [Google Scholar]
- (113).Pereyaslavets L; Kamath G; Butin O; Illarionov A; Olevanov M; Kurnikov I; Sakipov S; Leontyev I, et al. Accurate determination of solvation free energies of neutral organic compounds from first principles. Nature Communications 2022, 13, 414. [DOI] [PMC free article] [PubMed] [Google Scholar]
- (114).Chen Y; Krämer A; Charron NE; Husic BE; Clementi C; Noé F Machine learning implicit solvation for molecular dynamics. The Journal of Chemical Physics 2021, 155, 084101. [DOI] [PubMed] [Google Scholar]
- (115).Gebhardt J; Kiesel M; Riniker S; Hansen N Combining Molecular Dynamics and Machine Learning to Predict Self-Solvation Free Energies and Limiting Activity Coefficients. Journal of Chemical Information and Modeling 2020, 60, 5319–5330. [DOI] [PubMed] [Google Scholar]
- (116).Rauer C; Bereau T Hydration free energies from kernel-based machine learning: Compound-database bias. The Journal of Chemical Physics 2020, 153. [DOI] [PubMed] [Google Scholar]
- (117).Hutchinson ST; Kobayashi R Solvent-Specific Featurization for Predicting Free Energies of Solvation through Machine Learning. J Chem Inf Model 2019, 59, 1338–1346, J Chem Inf Model. 2019 Apr 22;59(4):1338–1346. doi: 10.1021/acs.jcim.8b00901.Epub 2019 Mar 18. [DOI] [PubMed] [Google Scholar]
- (118).Mobley DL; Guthrie JP FreeSolv: a database of experimental and calculated hydration free energies, with input files. Journal of Computer-aided Molecular Design 2014, 28, 711–720, 24928188[pmid]. [DOI] [PMC free article] [PubMed] [Google Scholar]
- (119).Karpatne A; Atluri G; Faghmous JH; Steinbach M; Banerjee A; Ganguly A; Shekhar S; Samatova N; Kumar V Theory-guided data science: A new paradigm for scientific discovery from data. IEEE Transactions on knowledge and data engineering 2017, 29, 2318–2331. [Google Scholar]
- (120).Willard J; Jia X; Xu S; Steinbach M; Kumar V Integrating scientific knowledge with machine learning for engineering and environmental systems. ACM Computing Surveys 2022, 55, 1–37. [Google Scholar]
- (121).Karpatne A; Kannan R; Kumar V Knowledge Guided Machine Learning: Accelerating Discovery Using Scientific Knowledge and Data; CRC Press, 2022. [Google Scholar]
- (122).Daw A; Karpatne A; Watkins W; Read J; Kumar V Physics-guided neural networks (PGNN): An application in lake temperature modeling. arXiv preprint arXiv:1710.11431 2017, [Google Scholar]
- (123).Jia X; Willard J; Karpatne A; Read JS; Zwart JA; Steinbach M; Kumar V Physics-guided machine learning for scientific discovery: An application in simulating lake temperature profiles. ACM/IMS Transactions on Data Science 2021, 2, 1–26. [Google Scholar]
- (124).Elhamod M; Bu J; Singh C; Redell M; Ghosh A; Podolskiy V; Lee W-C; Karpatne A CoPhy-PGNN: Learning physics-guided neural networks with competing loss functions for solving eigenvalue problems. ACM Transactions on Intelligent Systems and Technology 2022, 13, 1–23. [Google Scholar]
- (125).Karniadakis GE; Kevrekidis IG; Lu L; Perdikaris P; Wang S; Yang L Physics-informed machine learning. Nature Reviews Physics 2021, 3, 422–440. [Google Scholar]
- (126).Physics-informed neural networks: A deep learning framework for solving forward and inverse problems involving nonlinear partial differential equations. Journal of Computational physics 2019, 378, 686–707. [Google Scholar]
- (127).Nicholls A; Mobley DL; Guthrie JP; Chodera JD; Bayly CI; Cooper MD; Pande VS Predicting Small-Molecule Solvation Free Energies: An Informal Blind Test for Computational Chemistry. J. Med. Chem 2008, 51, 769–779. [DOI] [PubMed] [Google Scholar]
- (128).Roux B; Simonson T Implicit solvent models. Biophysical chemistry 1999, 78, 1–20. [DOI] [PubMed] [Google Scholar]
- (129).Mobley D; Gilson M Predicting Binding Free Energies. Annual Review of Biophysics 2017, 46, null. [DOI] [PMC free article] [PubMed] [Google Scholar]
- (130).Kollman PA; Massova I; Reyes C; Kuhn B; Huo S; Chong L; Lee M; Lee T, et al. Calculating structures and free energies of complex molecules: combining molecular mechanics and continuum models. Accounts of chemical research 2000, 33, 889–897. [DOI] [PubMed] [Google Scholar]
- (131).Yin J; Henriksen N; Muddana H; Gilson M Bind3P: Optimization of a Water Model Based on Host–Guest Binding Data. Journal of Chemical Theory and Computation 2018, 14, 3621–3632, [DOI] [PMC free article] [PubMed] [Google Scholar]; SOLVATION FREE ENERGY: Overall, the OPC model yields the most accurate binding enthalpies with an RMSE of 2.3 kcal/ mol and R2 of 0.4, despite providing middling results for the binding free energies.
- (132).Gao K; Yin J; Henriksen N; Fenley A; Gilson M Binding Enthalpy Calculations for a Neutral Host–Guest Pair Yield Widely Divergent Salt Effects across Water Models. J. Chem. Theory Comput 2015, 11, 4555–4564. [DOI] [PMC free article] [PubMed] [Google Scholar]
- (133).Izadi S; Aguilar B; Onufriev AV Protein–Ligand Electrostatic Binding Free Energies from Explicit and Implicit Solvation. J. Chem. Theory Comput 2015, 11, 4450–4459. [DOI] [PMC free article] [PubMed] [Google Scholar]
- (134).Folescu DE; Onufriev AV A Closed-Form, Analytical Approximation for Apparent Surface Charge and Electric Field of Molecules. ACS Omega 2022, 7, 26123–26136. [DOI] [PMC free article] [PubMed] [Google Scholar]
- (135).Piana S; Donchev A; Robustelli P; Shaw D Water dispersion interactions strongly influence simulated structural properties of disordered protein states. The journal of physical chemistry. B 2015, 119, 5113–5123. [DOI] [PubMed] [Google Scholar]
- (136).van der Spoel D; van Maaren P; Berendsen H A systematic study of water models for molecular simulation: Derivation of water models optimized for use with a reaction field. The Journal of Chemical Physics 1998, 108, 10220–10230. [Google Scholar]
- (137).Manzoni F; Söderhjelm P Prediction of hydration free energies for the SAMPL4 data set with the AMOEBA polarizable force field. J Comput Aided Mol Des 2014, 28, 235–244. [DOI] [PubMed] [Google Scholar]
- (138).Tomasi J; Mennucci B; Cammi R Quantum Mechanical Continuum Solvation Models. Chem. Rev. (Washington, DC, U. S.) 2005, 105, 2999–3094. [DOI] [PubMed] [Google Scholar]
- (139).Fennell CJ; Kehoe CW; Dill KA Modeling aqueous solvation with semi-explicit assembly. Proc Natl Acad Sci U S A 2011, 108, 3234–3239. [DOI] [PMC free article] [PubMed] [Google Scholar]
- (140).Duvenaud DK; Maclaurin D; Iparraguirre J; Bombarell R; Hirzel T; Aspuru-Guzik A; Adams RP Convolutional networks on graphs for learning molecular fingerprints. Advances in neural information processing systems 2015, 28. [Google Scholar]
- (141).Srivastava N; Hinton GE; Krizhevsky A; Sutskever I; Salakhutdinov R Dropout: a simple way to prevent neural networks from overfitting. Journal of Machine Learning Research 2014, 15, 1929–1958. [Google Scholar]
- (142).Artrith N; Butler KT; Coudert FX; Han S; Isayev O; Jain A; Walsh A Best practices in machine learning for chemistry. Nat Chem 2021, 13, 505–508. [DOI] [PubMed] [Google Scholar]
- (143).Kingma DP; Ba J Adam: A Method for Stochastic Optimization. 2017.
- (144).Swanson JMJ; Adcock SA; McCammon JA Optimized Radii for Poisson-Boltzmann Calculations with the AMBER Force Field. J. Chem. Theory Comput 2005, 1, 484–493. [DOI] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
