Abstract
Adaptive Multilevel Splitting (AMS) is a replica-based rare event sampling method that has been used successfully in high-dimensional stochastic simulations to identify trajectories across a high potential barrier separating one metastable state from another, and to estimate the probability of observing such a trajectory. An attractive feature of AMS is that, in the limit of a large number of replicas, it remains valid regardless of the choice of reaction coordinate used to characterize the trajectories. Previous studies have shown AMS to be accurate in Monte Carlo simulations. In this study, we extend the application of AMS to molecular dynamics simulations and demonstrate its effectiveness using a simple test system. Our conclusion paves the way for useful applications, such as molecular dynamics calculations of the characteristic time of drug dissociation from a protein target.
Introduction
In high-dimensional stochastic systems, one commonly encounters potential or entropic barriers so high that observing even a single crossing event would require a prohibitive amount of computational effort. The problem of sampling such rare events is of particular interest in molecular dynamics (MD), which involves extremely many degrees of freedom and requires very small time steps so that events of interest are frequently out of reach for equilibrium simulations. For example, the dissociation of small molecules from a substrate can have characteristic times on the order of seconds, compared to the usual MD time step size of 1 or 2 fs.
Various sampling techniques have been developed to mitigate high potential barriers between metastable states, allowing the sampling of trajectories that cross these barriers. One such technique is Adaptive Multilevel Splitting (AMS) [2, 3], which has been successfully tested on Monte Carlo simulations. Thus, one interest of the current work is to test for the first time the validity of AMS applied to phase space Langevin dynamics. The AMS algorithm is less sensitive to the choice of the reaction coordinate than other methods. In particular, it can be shown that the estimator of the (very small) probability to observe a reactive trajectory is unbiased whatever the choice of the reaction coordinate. Due to its versatility, AMS has promising prospects in biological MD simulations, where prior information about trajectories is often lacking due to experimental difficulties or sheer complexity of the system.
We applied the AMS algorithm to a simple MD test case involving an ion in a harmonic well, using the software NAMD [1], developed by the Theoretical and Computational Biophysics Group in the Beckman Institute for Advanced Science and Technology at the University of Illinois at Urbana-Champaign. In order to implement the algorithm, alterations to the current released version (2.9) of the software were necessary, as described herein. Direct long-time simulations of the test case, as well as an analytic derivation of key physical quantities, were performed along with AMS simulations for comparison. We found that AMS faithfully yielded results in agreement with measurements from the direct simulations and values obtained from the analytic calculation. In one of the simulations, AMS was able to sample events that were too rare for direct simulation to reproduce, and at a measured rate that agreed with the analytic calculation using a simple model. This study is an indication that AMS is a feasible sampling method in MD simulations, and is a precursor to simulations of more complicated biological systems and efforts to streamline the software for better efficiency.
After a brief description of the AMS algorithm, we present the software modifications which were necessary to implement AMS on NAMD. This is an opportunity to discuss the complexity of the algorithm, in terms of processor time and memory requirements. Finally, we discuss numerical results from the test simulations. The appendix is devoted to analytic computations on a simple model for the sake of comparison with the numerical results.
Description of AMS
Consider (Xt)t≥0 a Markov process in ℝd \ (A∪B). Assume that the process follows almost surely continuous paths and is ergodic with respect to some equilibrium measure μ. In our simulations, Xt is the vector of current positions and velocities of a molecular system obeying Langevin dynamics, and μ is the Boltzmann distribution. The aim of AMS is to understand reactive trajectories, that is, paths of the process which start in some set A ⊂ ℝd and then reach another set B ⊂ ℝd without first returning to A. Define a reaction coordinate ξ: ℝd → ℝ which measures progress from A to B. Thus, it can be written
| (1) |
for some zA < zB. We assume ξ is differentiable and ∇ξ (x) ≠ 0 for all x ∈ ℝd \ (A ∪ B), so that the level sets of ξ are submanifolds of ℝd \ (A ∪ B). We will refer to the submanifold {x ∈ ℝd : ξ(x) = z} as the level z.
The AMS algorithm uses N replicas of the process along with a splitting technique to efficiently generate reactive paths. The algorithm, rephrased from [3], is as follows:
Algorithm 1. Pick a number N of replicas, and choose zmin between zA and zB. Set the total number of branching points, M, to zero: M = 0. Then iterate the following:
Initialization Step:
-
1.
Generate N independent samples of the (normalized) equilibrium distribution μ(A)−1μ|A restricted to A.
-
2.
Starting at these N sample points, evolve N replicas independently until each replica reaches the level zmin.
-
3.
Continue evolving the N replicas until each replica reaches A or B.
Branching Step:
-
4.Let zk be the maximum level reached by replica k, and let
-
5.
Kill the existing k*th replica. Pick another replica uniformly at random, say replica k′. Let x ∈ ℝd be the point at which replica k′ first reaches the level z*.
-
6.
Start a new k*th replica at the point x, and evolve the replica until it reaches either A or B. Update the total number of branching points: M = M + 1.
-
7.
If all the replicas have reached B, stop. Otherwise, return to Step 4.
The parameter zmin is chosen so that the time for the process to go from zA to zmin is easily accessible by direct simulation, while at the same time a replica starting at zmin can evolve for some time without immediately returning to zA. Choosing a value too far from zA would cause the initial simulation from zA to zmin to become prohibitively long. On the other hand, choosing a value too close to zA would cause the measurements of T1, T2, and p to be small, resulting in large statistical errors in the final result. At this time, the effects of the choice of zmin on AMS results have not been quantified, nor is there a known a priori criterion for selecting zmin. It is recommended for now that a trial-and-error approach be adopted to select a value such that the process from zA to zmin occurs within a reasonable computational time, given the computational resources available. For example, one may start by setting zmin to be two standard deviations from the mean reaction coordinate value over the initial equilibration phase, and proceed by adjusting the chosen value according to the time taken to generate a small number of initial trajectories. The effect of the value of zmin on AMS results can be evaluated by repeated AMS runs using different values of zmin.
Future work may take advantage of parallelism by using a variant of AMS [2] in which i ≥ 1 replicas are killed, as explained in the following remark:
Remark 1. In the Branching Step, instead of killing and restarting just one replica, we can instead choose the i lowest zk values and kill the corresponding replicas. Each of the killed replicas would then be restarted at a randomly chosen surviving replica, as per Step 6. It has been shown that the i = 1 case produced the best results [4]; however, the gain in computational speed from running i trajectories in parallel at each iteration may justify such a compromise in accuracy.
Let p be the probability that, starting at the level zmin as in Step 2 above, the process reaches B before returning to A. After the algorithm is complete, we have the following estimate pAMS of p:
| (2) |
where by definition M is the total number of branching points. Under ideal conditions, pAMS is an unbiased estimate of p [4, 5], and moreover, the variance of pAMS, Var(pAMS), is, asymptotically as N → ∞ [4, 7]:
| (3) |
At this point, it is worth cautioning that the above formula for Var(pAMS) is in general not exact. However, the estimate can be expected to be good if the reaction coordinate is a good approximation of the committor function [4, 6]. Furthermore, one output of the AMS algorithm is the density of configurations along reactive paths, from which the committor function can be computed, so that it is possible to adaptively adjust the reaction coordinate to obtain a satisfactory approximation [3].
AMS can also be used to estimate τ, the average time to go from ∂A to B, which is done as follows. Let T1 be the average time for a replica in the Initialization Step to go from the level zA to the level zmin; similarly let T2 be the average time for a replica in the Initialization Step to go from zmin to zA without reaching zB. T1 and T2 can be sampled from a direct simulation. Let T3 be the average time for the process to go from the level zmin to zB without ever returning to zA; T3 can be estimated by averaging over the reactive trajectories at the end of the algorithm. We then obtain the following estimate τAMS of τ:
| (4) |
See [3] for details and a discussion.
Software Modifications
To establish the feasibility of applying the AMS methodology to complex MD simulations, the initial implementation described herein is largely focused on resolving several technical challenges, namely, rapid prototype development, simulation scalability, data management, and method parallelization. A general outline of the AMS implementation and organization is shown schematically in Fig. 1. Under this scheme, the AMS control code works in conjunction with NAMD to set up, run, and analyze each simulation step. A pool of NAMD instances utilize a shared file system to allow cross-process communication, enabling dynamic initialization and termination of simulations as guided by the AMS control logic.
Figure 1.
Schematic of AMS implementation. The AMS control logic running within each NAMD instance monitors global progress using a shared file system to communicate information between processes. Upon selecting a system replica for restart, the associated positional, velocity, and periodic cell data is loaded into an available NAMD instance and the simulation is launched. During the course of each simulation, NAMD continually saves the simulation trajectory to disk and updates multiple elements of shared data used to drive the AMS control logic.
As the centerpiece of the described implementation, the NAMD software package [1] was chosen for its highly efficient, scalable, and feature-rich MD engine that can run on a variety of platforms and is maintained at most supercomputing centers around the world. Two NAMD features of primary importance to AMS are the native “Colvars” implementation [8] and the embedded Tcl interpreter. The colvars module is leveraged to define and monitor the AMS reaction coordinate using the built-in collective variables, which are easily defined and can be combined to describe complex reaction coordinates. Access to the Tcl scripting interface, a feature unique to NAMD, allows a great deal of flexibility to rapidly develop and debug the AMS control code without requiring detailed knowledge of the internal workings of NAMD. Separating the AMS control from the NAMD “black box” requires only minor modifications to the NAMD source code–adding mechanisms for reading/writing of restart data and sequential trajectory files. Although these features were added specifically to enable the AMS method as described herein, they are of general utility to NAMD users and will be included in the NAMD 2.10 software release.
From the outset, coupling the AMS methodology to MD simulations raised significant concerns regarding data management; MD is data-intensive (generates large positional trajectories and restart files), while AMS is highly duplicative (reactive trajectories are “copied” up to the branch point). Accordingly, the storage requirements of an AMS run are mitigated using two concurrent approaches. The first approach is centered on minimizing the number of simulation restart files that are stored. As described in the Branching Step of the algorithm, simulations are restarted at the first crossing of a particular isosurface of the reaction coordinate. In practical terms, this specification restricts potential restarts to frames in which the z-value is a new maximum observed value, as shown in Fig. 2. Each set of restart data (position, velocity, periodic cell) is stored using the frame number, preserving the timing information of the simulation. The largest reduction in storage requirements can be realized by suppressing the positional trajectory output (DCD files) without compromising the kinetic analysis (pAMS, τAMS), albeit at the loss of continuity in the atomic detail for each reactive trajectory.
Figure 2.
Illustration of a typical simulation trajectory. Each circle represents a point at which the reaction coordinate (z) is measured. In the interest to data economy, restart data is only written to memory when the measured reaction coordinate achieves a new global maximum (filled circles), representing a valid restart point for subsequent simulations.
The second approach employs reference counting, a computer science framework for managing objects in memory, to curate the simulation data. As implied in Fig. 1, the simulation data (trajectories, restart data) are stored and manipulated as a collection of individual files. Each system replica maintains an ordered list of references to frames within a particular simulation file that, when stitched together, defines the reactive trajectory. The number of references to each file are maintained as part of the shared data, and reference counts are incremented or decremented as trajectories are duplicated or discarded, respectively. When the reference count for a particular file reaches zero, the data is no longer relevant to any of the surviving reactive trajectories, and therefore, deleted from the file system. Using this framework, duplicate data is minimized and simulation data that is no longer relevant to the remaining reactive trajectories is immediately discarded.
Finally, a critical component of the present implementation is the “pseudo-parallelization” of the AMS method. The original codes used demonstrate the AMS method for 1D and 2D potential energy surfaces operated in a purely serial mode, waiting for each simulation to reach a termination criterion (return to state A or advance to state B) before starting a new replica (as in Step 6 of the algorithm). In the present implementation, we instead start a new replica as soon as the smallest maximum level among all running replicas surpasses the smallest maximum level among all stopped replicas. Thus, many replicas can be running concurrently in parallel. Fig. 3 depicts this strategy in which the least advanced trajectory, shown in green, represents a currently running replica. Once this replica has surpassed the threshold demarcated by the orange line, the least-progressed stopped replica (blue), can be restarted. Although the degree of parallelism and replica start times are unpredictable (an outcome of the stochastic MD process), in our experience, this design allows for significant parallelization in two key areas: as simulations rapidly progress along the reaction coordinate in areas of low energy, and when simulations make any degree of progress along the reaction coordinate (e.g., high energy) but require a non-trivial amount of time to return to state A.
Figure 3.
Illustration of “pseudo-parallelization” using a chart of maximum z attained in an example AMS simulation with seven replicas. The letters “R” and “S” label running and stopped replicas, respectively. When all running simulations have surpassed the level of the least-progressed stopped replica (shown in blue), an event denoted by the green replica crossing the orange threshold, the blue-colored replica can begin running even before any running replicas terminate.
Simulations and Results
The test system consists of a 50 Å × 50 Å × 50 Å box of explicit water with 0.15 M potassium chloride in solution such that the net charge is zero. We pick one particular K+ ion, positioned initially at the origin. Henceforth, we refer only to the chosen ion and omit all other ions from our discussion. Our objective is to use AMS to evaluate the characteristic time taken for the ion to migrate from a point of distance zA from the origin to a point zB away from the origin, under the influence of a harmonic well potential centered on the origin. For this purpose, we identify the reaction coordinate z with the distance from the origin r.
CHARMM parameters for ion interactions were taken from Roux and coworkers [10] while the water molecules were characterized by the TIP3P water model [9]. Simulations were run with 1-fs time steps. Long range electrostatic forces were calculated using the particle mesh-Ewald (PME) method with a mesh density of about 1.5 Å−3. Van der Waals forces were calculated using a 12 Å cutoff and a switching function starting at 10 Å. Force evaluations were performed at every time step. Periodic boundary conditions were imposed on the faces of the waterbox and Langevin dynamics was simulated with a temperature of 300 K and damping coefficient of 1 ps−1. Pressure was maintained at 1 atm using a Nosé-Hoover Langevin piston with a damping timescale of 50 fs and a period of 200 fs.
The simulations were carried out in a series of steps. First, the system was energy-minimized over 1000 time steps before being equilibrated for 5 ns with the ion fixed at the origin. Next, N = 100 replicas of the system were initialized and run independently, with the ion free to diffuse but under the influence of a spherical harmonic potential . Due to the spherical symmetry of the system, we can omit Step 1 of the algorithm, starting with Step 2 instead, without fear of introducing bias. Each replica is run until the ion reaches zmin. Step 2 and the subsequent step of the algorithm were performed for three different values of k, with k = 0.01 (kcal/mol) Å−2, 0.02 (kcal/mol) Å−2, 0.08 (kcal/mol) Å−2. After the preparation steps above, the resulting states of the replicas were then fed into three sets of simulations.
The first set of simulations follows the AMS algorithm described above. The probability pAMS (Eq. (2)) calculated by the AMS simulation set corresponds to that of the ion, initially at zmin, diffusing to zB without first visiting the sphere A = {r : r < zA}. The simulation parameters, number of AMS iterative steps M, and the corresponding pAMS values with error estimates given by the square root of the variance (Eq. (3)) are tabulated in Table 1.
Table 1.
Parameters and results of AMS simulations.
| k ((kcal/mol) Å−2) | zA(Å) | zmin(Å) | zB(Å) | M | pAMS |
|---|---|---|---|---|---|
| 0.01 | 8 | 10 | 22 | 253 | 0.079 ± 0.013 |
| 0.02 | 8 | 12 | 18 | 202 | 0.13 ± 0.02 |
| 0.08 | 5 | 9 | 15 | 474 | (4.5 ± 1.2) × 10−4 |
The second set of simulations consisted of direct 10-ns equilibrium runs on each of the 100 replicas. The long sampling time allowed us to measure the times T1, T2 and T3 by averaging over the times taken for the ion to travel between zA and zmin and from zmin to zB. These time values are required to calculate the AMS prediction of τ as per Eq. (4). The trajectories obtained also provided direct measurements of τ and p, given by τsim and psim, respectively. τsim is the average of the measured times taken for the ion starting at zA to reach zB in the direct simulations. psim is measured using the formula where n(success) and n(fail) are, respectively, the number of trajectories starting from zmin that reach zB before zA, and the number of trajectories starting from zmin that reach zA before zB. With the exception of (*), the aforementioned quantities, listed in Table 2, were obtained through direct simulation. (*) was measured from reconstructions of the reactive trajectories generated by the AMS algorithm. In the cases where these direct measurements were possible, we were able to compare the p and τ values with those obtained from the AMS runs; however, we were unable to sample enough “success” events for the k = 0.08 (kcal/mol) Å−2 case. For the purpose of validating the AMS results in the latter case, we performed an analytic calculation on a simple model, described in Appendix A. As a verification of the value of β in the model, we measured in the direct simulations the time-averaged equilibrium ion distribution by histogramming the position of the ion at every 100-fs interval over the entire trajectories of all the replicas. The equilibrium distributions obtained for each k are shown in Figure 4 with the theoretical expected Boltzmann distribution given by pB ∝ r2 exp(−kr2/2kBT). Note that they compare favorably with the Boltzmann distribution curve, justifying the use of β = 1/kBT in the analytic calculations.
Table 2.
Direct measurement of variables required for AMS, with the exception of T3 for k = 0.08 (kcal/mol) Å−2. The latter was calculated by reconstructing the reactive path trajectories obtained from the AMS algorithm itself. Unexpectedly, it was found for the smallest k value that T2 < T1 by a small margin. This anomaly is probably due to statistical fluctuations, since the potential in the region z < zmin is almost flat in the small k limit.
| k ((kcal/mol) Å−2) | n(success) | n(failure) | psim | T1 (ns) | T2 (ns) | T3 (ns) | τsim (ns) |
|---|---|---|---|---|---|---|---|
| 0.01 | 105 | 1136 | 0.092 ± 0.009 | 0.035 ± 0.005 | 0.051 ± 0.008 | 0.14 ± 0.01 | 0.94 ± 0.09 |
| 0.02 | 112 | 734 | 0.13 ± 0.01 | 0.13 ± 0.01 | 0.046 ± 0.004 | 0.060 ± 0.005 | 1.0 ± 0.1 |
| 0.08 | - | - | - | 0.34 ± 0.03 | 0.021 ± 0.001 | 0.54 ± 0.05* | - |
Figure 4.
Normalized equilibrium distribution curves obtained from simulation and Boltzmann distribution, with (a) k = 0.01 (kcal/mol) Å−2, (b) k = 0.02 (kcal/mol) Å−2, (c) k = 0.08 (kcal/mol) Å−2. The sudden rise at the tail end near 20 Å in (a) and (b) is an artifact of the periodic boundary conditions used in the simulation.
The analytic model also requires the diffusion coefficient of the ion in water, D, to be specified. In the third set of simulations, we measure the local diffusion coefficient at various points in the system for the case k = 0.08 (kcal/mol) Å−2, using an existing method, which is described together with the results in Appendix B. In the analytic calculation in Appendix A, we assume the result, D = 254 ± 12 Å2/ns, to be constant in space and valid for the other values of k. The uncertainty in τanalytic is due to the uncertainty in D being carried forward.
Using the relevant parameters, we obtain theoretical values, panalytic and τanalytic, for p and τ, respectively, with the analytic derivation given in Appendix A. The estimates from the AMS calculation τAMS of τ are also obtained using Eq. (4) with values for T1, T2 and T3 from Table 2. Table 3 lists the results from the AMS calculation, direct simulations, and theoretical calculations for comparison. Table 3 shows that the AMS results compare favorably with those of the direct simulation and analytic calculation. Values for p agreed within the error bounds. Minor discrepancies were found in the values of τ, suggesting that the error bounds have been underestimated. It is recommended in such a case that a follow-up simulation be run with a new reaction coordinate chosen based on the distribution of configurations obtained in the current simulation.
Table 3.
Comparison of p and τ values obtained from AMS, direct simulation, and analytic calculations.
| k ((kcal/mol) Å−2) | pAMS | psim | panalytic | τAMS (ns) | τsim (ns) | τanalytic (ns) |
|---|---|---|---|---|---|---|
| 0.01 | 0.079 ± 0.013 | 0.092 ± 0.009 | 0.084 | 1.3 ± 0.3 | 0.94 ± 0.09 | 0.99 ± 0.05 |
| 0.02 | 0.13 ± 0.02 | 0.13 ± 0.01 | 0.13 | 1.5 ± 0.2 | 1.0 ± 0.1 | 1.17 ± 0.06 |
| 0.08 | (4.5 ± 1.2) × 10−4 | - | 3.7 × 10−4 | 800 ± 200 | - | 1040 ± 50 |
Conclusion
Our findings show that AMS can be feasibly applied in MD simulations, providing accurate measurements that agree with both direct simulations and analytic calculations. The accuracy of the AMS measurements in these test cases are impressive in light of the savings in computational cost. Each AMS simulation took about 3 days to complete, in contrast to one direct simulation, which required about 0.5 days on each of 100 replicas, for an equivalent processing time of around 50 days.
Furthermore, there is still much room for improvement in the efficiency of the NAMD implementation of AMS. The AMS algorithm used in the reported test cases did not utilize the “pseudo-parallelization” scheme. The present implementation also is an early Tcl-based prototype which relies on a cumbersome file read/write system. Streamlining of the code, together with enhancement in parallelization through “pseudo-parallelization” or a parallel variant of AMS (refer to Remark 1), will significantly increase gains in efficiency. These improvements will eventually be required as AMS is used for practical applications in biology, which are more challenging than the simple test cases in this study.
The complexity of biological systems may require more than a straightforward increase in software efficiency. For example, one application of AMS in MD would be the measurement of dissociation rates, such as that of a drug molecule from its protein target. However, protein mechanics can be complex, sometimes involving coordinated movements along several degrees of freedom. It is not yet clear what the limitations of AMS are when applied to such systems. One might expect the need to identify sophisticated reaction coordinates, or to use multiple AMS runs with different reaction coordinates, when handling complex systems.
Another challenge that may be encountered in complex systems is the sampling of the initial state. Although the applications that we envision AMS being applied to typically have initial states that are easy to sample, this may not be the case in general. Apart from using advanced sampling techniques like umbrella sampling to obtain a good representation of the distribution within the initial state, one could also redefine the initial state to be the region of phase space which is covered over a reasonable time period. In the case of MD simulations, transitions that occur on the microsecond and above time scales are regarded as challenging to simulate, hence an equilibrium sampling over tens of nanoseconds may be regarded as a suitable initial metastable state.
Appendix
A. Analytic Results of Test System
In this section we derive analytic expressions for p and τ in the case where ξ(x) = |x| and the process (Xt)t≥0 evolves in ℝ3 according to the SDE:
Thus, the process is governed by zero-mass Langevin dynamics with the friction coefficient γ and harmonic potential energy V (x) = −k|x|2/2. We also recall the definition of A and B in Eq. (1).
Furthermore, let L be the generator of the process (Xt)t≥0, defined for suitable functions f by
We first prove an analytic formula for τ:
Theorem A.1. Let u0(r) be the average time for the process (Xt)t≥0 to reach the level zB, starting at the level r. Then,
where D is the diffusion coefficient, given by
In particular,
where τ is the average time for the process to go from ∂A to B.
Proof. Pick z ∈ (0, zB), and let the process (Xt)t≥0 be reflected at the level z and absorbed at the level zB. Define uz(x) as the average time for (Xt)t≥0 to be absorbed, given that X0 = x and z < |x| ≤ zB. It is well known that uz(x) is the solution to:
Putting this equation in spherical coordinates, applying spherical symmetry and using D = (γβ)−1, we get
where now uz is a function of r. Re-writing the above expression gives
Using the integrating factor r2 exp(−βkr2) and the reflecting boundary condition we get
Integrating again, using the absorbing boundary condition and finally letting z → 0, we obtain:
We turn now to an analytic expression for p:
Theorem A.2. Let υ(r) be the probability that the process (Xt)t≥0 reaches the level zB before zA, starting at the level r. Then
In particular,
where p is the probability for the process to reach B before A, starting at the level zmin.
Proof. Let υ(x) be the probability for the process (Xt)t≥0 to hit the level zB before zA, given that X0 = x and zA ≤ |x| ≤ zB. It is well known that υ is the solution to
Using spherical coordinates as above we get
Thus
and so
with C1 a constant. Finally
with C2, C3 constants. Using the boundary conditions, we obtain
Remark 2. We remark that by spherical symmetry, the time τ and the probability p do not depend on the initial distributions of the process on ∂A and at the level zmin, respectively.
B. Determination of Diffusion Coefficient D for Analytic Model
The analytic model requires the local diffusion coefficient D as one of two parameters. D was measured using a formula due to Roux et. al. [11] and simplified by Hummer [12], given as follows:
| (5) |
where X is the Cartesian coordinates of the ion, 〈…〉 denotes ensemble average (in practice the quantity measured as an average over time) and D is the local diffusion coefficient at position 〈X〉.
In accordance with the procedure described in Hummer, a potassium ion was allowed to diffuse in the system under the influence of both the harmonic well potential of constant k = 0.08 kcal/mol, and an additional restraining harmonic potential with constant kr centered at points of radius r0 = 0, 10, 20 Å away from the origin. It is later found that D does not depend on the local potential gradient, hence we assume that the value of D obtained is also valid for other k values. Starting from a state with the ion near r0, the system was run at equilibrium for 10 ns with data taken every 10-fs interval. Eq. (5) was then used to calculate the value of D at the respective points. The runs were repeated for kr = 0.1, 0.3, 0.6 kcal/mol Å2. The results are as follows:
| kr ((kcal/mol) Å−2) | r0 (Å) | D (Å2/ns) |
|---|---|---|
| 0.1 | 0 | 256 |
| 0.1 | 10 | 246 |
| 0.1 | 20 | 245 |
| 0.3 | 0 | 272 |
| 0.3 | 10 | 269 |
| 0.3 | 20 | 256 |
| 0.6 | 0 | 247 |
| 0.6 | 10 | 234 |
| 0.6 | 20 | 258 |
Taking the mean and standard deviation gives D = 254 ± 12 Å2/ns.
Footnotes
The Laboratoire International Associé between the Centre National de la Recherche Scientifique (CNRS) and the University of Illinois at Urbana-Champaign (UIUC) is gratefully acknowledged. We acknowledge the financial support of Sanofi R&D and beneficial discussions with Dr. Marc Bianciotto, Dr. Claire Minoletti and Dr. Hervé Minoux, from Sanofi-Aventis. This work has also received support from the EU/EFPIA Innovative Medicines Initiative Joint Undertaking, K4DD grant n○ 115366. The work of T. Lelièvre is supported by the European Research Council under the European Union’s Seventh Framework Programme (FP/2007–2013) / ERC Grant Agreement number 614492. The work of C. Mayne and I. Teo was supported by the National Institutes of Health (NIH) Grant 9P41GM104601. Last but not least, we are grateful to Dr. Mahmoud Moradi for the invaluable discussions about technical aspects of the NAMD colvars module, and to Dr. James Phillips for his assistance in modifying the NAMD code for implementation of the AMS algorithm.
References
- 1.Phillips JC, Braun R, Wang W, Gumbart J, Tajkhorshid E, Villa E, Chipot C, Skeel RD, Kale L, Schulten K. Scalable molecular dynamics with NAMD. J. Comput. Chem. 2005;26:1781–1802. doi: 10.1002/jcc.20289. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Cérou F, Guyader A. Adaptive multilevel splitting for rare event analysis. Stoch. Anal. Appl. 2007;25(2):417–443. [Google Scholar]
- 3.Cérou F, Guyader A, Lelièvre T, Pommier D. A multiple replica approach to simulate reactive trajectories. J. Chem. Phys. 2011;134(5):054108. doi: 10.1063/1.3518708. [DOI] [PubMed] [Google Scholar]
- 4.Guyader A, Hengartner N, Matzner-Løber E. Simulation and Estimation of Extreme Quantiles and Extreme Probabilities. 2010 [Google Scholar]
- 5.Rousset M, et al. Adaptive Multilevel Splitting: unbiased estimators and new algorithmic variants. in preparation [Google Scholar]
- 6.Bréhier C-E, Lelièvre T, Rousset M. Analysis of Adaptive Multilevel Splitting algorithms in an idealized case. http://hal.archives-ouvertes.fr/hal-00987297. [Google Scholar]
- 7.Cérou F, Del Moral P, Furon T, Guyader A. Sequential Monte Carlo for rare event estimation. Stat. Comput. 2012;22(3):795–808. [Google Scholar]
- 8.Fiorin G, Klein ML, Hénin J. Using collective variables to drive molecular dynamics simulations. Molecular Physics. 2013;111(22–23):3345–3362. [Google Scholar]
- 9.Jorgensen WL, Chandrasekhar J, Madura JD, Impey RW, Klein ML. Comparison of simple potential functions for simulating liquid water. J. Chem. Phys. 1983;79(2):926–935. [Google Scholar]
- 10.Beglov D, Roux B. Finite Representation of an Infinite Bulk System: Solvent Boundary Potential for Computer Simulations. J. Chem. Phys. 1994;100(12):9050–9063. [Google Scholar]
- 11.Woolf TB, Roux TBB. Conformational Flexibility of o-Phosphorylcholine and o-Phosphorylethanolamine: A Molecular Dynamics Study of Solvation Effects. J. Am. Chem. Soc. 1994;116(13):5916–5926. [Google Scholar]
- 12.Hummer G. Position-dependent diffusion coefficients and free energies from Bayesian analysis of equilibrium and replica molecular dynamics simulations. New J. Phys. 2005;7(1):34. [Google Scholar]




