Abstract
We report a single-copy tempering method for simulating large complex systems. In a generalized ensemble, the method uses runtime estimate of the thermal average energy computed from a novel integral identity to guide a continuous temperature-space random walk. We first validated the method in a two-dimensional Ising model and a Lennard-Jones liquid system. It was then applied to folding of three small proteins, trpzip2, trp-cage, and villin headpiece in explicit solvent. Within 0.5∼1 microsecond, all three systems were reversibly folded into atomic accuracy: the alpha carbon root mean square deviations of the best folded conformations from the native states were 0.2, 0.4, and 0.4 Å, for trpzip2, trp-cage, and villin headpiece, respectively.
INTRODUCTION
Molecular simulations at room temperature usually suffer from a slow dynamics for large complex systems, such as proteins in explicit solvent. A promising solution to the problem is to use tempering methods, either single-copy based methods1, 2 such as simulated tempering, or multiple-copy based methods such as parallel tempering, also known as replica exchange.3 In either case, the system regularly changes its temperature in a way that is consistent to the underlying thermodynamics. The value of these methods lies in that they can efficiently overcome energy barriers by exploiting a fast dynamics at higher temperatures.
In traditional tempering methods, the temperature is a discrete random variable that can only assume a few predefined values. The success rate of transitions between two neighboring temperatures depends on the overlap of canonical energy distributions at the two temperatures and decays as the system size grows. Thus, to reach an optimal sampling efficiency for a large system, one needs to narrow down the temperature gap, and to increase the number of sampling temperatures. In simulated tempering, the number of weighting parameters to be estimated for simulation increases with the number of temperatures. In replica exchange, the node-node communication cost increases with the number of temperatures. Naturally, it is desirable to have an efficient tempering method that does not depend on a discrete-temperature setup.
In this paper, we report a single-copy tempering method in which the temperature is a continuous variable driven by a Langevin equation. It is based on an improved version of a previous method.2 By employing improved estimators for thermodynamic quantities, one cannot only realize an efficient tempering but also correctly calculate thermodynamic quantities for the entire temperature spectrum. The essential feature of the method is the calculation of the thermal average energy along the simulation trajectory. After the convergence of , the partition function and other thermodynamic quantities can be easily derived.
The paper is organized as follows. In Sec. 2, we give a theoretical derivation of the method. In Sec. 3, the method is verified on a two-dimensional Ising model and a Lennard-Jones liquid system, where the exact thermodynamic quantities are either known or accurately computable. In Sec. 4, we apply the method to the folding of three small proteins, trpzip2, trp-cage, and villin headpiece, in explicit solvent. The minimal alpha-carbon root mean square deviation (Cα-RMSDs) of the best folded conformations from the native states are 0.2, 0.4, and 0.4 Å, respectively (the last figure for villin headpiece is measured from an x-ray reference structure; and it should be 1.0 Å if it is measured from an NMR reference structure).
METHOD
Our method samples the system in a continuous temperature range and calculates thermodynamic properties as functions of the temperature. As we shall see, a random walk in temperature space only requires an estimate of the average energy at the current temperature in order to correctly populate the desired distribution. We present an efficient way for estimating the average energy and use it to perform a fast sampling along the temperature. In addition, an adaptive averaging scheme is used to improve convergence in early stages.
This section mainly concerns the detailed description of the method. A relatively self-contained outline is first presented in Sec. 2A. The rest of the section is organized as follows. In Sec. 2B, we review basics of sampling in a generalized ensemble where the temperature is a continuous random variable. In Secs. 2C, 2D, we present integral identities that help the ensemble to asymptotically reach the desired distribution. In Sec. 2E, we present an adaptive averaging scheme to accelerate initial convergence.
Brief outline of implementation
The method can be implemented as follows. It concerns the simulation of a system in a given temperature range (βmin,βmax) according to a predefined temperature distribution w(β), which is usually proportional to 1∕β for a molecular system (the choice is explained in Appendix C). Note, we work with the reciprocal temperature β=1∕(kBT), where kB is the Boltzmann constant and T is the regular temperature, β here is a variable that continuously changes within the temperature range.
In each simulation step, we first change system configuration according either to constant temperature molecular dynamics or to Monte Carlo methods. Next, we update statistics about the potential energy and then compute an estimated average energy . The energy fluctuation, measured from the difference between the instantaneous potential energy E and the averaged value at the current temperature, determines the amount of temperature change that the system can afford to maintain the distribution of the generalized ensemble. It thus can be used to drive a temperature-space random walk in a way that preserves the underlying thermodynamics at each individual temperature. Practically, we use the following Langevin equation to guide the temperature space random walk,
(1) |
where E is the current potential energy, and ξ is a Gaussian white noise that satisfies ⟨ξ(t)⋅ξ(t′)⟩=δ(t−t′). Note, here t is the time scale for integrating the Langevin equation, which does not necessarily coincide with the actual time in molecular dynamics. Apart from the random noise, the rate of change of the temperature is determined by the difference between the instantaneous energy E and the average energy at the current temperature . Thus, Eq. 1 tends to raise the temperature when the instantaneous energy E rises above the average value , or to lower the temperature when E falls under . A derivation of Eq. 1 can be found in Appendix A. In implementation, β is first converted to its reciprocal kBT=1∕β; then kBT is updated according to Eq. 1; finally the new β is computed from the reciprocal of the updated kBT. We use kBT instead of β as the variable of integration, because in this way, the magnitude from the random noise is proportional to the temperature value . The effect of the temperature change can be realized as a scaling of velocity or force in molecular dynamics. We repeat the process for each simulation step until the simulation ends.
In order to interpolate thermodynamic quantities on a continuous temperature range, we divide the entire range evenly into many small bins (βi,βi+1). The bin size is used for applying integral identities and is much smaller than the gap between neighboring sampling temperatures in traditional tempering methods, such as replica exchange or simulated tempering. During simulation, each bin i collects separate statistics on the potential energy and its variance along the trajectory for states with β∊(βi,βi+1). Statistics in different bins are later combined together to form unbiased estimates of the average energy. For a complex molecular system, the adaptive averaging in Sec. 2E can be used to improve the early convergence.
For statistical efficiency, , β∊(βi,βi+1), is not calculated from the average energy of the bin, but instead from a large temperature window (β−,β+) containing the bin as
(2) |
where Δβj=βj+1−βj is the bin width; ⟨E⟩j is the average energy from the jth bin (βj,βj+1), is a modulating factor to ensure an unbiased estimate, see Fig. 1b, and its computation is detailed in the next paragraph. The temperature window (β−,β+) is determined from the current temperature β in such a way that β is approximately at the center of the corresponding window (β−,β+).
The first step of computing is to determine two parameters a+ and a− by solving the following two equations:
(3) |
(4) |
Here, a+ and a− are parameters defined in the equation ϕ(β)=ϕs(β)+ϕt(β), whose two components ϕs(β) and ϕt(β) are defined as
(5) |
and
(6) |
respectively. Figure 1a schematically illustrates ϕs(β), ϕt(β), and ϕ(β). In Eq. 4, the inner bracket ⟨⋯⟩β′ is a configuration average of the energy fluctuation ⟨ΔE2⟩β′ at a fixed temperature β′. For β′∊(βj,βj+1), since the bin size is small, we compute the energy fluctuation from the states collected in the bin (βj,βj+1), and use it to approximate ⟨ΔE2⟩β′. The outer angular bracket ⟨⋯⟩(β−,β+) denotes an average over temperature β′ within the window (β−,β+), it is computed as a sum of the energy fluctuation from different bins within (β−,β+) with ϕ(β′) being the coefficient of combination. After the averaging, Eq. 4 is a simple linear equation of a+ and a−. By solving Eqs. 3, 4, a+ and a− are determined. In a physical solution, both a+ and a− are non-negative. If the linear equations lead to a negative value for either a+ or a−, zero is used instead. This measure ensures the robustness of our estimator and is thus useful in early stages when the energy fluctuation ⟨ΔE2⟩β′ is unreliable. The determination of a+ or a− completely specifies ϕs(β) and its derivative .
Generalized ensemble with a continuous temperature
We start our method by constructing a generalized ensemble in which the temperature β is a continuous variable in a given range (βmin,βmax). To sample the system correctly, we also need to make sure that the configurational distribution at a particular β is identical to that of the canonical ensemble at the same temperature. The aim of the method is to correctly populate states in the generalized ensemble and to extract thermodynamic properties for the entire temperature range.
First of all, the generalized ensemble is completely specified by an overall β-distribution p(β). Once p(β) is given, the complete distribution of atomic configuration X as well as the temperature β is also determined
(7) |
where E(X) is the potential energy of configuration X, and Z(β)=∫exp[−βE(X)]dX is the canonical partition function. Equation 7 is due to the requirement of preserving a canonical distribution at each temperature. It is easily verified that p(β) is recovered after we integrate p(β,X) over all configurations.
The usefulness of the joint distribution p(β,X) in Eq. 7 lies in that it specifies a temperature distribution under a fixed configuration X. For a fixed configuration X, we can perform temperature-space sampling according to p(β,X) and replace p(β) by any desired temperature distribution w(β), which is fixed during simulation, in Eq. 7. If the configurational space is sampled according to the Boltzmann distribution, the resulting overall temperature distribution after an infinitely long simulation trajectory, must be identical to the desired one w(β).
However, the exact Z(β) is usually unknown in advance. We therefore use a modified version of Eq. 7
(8) |
to conduct sampling in the temperature space, where an approximate partition function is used in place of Z(β). Note, if differs from Z(β) in Eq. 8, w(β) is no longer the overall temperature distribution, but only a parameter that specifies p(β,X), which is used in guiding the temperature-space sampling. The overall temperature distribution p(β) is calculated from integrating the joint distribution p(β,X) over configurations
(9) |
(10) |
In simulation, is adaptively adjusted and w(β) is fixed. Therefore, the overall distribution p(β) varies according to Eq. 10. Upon convergence, , the overall temperature distribution p(β) converges to the desired one w(β).
Given a configuration X, as well as the joint distribution p(β,X) Eq. 8, sampling along the temperature can be performed by the Langevin equation Eq. 1, in which the estimated average energy relates to the estimated partition function as . One can demonstrate its correctness by solving the corresponding Fokker–Planck equation, see Appendix A.
The remaining task is to make sure the convergence of the estimated partition function to the correct one Z(β). As evidenced by Eq. 10, the current overall temperature distribution p(β) is close to the desired one w(β) only if the estimated the partition function is sufficiently accurate. It is interesting to note that the Langevin equation Eq. 1 does not involve the estimated partition function itself, but its derivative instead. We should therefore exploit this feature and focus on a technique that adaptively improves the estimate, .
Unbiased estimate of a partition function
The asymptotic convergence of the partition function requires that approaches to Z(β) at any β in the range (βmin,βmax). In implementation, we divide the temperature range to many narrow bins (βi,βi+1). Thus we lower the requirement to that at any bin boundary βi, should equal to Z(βi). By choosing a small bin size, we can ensure that the deviation from to Z(β) is negligible for all practical purposes. Further, to remove the dependence on a reference value of the partition function, the convergence condition is unambiguously expressed as a condition on ratios
(11) |
Equation 11 can be rearranged as2
where we have used Eq. 10 on the second line; the inner bracket ⟨E⟩β denotes the average energy at a particular energy β; on the last line, we convert the integral over the small temperature bin (βi,βi+1) to a temperature average in the same bin, as represented by the outer bracket ⟨⋯⟩i, which is formally defined as .
During simulation, if we adaptively enforce the right hand side of the above equation to be zero, i.e.,
(12) |
the left hand side naturally vanishes as well, i.e., . In this way, the partition function as a function of the temperature can be obtained.2
We shall proceed by assuming a sufficiently small bin size and (i) being a constant within a bin i, and correspondingly, varies linearly with β, and (ii) p(β)∕w(β) can be treated as a constant. Equation 12 is then simplified as
(13) |
and the ratio of the partition function is estimated as
where Δβi=βi+1−βi.
A direct implication of Eq. 13 is that one can estimate from averages from statistics accumulated in bin i. Such an approach (which is similar to those used in the force averaging method4) is, although correct, ineffective when the bin size is small because the amount of statistics within a bin shrinks with the bin size. On the other hand, a large bin size can lead to a significant deviation from the desired temperature distribution, since we assumed a constant within a bin. This dilemma can be resolved by using integral identities that remove the bin size dependence.
Estimators based on integral identities
We now present a method for drawing an unbiased estimate from a large temperature window instead of a small temperature bin. The method removes bin size dependence by combining statistics from neighboring bins in a way that avoids systematic error. A similar technique of employing integral identities to improve statistics was previously used in improving statistical distributions.5, 6
We aim at transforming the right hand side of Eq. 13 from an average over a single bin (βi,βi+1) to an average over a larger temperature range (β−,β+) that encloses the bin. To do so, we use the following integral identity:
(14) |
where ϕ(β) is a function that vanishes at the two boundaries, i.e., ϕ(β−)=ϕ(β+)=0; on the second line, we convert the difference between the two boundaries to an integral within the temperature range; the identity ∂⟨E⟩β∕∂β=−⟨ΔE2⟩β from statistical mechanics is also used.
We choose ϕ(β) as a superposition of a smoothly varying ϕs(β) that spans over the entire window (β−,β+) and a localized function ϕt(β) limited within the bin (βi.βi+1), see Fig. 1 and Eqs. 5, 6, in which a+ and a− are two non-negative parameters that sum to unity, i.e., a++a−=1. Note at β=βi+1, the sudden jump in ϕs(β) is exactly cancelled by that in ϕt(β), thus we can ignore the δ-functions in and in actual computation.
The purpose of the decomposition of ϕ(β) into ϕs(β) and ϕt(β) is to use the localized function to create an integral exactly equal to the right hand side of Eq. 13. In this way, the integral over the small bin is transformed to another one, but over a larger temperature window
(15) |
In early stages of simulation, the energy fluctuation ⟨ΔE2⟩β in the second term of the right hand side of Eq. 15 can be unreliable for a complex system. To avoid direct inclusion of energy fluctuation, we choose the combination parameters a+ and a− in such a way that the fluctuation term vanishes,
(16) |
Equation 16 and a++a−=1 yield a solution of a+ and a−, which also determines . Thus we have,
(17) |
Since is a constant within a single bin, it can be factored out of the integral when integrating each individual bin. The integral is thus converted to a sum over averages,
(18) |
A comparison with Eq. 13 shows that Eq. 18 is merely a linear combination of ’s obtained from different bins. The auxiliary function serves as a set of coefficients of combination, whereas Eq. 16 ensures the asymptotical convergence.
The above technique of extracting an estimate from an integral identity can be employed in computing other thermodynamic quantities, such as the average energy, heat capacity, and energy histogram. Thus, we are able to calculate these quantities at a particular temperature, even though the simulation is performed in an ensemble where the temperature is continuous.
The result for the average energy is most easily obtained. Consider a limiting case where βi+1→βi (with ϕ(β) modified accordingly), Eq. 15 immediately becomes an unbiased estimate of the thermal average energy ⟨E⟩βi exactly at βi.
Similarly, the heat capacity CV(β) can be calculated from the energy fluctuation as β2⟨ΔE2⟩β. The energy fluctuation at a particular temperature β is computed as
(19) |
where we have substituted −⟨ΔE3⟩β for ∂⟨ΔE2⟩β∕∂β.
Estimating the energy histogram requires additional care in choosing the function ϕ(β), in order to ensure that the histogram is non-negative everywhere. It is shown in Appendix D that the optimal estimate of an energy histogram at temperature β is
(20) |
where on the first line, the energy histogram hβ(E,E+ΔE) at β is defined from the constant temperature energy distribution pβ(E) at the same temperature; on the second line, it is converted to an average over the joint temperature-energy p(β,E) distribution in the generalized ensemble, which can be measured from the temperature-energy histogram in simulation trajectory. Note, for simplicity, we have assumed that the energy bin size is small compared with the energy fluctuation of a typical Boltzmann distribution. Equation 20 resembles the result from the multiple histogram method,7 and therefore can be treated as its counterpart in a continuous temperature ensemble.
For a quantity whose statistics are not fully accumulated during simulation (e.g., it is calculated from periodically saved trajectory snapshots after simulation), we use the following reweighting formula to obtain its value at a particular temperature β
(21) |
where serves as a weighting function for borrowing statistics from β′ to β. Note, Eq. 21 is unbiased estimator even if contains error.
Adaptive averaging
We now introduce an adaptive averaging scheme for accelerating the convergence in initial stages of a simulation. Since we start from zero statistics, the error in initially estimated can lead to a slow random walk in the temperature space. The adaptive averaging scheme2 overcomes the problem by assigning larger weights toward recent statistics, and thus encourages a faster random walk in the temperature space.
Usually, for a statistical sample of size n, the average of a quantity A is calculated as an arithmetic mean , where is the sum of A and is the sample size n. In simulation, since statistics is collected along the trajectory, the sample size n increases as simulation progresses. To increase the weight toward recent Ai’s, we redefine and as
where γ(<1) is used to gradually damp out old statistics. Such an average can be easily implemented as recursions,
(22) |
However, if a constant γ is used, an average derived from Eq. 22 does not asymptotically reduce its error due to that the sample size S1 ultimately saturates to a fixed value 1∕(1−γ). Therefore, we use
(23) |
where Cγ is a numerical constant, to gradually reduce the difference between γ and 1.0. In this way, both a fast random walk in early stages and an asymptotical convergence can be achieved.
NUMERICAL RESULTS
Ising model
As the first example, we test our method on a 32×32 Ising model, which is a nontrivial system with exactly known thermodynamic properties.8 Results from the alternative method of estimating , described in Appendix B, are also included. Parameters common to the two methods were set to be the same.
The temperature range was β∊(0.35,0.55) or T=(1.818,2.857), which covered the critical temperature T≈2.27 of the phase transition, and the temperature distribution was defined by a constant w(β), or a flat-β histogram. The bin size for collecting statistics was δβ=0.0002, and thus there were 1000 bins in the entire temperature range. For each temperature bin (βi,βi+1), the temperature window for applying integral identities was given by (β−,β+)=(βi−Δβ,βi+1+Δβ) with Δβ=0.02 in evaluating Eq. 18; thus 201 bins were included in a window (except at temperatures near the boundaries, where the largest possible value of Δβ is used). The Langevin equation was integrated after every 100 Monte Carlo moves, with an integrating step Δt=2.0×10−5. The whole simulation stopped after 106 Monte Carlo moves per site.
The calculated partition function, average energy and heat capacity are shown in Fig. 2. The thin solid lines and dashed lines are for the method introduced in Sec. 2D (method 1), and that introduced in Appendix B (method 2), respectively. In most cases, both results for the partition function and the average energy coincide with exact results,8 see Figs. 2a, 2c, 2d. Even for the heat capacity, which is the second order derivative of the partition function and is harder to compute, the deviations in both cases are small. This indicates that our method is unbiased and it can produce exact thermodynamic quantities asymptotically.
We now show that the correct energy distribution can be reconstructed from Eq. 20. Since the energy levels are discrete and the energy bin size we used was equal to the smallest gap between energy levels, a normalized energy histogram is equivalent to the energy distribution. During simulation, the current potential energy was registered into the histogram every 100 Monte Carlo moves, thus there were roughly 105 samples in the entire histogram. The reconstructed energy distributions at a few representative temperatures are shown in Figure 3. The temperature window for estimating the energy distribution at β was (β−Δβ,β+Δβ) with Δβ=0.02. One can see a good agreement between the integral identity Eq. 20 (the thin solid line), and the exact distribution (thick solid line), which was computed from the exact density of states.9 For comparison, distributions constructed by averaging energy distributions of two adjacent bins are shown as dashed lines. It is apparent that the simple average yields more noisy results than the integral identity Eq. 20. The reduced error is due to Eq. 20 being able to access statistics from a much larger temperature window than from two neighboring bins.
Lennard-Jones system
As the second example, we test the method on an 864-particle Lennard-Jones liquid. In reduced units, the Lennard-Jones potential for a particle pair separated by distance r is
where the units of energy, mass, and length are 1.0. In the simulation, the density was 0.8; the cutoff was 2.5; and the temperature range was β=(0.48,1.02), corresponding to T=(0.98,2.08). We used Monte Carlo to generate configuration changes. In each step, a random particle was displaced randomly in each of x, y, and z directions according to a uniform distribution in (−0.1,0.1). After a Monte Carlo step, we applied the Langevin equation Eq. 1 with an integration time step Δt=0.0002. The system was simulated for 105 sweeps (a sweep=a step per particle). Coordinates were saved every ten sweeps for data analysis. The overall temperature distribution w(β) was proportional to 1∕β (the choice of optimal w(β) is discussed in Appendix C). The temperature bin for collecting statistics was δβ=0.0005. The window size Δβ for applying the integral identity Eq. 18 was 0.05.
The simulation results for the estimated average energy ⟨E⟩β and are shown in Fig. 4a. Due to a small bin size, the difference between the two is invisible. The dots on the figures represent results from independent constant temperature simulations, each of which uses the same amount of simulation time. A good agreement between the two indicates that our method can produce exact thermodynamic quantities for the entire temperature spectrum with less simulation time.
Furthermore, we reconstructed the radial distribution function g(r) at a particular temperature according to Eq. 21. The reconstruction was performed after simulation and was based on the saved coordinates along the trajectory. From Fig. 4b, we see a good agreement between reconstructed g(r)’s and those from constant temperature simulations at two different temperatures β=0.5 and 1.0, or T=2.0 and 1.0, respectively.
APPLICATIONS IN FOLDING SMALL PROTEINS
In this section, we report applications of the method in folding of several small proteins. The method was implemented in a modified GROMACS 4.0.5,10 using AMBER force field ports11 with TIP3P water model.12 In all cases, the particle-meshed Ewald method13 was used for handling long range electrostatic interaction, and the velocity-rescaling method was used as thermostat.14 For constraints, we used the SETTLE algorithm for water molecules,15 and the parallel LINCS algorithm16 for proteins. Since proteins drastically changed their configurations during simulations, dynamic load balancing was turned on when using domain decomposition.
For configuration-space sampling under a fixed temperature, the canonical ensemble, i.e., the constant (N,V,T) ensemble, was used. A 10 Å cutoff was used for Lennard-Jones interaction, electrostatic interactions and neighboring list. Since the temperature in our method is a variable, the temperature change was realized by scaling the force according to F′=(β∕β0)F=(T0∕T)F, with F′ and F being the scaled and the original force, respectively. In this way, the scaled potential energy transforms a canonical ensemble at T0 to another temperature T, for sampling in the configurational space. On the other hand, the thermostat temperature T0, which controls the kinetic energy, is unaffected and can be maintained at a fixed value T0=480 K. Three separate thermostats were applied to protein, solvent and ion groups, and the coupling time τT for the thermostats was 0.1 ps.
The time step for molecular dynamics was 0.002 ps. The center of mass motion was removed every step. Trajectory snapshots were saved every 2 ps. Since the force and energy calculation was much more time consuming than estimating , we applied the Langevin equation in every molecular dynamics step using an integration step of Δt=10−4. The parameter Cγ in the adaptive averaging Eq. 23 was 0.1 in all cases.
Trpzip-2
The first system is a 12 amino acid β-hairpin tryptophan zipper, whose Protein Data Bank (PDB) ID code is 1LE1, and its sequence is SWTWENGKWTWK.17 A unique feature of this short hairpin is that its four tryptophan side chains are locked against each other to stabilize the structure.
Previously, the system was intensively studied both in explicit solvent18 and in implicit solvent.19, 20 However, de novo folding in explicit solvent from an extended chain, see Fig. 5a, is much more challenging. We used AMBER99SB as the force field, which is a modified version of AMBER99 force field21 with updated φ and ψ torsions.22
We report four simulation trajectories. All of them reached atomic accuracy within in a time scale of 1 μs. In all the cases, we used a cubic 45×45×45 Å3 box, filled with 2968 water molecules as well as two Cl− ions. The temperature range was β∊(0.20,0.41), which corresponded to T=293.6–601.9 K temperature range. The grid spacing for Fourier transform was 1.15 Å, and the alpha parameter was 0.3123 Å−1 for the Ewald method. In application of integral identities equations 18, 19, 20, the temperature windows size was 8% of temperature value, e.g., at β=1.0∕(kB500 K)≈0.241 the temperature window is (β−Δβ,β+Δβ), with Δβ=4%×β≈0.010, which can be translated to T=(480.8,520.8) K. At boundaries, we used the largest possible size that allows a symmetrical window.
A typical folded structure is shown in Fig. 5b, with its root mean square deviations (RMSD) for alpha-carbon (Cα) and heavy atoms being 0.25 and 1.08 Å, respectively. The lowest Cα-RMSD and heavy atom RMSD found in the four trajectories are listed in Table 1, in which we also list the approximate first time of stably reaching the atomically accurate native structure (the criteria were Cα-RMSD<0.5 Å). It is interesting to note that even among structures with lowest RMSDs, tryptophan side chains can still adopt different conformations. For example, in trajectories 1 and 3, we found nativelike structures with one of the tryptophan side chain (TRP9 or TRP4) flipped 180° with respect to the native configuration. This suggests that the free energy change involved in flipping a TRP side chain is small.
Table 1.
Traj. ID | Lowest Cα-RMSD (Å) | Lowest heavy atom RMSD (Å) | First time of stably reaching atomic accuracy of the native state (ns) |
---|---|---|---|
1 | 0.20 | 1.00 | 20 |
2 | 0.25 | 0.84 | 520 |
3 | 0.20 | 0.88 | 530 |
4 | 0.25 | 1.31 | 1080 |
In Fig. 6, Cα-RMSD, radius of gyration, instantaneous temperature, and potential energy along trajectory are shown in (a)–(d), respectively, for the two independent simulations, trajectory 1 and 3. In trajectory 1, the folded state was reached within 20 ns. This suggests the possibility of fast folding. However, in other trajectories, it took longer for the system to reach the native structure. In trajectory 3, upon reaching of the native structure, the system lingered in a state of low-temperature, low-energy state, and small radius of gyration for about 100 ns. This corresponds to the fact that the folded state has a lower energy than the unfolded state and thus occupies a larger fraction in a low temperature Boltzmann distribution. This feature serves as a signature of the system reaching a native structure, and can be useful in folding prediction where the native structure is unknown.
Figure 7a shows the distribution along the Cα-RMSD at three different temperatures calculated from trajectory 3. The distribution demonstrates two well-separated peaks, corresponding to roughly defined folded and the unfolded states, respectively. The average Cα-RMSD from the native structure is roughly 0.8 and 5.5 Å for folded and unfolded states, respectively. As the temperature increases, the first peak gradually diminishes, whereas the second peak dominates.
The folding temperature can be computed by assuming a two-state model of folding. The folding fraction P was first calculated as the fraction of configurations with Cα-RMSD less 2.0 Å at different temperatures. The curve P(β) as a function of temperature β was then fitted against a two-state model in the range β∊(0.24,0.4), equivalentlyT∊(300.9 K,501.5 K),
(24) |
where the values of the three parameters were determined as P0=0.61, ΔE=38.96 kJ mol−1, and βm=0.350 kJ−1 mol by regression, see Fig. 7b. Here, ΔE is the energy difference between the folded state and the unfolded state; βm is the melting temperature. In the simplest two-state model, the folding fraction increases monotonically to unity as the temperature decreases to zero. Here, the maximal fraction was changed from 1.0 to an adjustable parameter P0, which helped fitting the calculated P(β) to the two-state model. Physically, such a modification implies the existence of many configurations with energy similar to that the native one but with different structures in our simulation trajectory.
If the volume change during folding is ignored, the enthalpy change between the folded state and the unfolded state is roughly equal to the energy change ΔH≈39.0 kJ mol−1. The entropy difference of the folded and unfolded states is ΔS=βmΔE=113 J mol−1. These values are relatively small compared with the experimental values ΔHexp=70.2 kJ mol−1, ΔSexp=203.3 J mol−1.17 However, the estimated folding temperature from our calculation 344 K is close to the experimental result 345 K.17 From trajectory 3, which yields the highest folding fraction at 300 K among four trajectories, the fraction of folded states at 300 K is roughly 55%, which still differs significantly from the experimental value 91%.17 For the other three trajectories, the calculated fractions are even smaller. The difference between our calculation and experiments were likely due to insufficient sampling and∕or force field inaccuracy.
Figure 8 shows the heat capacity versus temperature from Eq. 19. The difference between three independent trajectories was small, suggesting thermodynamic properties of the entire system, protein and water, reaching convergence. It is interesting that in our simulation, the heat capacity is not a constant, but inversely proportional to the temperature as CV≈4.2×103∕T kJ mol−1 K−1.
Figure 9a shows the joint distribution of the radius of gyration Rg (calculated from Cα atoms) and the Cα-RMSD. Generally the two measures are positively correlated. However, the smallest radius of gyration does not occur at the native structure for the hairpin, whose Rg is roughly around 5.8 Å. A non-native structure, on the other hand, had a smaller Rg around 4.9 Å, but a larger RMSD around 3 Å. Figure 9b shows the joint distribution of temperature and energy. A typical temperature fluctuation is only around 5 K, which is roughly the magnitude of temperature gap in other tempering methods based on a discrete temperature, such as replica exchange. On the other hand, the temperature window size used in our simulation was much larger. This suggests that our method could more efficiently use statistics to facilitate the temperature-space random walk.
Trp-cage
The second application is a 20 amino acid alpha helical protein, tryptophan cage (trp-cage).23 The PDB code is 1L2Y, and the amino acid sequence is NLYIQWLKDGGPSSGRPPPS. The system was extensively studied by experiments23, 24 as well as by various sampling techniques, either in explicit solvent25, 26 or in implicit solvent.20, 27 We again used AMBER99SB as the force field.22
We simulated the system in a cubic 46×46×46 Å3 box, filled with 3161 water molecules and two Cl− ions. The grid spacing for Fourier transform was 1.19 Å, and the alpha parameter was 0.3123 Å−1. The initial structure was an open chain see Fig. 10a, which was constructed by bending a fully extended chain to fit into the box.
We report three independent 1 μs simulation trajectories. The temperature range for trajectory 1 was β∊(0.24,0.41), or T∊(293.6 K,501.5 K). In trajectories 2 and 3, we used a larger temperature range β∊(0.20,0.41), which covered a 293.6 K∼601.9 K temperature range. The temperature bin size δβ was 0.0005, 0.0005, and 0.0002 respectively. In applying integral estimators, the temperature windows size was 10%, 10%, and 8% of temperature value for trajectories 1, 2, and 3, respectively. In all cases, we used the alternative estimator introduced in Appendix B.
All three simulations independently reached atomically accurate native configurations. A typical folded structure in trajectory 1 is shown in Fig. 10b. The Cα-RMSDs from the three trajectories were 0.43, 0.48, and 0.44 Å, respectively. The average Cα-RMSD for the native structure was around 0.8 Å. The lowest RMSDs for all heavy atoms were 1.34, 1.47, and 1.46 Å, respectively.
The Cα-RMSD, radius of gyration, instantaneous temperature, and potential energy along trajectory are shown in panels (a)–(d), respectively, of Fig. 11, for the three independent simulations. In each trajectory, there were two folding events reaching an atomic accuracy, e.g., for trajectory 2, the native structure was reached at 350 and 560 ns. As in the trpzip2 case, the system stayed around a low-temperature and low-energy state for 30–100 ns upon reaching the native state. For the folding speed, it appears that simulations 2 and 3, which used a higher roof temperature Tmax=600 K, tended to reach the native state sooner than simulation 1, whose roof temperature Tmax=500 K was lower. However, in simulation 1, the system was able to stay in the native states longer and performed a more detailed sampling at the low temperature end.
Trajectory 1 yielded the largest fraction of the folded state at 300 K, 19%, which is still much lower than experiment value, 70%.23, 28 By fitting the folding fraction, computed from the fraction of states with Cα-RMSD less than 2.2 Å cutoff, to the two-state formula Eq. 24, the parameters are P0=0.50, ΔE=21.1 kJ mol−1, and βm=0.367 kJ−1 mol. The enthalpy change is thus ΔH≈21 kJ mol−1, which is relatively small compared with the experimental values ΔHexp=56.2 kJ mol−1.28 The estimated folding temperature from our calculation is 328 K, which is slightly higher than the experimental result 315 K.28
Villin headpiece
Our last application was the villin headpiece, a 36 residue alpha-helical protein HP36. The PDB ID is 1VII, and the amino acid sequence is MLSDEDFKAVFGMTRSAFANLPLWKQQNLKKEKGLF. This system was the first protein partially folded in explicit solvent.29 Recently, a high resolution x-ray structure of slightly modified protein HP35, PDB ID 1YRF, was published.30 The sequence of HP35 is LSDEDFKAVFGMTRSAFANLPLWKQQHLKKEKGLF, in which the N-terminal methionine (MET) of the original sequence was chopped off and the 28th residue, an asparagine, was replaced by a hisidine. Both sequences were studied in literature, both by simulations,31, 32, 33, 34 and in experiments.35, 36
However, there is a significant difference between the NMR structure of HP36 and the x-ray structure of HP35, about 1.62 Å difference in terms of Cα-RMSD. Further, while molecular dynamics simulations on HP35 reached an atomic accuracy in implicit solvent32 as well as in explicit solvent,34 simulations on HP36, especially in explicit solvent, yielded relatively poorer results. The difference between the two structures could be due to (1) the intrinsic difference between HP35 and HP36 or (2) differences in the two experimental techniques, NMR versus x ray.30
In this simulation, we used AMBER03 (Ref. 37) as the force field, which was previously used to fold HP35 to an atomic accuracy in implicit solvent.32 Another reason of choosing AMBER03 instead of AMBER99SB was that the latter force field might slightly disfavor helical conformations.38 To reduce simulation size, we used a dodecahedron simulation box with edge length being 24.1 Å to accommodate the protein as well as 3343 water molecules, and two Cl− ions. The volume of the box was 53.3×53.3×37.6 Å3=1.069×105 Å3. The initial conformation of the protein was a fully extended chain, which was bent from a linear chain to fit into the box, see Fig. 12a. The grid spacing for Fourier transform was 1.19 Å and the alpha parameter was 0.3123 Å−1.
We report two independent simulation trajectories, each of 2 μs. The temperature range for trajectory 1 was β∊(0.18,0.41), or T∊(293.6 K,668.7 K); that for trajectory 2 was β∊(0.20,0.41), or T∊(293.6 K,601.9 K). The temperature bin size was 0.0002 in both cases. The integral estimator introduced in the Sec. 2 was used, and the temperature windows size was 8% of current temperature value.
The alpha-carbon root mean square deviation can be calculated from the NMR structure as well as the x-ray structure. In both cases, the N-terminal MET and leucine (LEU) as well as the C-terminal phenylalanine (PHE) are not included in our calculation.32 Due to the flexibility of the C-terminal and the N-terminal helix, we also calculate the Cα-RMSD for residues 9–32 of HP36 as in the literature,29, 31 denoted as RMSDcore here. Note, in terms of Cα-RMSDcore, the NMR reference structure differs from the x-ray reference structure by 0.87 Å.
The lowest Cα-RMSDs reached in the two simulation trajectories are listed in Table 2. The lowest Cα-RMSDcore from the NMR structure are 0.72 and 0.73 Å, for the two trajectories, respectively. These figures are smaller than those from a previous study, in which the lowest RMSDcore was around 1.5 Å.31 In both trajectories, the best folded structures are more similar to the x-ray structure than to the NMR structure. The average Cα-RMSD and Cα-RMSDcore from the NMR structure are 1.30 and 1.95 Å, respectively. In comparison, the two figures drop to 0.90 and 1.10 Å if the reference is switched to the x-ray structure.
Table 2.
Traj. ID | RMSDcore (NMR) | RMSD (NMR) | RMSDcore (x ray) | RMSD (x ray) | First time of reaching atomic accuracy of the native state (ns) |
---|---|---|---|---|---|
1 | 0.72 | 1.32 | 0.30 | 0.42 | 310 |
2 | 0.73 | 0.99 | 0.30 | 0.40 | 26 |
Table 2 also lists the first time for reaching the native structure. In both trajectories, the time (310 ns for trajectory 1 and 26 ns for trajectory 2) is significantly shorter than those in a very recent study (where the folding occurs in 5–6 μs).34
In Figs. 12b, 12c, we show the superposition of typical folded structures to the NMR reference structure and the x-ray reference structure, respectively, from simulation 1. Although generally deviations from the native structures are small, the position of the N-terminal helix differs appreciably from the NMR native structure [notice the difference in position of PHE 7 and PHE 11 in Fig. 12b], whereas the difference is much smaller compared with the x-ray structure [PHE 7 and PHE 11 of the two structures are superimposable in Fig. 12c].
Figures 13a, 13b show the Cα-RMSDcore (the stable region residues 9–32) versus the Cα-RMSD (entire chain) for the NMR and x-ray reference structures, respectively. It is clear that the best folded structures have larger deviations from the NMR structure than from the x-ray structure. Moreover, the Cα-RMSDcore is less consistent with the Cα-RMSD with respect to the NMR structure than to the x-ray structure, which can be a result of the extremely flexible N-terminal helix. In general, we found that the folded structures in our simulation are closer to the x-ray structure. In both case, we observe many metastable states around the native state, demonstrating an extremely rugged energy landscape.
The folding fraction, computed according to Cα-RMSD from the x-ray structure and using 3.0 Å as cutoff, was fitted against Eq. 24, and yielded P0=0.21, ΔE=20.9 kJ mol−1, and βm=0.349 kJ−1 mol. Assuming that the volume change during folding is negligible, we estimate the enthalpy change ΔH=21 kJ mol−1, and the folding temperature Tm=345 K. Although the folding temperature agrees well with the experimental value 342 K, the folding enthalpy is relatively small compared to ΔHexp=113 kJ mol−1.35
CONCLUDING DISCUSSIONS
In conclusion, we presented a single-copy enhanced sampling method for studying large complex biological systems. The method was validated in an Ising model as well as in a Lennard-Jones fluid, and was successfully applied to folding of three small proteins, trpzip2, trp-cage, and villin headpiece, in explicit solvent. In all three protein cases, we reversibly reached atomic accuracy of the native structures within a microsecond. Since our method is based on a single trajectory, it is computationally less demanding to reach a long time scale than tempering methods based on multiple copies, such as the replica exchange method.
In terms of sampling efficiency, in addition to the advantage from using a continuous temperature generalize ensemble, which eases the issue of low temperature-transition rate in large complex systems,2 a major improvement in this work is the employment of integral identity that efficiently estimates the average energy for the temperature space random walk. The integral identity draws estimates from a large temperature window instead of from a single bin to improve statistics. This strategy makes the method more robust and applicable to large and complex systems.
Additionally, the adaptive averaging scheme used for refreshing statistics can effectively boost the temperature space random walk in early stages, when the damping magnitude γ still differs significantly from unit. It could account for several fast folding (<50 ns) events we observed. However, in later stages of simulations, the boosting effect gradually weakens as the damping magnitude γ is switched toward 1.0 [see Eq. 23] to achieve an asymptotic convergence.
In folding applications, the calculated values for the folding enthalpy and folding temperature still showed discrepancies from experimental results. This was possibly due to errors in the force fields26, 34 as well as limited simulation time. These factors might lead to overpopulation of non-native conformations at room temperature. This observation is in line with a very recent long time replica exchange simulation on trp-cage with an aggregated 40×1 μs, which also revealed discrepancy between experimental and simulation results.26 Even in the presence of enhanced sampling, one would desire to perform much longer simulations to allow much more folding events to atomic accuracy. In this way, one can more accurately examine population ratios of the native conformation to various folding intermediate and non-native conformations.
ACKNOWLEDGMENTS
J.M. acknowledges support from National Institutes of Health (Contract No. R01-GM067801), National Science Foundation (Contract No. MCB-0818353), The Welch Foundation (Contract No. Q-1512), the Welch Chemistry and Biology Collaborative Grant from John S. Dunn Gulf Coast Consortium for Chemical Genomics, and the Faculty Initiatives Fund from Rice University. Computer time was provided by the Shared University Grid at Rice funded by NSF under Grant No. EIA-0216467, and a partnership between Rice University, Sun Microsystems, and Sigma Solutions, Inc. Use of VMD (Ref. 39) is gratefully acknowledged.
APPENDIX A: FOKKER–PLANCK EQUATION AND STATIONARY DISTRIBUTION
Though the temperature distribution of the generalized ensemble is uniquely specified by Eq. 8, various Langevin equations exist for sampling the distribution, e.g., both Eq. 1 and the one used in the previous study2 correctly populate the distribution.
Before applying a Langevin equation, one needs to find a proper “potential” V(β) for the temperature to guide its diffusion. The potential is the negative logarithm of the temperature distribution
The negative derivative of the potential naturally serves as the force of driving the temperature space random walk. According to Eq. 8, we have
Equation 1 is now simplified as
(A1) |
To show that Eq. A1 is correct, we introduce a new variable τ=1∕β=kBT and write down the Fokker–Planck equation that governs the distribution of τ or ρ(τ)
(A2) |
where we have used −τ2(∂∕∂τ)=∂∕∂β. A stationary solution of Eq. A2 can be found by solving the following equation:
whose solution is readily obtained,
Finally, we translate the distribution of τ back to that of β according to the probability invariance ρ(β)dβ=ρ(τ)dτ,
We thus conclude that ρ(β)=τ2ρ(τ)=p(β,X), which proves that Eq. 8 is the stationary solution of the Fokker–Planck equation.
APPENDIX B: ALTERNATIVE FITTING BASED ESTIMATOR
Here we introduce an alternative estimator for based on linear extrapolation. First, we generalize Eq. 15 to the following. For any function
(B1) |
where and k is an arbitrary constant, we have
(B2) |
where on the second line, the linear term in Eq. B1 vanishes after the integration. Equation 15 can be considered as a special case for k=0.
Equations B1, B2 allow a properly extrapolation for the average energy ⟨E⟩β from β to before applying the integral identity, or the temperature averaging. If we assume that ⟨E⟩β is roughly linear function of β in the window (β−,β+), the slope k of ⟨E⟩β versus β can be obtained from linear regression as
(B3) |
where the angular bracket ⟨⋯⟩+ denotes a temperature average over (β−,β+).
The advantage of using Eq. B2 is that it reduces the magnitude of the second integral and thus makes the estimator more robust. If the relation of ⟨E⟩β and β is perfectly linear, f′(β)=∂⟨E⟩β∕∂β−k vanishes everywhere, and thus the second integral yields zero.
In practice, we use a ϕ(β) parameterized with a+=a−=1∕2, and restrain the magnitude of the second integral within the average energy fluctuation in the window to ensure stability of the estimator.
The estimator introduced here can be thought as a first-order generalization of Eq. 15, and thus allows a large temperature window or application to larger systems. Besides, one can also apply additional constraints to the derivative to improve the accuracy and stability of the estimator. For example, in our case of estimating the average energy, the derivative k corresponds to the negative energy fluctuation as ∂⟨E⟩β∕∂β=−⟨ΔE2⟩β=−kBT2CV (where CV is the heat capacity), and its value can be restrained within a certain range. In the systems tested in this work, however, it did not produce significantly different estimates from the one introduced in the Sec. 2D.
APPENDIX C: CHOICE OF THE TEMPERATURE DISTRIBUTION ω(β)
Here, we discuss the optimal choice of w(β). For molecular systems, we invariably use w(β)≈p(β)∼1∕β basing on the following reason. The overall energy distribution p(E) of the generalized ensemble is a superposition of energy distributions of canonical ensembles from different temperatures. To ensure a fixed degree of overlap, the height of p(E) only needs to match that of a canonical distribution at β, whose average energy ⟨E(β)⟩ is roughly equal to E. The average height of a canonical ensemble is inversely proportional to its width , and thus,
To translate the energy distribution to the temperature space, we change variable from E to β. According to the probability invariance p(E)dE=p(β)dβ, and |dE∕dβ|=⟨ΔE2⟩β, we have
In the last step, we have used fact that kBβ2⟨ΔE2⟩β=C, where C is the heat capacity.
For a Lennard-Jones-like system, the heat capacity C is roughly a constant, accordingly the optimal w(β) is proportional to 1∕β. On the other hand, our protein simulations show that the heat capacity of the entire system, water and protein, roughly follows C∝β, see Fig. 8. According to this observation, the optimal w(β) should be proportional to . However, in our simulations, we still used 1∕β as w(β), this setup slightly biases the ensemble toward the high temperature end in order to encourage a faster motion at higher temperature and to help overcome broken ergodicity at lower temperature.
APPENDIX D: INTEGRAL IDENTITY FOR A NON-NEGATIVE QUANTITY
For a histogram-like quantity, regular integral identities such as Eq. 15 should be modified to ensure the output is non-negative. Here, we briefly sketch a technique for this purpose (a more detailed and general treatment is presented elsewhere6).
Suppose h(β) is a non-negative quantity, and we are interested in estimating its value at a particular temperature β*. Instead of applying an integral identity to h(β) itself, we introduce a smooth modulating function f(β), with f(β*)=1, and apply the identity to the product h(β)f(β),
(D1) |
If we choose
(D2) |
the second term on the right hand side of Eq. D1 vanishes because [h(β)f(β)]′=0 everywhere. Thus, the estimated value is always non-negative given that ϕ′(β)≥0. ϕ′(β) satisfies the normalization condition , and should be proportional to w(β)∕f(β) to minimize the error. Thus, the general expression for h(β*) is
(D3) |
In the case of energy histogram, h(β) can defined as an integral of the canonical energy distribution at β over a small energy bin ΔE,
where g(E′) is the density of states; on the second line, we have used the distribution function of the generalized ensemble pβ(E′)=w(β)g(E′)exp(−βE′)∕Z(β).
It can be straightforwardly verified that
where on the last line, we assumed that the energy bin size ΔE is sufficiently small such that a typical energy distribution does not vary drastically within a bin. Practically, it only requires the bin size ΔE to be much smaller than the fluctuation (or width) of a typical energy distribution. We finally reached
By substituting for Z(β), we recover Eq. 20.
References
- Lyubartsev A. P., Martsinovski A. A., Shevkunov S. V., and Vorontsovvelyaminov P. N., J. Chem. Phys. 96, 1776 (1992) 10.1063/1.462133 [DOI] [Google Scholar]; Marinari E. and Parisi G., Europhys. Lett. 19, 451 (1992) 10.1209/0295-5075/19/6/002 [DOI] [Google Scholar]; Zhang C. and Ma J., Phys. Rev. E 76, 036708 (2007) 10.1103/PhysRevE.76.036708 [DOI] [PMC free article] [PubMed] [Google Scholar]; Li H., Fajer M., and Yang W., J. Chem. Phys. 126, 024106 (2007) 10.1063/1.2424700 [DOI] [PubMed] [Google Scholar]; Gao Y. Q., J. Chem. Phys. 128, 064105 (2008); 10.1063/1.2825614 [DOI] [PubMed] [Google Scholar]; Kim J., Straub J. E., and Keyes T., Phys. Rev. Lett. 97, 050601 (2006). 10.1003/PhysRevLett.97.050601 [DOI] [PubMed] [Google Scholar]
- Zhang C. and Ma J., J. Chem. Phys. 130, 194112 (2009). 10.1063/1.3139192 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Geyer C. J., Proceedings of the 23rd Symposium on the Interface (American Statistical Association, New York, 1991); Hansmann U. H. E., Chem. Phys. Lett. 281, 140 (1997) 10.1016/S0009-2614(97)01198-6 [DOI] [Google Scholar]; Hukushima K. and Nemoto K., J. Phys. Soc. Jpn. 65, 1604 (1996) 10.1143/JPSJ.65.1604 [DOI] [Google Scholar]; Swendsen R. H. and Wang J. S., Phys. Rev. Lett. 57, 2607 (1986); 10.1103/PhysRevLett.57.2607 [DOI] [PubMed] [Google Scholar]; Kim J., Straub J. E., and Keyes T., J. Chem. Phys. 132, 224107 (2010). 10.1063/1.3432176 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Darve E. and Pohorille A., J. Chem. Phys. 115, 9169 (2001) 10.1063/1.1410978 [DOI] [Google Scholar]; Fasnacht M., Swendsen R. H., and Rosenberg J. M., Phys. Rev. E 69, 056704 (2004). 10.1103/PhysRevE.69.056704 [DOI] [PubMed] [Google Scholar]
- Adib A. B. and Jarzynski C., J. Chem. Phys. 122, 014114 (2005). 10.1063/1.1829631 [DOI] [PubMed] [Google Scholar]
- Zhang C. and Ma J., arXiv:1005.0170 (2010).
- Ferrenberg A. M. and Swendsen R. H., Phys. Rev. Lett. 61, 2635 (1988) 10.1103/PhysRevLett.61.2635 [DOI] [PubMed] [Google Scholar]; Ferrenberg A. M. and Swendsen R. H., Phys. Rev. Lett. 63, 1195 (1989). 10.1103/PhysRevLett.63.1195 [DOI] [PubMed] [Google Scholar]
- Ferdinand A. E. and Fisher M. E., Phys. Rev. 185, 832 (1969). 10.1103/PhysRev.185.832 [DOI] [Google Scholar]
- Beale P. D., Phys. Rev. Lett. 76, 78 (1996). 10.1103/PhysRevLett.76.78 [DOI] [PubMed] [Google Scholar]
- van der Spoel D., Lindahl E., Hess B., Groenhof G., Mark A. E., and Berendsen H. J., J. Comput. Chem. 26, 1701 (2005) 10.1002/jcc.20291 [DOI] [PubMed] [Google Scholar]; Hess B., Kutzner C., van der Spoel D., and Lindahl E., J. Chem. Theory Comput. 4, 435 (2008) 10.1021/ct700301q [DOI] [PubMed] [Google Scholar]; Lindahl E., Hess B., and van der Spoel D., J. Mol. Model. 7, 306 (2001) [Google Scholar]; Berendsen H. J. C. and van der Spoel D., Comput. Phys. Commun. 91, 43 (1995). 10.1016/0010-4655(95)00042-E [DOI] [Google Scholar]
- Sorin E. J. and Pande V. S., Biophys. J. 88, 2472 (2005). 10.1529/biophysj.104.051938 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Jorgensen J. C. W. L., Madura J. D., Impey R. W., and Klein M. L., J. Chem. Phys. 79, 926 (1983). 10.1063/1.445869 [DOI] [Google Scholar]
- Essman U., Perera L., Berkowitz M. L., Darden T., Lee H., and Pedersen L. G., J. Chem. Phys. 103, 8577 (1995). 10.1063/1.470117 [DOI] [Google Scholar]
- Bussi G., Donadio D., and Parrinello M., J. Chem. Phys. 126, 014101 (2007). 10.1063/1.2408420 [DOI] [PubMed] [Google Scholar]
- Miyamoto S. and Kollman P. A., J. Comput. Chem. 13, 952 (1992). 10.1002/jcc.540130805 [DOI] [Google Scholar]
- Hess B., J. Chem. Theory Comput. 4, 116 (2008). 10.1021/ct700200b [DOI] [PubMed] [Google Scholar]
- Cochran A. G., Skelton N. J., and Starovasnik M. A., Proc. Natl. Acad. Sci. U.S.A. 98, 5578 (2001). 10.1073/pnas.091100898 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Zhang J., Qin M., and Wang W., Proteins 62, 672 (2006) 10.1002/prot.20813 [DOI] [PubMed] [Google Scholar]; Pitera J. W., Haque I., and Swope W. C., J. Chem. Phys. 124, 141102 (2006). 10.1063/1.2190226 [DOI] [PubMed] [Google Scholar]
- Yang L., Shao Q., and Gao Y. Q., J. Phys. Chem. B 113, 803 (2009) 10.1021/jp803160f [DOI] [PubMed] [Google Scholar]; Snow C. D., Qiu L., Du D., Gai F., Hagen S. J., and Pande V. S., Proc. Natl. Acad. Sci. U.S.A. 101, 4077 (2004) 10.1073/pnas.0305260101 [DOI] [PMC free article] [PubMed] [Google Scholar]; Roitberg A. E., Okur A., and Simmerling C., J. Phys. Chem. B 111, 2415 (2007) 10.1021/jp068335b [DOI] [PMC free article] [PubMed] [Google Scholar]; Chen C. and Xiao Y., Bioinformatics 24, 659 (2008) 10.1093/bioinformatics/btn029 [DOI] [PubMed] [Google Scholar]; Yang W. Y., Pitera J. W., Swope W. C., and Gruebele M., J. Mol. Biol. 336, 241 (2004). 10.1016/j.jmb.2003.11.033 [DOI] [PubMed] [Google Scholar]
- Ulmschneider J. P., Ulmschneider M. B., and Di Nola A., J. Phys. Chem. B 110, 16733 (2006). 10.1021/jp061619b [DOI] [PubMed] [Google Scholar]
- Wang J., Cieplak P., and Kollman P. A., J. Comput. Chem. 21, 1049 (2000). [DOI] [Google Scholar]
- Hornak V., Abel R., Okur A., Strockbine B., Roitberg A., and Simmerling C., Proteins 65, 712 (2006). 10.1002/prot.21123 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Neidigh J. W., Fesinmeyer R. M., and Andersen N. H., Nat. Struct. Biol. 9, 425 (2002). 10.1038/nsb798 [DOI] [PubMed] [Google Scholar]
- Qiu L., Pabit S. A., Roitberg A. E., and Hagen S. J., J. Am. Chem. Soc. 124, 12952 (2002). 10.1021/ja0279141 [DOI] [PubMed] [Google Scholar]
- Zhou R., Proc. Natl. Acad. Sci. U.S.A. 100, 13280 (2003) 10.1073/pnas.2233312100 [DOI] [PMC free article] [PubMed] [Google Scholar]; Paschek D., Hempel S., and Garcia A. E., Proc. Natl. Acad. Sci. U.S.A. 105, 17754 (2008) 10.1073/pnas.0804775105 [DOI] [PMC free article] [PubMed] [Google Scholar]; Juraszek J. and Bolhuis P. G., Proc. Natl. Acad. Sci. U.S.A. 103, 15859 (2006); 10.1073/pnas.0606692103 [DOI] [PMC free article] [PubMed] [Google Scholar]; Kannan S. and Zacharias M., Proteins 76, 448 (2008). 10.1002/prot.22359 [DOI] [PubMed] [Google Scholar]
- Day R., Paschek D., and Garcia A. E., Proteins 78, 11 (2010). [DOI] [PMC free article] [PubMed] [Google Scholar]
- Yang L., Grubb M. P., and Gao Y. Q., J. Chem. Phys. 126, 125102 (2007) 10.1063/1.2709639 [DOI] [PubMed] [Google Scholar]; Simmerling C., Strockbine B., and Roitberg A. E., J. Am. Chem. Soc. 124, 11258 (2002) 10.1021/ja0273851 [DOI] [PubMed] [Google Scholar]; Chowdhury S., Lee M. C., Xiong G., and Duan Y., J. Mol. Biol. 327, 711 (2003) 10.1016/S0022-2836(03)00177-3 [DOI] [PubMed] [Google Scholar]; Snow C. D., Zagrovic B., and Pande V. S., J. Am. Chem. Soc. 124, 14548 (2002) 10.1021/ja028604l [DOI] [PubMed] [Google Scholar]; Pitera J. W. and Swope W., Proc. Natl. Acad. Sci. U.S.A. 100, 7587 (2003). 10.1073/pnas.1330954100 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Streicher W. W. and Makhatadze G. I., Biochemistry 46, 2876 (2007). 10.1021/bi602424x [DOI] [PubMed] [Google Scholar]
- Duan Y. and Kollman P. A., Science 282, 740 (1998). 10.1126/science.282.5389.740 [DOI] [PubMed] [Google Scholar]
- Chiu T. K., Kubelka J., Herbst-Irmer R., Eaton W. A., Hofrichter J., and Davies D. R., Proc. Natl. Acad. Sci. U.S.A. 102, 7517 (2005). 10.1073/pnas.0502495102 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Jayachandran G., Vishal V., and Pande V. S., J. Chem. Phys. 124, 164902 (2006). 10.1063/1.2186317 [DOI] [PubMed] [Google Scholar]
- Lei H., Wu C., Liu H., and Duan Y., Proc. Natl. Acad. Sci. U.S.A. 104, 4925 (2007). 10.1073/pnas.0608432104 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Ensign D. L., Kasson P. M., and Pande V. S., J. Mol. Biol. 374, 806 (2007) 10.1016/j.jmb.2007.09.069 [DOI] [PMC free article] [PubMed] [Google Scholar]; Snow C. D., Nguyen H., Pande V. S., and Gruebele M., Nature (London) 420, 102 (2002) 10.1038/nature01160 [DOI] [PubMed] [Google Scholar]; Shen M. Y. and Freed K. F., Proteins 49, 439 (2002) 10.1002/prot.10230 [DOI] [PubMed] [Google Scholar]; Fernández A., Shen M. Y., Colubri A., Sosnick T. R., Berry R. S., and Freed K. F., Biochemistry 42, 664 (2003) 10.1021/bi026510i [DOI] [PubMed] [Google Scholar]; Zagrovic B., Snow C. D., Shirts M. R., and Pande V. S., J. Mol. Biol. 323, 927 (2002). 10.1016/S0022-2836(02)00997-X [DOI] [PubMed] [Google Scholar]
- Freddolino P. L. and Schulten K., Biophys. J. 97, 2338 (2009). 10.1016/j.bpj.2009.08.012 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kubelka J., Eaton W. A., and Hofrichter J., J. Mol. Biol. 329, 625 (2003). 10.1016/S0022-2836(03)00519-9 [DOI] [PubMed] [Google Scholar]
- Wang M., Tang Y., Sato S., Vugmeyster L., McKnight C. J., and Raleigh D. P., J. Am. Chem. Soc. 125, 6032 (2003). 10.1021/ja028752b [DOI] [PubMed] [Google Scholar]
- Duan Y., Wu C., Chowdhury S., Lee M. C., Xiong G., Zhang W., Yang R., Cieplak P., Luo R., Lee T., Caldwell J., Wang J., and Kollman P., J. Comput. Chem. 24, 1999 (2003). 10.1002/jcc.10349 [DOI] [PubMed] [Google Scholar]
- Best R. B., Buchete N. V., and Hummer G., Biophys J 95, L07 (2008) 10.1529/biophysj.108.132696 [DOI] [PMC free article] [PubMed] [Google Scholar]; Best R. B. and Hummer G., J. Phys. Chem. B 113, 9004 (2009) 10.1021/jp901540t [DOI] [PMC free article] [PubMed] [Google Scholar]; Thompson E. J., DePaul A. J., Patel S. S., and Sorin E. J., PLoS ONE 5, e10056 (2010). 10.1371/journal.pone.0010056 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Humphrey W., Dalke A., and Schulten K., J. Mol. Graph. 14, 33 (1996). 10.1016/0263-7855(96)00018-5 [DOI] [PubMed] [Google Scholar]