Skip to main content
ACS AuthorChoice logoLink to ACS AuthorChoice
. 2025 Jul 18;21(15):7236–7248. doi: 10.1021/acs.jctc.5c00479

Kinetically Consistent Coarse Graining Using Kernel-Based Extended Dynamic Mode Decomposition

Vahid Nateghi 1, Feliks Nüske 1,*
PMCID: PMC12355707  PMID: 40679598

Abstract

In this paper, we show how kernel-based models for the Koopman generatorthe gEDMD methodcan be used to identify coarse-grained dynamics on reduced variables, which retain the slowest transition time scales of the original dynamics. The centerpiece of this study is a learning method to identify an effective diffusion in coarse-grained space, which is similar in spirit to the force matching method. By leveraging the gEDMD model for the Koopman generator, the kinetic accuracy of the CG model can be evaluated. By combining this method with a suitable learning method for the effective free energy, such as force matching, a complete model for the effective dynamics can be inferred. Using a two-dimensional model system and molecular dynamics simulation data of alanine dipeptide and the Chignolin mini-protein, we demonstrate that the proposed method successfully and robustly recovers the essential kinetic and also thermodynamic properties of the full model. The parameters of the method can be determined using standard model validation techniques.


graphic file with name ct5c00479_0030.jpg


graphic file with name ct5c00479_0028.jpg

1. Introduction

Stochastic simulations of large-scale dynamical systems are widely used to model the behavior of complex systems, with applications in computational physics, chemistry, materials science, and engineering. Many examples of such systems are high dimensional and subject to meta-stability, which means the system remains trapped in a set of geometrically similar configurations, while transitions to another such state are extremely rare. As a consequence, it becomes necessary to produce very long simulations in order to make statistically robust predictions. A prime example are atomistic molecular dynamics simulations (MD) of macro-molecules, where meta-stability is typically caused by high energetic barriers separating deep potential energy minima. As a result, it requires specialized high-performance computing facilities to reach the required simulation times, or it may just not be feasible at all.

Coarse graining (CG) describes the process of replacing the original dynamical system by a surrogate model on a (much) lower-dimensional space of descriptors, , in such a way that certain properties of the original dynamics are preserved. CG models can enable scientists to achieve much longer simulation times because of the reduced computational cost, while maintaining predictive capabilities of the full-order model. Setting up a CG model typically requires the following steps: first, the choice of a linear or nonlinear mapping (CG map) from full state space to a lower-dimensional space, where the latter serves as the state space of the surrogate model. Second, definition of a parametric model class for the surrogate dynamics. Finally, fitting the parameters of the selected model class using available data.

The first step is crucial to the CG model’s success, and has been a very active area of research for a long time, see refs for reviews on this topic. Traditionally, coarse grained coordinates have been based on molecular structure, e.g., by considering only alpha-carbons or reduced atom representations. More recently, CG projections into less interpretable and nonlinear spaces have also been considered, such as the latent space of a neural network transformation. , The selection and quality of the CG coordinate is not the central aspect of this study, we focus on model selection and parameter fitting instead. Therefore, we only show examples of low-dimensional CG coordinates that have already been validated, and that are not directly transferable. The problem of learning high-dimensional and fully transferrable CG models along with their collective variables is left for future studies.

CG models have often been parametrized using physically intuitive functional forms for the coarse-grained energy. More recently, much more general functional forms have been used for the CG parameters, which are then approximated by powerful model classes, such as deep neural networks or reproducing kernels, which is the approach we follow in this paper. We study CG for reversible stochastic differential equations (SDE) with a Boltzmann-type invariant distribution, such as Langevin dynamics. Theoretical frameworks to CG modeling are typically based on projections of dynamical evolution operators. This includes the Mori-Zwanzig formalism, , as well as the approaches by Gyöngy, Legoll and Lelièvre, and the averaging/homogenization framework. We follow Legoll and Lelièvre’s projection method, which means to parametrize the coarse-grained model as a reversible SDE, disregarding memory terms. The theoretical properties of this approach have been studied to quite some extent in the literature.

The success of machine learning (ML) in recent years has led to the development of many powerful learning schemes for the parameters of a CG model, see ref for a comprehensive overview. Examples are free energy learning, and force matching, among others. Many of these learning methods are geared toward ensuring thermodynamic consistency, which means that the marginalized Boltzmann distribution in CG space is preserved. Ensuring faithful reproduction of kinetic properties, such as time-correlation functions or transition time scales, is a much less developed topic. Besides the theoretical contributions noted above, several authors have focused on preserving specific dynamical observables or time-dependent distributions by incorporating these quantities into the learning process. Furthermore, several recent studies have considered integrated learning frameworks for CG coordinates and associated dynamics geared toward preserving transition rates, using autoencoders, normalizing flows or diffusion maps.

In this paper, we combine learning of a coarse-grained SDE with Koopman operator models in order to recover implied transition time scales associated with meta stable states. Transition time scales are derived from the leading spectrum of the Koopman generator. This connection has been at the heart of the Markov state modeling (MSM) approach ,, and many important developments based on it. The spectral matching approach, later formalized in ref , was the first to make use of this connection, by parametrizing the CG model as a linear expansion of fixed basis functions, and then solving a regression problem to recover the eigenvalues of the Koopman generator. The generator matrix can be estimated by a data-driven algorithm called generator EDMD (gEDMD).

We significantly improve on the idea of leveraging the Koopman generator for the identification of coarse grained models in the following ways:

  • Based on the projection approach, we formulate a stand-alone learning problem for the effective diffusion of a coarse-grained SDE. This formulation is analogous to the force matching approach for the coarse-grained energy. Just as force matching relies on measurements of the local mean force, our approach rests on a similar quantity called local diffusion. Combined with a suitable estimate for the free energy, the learned effective diffusion provides a closed-form expression for the CG dynamics.

  • We suggest to parametrize the diffusion by a basis of random Fourier features, which form a widely used approximation technique for reproducing kernels. Random features offer a compromise between representational power and computational efficiency. The only hyper-parameters to be tuned are those of the kernel function. Conveniently, we show that the same random feature basis can be used to train a kinetic model for the Koopman generator. The method is robust to statistical noise and ill-conditioning as it is based on a whitened and truncated basis set.

  • We show that gEDMD models can be leveraged to evaluate the kinetic consistency of the learned CG model on-the-fly by comparing its eigenvalues to those of the reference gEDMD matrix. Importantly, this assessment does not require simulations of the CG model.

  • We show that kinetic and also thermodynamic consistency are achieved by the method using three test cases, a two-dimensional model system and molecular dynamics simulations of the alanine dipeptide and the Chignolin mini-protein. For the molecular systems, we learn a CG model corresponding to overdamped Langevin dynamics. The results show that for systems close enough to the overdamped limit, this approximation leads to a uniform rescaling of the slow time scales, which can be explicitly corrected for.

The structure of the paper is as follows: we introduce the required background on SDEs, coarse graining, and Koopman operator learning in Section . Our learning framework is then presented in Section , while the numerical examples follow in Section . Additional information on simulation details and model selection is given in the Supporting Information.

2. Theory

In this section, we provide the necessary background on stochastic dynamics, data-driven modeling, and Koopman spectral theory. The important notation used in the manuscript is summarized in Table .

1. Overview of Notation.

Symbol Definition
X t stochastic process
Kt Koopman operator with lag time t
L generator of the Koopman operator
h reduced basis set from whitening transformation
, r generator matrix and reduced generator matrix
σα effective diffusion parametrized by α
α effective generator matrix for diffusion with parameters α
V, F potential and effective potential
fξlmf, aξloc local mean force and local diffusion
A ·|i,j B contraction of dimensions i and j of arrays A and B

2.1. Stochastic Processes

We consider a dynamical system described by a stochastic differential equation (SDE)

dXt=b(Xt)dt+σ(Xt)dWt 1

where b(Xt):RdRd is the drift vector field, σ(Xt):RdRd×d is the diffusion field, and W t is a d-dimensional Brownian motion. The diffusion covariance matrix is denoted as aRd×d :

2.1. 2

A standard example for eq , commonly used in molecular modeling, is overdamped Langevin dynamics

dXt=1γV(Xt)dt+2β1γ1dWt 3

where V:ΩR is the potential energy, β = (k B T)−1 and γ are constants corresponding to the inverse temperature and the friction, respectively. The invariant measure for X t in eq is the Boltzmann distribution μ ∝ exp­(−βV), and the dynamics are reversible with respect to μ. More generally, a reversible SDE with invariant measure μ ∝ exp­(−V) can be parametrized in terms of the generalized scalar potential Inline graphic, and the diffusion covariance a, as follows

dXt=[12a(Xt)V(Xt)+12·a(Xt)]dt+σ(Xt)dWt 4

We will only consider reversible SDEs in this paper, and make use of the parametrization in eq when formulating learning methods.

2.2. Koopman Generator and Spectral Decomposition

Koopman theory , lifts the dynamics in eq into an infinite-dimensional space of observable functions to express the dynamics linearly. More precisely, the family of Koopman operators Kt for stochastic dynamics is defined as

Ktψ(x)=Ex[ψ(Xt)]=E[ψ(Xt)|X0=x] 5

where ψ is a real-valued observable of the system, and E[·] denotes the expected value. The associated infinitesimal generator L is the time-derivative of the expectation value, which can be written as a linear differential operator:

Lψ(x)=b(x)·ψ(x)+12a(x):2ψ(x)=i=1dbi(x)xiψ(x)+12i,j=1daij(x)2xixjψ(x) 6

where a and b are the diffusion and drift terms defined above, ∇2[·] is the Hessian matrix of a function, and the colon: is a short-hand for the dot product between two matrices. For overdamped Langevin dynamics, eq simplifies to

Lψ(x)=1γV(x)·ψ(x)+1γβΔψ(x)

The key quantity of interest are the eigenvalues and eigenfunctions of the generator. The study of spectral components of the generator helps us identify the long-time dynamics of the system. In molecular dynamics, we expect to find a number of eigenvalues close to zero, followed by a spectral gap. These low-lying eigenvalues are indicating the number of metastable states of the system, which are the macro states the system stays in the longest. We write the eigenvalue problem for the generator as

Lψi=λiψi 7

The eigenvalues λ i of L must be non-negative, and the lowest eigenvalue λ1 = 0 is nondegenerate: 0 = λ1 < λ2 ≤ λ3 ≤··· We also refer to the eigenvalues as rates, and to their reciprocals as implied time scales

ti=1λi 8

2.3. Coarse Graining and Projection

One of the main motivations of this work is to learn an SDE representing the full dynamics () on a coarse grained space. Coarse graining (CG) is realized by mapping the state space Ω onto a lower-dimensional space Ω̂Rd by means of a smooth CG function ξ. We write ν ∝ exp­(−F) for the marginal distribution of the full-space invariant measure μ, where F is the free energy in the CG space.

To define dynamics in the CG space, we use the conditional expectation operator, , also called Zwanzig projector:

Pψ(z)=Eμ[ψ(x)|ξ(x)=z] 9

where z is a position in CG space. This operator calculates the average of a function ψ over all x ∈Ω whose projection onto CG space is the same point z ∈ Ω̂. Following the exposition in ref , one can define the projected generator

Lξ=PLP 10

which corresponds to the Markovian part in the Mori–Zwanzig decomposition. It turns out its action on a function ϕ = ϕ­(z) in CG space is given by

Lξ(ϕ)=P[Lξ]·zϕ+12P[ξTaξ]:z2ϕ 11

As one can see, Lξ is of the same form as the original generator L in eq , and indeed it is the generator of an SDE Z t on Ω̂

dZt=bξ(Zt)dt+σξ(Zt)dWt 12

The effective drift and diffusion coefficients are given in analytical form by

bξ(z)=P(Lξ)(z)aξ(z)=P(ξTaξ)(z) 13

and the practical task of coarse graining is to approximate them numerically.

2.4. Generator EDMD

Numerical approximations to the infinitesimal generator L can be obtained by a data-driven learning method called generator extended dynamic mode decomposition (gEDMD). Given a finite set of scalar basis functions ψ­(x) = {ψ1(x),..., ψ n (x)}, and training data {x l } l = 1 sampled from the invariant measure μ, we form the matrices

Ψ=[ψi(xl)]i,l,LΨ=[Lψi(xl)]i,l

using the analytical formula () to evaluate the second of these matrices. The solution of a linear regression problem leads to the matrix approximation

L^=G^1A^ 14

where

A^ij=1mk=1mψi(xk)Lψj(xk),G^ij=1mk=1mψi(xk)ψj(xk) 15

These matrices are empirical estimators of the following mass, stiffness, and generator matrices

Aij=ψi,Lψjμ,Gij=ψi,ψjμ,L=G1A 16

The empirical mass matrix Ĝ is often ill-conditioned. A standard approach to circumvent this is to perform a whitening transformation based on removing small eigenvalues

2.4. 17

in which rn. Here, R is a transformation matrix mapping the original basis to the reduced basis

2.4. 18

Dominant eigenvalues of the generator can be computed by diagonalizing the matrix or r .

For arbitrary stochastic dynamics, the computation of A involves a second-order differentiation as shown in eq . However, if the stochastic dynamics are reversible, only first-order derivatives are required to compute the matrix A, as the generator satisfies the following integration-by-parts formula

2.4. 19

Importantly, if the basis functions are actually defined in a CG space Ω̂, that is ψ i (x) = ψ i (ξ­(x)), then by the chain rule the matrix A can be written as

2.4. 20

We refer to the matrix

2.4. 21

as local diffusion, and note that it is independent of the basis functions. It can therefore be computed a priori in numerical calculations.

2.5. Random Fourier Features

The gEDMD algorithm requires choosing a set of basis functions ψ­(x). In this work, we use random Fourier features (RFFs), which are defined as

2.5. 22

The vectors ω1,..., ω n are random frequency vectors drawn from a spectral distribution ρ. RFFs provide a low-rank approximation to a reproducing kernel function, and can therefore generate a powerful basis without the need for manual basis set design. The precise relation between kernel-based gEDMD and random features was presented in ref . In the following applications, we use the spectral measure associated to a Gaussian squared exponential kernel with bandwidth parameter γ

k(xi,xj)=exp(xixj)22γ2) 23

or to a periodic Gaussian kernel on periodic domains, such as dihedral coordinates.

3. Methods

We now turn to the suggested framework for learning CG dynamics based on the projection formalism and gEDMD models. We recall that the dynamical equation in CG space is given by (), where the drift can be written as follows because of reversibility

bξ=12aξzF+12·aξ 24

3.1. Diffusion Learning

By eq , the analytical effective diffusion a ξ is the best-approximation of the local diffusion a loc by a (matrix-valued) function on the CG space. Hence, we can solve the following data-based minimization problem

aξ=argmina=a(z)1mi=1ma(ξ(xi))alocξ(xi)F2 25

where ∥·∥ F is the Frobenius norm for matrices. We parametrize the diffusion field a ξ element-wise as a linear combination of the reduced RFF basis

3.1. 26

where we view the coefficient array α as a third-order tensor of dimension d × d × r, and the symbol ·|i,j denotes contraction over indices i and j of two arrays. The parametrization must be symmetric, i.e., a ij = a ji , and we may also choose to set specific elements to zero, for example to enforce a diagonal diffusion field. With the parametrization (), the minimization problem () becomes a regression problem that can be directly solved, potentially after regularization. The complexity of the algorithm is governed by the cost of building r and learning the diffusion coefficients α. For a diagonal diffusion field, these costs can be estimated as O(mdp2) and O(mr2+dmr) , respectively, where m is again the number of samples, d the dimension of the CG space, p is the number of random Fourier features, and r is the rank of r . The critical parameters are therefore the number of random features p and the effective basis set size r.

3.2. Recovery of Spectral Properties

After solving the minimization problem (), we can make use of the gEDMD method to assess the dynamical properties of the learned SDE in CG space. Using the integration-by-parts formula (), the elements of the reduced generator matrix corresponding to the diffusion field () with coefficient array α are3.2.

In matrix notation, this leads to the following explicit formula for the parametrized generator matrix, which can be computed directly without resorting to numerical simulations of the CG dynamics

3.2. 27

Properties inferred from the matrix r can be compared to those obtained from the original gEDMD matrix r estimated off the full-space simulation data. For example, diagonalization of both r and r leads to estimates λ i and λ i for the dominant generator eigenvalues, which can be systematically compared. We mainly resort to comparing dominant eigenvalues in the examples below, but we point out that a more detailed assessment is possible: for instance, by computing matrix exponentials exp­(t r ) and exp­(t r ), time-correlation functions can also be evaluated.

3.3. Learning the Effective Potential

We have seen that the accuracy of the effective diffusion field largely determines the dynamical properties of the coarse grained dynamics. In order to run simulations of the CG dynamics, and to ensure thermodynamic consistency, the effective potential F also must be learned in a parametric form. This is not the main focus of our study, hence we just point out a few options. A well-known and generally applicable technique is force matching, which is based on the following minimization problem for the effective force

zF=argming=g(z)1mi=1mg(ξ(xi))flmfξ(xi)2 28

where f lmf is called local mean force and defined as follows

3.3. 29

We point out the similarity to (), which also led us to the name local diffusion for a loc . The effective potential can be parametrized as a linear combination of basis functions, such as random features, or as a deep neural network. In low-dimensional CG spaces, it also possible to approximate the projected invariant distribution ν as a linear combination of kernel functions centered at the data sites, known as kernel density estimate (KDE). Since we only consider low-dimensional CG spaces here, we opt for the KDE option in the examples below.

3.

3.4. Overdamped Models for Molecular Systems

In practical MD simulations, computation of the local diffusion () requires knowledge of the full-state diffusion tensor, which depends on the thermostat used to drive the molecular dynamics. If the full-state dynamics are just overdamped Langevin dynamics (), the local diffusion reduces to the following simple form:

3.4. 30

where M is the diagonal mass matrix of all atoms.

Very often, however, one can apply an overdamped approximation. If the full-state dynamics are underdamped Langevin, then averaging theory shows that for small friction γ, and under a rescaling of time, the position space dynamics are close to the overdamped process (). In practice, we observe empirically that even if the friction is not asymptotically small, and even for thermostats different from underdamped Langevin, one can find a rescaling of time such that the position process is similar to an overdamped process.

We therefore apply the overdamped approximation of the local diffusion () in the molecular examples in Sections and . This simplification is also convenient as a mature theory of the projection formalism for underdamped dynamics is still under construction, see ref for some preliminary results.

As the overdamped approximation is expected to hold after a rescaling of time, the time scales of the resulting CG model will be faster than those of the original dynamics. To account for this rescaling, we can make use of the existing simulation data to also compute a standard kinetic model for the Koopman operator, for example a Markov state model T t at a suitable lag time t > 0. It is sufficient to construct the MSM in CG space, hence the definition of appropriate MSM states is not challenging. By comparing the MSM time scales to those of the learned generator r , the rescaling of time can be practically computed.

4. Examples

To show the effectiveness of the proposed method, we apply it to a two-dimensional model system defined by the Lemon-slice potential, and to MD simulation data of the alanine dipeptide and of the mini-protein Chignolin, which are widely used test cases in molecular dynamics.

4.1. Lemon-Slice Potential

4.1.1. System Introduction

The Lemon-slice system is governed by overdamped Langevin dynamics in eq with the following potential V

V(x,y)=V(r,ϕ)=cos(4ϕ)+10(r1)2 31

where r and ϕ are polar coordinates. The energy landscape of the system is shown in Figure a. To form the SDE for this example, we consider a diagonal state-dependent diffusion field σ­(x) defined as

σ(x)=[2β(sin(ϕ)+1.5)002β(sin(ϕ)+1.5)] 32

where β = 1 is the inverse temperature. Using the Euler–Maruyama scheme at discrete integration time step dt = 10–3 for integration of the SDE, we collect the training data for learning. For the sake of validation and showing the robustness of the method, we produce 5 independent experiments, each with length of m = 105 time steps. We further down sample them to 1000 samples each for learning effective force and diffusion.

1.

1

Approximation of generator for the Lemon slice system. Potential field in (a). Membership analysis in (b) using 1000 samples. The dominant eigenvalues of the reference generator r and the learned generator α built upon the learned effective diffusion, using Gaussian and periodic Gaussian kernels, in (c). The relative error of these eigenvalues compared to the reference is shown in (d).

As shown in previous studies, the polar angle ϕ is a suitable CG coordinate for this system, as it resolves all four metastable states

ξ(x,y)=ϕ 33

For this system, analytical expressions for the effective drift and diffusion along ξ can be obtained by a slight modification of the results in ref , and serve as reference values.

We apply our learning method with random Fourier features on the reaction coordinate ξ, to identify the generator eigenvalues and metastable states and, subsequently, to identify an effective dynamics along ξ using Algorithm 1. As the polar angle is a periodic reaction coordinate (RC), we use the spectral measures associated to both a periodic and nonperiod Gaussian kernel and compare them. The number of random features and the kernel bandwidth in either versions of Gaussian kernel are optimized using cross validation based on the VAMP-score. Details on the VAMP-score analysis are reported in the Supporting Information.

4.1.2. Meta-Stability Analysis

Figure c shows the leading eigenvalues obtained from the generator matrix r . As one notices, there are four dominant eigenvalues followed by a gap. These four eigenvalues are corresponding to the four minima in the potential field. Having determined the eigenvectors of the generator, we can perform robust Perron Cluster Cluster Analysis (PCCA+) algorithm to assign to each sample point its membership to each metastable state. Figure b shows that the four potential minima are perfectly recovered in this way. A comparison of the leading eigenvalues of the reference model r and the learned matrix α for the optimal parameters α is shown in Figure c. Both choices of the kernel function lead to satisfactory results, the periodic kernel provides slightly higher accuracy in approximation of the generator eigenvalues. Note that the kernel bandwidth is tuned for each kernel function separately.

4.1.3. Analysis of the CG Dynamics

The learned generator providing the eigenvalues reported above is built upon the effective diffusion shown in Figure b, which is almost perfectly following the reference. Furthermore, we perform the force matching as well and obtain the effective force in the CG space shown in Figure a. From the effective force and diffusion, the effective drift can be obtained according to eq , which is also compared against the analytical expression in Figure c, likewise showing very good agreement.

2.

2

Application of Algorithm 1 to identify angular dynamics for the Lemon-slice system. Effective force in (a), effective diffusion in (b), effective drift in (c), and integration of an example trajectory, using both the reference and learned SDE in (d).

With the effective drift and diffusion fields, we are able to simulate the learned SDE governing the CG coordinate. We use the Euler-Maruyama scheme to integrate the learned and reference SDEs with integration time step of dt = 10–3. Figure d shows two trajectories of the CG coordinate ϕ for both dynamics for 104 time steps, using the same Brownian motion for both trajectories. The propagated learned system follows the reference closely, with both systems staying long times in each metastable state, and rarely swapping in between those. Combined, the results above demonstrate that the proposed method can approximate the full system’s metastable sets well, and identify a suitable SDE for CG dynamics which is accurate even on the level of individual trajectories.

As a final analysis, we compare the properties of the learned CG model with variable diffusion to those of a CG dynamics with constant diffusion, in order to demonstrate the necessity of allowing a state-dependent diffusion. We set the effective diffusion for the constant model to a=2β=2 . We propagate the corresponding SDEs for a sufficiently large span of time, and estimate a new generator EDMD model based on these simulations. Figure , shows the eigenvalues of the generator for these cases compared to the learned generator built upon the original data set. The result shows that learning a state-dependent diffusion is necessary to recover the original system’s leading eigenvalues.

3.

3

Dominant eigenvalues of the generator, using models built on simulation data of the learned coarse grained dynamics with state-dependent diffusion (SDD, orange) and with constant diffusion (CD, green). As a comparison, we show the eigenvalues of the generator r using the original data set (blue). Note that the first eigenvalue is omitted as it is zero.

4.2. Alanine Dipeptide

4.2.1. System Introduction

Alanine dipeptide is a model system widely used in method development for simulation studies of macro-molecules. Figure shows the graphical representation of Alanine dipeptide. It is well-known that the dynamical behavior of the molecule can be expressed in terms of the backbone dihedral angles ϕ and ψ, which constitute the two-dimensional reaction coordinate space defining the CG map ξ:

ξ(x)=[ϕ(x)ψ(x)] 34

We generated a 500 ns simulation of the system in explicit water, the details of the simulation settings are summarized in the Supporting Information.

4.

4

Graphical representation of the alanine dipeptide molecule on the left, and the reference free energy profile in two-dimensional dihedral angle space on the right.

The familiar free energy landscape of the system with respect to these two angles is shown in Figure , displaying four minima, two on the left side, usually denoted (P II, αR), and two in the central part, called (αD, αL).

We apply the gEDMD algorithm with random Fourier features to find the metastable sets, and then use Algorithm 1 to learn the effective force and a state-dependent effective diffusion field in the dihedral angle space. Because of the periodicity of the CG coordinates, ϕ and ψ, the spectral measure corresponds to a periodic Gaussian kernel. Similar to the previous example, we tune the bandwidth of the kernel function as well as the size of random features using the VAMP-score.

4.2.2. Meta-Stability Analysis

Figure a shows the leading finite time scales by taking reciprocals of the first three nonzero eigenvalues of the generator obtained from the gEDMD matrix r (error bars in the figure are generated by analyzing 5 independent subsampled sets of the original data set, each comprising 50,000 samples). The figure indicates the three dominant time scales which are corresponding to the four minima in the free energy landscape followed by a gap. In addition, we also show the time scales corresponding to the generator L α based on the optimal effective diffusion, which agree well with the reference. Note that the generator time scales shown have been rescaled after comparison to a Markov state model T t trained on the original simulation data, as described in Section . This comparison showed that the time scales of the generator models r and α were smaller than those of the MSM model by a uniform factor of about 100, meaning that the dynamics in CG space based on the overdamped assumption is accelerated by a factor 100 for this example. After applying the uniform rescaling, the generator time scales match those of the MSM analysis very well.

5.

5

Approximation of generator for alanine dipeptide. The dominant time scales corresponding to the reference generator r and the learned generator α built upon the learned effective diffusion on the left, and the relative error of these time scales on the right. The time scales of the MSM model are shown as black dashed lines for comparison. Note that time scales of the generators are rescaled by a factor of 100 to account for the overdamped approximation. The first time scale (l = 2) corresponds to the transition between the left-hand side and the central part, the second one (l = 3) corresponds to the transition between P II and α R , and the third one (l = 4) corresponds to the transition between αD and αL.

4.2.3. Analysis of the CG Dynamics

For this 2-dimensional coarse graining, we can express the diffusion field as a 2 × 2 full matrix. For simplicity, however, we assume that the learned diffusion is a diagonal matrix. Figure shows the first and second diagonal terms of the learned diffusion field based on 50000 samples of the available data set. To learn the effective potential, we found that the KDE method works best. The reference and learned effective free energy surfaces are depicted in Figure c,d, respectively. It it noticeable that the learned free energy surface correctly captures all energetic minima and barriers up to some minor spurious behavior close to the transition regions. We emphasize once again that this approximation could probably be improved further by using a more accurate learning method.

6.

6

First (a) and second (b) diagonal terms of the learned diffusion covariance matrix, the reference free energy surface (c) and the free energy surface learned via KDE (d).

From the effective force and diffusion, one can compute the effective drift from which the SDE governing the dynamics in the CG space can be formed. We integrate the learned SDE for 5 × 105 integration steps, with an effective (rescaled) time step of 0.1 ps, corresponding to an effective total simulation time of 50 ns. Figure b shows the estimated free energy surface obtained from a histogram of the propagated data set which is somewhat less accurate than the learned potential. Since we are mainly interested in kinetic properties, we estimate a new gEDMD model on the propagated data set for the CG dynamics. We find that the four metastable states are correctly reproduced by a PCCA+ analysis of the propagated coarse grained SDE, as shown in the left panel of Figure . In addition, we show the resulting transition time scales on the right of Figure , compared to the ones corresponding to the learned generator built upon the original data set, as well as the rescaled MSM time scales. The results confirm that the two-dimensional CG dynamics with learned effective diffusion accurately recover the metastable states and transition time scales of the original dynamics, while adequately recovering their thermodynamic properties.

7.

7

Left: Free energy surface learned via KDE. Right: estimated free energy surface from histogramming the simulated CG dynamics.

8.

8

Kinetic consistency of the CG dynamics for alanine dipeptide. Left: PCCA+ membership analysis applied to simulation data of the CG dynamics. Right: slowest finite time scales calculated using an approximation of the generator from the reference data set (blue) and the propagated CG dynamics with state-dependent diffusion (SDD, orange) as well as constant diffusion (CD, green), compared to those obtained via a Markov state model (black).

As a final analysis, we also generate a trajectory of the coarse grained SDE, but with the diffusion set to a constant. We choose the value of constant diffusion according to the average of the learned diffusion on the original data set, resulting in a ≈ 30.25 ps–1. We also estimate a gEDMD model for these dynamics, and report the transition time scales in Figure . The result shows the necessity of learning a state-dependent diffusion field.

4.3. Chignolin

4.3.1. System Introduction

Finally, we apply the proposed method to the 025 mutant of Chignolin (CLN025), which is a mini-protein consisting of 10 amino acids. Figure shows the graphical representation of the molecule. The data for this example was obtained via simulation in OpenMM based on AMBER99 SB-ILDN force field, see ref for details of the setup. The data set consists of 20 independent trajectories each for 5 μs.

9.

9

Graphical representation of CLN025 on the left, and the reference free energy surface in the two-dimensional TICA space on the right. The left-hand side minimum corresponds to the folded state, the bottom right minimum corresponds to the unfolded state and the top one associates to the misfolded state.

For this example, we need to find a coarse graining function in a data-driven manner. To obtain the CG space, we start with a 45-dimensional feature space comprising the C α distances of all residues. A straightforward linear method to find the CG coordinates is Time-Lagged Independent Component Analysis (TICA). As a result of TICA, we select the first 2 dominant components to constitute the RC space:

ξ(x)=[TIC1(x)TIC2(x)] 35

By projecting the atomistic positional information of the system onto this 2-dimensional TICA space and computing the histogram of the data, the free energy surface can be obtained, as shown in Figure . As shown in previous studies, the two-dimensional TICA space adequately captures the slow dynamics. In particular, the free energy surface shows three minima, representing the three conformational states of folded, unfolded and misfolded.

4.3.2. Meta-Stability Analysis

To find the time scales of the system, we applied the gEDMD method with random Fourier features as before, and computed the eigenvalues of the generator model r . We performed the same analysis as for the previous example to tune the kernel bandwidth and the number of random features based on the VAMP-score, see the Supporting Information for details. Figure shows the corresponding time scales of the system, which are the inverse of the generator’s eigenvalues. The figure indicates the two leading time scales of the system corresponding to the three metastable sets, followed by a spectral gap. Moreover, we show that the time scales of the CG generator L α for the optimal effective diffusion are very similar, the relative errors shown on the right of the same figure are sufficiently small. Also, we observe that the gEDMD time scales are once again uniformly rescaled compared to the leading time scales of an MSM estimated on the original data, see the previous example and Section . The rescaling factor is quite drastic this time, reducing microsecond time scales of the full system to less than pico-seconds for the CG dynamics. Nevertheless, as the rescaling is again uniform, the original time scales can be recovered by rescaling time. Error-bar figures were again generated by analyzing 5 independent subsampled sets, each comprising 1.6 × 105 samples.

10.

10

Approximation of generator for Chignolin. The slowest finite time scales corresponding to the reference generator r and the learned generator α built upon the learned effective diffusion on the left, and the relative error on the right. The time scales of the MSM model on the original simulation data are shown as black dashed lines for comparison. Note that time scales of the generators are rescaled by a factor of 106. The first time scale (l = 2) corresponds to the folded-unfolded transition and the second one (l = 3) corresponds to the unfolded-misfolded transition.

4.3.3. Analysis of the CG Dynamics

Following the same procedure as in the previous examples, we learned a 2 × 2 diffusion matrix in the CG space, but this time, we tested out a full nondiagonal diffusion field. Figure shows the four elements of the learned diffusion matrix. In addition, the left panel of Figure depicts the free energy surface learned by the KDE method, which is in satisfactory agreement with the reference one in Figure .

11.

11

(a–d) Components of the learned diffusion covariance matrix for Chignolin in its two-dimensional TICA space (note that the off-diagonal elements are symmetric).

12.

12

Free energy surface in the two-dimensional TICA space for Chignolin, as learned by the KDE estimator on the left, and obtained from a histogram of the CG dynamics on the right.

From the effective diffusion and potential energy, we compute the effective drift according to eq . We integrate the learned SDE for 5 × 105 integration steps, with an effective (rescaled) time step of dt = 20 ps, corresponding to an effective total simulation time of 10 μs. The right panel in Figure shows the estimated free energy surface obtained from a histogram of the propagated CG dynamics. Once again, we find it in satisfactory agreement with the learned and the reference free energy in the CG space. Its accuracy could likely be improved by applying a more accurate learning method.

As we are mainly interested in kinetic properties, we compute a new gEDMD model on the propagated CG dynamics, and recompute the associated eigenvalues and eigenvectors. The result of a PCCA+ analysis indicates that the correct metastable sets are recovered, as shown in the left panel in Figure . Likewise, the leading implied time scales estimated from the simulated CG dynamics are in good agreement with those of the original gEDMD model r and the rescaled MSM time scales, both estimated from the original simulation data, as shown in the right panel of Figure .

13.

13

Kinetic consistency of the learned CG model for Chignolin. Left: PCCA+ states obtained from simulating the learned CG model. Right: Slowest finite time scales of the system calculated using an approximation of the generator from the reference data set (blue) and from the propagated CG dynamics (state-dependent diffusion in orange, constant diffusion in green). We also compare to rescaled time scales from a Markov state model on the original simulation data (black).

Similar to the previous example, we also generate a separate trajectory based on a constant diffusion according to the average of the learned diffusion. We find that transition time scales for the constant diffusion are not well fitted to the reference. Due to taking the average, too much detailed information about the diffusion field is lost, leading to different time scales. This result confirms the need to learn a state-dependent diffusion field in the CG space to achieve kinetic consistency.

5. Discussion

We presented a novel approach to learn kinetically consistent coarse grained models for stochastic dynamics. We have introduced a learning method for the effective diffusion field in CG space, and shown how the kinetic properties of the CG dynamics can be evaluated by exploiting models for the Koopman generator (gEDMD algorithm). We have also shown that random Fourier features provide an efficient and flexible parametrization for both the effective diffusion and the gEDMD model. By means of three examples, a two-dimensional model potential and two data sets of molecular dynamics simulations, we showed that the effective dynamics in low-dimensional reaction coordinate spaces are able to reproduce both thermodynamic and kinetic quantities of the full dynamics accurately.

For the molecular examples, we have relied on the overdamped assumption to parametrize reversible CG dynamics. We have seen that this assumption leads to a uniform acceleration of the CG dynamics compared to the full system. The rescaling factor can be estimated numerically by comparing the gEDMD model to a kinetic model that does not rely on the overdamped assumption. We used MSMs in this paper, but note that a more general EDMD model (e.g., using random features) would work just as well.

In this study, we used long equilibrium simulations to train CG models. However, one of the appealing aspects of the generator EDMD approach is that it only requires Boltzmann samples. As has been pointed out in previous studies, these samples can also be obtained from biased sampling simulations, , or by employing generative models.

Among other topics, future work will focus on applying the formalism to higher-dimensional and more transferrable CG coordinates, for example C-alpha models. We do not anticipate a principal limitation to applying our method in higher-dimensional spaces. Learning the effective diffusion and the gEDMD model, which is crucial to validate kinetic consistency of the CG model, might require more careful parameter choices in higher-dimensional spaces. This is currently under investigation. Another topic is the construction of CG models that can explicitly account for the underdamped structure of the full system, or that can incorporate memory terms, which were entirely disregarded in our study. Moreover, one can also try to simultaneously optimize the CG mapping ξ along with the parameters of the CG model, for instance by balancing the VAMP score versus the complexity of the CG model.

Supplementary Material

ct5c00479_si_001.pdf (252.5KB, pdf)

Acknowledgments

The authors thank the Theoretical and Computational Biophysics Group at Freie Universität Berlin for sharing the simulation data of the Chignolin mini-protein.

Codes and data to reproduce the results and figures shown in this manuscript are available from the following public repository: 10.5281/zenodo.15209618.

The Supporting Information is available free of charge at https://pubs.acs.org/doi/10.1021/acs.jctc.5c00479.

  • VAMP-score; simulation settings for alanine dipeptide; references; and additional figures and tables (PDF)

Open access funded by Max Planck Society.

The authors declare no competing financial interest.

Published as part of Journal of Chemical Theory and Computation special issue “Markov State Modeling of Conformational Dynamics”.

Footnotes

1

The image is generated using Protein Data Bank in Europe platform.

References

  1. Frenkel, D. ; Smit, B. . Understanding molecular simulation: from algorithms to applications, 3rd ed.; Elsevier: 2023. [Google Scholar]
  2. Onuchic J. N., Luthey-Schulten Z., Wolynes P. G.. Theory of protein folding: the energy landscape perspective. Annu. Rev. Phys. Chem. 1997;48:545–600. doi: 10.1146/annurev.physchem.48.1.545. [DOI] [PubMed] [Google Scholar]
  3. Karplus M., Petsko G. A.. Molecular dynamics simulations in biology. Nature. 1990;347:631–639. doi: 10.1038/347631a0. [DOI] [PubMed] [Google Scholar]
  4. Das P., Moll M., Stamati H., Kavraki L. E., Clementi C.. Low-dimensional, free-energy landscapes of protein-folding reactions by nonlinear dimensionality reduction. Proc. Natl. Acad. Sci. U. S. A. 2006;103:9885–9890. doi: 10.1073/pnas.0603553103. [DOI] [PMC free article] [PubMed] [Google Scholar]
  5. Clementi C.. Coarse-grained models of protein folding: toy models or predictive tools? Curr. Opin. Struct. Biol. 2008;18:10–15. doi: 10.1016/j.sbi.2007.10.005. [DOI] [PubMed] [Google Scholar]
  6. Rohrdanz M. A., Zheng W., Clementi C.. Discovering Mountain Passes via Torchlight: Methods for the Definition of Reaction Coordinates and Pathways in Complex Macromolecular Reactions. Annu. Rev. Phys. Chem. 2013;64:295–316. doi: 10.1146/annurev-physchem-040412-110006. [DOI] [PubMed] [Google Scholar]
  7. Wang J., Ferguson A.. Nonlinear machine learning in simulations of soft and biological materials. Mol. Simul. 2018;44:1090–1107. doi: 10.1080/08927022.2017.1400164. [DOI] [Google Scholar]
  8. Sidky H., Chen W., Ferguson A. L.. Machine learning for collective variable discovery and enhanced sampling in biomolecular simulation. Mol. Phys. 2020;118:1737742. doi: 10.1080/00268976.2020.1737742. [DOI] [Google Scholar]
  9. Sidky H., Chen W., Ferguson A. L.. Molecular latent space simulators. Chemical Science. 2020;11:9459–9467. doi: 10.1039/D0SC03635H. [DOI] [PMC free article] [PubMed] [Google Scholar]
  10. Wu H., Noé F.. Reaction coordinate flows for model reduction of molecular kinetics. J. Chem. Phys. 2024;160:044109. doi: 10.1063/5.0176078. [DOI] [PubMed] [Google Scholar]
  11. John S., Csányi G.. Many-body coarse-grained interactions using Gaussian approximation potentials. J. Phys. Chem. B. 2017;121:10934–10949. doi: 10.1021/acs.jpcb.7b09636. [DOI] [PubMed] [Google Scholar]
  12. Zhang L., Han J., Wang H., Car R., E W.. DeePCG: Constructing coarse-grained models via deep neural networks. J. Chem. Phys. 2018;149:034101. doi: 10.1063/1.5027645. [DOI] [PubMed] [Google Scholar]
  13. Wang J., Olsson S., Wehmeyer C., Pérez A., Charron N. E., Fabritiis G. d., Noé F., Clementi C.. Machine Learning of Coarse-Grained Molecular Dynamics Force Fields. ACS Cent. Sci. 2019;5:755–767. doi: 10.1021/acscentsci.8b00913. [DOI] [PMC free article] [PubMed] [Google Scholar]
  14. Mori H.. Transport, Collective Motion, and Brownian Motion. Prog. Theor. Phys. 1965;33:423–455. doi: 10.1143/PTP.33.423. [DOI] [Google Scholar]
  15. Zwanzig R.. Nonlinear generalized Langevin equations. J. Stat. Phys. 1973;9:215–220. doi: 10.1007/BF01008729. [DOI] [Google Scholar]
  16. Gyöngy I.. Mimicking the one-dimensional marginal distributions of processes having an ito differential. Probability Theory and Related Fields. 1986;71:501–516. doi: 10.1007/BF00699039. [DOI] [Google Scholar]
  17. Legoll F., Lelièvre T.. Effective dynamics using conditional expectations. Nonlinearity. 2010;23:2131–2163. doi: 10.1088/0951-7715/23/9/006. [DOI] [Google Scholar]
  18. Pavliotis G. A., Stuart A. M.. Parameter Estimation for Multiscale Diffusions. J. Stat. Phys. 2007;127:741–781. doi: 10.1007/s10955-007-9300-6. [DOI] [Google Scholar]
  19. Zhang W., Hartmann C., Schütte C.. Effective dynamics along given reaction coordinates, and reaction rate theory. Faraday Discuss. 2016;195:365–394. doi: 10.1039/C6FD00147E. [DOI] [PubMed] [Google Scholar]
  20. Nüske F., Koltai P., Boninsegna L., Clementi C.. Spectral properties of effective dynamics from conditional expectations. Entropy. 2021;23:134. doi: 10.3390/e23020134. [DOI] [PMC free article] [PubMed] [Google Scholar]
  21. Zhang W., Schütte C.. On Finding Optimal Collective Variables for Complex Systems by Minimizing the Deviation between Effective and Full Dynamics. Multiscale Modeling & Simulation. 2025;23:924–958. doi: 10.1137/24M1658917. [DOI] [Google Scholar]
  22. Jin J., Pak A. J., Durumeric A. E. P., Loose T. D., Voth G. A.. Bottom-up Coarse-Graining: Principles and Perspectives. J. Chem. Theory Comput. 2022;18:5759–5791. doi: 10.1021/acs.jctc.2c00643. [DOI] [PMC free article] [PubMed] [Google Scholar]
  23. Schneider E., Dai L., Topper R. Q., Drechsel-Grau C., Tuckerman M. E.. Stochastic neural network approach for learning high-dimensional free energy surfaces. Physical review letters. 2017;119:150601. doi: 10.1103/PhysRevLett.119.150601. [DOI] [PubMed] [Google Scholar]
  24. Noid W. G., Chu J.-W., Ayton G. S., Krishna V., Izvekov S., Voth G. A., Das A., Andersen H. C.. The multiscale coarse-graining method. I. A rigorous bridge between atomistic and coarse-grained models. J. Chem. Phys. 2008;128:244114. doi: 10.1063/1.2938860. [DOI] [PMC free article] [PubMed] [Google Scholar]
  25. Rudzinski J. F., Kremer K., Bereau T.. Communication: Consistent interpretation of molecular simulation kinetics using Markov state models biased with external information. J. Chem. Phys. 2016;144:051102. doi: 10.1063/1.4941455. [DOI] [PubMed] [Google Scholar]
  26. Rudzinski J. F., Kloth S., Wörner S., Pal T., Kremer K., Bereau T., Vogel M.. Dynamical properties across different coarse-grained models for ionic liquids. J. Phys.: Condens. Matter. 2021;33:224001. doi: 10.1088/1361-648X/abe6e1. [DOI] [PubMed] [Google Scholar]
  27. Martino S. A., Morado J., Li C., Lu Z., Rosta E.. Kemeny Constant-Based Optimization of Network Clustering Using Graph Neural Networks. J. Phys. Chem. B. 2024;128:8103–8115. doi: 10.1021/acs.jpcb.3c08213. [DOI] [PMC free article] [PubMed] [Google Scholar]
  28. Wang, Y. ; Voth, G. A. . Adversarial Training for Dynamics Matching in Coarse-Grained Models, 2025. http://arxiv.org/abs/2504.06505, arXiv:2504.06505.
  29. Sule, S. ; Mehta, A. ; Cameron, M. K. . Learning collective variables that preserve transition rates. 2025; http://arxiv.org/abs/2506.01222, arXiv:2506.01222.
  30. Prinz J.-H., Wu H., Sarich M., Keller B., Senne M., Held M., Chodera J. D., Schütte C., Noé F.. Markov models of molecular kinetics: Generation and validation. J. Chem. Phys. 2011;134:174105. doi: 10.1063/1.3565032. [DOI] [PubMed] [Google Scholar]
  31. Davies E. B.. Metastable States of Symmetric Markov Semigroups II. Journal of the London Mathematical Society. 1982;s2–26:541–556. doi: 10.1112/jlms/s2-26.3.541. [DOI] [Google Scholar]
  32. Dellnitz M., Junge O.. On the approximation of complicated dynamical behavior. SIAM J. Numer. Anal. 1999;36:491–515. doi: 10.1137/S0036142996313002. [DOI] [Google Scholar]
  33. Schütte C., Fischer A., Huisinga W., Deuflhard P.. A direct approach to conformational dynamics based on hybrid Monte Carlo. J. Comput. Phys. 1999;151:146–168. doi: 10.1006/jcph.1999.6231. [DOI] [Google Scholar]
  34. Klus S., Nüske F., Koltai P., Wu H., Kevrekidis I., Schütte C., Noé F.. Data-Driven Model Reduction and Transfer Operator Approximation. J. Nonlinear Sci. 2018;28:985–1010. doi: 10.1007/s00332-017-9437-7. [DOI] [Google Scholar]
  35. Sarich M., Noé F., Schütte C.. On the approximation quality of Markov state models. Multiscale Model. Simul. 2010;8:1154–1177. doi: 10.1137/090764049. [DOI] [Google Scholar]
  36. Bowman, G. R. ; Pande, V. S. ; Noé, F. , Eds.; An Introduction to Markov State Models and Their Application to Long Timescale Molecular Simulation; Springer: Netherlands, 2014; Vol. 797. [Google Scholar]
  37. Noé F., Nüske F.. A variational approach to modeling slow processes in stochastic dynamical systems. Multiscale Modeling and Simulation. 2013;11:635–655. doi: 10.1137/110858616. [DOI] [Google Scholar]
  38. Mardt A., Pasquali L., Wu H., Noé F.. VAMPnets for deep learning of molecular kinetics. Nat. Commun. 2018;9:5. doi: 10.1038/s41467-017-02388-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
  39. Wu H., Noé F.. Variational approach for learning Markov processes from time series data. J. Nonlinear Sci. 2020;30:23–66. doi: 10.1007/s00332-019-09567-y. [DOI] [Google Scholar]
  40. Nüske F., Boninsegna L., Clementi C.. Coarse-graining molecular systems by spectral matching. J. Chem. Phys. 2019;151:044116. doi: 10.1063/1.5100131. [DOI] [PubMed] [Google Scholar]
  41. Klus S., Nüske F., Peitz S., Niemann J.-H., Clementi C., Schütte C.. Data-driven approximation of the Koopman generator: Model reduction, system identification, and control. Physica D: Nonlinear Phenomena. 2020;406:132416. doi: 10.1016/j.physd.2020.132416. [DOI] [Google Scholar]
  42. Rahimi, A. ; Recht, B. . Random features for large-scale kernel machines. In Advances in Neural Information Processing Systems; Curran Associates Inc.: 2007; Vol. 20. [Google Scholar]
  43. Pavliotis, G. A. Stochastic processes and applications; Texts in Applied Mathematics; Springer: 2014; Vol. 60. [Google Scholar]
  44. Koopman B. O.. Hamiltonian systems and transformation in Hilbert space. Proc. Natl. Acad. Sci. U S A. 1931;17:315. doi: 10.1073/pnas.17.5.315. [DOI] [PMC free article] [PubMed] [Google Scholar]
  45. Mezić I.. Spectral Properties of Dynamical Systems, Model Reduction and Decompositions. Nonlinear Dynamics. 2005;41:309–325. doi: 10.1007/s11071-005-2824-x. [DOI] [Google Scholar]
  46. Lelièvre T., Stoltz G.. Partial differential equations and stochastic methods in molecular dynamics. Acta Numerica. 2016;25:681–880. doi: 10.1017/S0962492916000039. [DOI] [Google Scholar]
  47. Nüske F., Klus S.. Efficient approximation of molecular kinetics using random Fourier features. J. Chem. Phys. 2023;159:074105. doi: 10.1063/5.0162619. [DOI] [PubMed] [Google Scholar]
  48. Duvenaud, D. Automatic model construction with Gaussian processes, PhD Thesis, 2014.
  49. Parzen E.. On Estimation of a Probability Density Function and Mode. Annals of Mathematical Statistics. 1962;33:1065–1076. doi: 10.1214/aoms/1177704472. [DOI] [Google Scholar]
  50. Lelièvre, T. ; Rousset, M. ; Stoltz, G. . Free Energy Computations; Imperial College Press: 2010. [Google Scholar]
  51. Duong M. H., Lamacz A., Peletier M. A., Schlichting A., Sharma U.. Quantification of coarse-graining error in Langevin and overdamped Langevin dynamics. Nonlinearity. 2018;31:4517–4566. doi: 10.1088/1361-6544/aaced5. [DOI] [Google Scholar]
  52. Deuflhard P., Weber M.. Robust Perron cluster analysis in conformation dynamics. Linear Algebra Appl. 2005;398:161–184. doi: 10.1016/j.laa.2004.10.026. [DOI] [Google Scholar]
  53. Honda S., Akiba T., Kato Y. S., Sawada Y., Sekijima M., Ishimura M., Ooishi A., Watanabe H., Odahara T., Harata K.. Crystal Structure of a Ten-Amino Acid Protein. J. Am. Chem. Soc. 2008;130:15327–15331. doi: 10.1021/ja8030533. [DOI] [PubMed] [Google Scholar]
  54. Charron, N. E. ; Musil, F. ; Guljas, A. ; Chen, Y. ; Bonneau, K. ; Pasos-Trejo, A. S. ; Venturin, J. ; Gusew, D. ; Zaporozhets, I. ; Krämer, A. . et al. Navigating protein landscapes with a machine-learned transferable coarse-grained model. arXiv preprint arXiv:2310.18278 2023,. [DOI] [PMC free article] [PubMed]
  55. Pérez-Hernández G., Paul F., Giorgino T., Fabritiis G. D., Noé F.. Identification of slow molecular order parameters for Markov model construction. J. Chem. Phys. 2013;139:015102. doi: 10.1063/1.4811489. [DOI] [PubMed] [Google Scholar]
  56. Lücke M., Nüske F.. tgEDMD: Approximation of the Kolmogorov Operator in Tensor Train Format. J. Nonlinear Sci. 2022;32:44. doi: 10.1007/s00332-022-09801-0. [DOI] [Google Scholar]
  57. Devergne, T. ; Kostic, V. ; Parrinello, M. ; Pontil, M. . From Biased to Unbiased Dynamics: An Infinitesimal Generator Approach, 2024; http://arxiv.org/abs/2406.09028, arXiv:2406.09028. [Google Scholar]
  58. Moqvist S., Chen W., Schreiner M., Nüske F., Olsson S.. Thermodynamic Interpolation: A Generative Approach to Molecular Thermodynamics and Kinetics. J. Chem. Theory Comput. 2025;21:2535–2545. doi: 10.1021/acs.jctc.4c01557. [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

ct5c00479_si_001.pdf (252.5KB, pdf)

Data Availability Statement

Codes and data to reproduce the results and figures shown in this manuscript are available from the following public repository: 10.5281/zenodo.15209618.


Articles from Journal of Chemical Theory and Computation are provided here courtesy of American Chemical Society

RESOURCES