Abstract
We consider complex dynamical systems showing metastable behavior, but no local separation of fast and slow time scales. The article raises the question of whether such systems exhibit a low-dimensional manifold supporting its effective dynamics. For answering this question, we aim at finding nonlinear coordinates, called reaction coordinates, such that the projection of the dynamics onto these coordinates preserves the dominant time scales of the dynamics. We show that, based on a specific reducibility property, the existence of good low-dimensional reaction coordinates preserving the dominant time scales is guaranteed. Based on this theoretical framework, we develop and test a novel numerical approach for computing good reaction coordinates. The proposed algorithmic approach is fully local and thus not prone to the curse of dimension with respect to the state space of the dynamics. Hence, it is a promising method for data-based model reduction of complex dynamical systems such as molecular dynamics.
Keywords: Metastability, Reaction coordinate, Coarse graining, Effective dynamics, Whitney embedding theorem, Transfer operator
Introduction
With the advancement of computing power, we are able to simulate and analyze more and more complicated and high-dimensional models of dynamical systems, ranging from astronomical scales for the simulation of galaxies, over planetary and continental scales for climate and weather prediction, down to molecular and subatomistic scales via, e.g., molecular dynamics (MD) simulations aimed at gaining insight into complex biological processes. Particular aspects of such processes, however, can often be described by much simpler means than the full process, thus reducing the full dynamics to some essential behavior or effective dynamics in terms of some essential observables of the system. Extracting these observables and the related effective dynamics from a dynamical system, though, is one of the most challenging problems in computational modeling (Froyland et al. 2014).
One prominent example of dynamical reduction is arguably given by a variety of multiscale systems with explicit fast–slow time scale separation, mostly singularly perturbed systems, where either the fast component is considered in a quasi-stationary regime (i.e., the slow components are fixed and assumed not to change for the observation period), or the effective behavior of the fast components is injected into the slow processes, e.g., by averaging or homogenization (Pavliotis and Stuart 2008). Much of the recent attention has been directed to the case where the deduction of the slow (or fast) effective dynamics is not possible by purely analytic means, due to the lack of an analytic description of the system, or because the complexity of the system renders this task unfeasible (Froyland et al. 2014, 2016; Coifman et al. 2008; Dsilva et al. 2016; Nadler et al. 2006; Singer et al. 2009; Crosskey and Maggioni 2017; Vanden-Eijnden 2007; Kevrekidis and Samaey 2009). However, all of these approaches still depend on some local form of time scale separation between the “fast” and the “slow” components of the dynamics.
The focus of this work is on specific multiscale systems without local dynamical slow–fast time scale separation, but for which a reduction to an effective dynamical behavior supported on some low-dimensional manifold is still possible. The dynamical property lying at the heart of our approach is that there is a time scale separation in the global kinetic behavior of the process, as opposed to the aforementioned slow–fast behavior encoded in the local dynamics. Here, global kinetic behavior means that the multiple scales show up if we consider the Fokker–Planck equation associated with the dynamics, say , where the Fokker–Planck operator will have several small eigenvalues, while the rest of its spectrum is significantly larger. Such dynamical systems exhibit metastable behavior, and the slow time scales are the time scales of statistical relaxation between the main metastable sets, while there is no time scale gap for the local dynamics within each of the metastable regions (Bovier et al. 2002; Schütte and Sarich 2014).
Global time scale separation induced by metastability has been analyzed for deterministic (Dellnitz and Junge 1999) and stochastic dynamical systems (Schütte et al. 1999; Huisinga et al. 2004) for more than a decade. A typical trajectory of a metastable dynamical system will spend most time within the metastable sets, while rare transitions between these sets happen as sudden “jumps” roughly along low-dimensional transition pathways that connect the metastable sets (Dellago and Bolhuis 2009; Noé et al. 2009; E and Vanden-Eijnden 2010). For an example, see Fig. 1.
Fig. 1.
a Curved double-well potential with two metastable sets (areas encircled by light gray lines) around the global minima and (1, 0). In a typical trajectory (red line), transitions between the metastable sets are rare events and generally happen along the transition path (dashed white line). b The -component of a longer trajectory that shows multiple rare transitions (or events) (Color figure online)
The tool to describe the global kinetic behavior of a metastable system is the so-called transfer operator (the evolution operator of the Fokker–Planck equation), which acts on functions on the state space. The time scale separation we rely on here implies a spectral gap for this operator. This fact has been exploited to find low-dimensional representations of the global kinetics in the form of Markov chains whose (discrete) states represent the metastable sets, while the transition probabilities between the states approximate the jump statistics between the sets on long time scales. Under the name “Markov state models” (MSMs), this approach has led to a variety of methods (Bowman et al. 2014; Schütte and Sarich 2014) with broad application, e.g., in molecular dynamics, cf. Schütte et al. (1999), Pande et al. (2010), Schütte et al. (2011) and Chodera and Noé (2014). This reduction comes with a price: Since the relaxation kinetics is described just by jumps between the metastable sets in a (finite) discrete state space, any information about the transition process and its dynamical features is lost. A variety of approaches have been developed for complementing the MSM approach appropriately (Metzner et al. 2009), but a continuous (in time and space) low-dimensional effective description based on MSMs allowing to understand the transition mechanism is infeasible.
In another branch of the literature, again heavily influenced by molecular dynamics applications, model reduction techniques have been developed that assume the existence of a low-dimensional reaction coordinate or order parameter in order to construct an effective dynamics or kinetics: Examples are free energy-based techniques (Torrie and Valleau 1977; Laio and Gervasio 2008), trajectory-based sampling techniques (Faradjian and Elber 2004; Becker et al. 2012; Moroni et al. 2004; Pérez-Hernández et al. 2013), methods based on diffusive processes (Best and Hummer 2010; Zhang et al. 2016; Pavliotis and Stuart 2008), and many more that rely on the assumption that the reaction coordinates are known. The problem of actually constructing good reaction coordinates remains an area of ongoing research (Li and Ma 2014), to which this paper contributes. Typically, reaction coordinates are either postulated using system-specific expert knowledge (Camacho and Thirumalai 1993; Socci et al. 1996), an approximation to the dominant eigenfunctions of the transfer operator is sought (Schütte and Sarich 2014; Chodera and Noé 2014; Pérez-Hernández et al. 2013), or machine learning techniques are proposed (Ma and Dinner 2005). Froyland et al. (2014) show that these eigenfunctions are indeed optimal—in the sense of optimally representing the slow dynamics—but for high-dimensional systems computational reaction coordinate identification still is often infeasible. In the context of transition path theory (Vanden-Eijnden 2006), the committor function is known to be an ideal (Lu and Vanden-Eijnden 2014) reaction coordinate. In Pozun et al. (2012), the authors construct a level set of the committor using support vector machines, but the computation of reaction coordinates is infeasible for high-dimensional systems. The main problem in computing reaction coordinates for high-dimensional metastable systems results from the fact that all of these algorithms try to solve a global problem in the entire state space that cannot be decomposed easily into purely local computations.
In this article, we elaborate on the definition, existence and algorithmic identification of reaction coordinates for metastable systems: We define reaction coordinates as a small set of nonlinear coordinates on which a reduced system (Legoll and Lelièvre 2010; Zhang et al. 2016) can be defined having the same dominant time scales (in terms of transfer operator eigenvalues) as the original system. We then consider a low-dimensional state space on which the reduced dynamics is a Markov process. Thus, our approach utilizes concepts and transfer operator theory developed previously, but in our case the projected transfer operator is still infinite-dimensional, in stark contrast to its reduction to a stochastic matrix in the MSM approach.
The contribution of this paper is twofold: First, we develop a conceptual framework that identifies good reaction coordinates as the ones that parametrize a low-dimensional transition manifold in the function space , which is the natural state space of the Fokker–Planck equation associated with the dynamics. The property which defines is that, on moderate time scales , the transition density functions of the dynamics concentrate around . We provide evidence that such an indeed exists due to metastability and the existence of transition pathways. Crucially, the dimension of is often lower than the number of dominant eigenfunctions.
Second, we present an algorithm to construct approximate reaction coordinates. Our algorithm is data-driven and fully local, thus circumventing the main problem of previously proposed algorithms: In order to compute the value of the desired reaction coordinate at a location x in the state space , only the ability to simulate short trajectories initialized at x is needed. In particular, we assume no a priori knowledge of metastable sets, no global equilibration, and we do not need to resolve the slow time scales numerically. The algorithm is built on two pillars:
The simulation time scale t can be chosen a lot smaller than the dominant time scales of the system, such that it is feasible to simulate many short trajectories of length t.
We utilize embedding techniques inspired by the seminal work of Whitney (1936) and the recent work Dellnitz et al. (2016) that allows one to take almost any mapping into a Euclidean space of more than twice the dimension of the manifold and to obtain a one-to-one image of it.
These two pillars together with the low-dimensionality of imply that we can represent the image of the reaction coordinate in a space with moderate (finite) dimension. Then, we can use established manifold learning techniques (Nadler et al. 2006; Coifman et al. 2008; Singer et al. 2009) to obtain a parametrization of the manifold in the embedding space and pull this parametrization back to the original state space, hence obtaining a reaction coordinate.
The locality of the algorithm also implies that reaction coordinates are only computed in the region of state space where sampled points are available. This is a common issue with manifold learning algorithms; here, it manifests as the transition manifold being reliably learned only in regions we have good sampling coverage of. However, recently several methods have appeared in the literature that allow a fast exploration of the state space. These methods do not provide equilibrium sampling, but instead try to rapidly cover the essential part of the state space with sampling points. This can be achieved with enhanced sampling methods such as umbrella sampling (Kumar et al. 1992; Torrie and Valleau 1977), metadynamics (Laio and Gervasio 2008; Laio and Parrinello 2002), blue-moon sampling (Ciccotti et al. 2005), adaptive biasing force method (Darve et al. 2008) or temperature-accelerated molecular dynamics (Maragliano and Vanden-Eijnden 2006), as well as trajectory-based techniques such as milestoning (Faradjian and Elber 2004), transition interface sampling (Moroni et al. 2004) or forward flux sampling (Becker et al. 2012). Alternatively, several techniques such as the equation-free approach (Kevrekidis and Samaey 2009), the heterogeneous multiscale method (HMM) (E and Engquist 2003) and methods based on diffusion maps (Chiavazzo et al. 2016) have been developed to utilize short unbiased MD trajectories for extracting information that allows much larger time steps. This can be combined with reaction coordinate-based effective dynamics (Zhang et al. 2016; Zhang and Schuette 2017).
In principle, the method we present in this article may be combined with any enhanced sampling technique in order to generate sampling points that cover a large part of the state space. For simplicity, we will use long MD trajectories to generate our sampling points, but we do not require that the points are distributed according to an equilibrium distribution.
The paper is organized as follows: Sect. introduces transfer operators, which describe the global kinetics of the stochastic process. Based on these transfer operators, we define metastability, i.e., the existence of dominant time scales. In Sect. 3, we describe the model reduction techniques Markov state modeling and coordinate projection that are designed to capture the dominant time scales of metastable systems. Furthermore, we characterize good reaction coordinates. In the first part of Sect. 4, we show that our dynamical assumption ensures the existence of good reaction coordinates, and then in the second part we describe our approach to compute them. Several numerical examples are given in Sect. . Concluding remarks and an outlook are provided in Sect. 6.
Transfer Operators and Their Properties
As mentioned in Introduction, global properties of dynamical systems such as metastable sets or a partitioning into fast and slow subprocesses can be obtained using transfer operators associated with the system and their eigenfunctions. In this section, we will introduce different transfer operators needed for our considerations.
Transfer Operators
In what follows, denotes probabilities conditioned on the event and the expectation value. Furthermore, is a stochastic process defined on a state space .
Definition 2.1
(Transition density function) Let be any measurable set, then the transition density function of a time-homogeneous stochastic process is defined by
That is, is the conditional probability density of given that .
With the aid of the transition density function, we can now define transfer operators. Note, however, that the transition density is in general not known explicitly and needs to be estimated from simulation data. In what follows, we assume that there is a unique equilibrium density that is invariant under , that is, it satisfies
a.e. on . Let denote the associated invariant measure .
Definition 2.2
(Transfer operators) Let be a probability density,1 be a probability density with respect to the equilibrium density , and be an observable of the system. For a given lag time t:
- The Perron–Frobenius operator is defined by the unique linear extension of
to . - The Perron–Frobenius operator with respect to the equilibrium density is defined by the unique linear extension of
to . - The Koopman operator is defined by
1
All these are well-defined non-expanding operators on the respective spaces.
The equilibrium density satisfies , that is, is an eigenfunction of with associated eigenvalue . The definition of relies on , and we have .
Instead of their natural domains from Definition 2.2, all our transfer operators are considered on the following Hilbert spaces: , , and . They are still well-defined non-expansive operators on these spaces (Baxter and Rosenthal 1995; Schervish and Carlin 1992; Klus et al. 2017).
Furthermore, we will need the notion of reversibility for our considerations. Reversibility means that the process is statistically indistinguishable from its time-reversed counterpart.
Definition 2.3
(Reversibility) A system is said to be reversible if the detailed balance condition
is satisfied for all .
In what follows, we will assume that the system is reversible.
One prominent example for a class of SDEs satisfying uniqueness of the equilibrium density and reversibility is given by
| 2 |
Here, V is called the potential, is the non-dimensionalized inverse temperature, and is a standard Wiener process. The process generated by (2) is ergodic and thus admits a unique positive equilibrium density, given by , under mild growth conditions on the potential V (Mattingly and Stuart 2002; Mattingly et al. 2002). Note that the subsequent considerations hold for all stochastic processes that satisfy reversibility and ergodicity with respect to a unique positive invariant measure and are not limited to the class of dynamical systems given by (2). See Schütte and Sarich (2014) for a discussion of a variety of stochastic dynamical systems that have been considered in this context.
As a result of the detailed balance condition, the Koopman operator and the Perron–Frobenius operator with respect to the equilibrium density become identical and we obtain
i.e., all the transfer operators become self-adjoint on the respective Hilbert spaces from above. Here and denote the natural scalar products on the weighted spaces and , respectively.
Spectral Decomposition
Due to the self-adjointness, the eigenvalues of and are real-valued and the eigenfunctions form an orthogonal basis with respect to and , respectively. In what follows, we assume that the spectrum of is purely discrete given by (infinitely many) isolated eigenvalues. This assumption is made for the sake of simplicity. It is actually not required for the rest of our considerations; it would be sufficient to assume that the spectral radius R of the essential spectrum of is strictly smaller than 1, and some isolated eigenvalues of modulus larger than R exist. It has been shown that this condition is satisfied for a large class of metastable dynamical systems, see Schütte and Sarich (2014), Sec. 5.3 for details. For example, the process generated by (2) has purely discrete spectrum under mild growth and regularity assumptions on the potential V.
Under this condition, ergodicity implies that the dominant eigenvalue is the only eigenvalue with absolute value 1 and we can thus order the eigenvalues so that
The eigenfunction of corresponding to is the constant function . Let be the normalized eigenfunctions of , i.e., , then any function can be written in terms of the eigenfunctions as . Applying thus results in
For more details, we refer to Klus et al. (2017) and references therein.
Implied Time Scales
For some , we call the dominant eigenvalues of the dominant spectrum of , i.e.,
Usually, d is chosen in such a way that there is a spectral gap after , i.e., . The (implied) time scales on which the associated dominant eigenfunctions decay are given by
| 3 |
If is a semigroup of operators, then there are with such that holds. Assuming there is a spectral gap, the dominant time scales satisfy . These are the time scales of the slow dynamical processes, also called rare events, which are of primary interest in applications. The other fast processes are regarded as fluctuations around the relative equilibria (or metastable states) between which the relevant slow processes travel.
Projected Transfer Operators and Reaction Coordinates
The purpose of dimension reduction in molecular dynamics is to find a reduced dynamical model that captures the dominant time scales of the system correctly while keeping the model as simple as possible. In this section, we will introduce two different projections and the corresponding projected transfer operators. The goal is to find suitable projections onto the slow processes.
Galerkin Projections and Markov State Models
One frequently used approach to obtain a reduced model is Markov state modeling. The goal is to find a model that is as simple as possible and yet correctly reproduces the dominant time scales. Given a fixed , most authors (Noé et al. 2009; Pande et al. 2010) refer to a Markov state model (MSM) as a matrix such that
| 4 |
and it has been studied in detail under which condition this can be achieved (Djurdjevac et al. 2012; Sarich et al. 2010).
There are different ways of constructing an MSM, maybe the most intuitive one is also the simplest: Let the entries of be the transition rates between metastable sets. A typical molecular system with d dominant time scales will have metastable sets (also called cores), and its dynamics is characterized by transitions between these sets and fluctuations inside the sets (see Fig. 1 for an illustration). Since the fluctuations are on faster time scales, we neglect them by setting (Schütte et al. 1999)
| 5 |
where denotes the probability measure conditioned to the initial condition being distributed according to . Thus, is the probability that the process in equilibrium jumps to the metastable set in time t, given that it started in the metastable set . Note that (5) can be equivalently rewritten as
| 6 |
where is the characteristic function of the set .
Equation (6) readily suggests that is a projection of the transfer operator , namely its Galerkin projection onto the space spanned by the characteristic functions (Schütte et al. 1999).
Definition 3.1
(Galerkin projection) Given a set of basis functions , let and . The projection to or, equivalently, to , is defined as
The residual projection is given by , where is the identity. The Galerkin projection of to is given by the linear operator satisfying
Equivalently, . We also denote the extension of to the whole , given by , by . Furthermore, we denote the matrix representation of with respect to the basis by as well. Either it will be clear from the context which of the objects is meant or it will not matter; e.g., the dominant spectrum is the same for all of them.
We see that is the matrix representation of the Galerkin projection with respect to the basis functions , . More general MSMs can be built by Galerkin projections of the transfer operator to spaces spanned by other—not necessarily piecewise constant—basis functions (Weber 2006; Schütte et al. 2011; Weber et al. 2017; Klus et al. 2016, 2017; Pérez-Hernández et al. 2013; Noé and Nüske 2013). However, in some of these methods, one also often loses the interpretation of the entries of the matrix as probabilities.
Ultimately, the best MSM in terms of approximation quality in (4) is given by the Galerkin projection of onto the space spanned by its dominant eigenfunctions . This space is invariant under since and the dominant eigenvalues (and hence the time scales) are the same for the MSM and for . Due to the curse of dimensionality, however, the computation of the eigenfunctions is in general infeasible for high-dimensional problems.
Remark 3.2
There are quantitative results assessing the error in (4) of the MSM in terms of the projection errors , , cf. Schütte and Sarich (2014), Section 5.3. One can obtain a weaker, but similar result from our Lemma 3.5 in the next section.
Coordinate Projections and Effective Transfer Operators
While the MSMs from above successfully reproduce the dominant time scales of the original system, they often discard all other information about the system, such as the transition paths between metastable sets. Minimal coordinates that describe these transitions are called reaction coordinates, and reducing the dynamics onto these coordinates yields effective dynamics (Legoll and Lelièvre 2010; Zhang et al. 2016). The goal of the previous section—namely to retain the dominant time scales of the original dynamics in a reduced model—can now be reformulated for this lower-dimensional effective dynamics or, equivalently, for its (effective) transfer operator.
Let be a function, where . Let be the z-level set of . The so-called coarea formula (Federer 1969, Section 3.2), which can be considered as a nonlinear variant of Fubini’s theorem, splits integrals over into consecutive integrals over level sets of and then over the range of . For , we have2
| 7 |
where and is the surface measure on . The coordinate projection, defined next, averages a given function along the level sets of a coordinate function .
Definition 3.3
(Coordinate projection) For , we define
| 8 |
| 9 |
where is a probability measure on with density with respect to . Here, is just the normalization constant so that becomes a probability measure. The residual projection is given by .
To get a better feeling for the action of , note that is the expectation of with respect to conditional to , i.e.,
Or, in other words, is the marginal of conditional to . Note, in particular, that is itself a function on , but it is constant on the level sets of , and thus let us set for . It follows from the coarea formula (7) and (9) that
| 10 |
Next, we state some properties of the coordinate projection.
Proposition 3.4
The coordinate projection has the following properties.
is a linear projection, i.e., .
is self-adjoint with respect to .
is orthogonal, hence non-expansive, i.e., .
Proof
See Appendix A.
We use the coordinate projection to describe the dynamics-induced propagation of reduced distributions with respect to the variable . To this end, we define the effective transfer operator by
| 11 |
We immediately obtain from the self-adjointness of (see Sect. 2) and Proposition 3.4 (b) that is a self-adjoint operator on . Moreover, and Proposition 3.4 (c) imply that . Thus, the spectrum of the effective transfer operator lies in the interval , too.
Returning to the purpose of these constructions, we call a good reaction coordinate if
| 12 |
While the previously introduced Markov state model obtained by the Galerkin projection was approximating the dominant spectrum of the original transfer operator by a finite-dimensional operator (i.e., a matrix), the effective transfer operator still acts on an infinite-dimensional space. The reduction lies in the fact that operates on functions over , but the effective transfer operator operates essentially on functions over , although we embed those into through the level sets of .
As mentioned above, a Galerkin projection of the transfer operator onto its dominant eigenfunctions is a perfect MSM. In the same vein, we ask here how we can characterize a good reaction coordinate. We can make use of the following general result.
Lemma 3.5
Let be a Hilbert space with scalar product and associated norm ; let be some orthogonal projection on a linear subspace of , with . Let be a self-adjoint non-expansive linear operator, and u with its eigenvector, i.e., for some . If , then has an eigenvalue with .
Proof
Using , we have
where since Q and T are non-expanding. Thus, satisfies , and the orthogonality of Q gives . Now, any orthogonal projection is self-adjoint, as shown in the proof of Proposition 3.4; hence, the operator QTQ is self-adjoint, too, and thus normal. From the theory of pseudospectra for normal operators (Trefethen and Embree 2005, Theorems 2.1, 2.2, and §4), we know that if , then has an eigenvalue in the -neighborhood of .
With , , and we immediately obtain the following result.
Corollary 3.6
As before, let and , , denote the dominant eigenvalues and eigenfunctions of , respectively. For any given i, if , then there is an eigenvalue of with .
Corollary 3.6 implies that if the projection error of all dominant eigenfunctions is small, then is a good reaction coordinate in the sense of (12). Very similar results are available for approximation of the eigenvalues of the infinitesimal generator of the Fokker–Planck equation associated with the transfer operator if the dynamical system under consideration is continuous in time (Zhang and Schuette 2017).
Under which conditions is the projection error small? Let us consider the case where there are , , such that . We then say that is a function of or that parametrizes . If parametrizes perfectly, the projection error obviously vanishes. Thus, trivially, by choosing , we obtain a perfect reaction coordinate since with we have . However, the eigenfunctions are global objects, i.e., their computation is prohibitive in high dimensions. Since we are aiming at computing a reaction coordinate, we have to answer the question of whether there is a reaction coordinate that can be evaluated based on local computations only, while it parametrizes the dominant eigenfunctions of well enough such that it leads to a small projection error. We will see next that this question can be answered by utilizing a common property of most metastable systems: The transitions between the metastable sets happen along the so-called reaction pathways, which imply the existence of transition manifolds in the space of transition densities. A suitable parametrization of this manifold results in a parametrization of the dominant eigenfunctions with a small error.
Identifying Good Reaction Coordinates
The goal is now to find a reaction coordinate that is as low-dimensional as possible and results in a good projected transfer operator in the sense of (12). As we saw in the previous section, the condition is sufficient. Thus, the idea to numerically seek that parametrizes the dominant eigenfunctions of in the -norm seems natural since this would lead to small projection error .
In fact, eigenfunctions of transfer operators have been used before to compute reduced dynamics and reaction coordinates: In Froyland et al. (2014), methods to decompose multiscale systems into fast and slow processes and to project the dynamics onto these subprocesses based on eigenfunctions of the Koopman operator are proposed. In McGibbon et al. (2017), the dominant eigenfunctions of the transfer operator , which due to the assumed reversibility of the system is identical to , are shown to be good reaction coordinates. Also, committor functions (introduced in Appendix B), which are closely related to the dominant eigenfunctions, have been used as reaction coordinates in Du et al. (1998) and Lu and Vanden-Eijnden (2014).
However, we propose a fundamentally different path in defining and finding reaction coordinates, as working with dominant eigenfunctions has two major disadvantages:
The eigenproblem is global. Thus, if we wish to learn the value of an eigenfunction at only one location , we need an approximation of the transfer operator that has to be accurate on all of . The computational effort to construct such an approximation grows exponentially with ; this is the curse of dimensionality. There have been attempts to mitigate this (Weber 2006; Junge and Koltai 2009; Weber 2012), but we aim to circumvent this problem entirely. Given two points , we will decide whether is close to or not by using only local computations around x and y (i.e., samples from the transition densities and for moderate t).
The number of dominant eigenfunctions equals the number of metastable states, and this number can be much larger than the dimension of the transition manifold. This fact is illustrated in Example 4.1.
Example 4.1
Let us consider a diffusion process of the form (2) with the circular multi-well potential shown in Fig. 2. Choosing a temperature that is not high enough for the central potential barrier to be overcome easily, transitions between the wells typically happen in the vicinity of a one-dimensional reaction pathway, the unit circle. The number of dominant eigenfunctions, however, corresponds to the number of wells. Nevertheless, projecting the system onto the unit circle would retain the dominant time scales of the system, cf. Sect. .
Fig. 2.
a Potential with seven wells and thus seven dominant eigenvalues, but only a one-dimensional reaction coordinate. The reaction pathway is marked by a dashed white line. b Dominant eigenvalues of for and . The spectral gap is clearly visible
Parametrization of Dominant Eigenfunctions
If the dominant eigenfunctions do not depend fully on the phase space , a lower-dimensional and ultimately easier to find reaction coordinate suffices for keeping the eigenvalue approximation error (12) small. It is easy to see that if there exists a function for some k so that the eigenfunctions are constant on the level sets of , i.e., there exist functions , , such that , then the projection error is zero. A quantitative generalization of this is the statement that if the eigenfunctions are almost constant on level sets of a , then the projection error is small.
Lemma 4.2
Assume that there exists a function for some k and functions , , with
| 13 |
Then .
Proof
Assuming (13) holds, there exists a function with so that
Thus, we have
For the projection error, we then obtain
Remark 4.3
From the proof we see that the pointwise condition (13) can be replaced by the much weaker condition
for all level sets of .
From here on, we address the following two central questions:
In which dynamical situations can we expect to find low-dimensional reaction coordinates?
How can we computationally exploit the properties of the dynamics to obtain reaction coordinates?
Let us start with the first question. We will address the second question in Sects. 4.2 and 4.3. Experience shows (E et al. 2002; Ren et al. 2005; E and Vanden-Eijnden 2006; Schütte and Sarich 2014) that transitions metastable states tend to happen along the so-called reaction pathways, which are the low-dimensional dynamical backbone in the high-dimensional state space, connecting the metastable states via saddle points of the potential V (Freidlin and Wentzell 1998).
From now on, we observe the system at an intermediate time scale (where and are the implied time scales from Sect. 2.3) and thus assume that the process has already left the transition region (if it started there), equilibrated to a quasi-stationary distribution inside some metastable wells, but has not had enough time to equilibrate globally. At this time scale, starting in some , the transition density is observed to approximately depend only on progress along these reaction paths; see Fig. 3 for an illustration. This means that the density on the fiber perpendicular to the transition pathway is approximately the same as for some on the transition pathway. As this pathway is low-dimensional, this means that the image of the map
is almost a low-dimensional manifold in .
Fig. 3.
a, b The transition densities and are “similar” to for some on the transition path (dashed line) that connects the metastable sets and . c The mapping can be thought of as mapping all points that are “similar” under to the same point in . The image of thus forms a r-dimensional manifold in
The existence of this low-dimensional structure in the space of probability densities is exactly the assumption we need to ensure that the dominant eigenfunctions are low-dimensionally parametrizable, and thus that a low-dimensional reaction coordinate exists. This assumption is made precise in Definition 4.4. To summarize, we will see that is a good reaction coordinate if for .
Definition 4.4
(-reducibility and transition manifold) We call the process -reducible, if there exists a smooth closed r-dimensional manifold such that for and all
| 14 |
holds. We call the transition manifold and the map ,
| 15 |
the mapping onto the transition manifold. We can set , where denotes the closure of the set .3
Remark 4.5
While it is natural to motivate -reducibility by the existence of reaction pathways in phase space, it is not strictly necessary. There exist stochastic systems without low-dimensional reaction pathways whose densities still quickly converge to a transition manifold in . Future work includes the identification of necessary and sufficient conditions for the existence of transition manifolds (see the first point in Conclusion). We also further elaborate on the connection between reaction pathways and transition manifolds in Appendix B.
Remark 4.6
We recall from Sect. that the Perron–Frobenius operator is also naturally defined on the space (Schervish and Carlin 1992). Further, with the Dirac distribution centered in , denoted by , we formally have . Hence, the choice of norm in Definition 4.4 is natural. It should also be noted that since is a probability measure, the Hölder inequality yields . Using this we have
which shows that if and are close in the norm, they are also close in the norm. We require the closeness of the respective in the norm for our theoretical considerations below, but otherwise we will think of them as functions in .
Note that we only need to evolve the system at hand for a moderate time , which has to be merely sufficiently large to damp out the fast fluctuations in the metastable states. This will be an important point later, allowing for numerical tractability.
Next, we show that -reducibility implies that dominant eigenfunctions are almost constant on the level sets of .
Lemma 4.7
If is -reducible, then for an eigenfunction of with and points with we have
Proof
First note that for the transition densities it holds that
| 16 |
With this we can show the assertion:
Applying (16), for some function with , we get
where in the last equation, we again used that due to reversibility and that is an eigenfunction. Thus, for the difference, we have
Assuming that the eigenfunctions are normalized (which we do from now on), i.e., , and that is sufficiently small, Lemma 4.7 implies that the dominant eigenfunctions (i.e., ) are almost constant on the level sets of . This can now be used to show that the are not fully dependent on , but only on the level sets of (up to a small error), in a sense similar to Lemma 4.2.
Corollary 4.8
Let be -reducible. Then there exists a function such that
Proof
Fix , and let . Define the function by
Since by Lemma 4.7 it holds that if , we have that
thus, our choice of gives
Embedding the Transition Manifold
In light of Corollary 4.8, one could say that is an “-valued reaction coordinate.” However, as we have no access to so far, and a -valued reaction coordinate is more intuitive, we aim to obtain a more useful representation of the transition manifold through embedding it into a finite, possibly low-dimensional Euclidean space.
We will see that we are very free in the choice of the embedding mapping, even though the manifold is not known explicitly. (We only assumed that it exists.) To achieve this, we will use an infinite-dimensional variant of the weak Whitney embedding theorem (Sauer et al. 1991; Whitney 1936), which, roughly speaking, states that “almost every bounded linear map from to will be one-to-one on and its image.” We first specify what we mean by “almost every” in the context of bounded linear maps, following the notions of Sauer et al. (1991).
Definition 4.9
(Prevalence) A Borel subset of a normed linear space is called prevalent if there is a finite-dimensional subspace of such that for each , belongs to for (Lebesgue) almost every e in .
As the infinite-dimensional embedding theorem from Hunt and Kaloshin (1999) is applicable not only to smooth manifolds, but to arbitrary subsets of fractal dimension, it uses the concepts of box-covering dimension and thickness exponent from fractal geometry. Intuitively, describes the exponent of the growth rate in the number of boxes of decreasing side length that are needed to cover , and describes how well can be approximated using only finite-dimensional linear subspaces of . As these concepts coincide with the traditional measure of dimensionality in our setting, we will not go into detail here and point to Hunt and Kaloshin (1999) for a precise definition.
The general infinite-dimensional embedding theorem reads:
Theorem 4.10
(Hunt and Kaloshin 1999, Theorem 3.9) Let be a Banach space and be a compact set with box-counting dimension d and thickness exponent . Let be an integer, and let be a real number with
Then for almost every (in the sense of prevalence) bounded linear function there exists such that for all ,
| 17 |
where denotes the Euclidean 2-norm.
Note that (17) implies Hölder continuity of on and in particular that is one-to-one on and its image. Using that the box-counting dimension of a smooth r-dimensional manifold is simply r and that the thickness exponent is bounded from above by the box-counting dimension, thus , see Hunt and Kaloshin (1999), we get the following infinite-dimensional embedding theorem for smooth manifolds.
Corollary 4.11
Let be a Banach space, let be a smooth manifold of dimension r, and let . Then almost every (in the sense of prevalence) bounded linear function is one-to-one on and its image in .
Thus, since the transition manifold is assumed to be a smooth r-dimensional manifold in , an arbitrarily chosen bounded linear map can be assumed to be one-to-one on and its image. In particular, is again an r-dimensional manifold (although not necessarily smooth). With this insight, we can now construct a reaction coordinate in Euclidean space:
Corollary 4.12
Let be -reducible, and let be one-to-one on and its image. Define by
| 18 |
Then there exists a function so that
| 19 |
Proof
As is one-to-one on and its image, it is invertible on . With chosen as in the proof of Corollary 4.8, define by
| 20 |
Then
Since is an r-dimensional manifold, is effectively an r-dimensional reaction coordinate. Thus, if the right-hand side of (19) is small, the are “almost parametrizable” by the r-dimensional reaction coordinate . Using Lemma 4.2, we immediately see that this results in a small projection error , and due to Corollary 3.6 in a good transfer operator approximation; hence, is a good reaction coordinate.
The reaction coordinate remains an “ideal” case, because we have no access to the map and hence to , but only to . We show the construction of the ideal reaction coordinate in Fig. 4.
Fig. 4.
Summary of the construction of the ideal reaction coordinate
Remark 4.13
The recent work of Dellnitz et al. (2016) uses similar embedding techniques to identify finite-dimensional objects in the state space of infinite-dimensional dynamical systems. They utilize the infinite-dimensional delay-embedding theorem of Robinson (2005), a generalization of the well-known Takens embedding theorem (Takens 1981), to compute finite-dimensional attractors of delay differential equations by established subdivision techniques (Dellnitz and Hohmann 1997).
Numerical Approximation of the Reaction Coordinate
Approximate Embedding of the Transition Manifold We now elaborate how to construct a good reaction coordinate numerically. To use the central definition (18) in practice, two points have to be addressed:
How to choose the embedding ?
How to deal with the fact that we do not know ?
For the choice of , we restrict ourselves to linear maps of the form
| 21 |
with arbitrarily chosen linearly independent functions , where . In practice, we will choose the as linear functions themselves, i.e., for some, usually randomly drawn, . Note that then , but this is not a problem because we will embed the functions , and can be shown to decay exponentially as , cf. Bittracher et al. (2015), Theorem C.1. Thus, will exist. For linearly independent , these maps are still generic in the sense of the Whitney embedding theorem and thus still embed the transition manifold .
A natural choice for the approximation of the unknown map is the mapping to the transition probability density,
| 22 |
as . With this, we consider
| 23 |
The values on the right-hand side can in turn be approximated by a Monte Carlo quadrature, using only short-time trajectories of the original dynamics:
| 24 |
where the are independent realizations of with starting point , in practice realized by a stochastic integrator (e.g., Euler–Maruyama).
The Computationally Infeasible Reaction Coordinate Note that is not yet an r-dimensional reaction coordinate, since is only approximately an r-dimensional manifold; more precisely, it lies in the -neighborhood of an r-dimensional submanifold of . Hence, also is only approximately an r-dimensional manifold; see the magenta regions in Fig. 5.
Fig. 5.
How to make a good r-dimensional reaction coordinate out of ? Given the analysis from the previous section, we would like to make the level sets of also the level sets of (red line segment). Unfortunately, we have no access to these (Color figure online)
The question now is how we can reduce to an r-dimensional good reaction coordinate. Since we know from above that is a good reaction coordinate, let us see what would be needed to construct it.
The property of that we want is that it is constant along level sets of , i.e., (because this implies that it is a good reaction coordinate, cf. Corollary 4.12). Hence, if we could identify as an r-dimensional manifold in , we would project along onto —assuming that and intersect in —to obtain as the resulting point (see Fig. 5, where we would project along the red line on the right). Unfortunately, we have no access to (not to mention that and need not intersect in ) and hence to its level sets . Thus, this strategy seems infeasible.
A Computationally Feasible Reaction Coordinate What helps us at this point is that there is a certain amount of arbitrariness in the definition of . Recalling Definition 4.4, what we are given is , and we construct as a projection of onto the r-dimensional manifold by the closest-point projection ; i.e., . This choice of is convenient, because we can show
| 25 |
which is used in Lemma 4.7. Other choices of could, however, yield a similarly practicable -bound in (25). Our strategy will be to choose a specific r-dimensional reaction coordinate and to show that in general it can be expected to be a good reaction coordinate.
Let us recall that, by assumption, the set is contained in the -neighborhood of an unknown smooth r-dimensional manifold . Thus, a generic smooth map will embed into , forming a diffeomorphism from to . Thus, is going to map to an -neighborhood of . This means the r-dimensional manifold structure of should still be detectable and can be identified with standard manifold learning tools. We use the diffusion maps algorithm (see Sect. 4.4), which gives us a map (the diffusion map). Then we define as
| 26 |
This is depicted on the right-hand side of Fig. 6, where the dashed red line shows the level set .
Fig. 6.
Realized reaction coordinate
Next, we consider the set . It holds that . Recall that is one-to-one; thus, intersects in exactly one point. We define this one point as , and thus, is the projection onto along . We see that is well defined and that .
At this point we assume that is sufficiently well behaved in a neighborhood of and it does not “distort transversality” of intersections such that the diameter of is with a moderate constant in . We will investigate a formal justification of this fact in a future work, here we assume it holds true, and we will see in the numerical experiments that the assumption is justified. This assumption implies that for , i.e., for . Now, however, Lemma 4.7 implies that is almost constant (up to an error ) on level sets of , which, in turn, by Lemma 4.2 and Corollary 3.6 shows that is a good reaction coordinate.
Identification of Through Manifold Learning
In this section, we describe how to identify numerically. The task is as follows: Given that we have computed for a number of sample points , we would like to identify the r-dimensional manifold , noting the points are in a -neighborhood of (see Sect. 4.3). Additionally, we would like an r-dimensional coordinate function that parametrizes (so that the level sets of are transversal to ).
This is a default setting for which manifold learning algorithms can be applied. Many standard methods exist; we name multidimensional scaling (Kruskal 1964b, a), isomap (Tenenbaum et al. 2000), and diffusion maps (Coifman and Lafon 2006) as a few of the most prominent examples. Because of its favorable properties, we choose the diffusion maps algorithm here and summarize it briefly for our setting in what follows. For details, the reader is referred to Coifman and Lafon (2006), Nadler et al. (2006), Coifman et al. (2008) and Singer et al. (2009).
Given sample points , diffusion maps proceed by constructing a similarity matrix with
where is the Euclidean norm in , is a scale factor, and is a kernel function which is most commonly chosen as with a suitably chosen cutoff R that sparsifies W and ensures that only local distances enter the construction. With D being the diagonal matrix containing the row sums of W, the similarity matrix is then normalized to give . Finally, the stochastic matrix is constructed, where is the diagonal matrix containing the row sums of . P is similar to the symmetric matrix ; thus, it has an orthonormal basis of eigenvectors with real eigenvalues . Since P is also stochastic, . The diffusion map is then given by
| 27 |
Using properties of the Laplacian eigenproblem on , one can show that indeed parametrizes the r-dimensional manifold for suitably chosen (Coifman and Lafon 2006).
Remark 4.14
The diffusion maps algorithm will only reliably identify based on the neighborhood relations between the embedded sample points , if the points cover all parts of sufficiently well. In particular, as and thus vary strongly with x traversing the transition regions, a good coverage of those regions is required.
For the various low-dimensional academic examples Sect. 5, this is ensured by choosing the to be a dense grid of points in . For the high-dimensional example in Sect. 5.2, the evaluation points are generated as a subsample from a long equilibrated trajectory, essentially sampling . Both of these ad hoc methods are likely to be unapplicable in realistic high-dimensional systems with very long equilibration times. However, as we mentioned in Introduction, there exist multiple statistical and dynamical approaches to this common problem of quickly sampling the relevant parts of phase space, including the transition regions. Each of these sampling methods can be easily integrated into our proposed algorithm as a preprocessing step.
Fundamentally though, the central idea of our method does not depend crucially on the applicability of diffusion maps. Rather, the latter can be considered an optional post-processing step. Using the -dimensional reaction coordinate
i.e., (26) without the manifold learning step, may in practice already represent a sufficient dimensionality reduction.
In addition, situations may occur where the a priori generation of evaluation points is not possible or desired. One of the final goals and the work currently in progress is the construction of an accelerated integration scheme that generates significant evaluation points and their reaction coordinate value “on the fly.” This is related to the effective dynamics mentioned in fifth point in Conclusion. However, this also requires us to be able to evaluate the reaction coordinate at isolated points, independent of each other, and thus also necessitates the use of the above instead of .
Numerical Examples
Based on the results from the previous sections, we propose the following algorithm to compute reaction coordinates numerically:
Let , , be the points for which we would like to evaluate . Here, we assume the points satisfy the requirements addressed in Remark 4.14.
Choose linearly independent functions , . The essential boundedness of the is not necessary, but should not grow faster than a polynomial as .
In each point , start M simulations of length t and estimate using (23) and (24), to obtain the point . We discuss the appropriate choice of M and t in Sect. 5.1.
Apply the diffusion maps technique from Sect. 4.4 for the point cloud , and obtain , a parametrization of the point in its r essential directions of variation.
- By (27), we define the reaction coordinate as
28
The numerical effort of this algorithm depends strongly on the third step. Given evaluation points, and a choice of M trajectories per point, the cost is mainly given by , where c(t) is the effort of a single numerical realization of the dynamics up to time t. The high-dimensional phase space only enters the algorithm as the domain of the observables . The cost of evaluating those typically very simple functions4 at the end points of the trajectory is negligible. The cost of the method is thus essentially independent of n.
In order to demonstrate the efficacy of our method, we compute the reaction coordinates for three representative problems, namely a simple curved double-well potential, a multi-well potential defined on a circle, both in low and high dimensions, and two slightly different quadruple-well potentials stressing the difference between a one- and a two-dimensional reaction coordinate.
Curved Double-Well Potential
As a first verification, we consider a system with an analytically known reaction coordinate that is then used for comparison. Consider the two-dimensional drift-diffusion process (2) with potential
and inverse temperature . This potential already served as a motivational example for the nature of reaction coordinates in Introduction and is shown in Fig. 1. The system possesses two metastable sets around the minima and , which are connected by the transition path . The implied time scales, defined in (3), can be computed from the eigenvalues using a standard Ulam-type Galerkin discretization (Klus et al. 2016, 2017) of the transfer operator and are shown in Fig. 7a.5 We observe a significant gap between and and thus identify as the last slow and as the first fast time scales. Choosing the lag time then satisfies . A visual inspection of a typical trajectory of length t starting in one of the two metastable sets as shown in Fig. 7b confirms that the respective set is sampled, yet a transition to the other set is a rare event.
Fig. 7.
a Implied time scales of the double-well system. b Trajectory of length
The low dimension of the system allows us to compute the reaction coordinate on a full regular grid over the phase space. We choose a grid in the rectangular region and denote the set of grid points by . For this system, we expect a one-dimensional transition path and thus a one-dimensional reaction coordinate . That is, and . Thus, we choose three linear observables in our embedding function (21), e.g.,
| 29 |
whose coefficients were drawn uniformly from . The expectation value in (23) is approximated by a Monte Carlo quadrature using sample trajectories for each grid point, cf. (24). The parameter M was chosen such that the error in (24), commonly defined as the variance of the Monte Carlo sum, is sufficiently low. The resulting embedding of the grid points x into is shown in Fig. 8. The transition path seems to be already parametrized well by the individual components of .
Fig. 8.
Individual components of the embedding on the grid points
For this example, the image of under should form a compact neighborhood of the one-dimensional manifold , as described in Sect. 4.3. The one-dimensional structure in is clearly visible, see Fig. 9a. To identify the one-dimensional coordinate along this set the diffusion map algorithm is used. Let denote the first diffusion map coordinate on the embedded grid points, also visualized in Fig. 9a. The final reaction coordinate, shown in Fig. 9b, is then given by
Fig. 9.
a Embedded grid points colored according to the first diffusion map coordinate. b The final reaction coordinate
Legoll and Lelièvre (2010) show that the effective dynamics based on the reaction coordinate
accurately reproduces the long-time dynamics of the full process—although they do not use dominant eigenvalues of the transfer operator in their argumentation. It is easy to verify that the level sets of traverse the transition path orthogonally. Figure 10 shows the comparison of the level sets of and . While the two reaction coordinates have different absolute values, their contour lines coincide well. As the projection operator only depends on the level sets of , the projected transfer operators and should be similar as well.
Fig. 10.
Selected contour lines (black) of the newly identified reaction coordinate and the reference reaction coordinate
Projected Eigenvalue Error To conclude this example, we compute the dominant spectrum of the projected transfer operator and compare it to the spectrum of the full transfer operator. To discretize , we use a simple Ulam-type discretization scheme based on a long equilibrated trajectory of the full dynamics. Recall from Sect. that, although formally acts as an operator on functions over , it is constant along level sets of and thus can be treated as an operator on functions over . For completeness, we state the rough outline of an algorithm that we used to approximate . An introduction to Ulam- and other Galerkin-type discretization schemes for transfer operators can be found, e.g., in Klus et al. (2016).
Compute points , a discrete trajectory with step size of the full phase space dynamics that adequately samples the invariant density .
Compute the reaction coordinate on the points .
- Divide the neighborhood of into boxes or other suitable discretization elements and sample the boxes from the trajectory, i.e., compute
- Count the time-t-transitions within between the boxes (where t is a multiple of ), i.e., compute the matrix
After row-normalization, the eigenvalues of approximate the point spectrum of .
Remark 5.1
Note that the equilibrated trajectory is typically unavailable for more complex systems. In practice, one would replace steps 1 and 2 by directly computing a reduced trajectory whose statistics approximate that of . The formulation of a reduced numerical integration scheme to realize this is a work currently in progress (see the fifth point in Conclusion).
For our example system, we compute as a step trajectory with step size using the Euler–Maruyama scheme. However, to reduce the numerical effort, is computed only on a subsample of ( points) and extended to by nearest-neighbor interpolation. On , the image of the is contained in the interval , which we discretize into subintervals of equal length. The spectrum of the full transfer operator was computed using the standard Ulam method over a uniform box discretization of the domain . With the choice for the lag time, the spectral gap is clearly visible.
We observe in Fig. 11 that the eigenvalues of and are in excellent agreement. Not only the dominant eigenvalues are approximated well (as predicted by Lemma 3.5), but also the further subdominant eigenvalues that are not covered by our theory. In particular, the reaction coordinate provides a better approximation to the spectrum of than other, manually chosen reaction coordinates: Fig. 11 also shows the eigenvalues of the projected transfer operator associated with the reaction coordinates
We see that these are consistently outperformed by the computed reaction coordinate (although it appears that already is quite a good reaction coordinate).
Fig. 11.
a Comparison of the two dominant and first four non-dominant eigenvalues of the full transfer operator and the projected transfer operators . b Detailed comparison of the second eigenvalue of the various transfer operators
Circular Potential
Let us now compute the reaction coordinates for the multi-well diffusion process described in Example 4.1. The corresponding k-well potential is defined as
We use , for which the potential is shown in Fig. 2a. The potential as well as the dominant eigenvalues of the corresponding transfer operator clearly indicates the existence of seven metastable sets, yet a typical long-time trajectory, shown in Fig. 12a, suggests a one-dimensional transition path, the unit circle . We demonstrate that with our method, a reaction coordinate of minimal dimension can be computed.
Fig. 12.
a Long-time trajectory of the diffusion process with the circular seven-well potential. b The contour lines of (black) and (red) show that is almost constant on the metastable sets, but resolves the transition regions well (Color figure online)
We again choose the inverse temperature and perform the same analysis as in the previous subsection. For this system, a time scale gap between and can be found. We thus choose the intermediate time scale . Since we again expect a one-dimensional transition path, the three observables (29) are used for the embedding of . We use the grid points of a grid, denoted again by , over the region as our test points.
The individual components of the embedding are shown in Fig. 13. The embedded grid points, shown as the individual points in Fig. 14a, seem to concentrate around a one-dimensional circular manifold and thus reveal the one-dimensional nature of the reaction coordinate. Although slightly unintuitive, the diffusion maps algorithm now identifies two significant diffusion map components, as shown in Fig. 14a. The reason is that the circular manifold cannot be embedded into , so that a two-component coordinate is necessary to parametrize it. Figure 12b shows some contour lines (of equidistant values) of the two components of . We see that is almost constant on the seven metastable sets, but resolves the transition regions well.
Fig. 13.
Individual components of the embedding on the grid points
Fig. 14.
Left column: The embedded grid points . The coloring shows the a first and c second significant diffusion map on the points. Right column: The b first and d second components of the final reaction coordinate
Parametrization of the Dominant Eigenfunctions Next, we experimentally investigate how well the dominant eigenfunctions of can be parametrized by the numerically computed reaction coordinate . If the eigenfunctions are almost functions of , then by Lemma 4.2 and Corollary 3.6 the reaction coordinate is suitable to reproduce all the dominant time scales. To this end, we compute the dominant eigenfunctions , by the Ulam-type Galerkin method (as in the previous example) and plot against . Note that due to the reasons discussed above, the range of is a one-dimensional manifold in . If can be parametrized by , we expect that , whenever . The result is shown in Fig. 15. We clearly see the functional dependency of the first seven (i.e., the dominant) eigenfunctions on the reaction coordinate.
Fig. 15.
Black dots: The values of the first nine eigenfunctions of plotted against , . The blue markers indicate the in the bottom plane. The seven dominant eigenfunctions ( to ) seem to have a smooth dependency on . In contrast, the values of the non-dominant and vary substantially over individual level sets of
Circular Potential in Higher Dimensions The identification of reaction coordinates is not limited to two dimensions. To show that our method can effectively find the reaction coordinates in high-dimensional systems, we extend the 7-well potential to ten dimensions by adding a quadratic term in :
We expect the one-dimensional circle to be the transition path and accordingly choose a three-dimensional linear observable , where the coefficients were again drawn uniformly from .
In ten dimensions, the computation of the reaction coordinate on all points of a regular grid is no longer possible due to the curse of dimensionality, and neither is the visualization of this grid. Instead, we compute on points sampled from the invariant measure and plot only the first three coordinates. Let this point cloud be called .
Performing the standard procedure, i.e., embedding into and identifying the one-dimensional core using diffusion maps, a two-component reaction coordinate is identified. Coloring the first three dimensions of by (Fig. 16a, b), we see that the expected reaction pathway is indeed parametrized. This pathway as well as the seven metastable states can also be recognized in a plot of the components of plotted against each other, indicating that the information about the dominant eigenfunctions, thus the long-time jump process, is indeed retained by .
Fig. 16.
a, b Two components and on the sampling points . The picture shows the first three dimensions of x, but is qualitatively the same when replacing by . c The values of and on plotted against each other
Two Quadruple-Well Potentials
Our theory is based on the existence of an r-dimensional transition manifold in around which the transition probability functions concentrate. In Appendix B, we argue that the existence of an r-dimensional transition path suffices to ensure the existence of . Here we illustrate how the existence of the transition path is reflected in the embedding procedure.
For this we consider the “hilly” and “flat” quadruple-well potentials
and
Both systems possess metastable sets around the four minima , but confines its dynamics outside of the metastable sets onto a one-dimensional transition path, whereas does not impose such restrictions on the dynamics (see Fig. 17). For both potentials the time lies inside the slow–fast time scale gap. Assuming a one-dimensional transition manifold (wrongfully for ), we use the three linear observables (29). A grid on is used as evaluation points for . The embedding of these points by is shown in Fig. 18. We observe a one-dimensional structure in the case of the “hilly” potential , whereas the embedding points of the “flat” potential lie on a seemingly two-dimensional manifold. As these embeddings are approximately one-to-one with the respective transition manifolds , we conclude that in the case of the manifold must be one-dimensional, whereas for it is two-dimensional.
Fig. 17.
Two quad-well potentials and possess qualitatively different transition regions
Fig. 18.
Embedding of the grid points for the a “hilly” and b “flat” four-well potential. A one-dimensional structure is only visible in a, i.e., in the presence of a one-dimensional transition path
Conclusion
Our main contributions in this paper are as follows:
We developed a mathematical framework to characterize good reaction coordinates for stochastic dynamical systems showing metastable behavior, but no local separation of fast and slow time scales.
We showed the existence of good low-dimensional reaction coordinates under certain dynamical assumptions on the system.
We proposed an algorithmic approach to numerically identify good reaction coordinates and the associated low-dimensional transition manifold based on local evaluation of short trajectories of the system only.
Our numerical examples show how the procedure works, that it can be used in higher dimensions, and the examples give further evidence that the dynamical assumptions from (b) are valid in many realistic cases. The application of our approach to relevant biomolecular problems, e.g., in protein folding, is ongoing work.
Apart from the application to actual molecular systems, there are several open questions and challenges, which we will address in the future:
A rigorous mathematical justification for the dynamical assumption in Definition 4.4 in terms of the potential V and the noise intensity in (2) would be desirable. This seems to be a demanding task, as the interplay between potential landscape and the thermal forcing is non-trivial. For the problem can be handled by large deviation approaches; however, understanding increasing is challenging: The strength of noise increases, and additional transitions between metastable sets become more probable, as the barriers in the potential landscape become less significant, and thus the reaction coordinate may increase in dimension.
Also related to the previous point, the choice of the correct lag time t is crucial. Choosing the time too small, the concentration of the transition densities near a low-dimensional manifold in may not have happened yet, but a too large lag time has severe consequences for the numerical expenses. If no expert knowledge of a proper lag time t is available, it has to be identified in a preprocessing step, for example using Markov state model techniques (Bowman et al. 2014).
As discussed in the last part of Sect. 4.3 and in Fig. 6, we need the embedding not to distort transversality close to the transition manifold too much, such that the realized reaction coordinate is indeed a good one. Theoretical bounds shall be developed. This problem seems to be coupled with the problem of how to control the condition number of the embedding and its numerical realization.
The dimension r of the reaction coordinate may not be known in advance; hence, we need an algorithmic strategy to identify this on the fly. Fortunately, once the sampling has been made, the evaluation of the embedding mapping , and finding intrinsic coordinates on the set of data points embedded in , has a negligible numerical effort; hence, different embedding dimensions k can be probed via (21). Theorem 4.10 suggests that if the identified dimension of the reaction coordinate is smaller than k / 2, then a reaction coordinate of sufficient dimension is found.
To benefit from the dimensionality reduction of the reaction coordinate , the dynamics that generates the reduced transfer operator has to be described in closed form. We are planning to employ techniques based on the Kramers–Moyal extension (Zhang et al. 2016) to again receive an SDE for a stochastic process on .
The embedding mapping is evaluated by Monte Carlo quadrature (24). Although Monte Carlo quadrature is known to have a convergence rate independent of the underlying dimension n of , there is still an impact of the dimension on the practical accuracy. This we shall investigate as well.
Acknowledgements
This research has been partially funded by Deutsche Forschungsgemeinschaft (DFG) through grant CRC 1114 “Scaling Cascades in Complex Systems,” Project B03 “Multilevel coarse graining of multi-scale problems” and the Einstein Foundation Berlin (Einstein Center ECMath).
A Properties of
Proof of Proposition 3.4
(a) This property has been shown by Zhang et al. (2016) as well, and we include the short reasoning for completeness. The linearity of is obvious. The property follows from (8) by noting that is constant on and that is a probability measure for every z.
(b) From (10) we have for that
| 30 |
where follows from the linearity of and the fact that ; thus, . Expression (30) is symmetric in f and g; hence, it follows that .
(c) We first prove that is an orthogonal projection:
Thus,
and the claim follows.
B On the Existence of Reaction Coordinates
To motivate the existence of low-dimensional reaction coordinates, let us assume that the dynamics of consideration has metastable regions . Let . For a selected lag time we make the following two assumptions:
- Fast local equilibration: If x is in (or close to) then we have
where is the quasi-stationary density of the metastable core :
with being the (random) exit time from the set . Slow transitions: The typical transition time to reach when starting in is larger than t. In other words, t is such that if the process transitions from x to some , it did not transition through some other with high probability.
These two assumptions essentially say that t is much larger than the fast time scales of the system, but smaller than the dominant time scales. It follows that, for any ,
where by assumption 2) the coefficients are given by the committor functions
We say that is an r-dimensional structure in if there is a function that jointly parametrizes all the committor functions, i.e., with . If this is the case, then
and clearly since . Moreover, since we can explicitly construct as . This obviously parametrizes , and it also parametrizes since .
However, r may also be smaller than d. As an example, consider the potential with 4 minima as shown in Fig. 19 on the left. At low temperatures, the “hilly” potential energy landscape confines all transitions between the minima to a narrow region close to the red square connecting the four minima. Figure 19 shows the level sets of , the level sets of the other committors are given by the rotational symmetry of the problem. All four committors can be jointly parametrized by a single coordinate which describes clockwise movement along the red square and is constant orthogonal to it. Therefore, . Figure 19 on the right shows the situation with a “flat” energy landscape. Transition paths between the minima are no longer confined to a one-dimensional structure, and the committor level sets are more complicated. We can no longer parametrize all four committors with a single coordinate , so . On the other hand, so .
Fig. 19.
A potential energy landscape with four minima (black contours) and level sets of (colored contours). Left: The “hilly” landscape structure confines transition pathways to a narrow region close to the red square connecting the four minima. As a result, all committor level sets are orthogonal to this main transition path. Right: “Flat” landscape structure with more complicated level sets of the committors (Color figure online)
This structural difference of the potentials can also be seen when applying our algorithm to construct the reaction coordinate , see Fig. 18 and Sect. 5.3.
Footnotes
We denote by the space (equivalence class) of q-integrable functions with respect to the Lebesgue measure. denotes the same space of function, now integrable with respect to the measure .
The coarea formula holds for functions, but , since is a probability measure. (That is, it is finite.)
If it is necessary to break ties in (15), we can do so by taking any of the minimizers. The mapping can be shown to be smooth (Bittracher et al. 2015, Theorem C.1); hence, is a smooth manifold satisfying (14).
In our examples, we used linear functions with great success.
In realistic, high-dimensional systems, the computation of the dominant eigenvalues using grid-based methods is likely infeasible. In these situations, the implied time scales have to be estimated, for example using standard Markov state model techniques (Bowman et al. 2014).
References
- Baxter JR, Rosenthal JS. Rates of convergence for everywhere-positive Markov chains. Stat. Probab. Lett. 1995;22(4):333–338. doi: 10.1016/0167-7152(94)00085-M. [DOI] [Google Scholar]
- Becker NB, Allen RJ, ten Wolde PR. Non-stationary forward flux sampling. J. Chem. Phys. 2012;136(17):174118. doi: 10.1063/1.4704810. [DOI] [PubMed] [Google Scholar]
- Best RB, Hummer G. Coordinate-dependent diffusion in protein folding. Proc. Natl. Acad. Sci. 2010;107(3):1088–1093. doi: 10.1073/pnas.0910390107. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Bittracher A, Koltai P, Junge O. Pseudogenerators of spatial transfer operators. SIAM J. Appl. Dyn. Syst. 2015;14(3):1478–1517. doi: 10.1137/14099872X. [DOI] [Google Scholar]
- Bovier A, Gayrard V, Klein M. Metastability in reversible diffusion processes II. Precise asymptotics for small eigenvalues. J. Eur. Math. Soc. 2002;7:69–99. [Google Scholar]
- Bowman, G.R., Pande, V.S., Noé, F. (eds): An Introduction to Markov State Models and Their Application to Long Timescale Molecular Simulation, volume 797 of Advances in Experimental Medicine and Biology. Springer, Berlin (2014)
- Camacho CJ, Thirumalai D. Kinetics and thermodynamics of folding in model proteins. Proc. Natl. Acad. Sci. 1993;90(13):6369–6372. doi: 10.1073/pnas.90.13.6369. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Chiavazzo, E., Coifman, R.R., Covino, R., Gear, C.W., Georgiou, A.S., Hummer, G., Kevrekidis, I.G.: iMapD: intrinsic map dynamics exploration for uncharted effective free energy landscapes. arXiv preprint arXiv:1701.01513 (2016) [DOI] [PMC free article] [PubMed]
- Chodera JD, Noé F. Markov state models of biomolecular conformational dynamics. Curr. Opin. Struct. Biol. 2014;25:135–144. doi: 10.1016/j.sbi.2014.04.002. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Ciccotti G, Kapral R, Vanden-Eijnden E. Blue moon sampling, vectorial reaction coordinates, and unbiased constrained dynamics. ChemPhysChem. 2005;6(9):1809–1814. doi: 10.1002/cphc.200400669. [DOI] [PubMed] [Google Scholar]
- Coifman RR, Lafon S. Diffusion maps. Appl. Comput. Harmon. Anal. 2006;21(1):5–30. doi: 10.1016/j.acha.2006.04.006. [DOI] [Google Scholar]
- Coifman RR, Kevrekidis IG, Lafon S, Maggioni M, Nadler B. Diffusion maps, reduction coordinates, and low dimensional representation of stochastic systems. Multiscale Model. Simul. 2008;7(2):842–864. doi: 10.1137/070696325. [DOI] [Google Scholar]
- Crosskey M, Maggioni M. ATLAS: a geometric approach to learning high-dimensional stochastic systems near manifolds. Multiscale Model. Simul. 2017;15(1):110–156. doi: 10.1137/140970951. [DOI] [Google Scholar]
- Darve E, Rodríguez-Gómez D, Pohorille A. Adaptive biasing force method for scalar and vector free energy calculations. J. Chem. Phys. 2008;128(14):144120. doi: 10.1063/1.2829861. [DOI] [PubMed] [Google Scholar]
- Dellago C, Bolhuis PG. Transition path sampling and other advanced simulation techniques for rare events. In: Holm C, Kremer K, editors. Advanced Computer Simulation Approaches for Soft Matter Sciences III. Berlin: Springer; 2009. pp. 167–233. [Google Scholar]
- Dellnitz M, Hohmann A. A subdivision algorithm for the computation of unstable manifolds and global attractors. Numer. Math. 1997;75(3):293–317. doi: 10.1007/s002110050240. [DOI] [Google Scholar]
- Dellnitz M, Junge O. On the approximation of complicated dynamical behavior. SIAM J. Numer. Anal. 1999;36(2):491–515. doi: 10.1137/S0036142996313002. [DOI] [Google Scholar]
- Dellnitz M, von Molo MH, Ziessler A. On the computation of attractors for delay differential equations. J. Comput. Dyn. 2016;3(1):93–112. [Google Scholar]
- Djurdjevac N, Sarich M, Schütte C. Estimating the eigenvalue error of Markov state models. Multiscale Model. Simul. 2012;10(1):61–81. doi: 10.1137/100798910. [DOI] [Google Scholar]
- Dsilva CJ, Talmon R, Gear CW, Coifman RR, Kevrekidis IG. Data-driven reduction for a class of multiscale fast–slow stochastic dynamical systems. SIAM J. Appl. Dyn. Syst. 2016;15(3):1327–1351. doi: 10.1137/151004896. [DOI] [Google Scholar]
- Du R, Pande VS, Grosberg AY, Tanaka T, Shakhnovich ES. On the transition coordinate for protein folding. J. Chem. Phys. 1998;108(1):334–350. doi: 10.1063/1.475393. [DOI] [Google Scholar]
- E, W., Engquist, B.: The heterogenous multiscale method. Commun. Math. Sci. 1(1), 87–132 (2003)
- E, W., Vanden-Eijnden, E.: Towards a theory of transition paths. J. Stat. Phys 123(3), 503–523 (2006)
- E, W., Vanden-Eijnden, E.: Transition-path theory and path-finding algorithms for the study of rare events. Annu. Rev. Phys. Chem 61(1), 391–420 (2010) [DOI] [PubMed]
- E, W., Ren, W., Vanden-Eijnden, E.: String method for the study of rare events. Phys. Rev. B 66(5), 052301 (2002) [DOI] [PubMed]
- Faradjian AK, Elber R. Computing time scales from reaction coordinates by milestoning. J. Chem. Phys. 2004;120:10880–10889. doi: 10.1063/1.1738640. [DOI] [PubMed] [Google Scholar]
- Federer H. Geometric Measure Theory. New York: Springer; 1969. [Google Scholar]
- Freidlin M, Wentzell AD. Random Perturbations of Dynamical Systems. New York: Springer; 1998. [Google Scholar]
- Froyland G, Gottwald G, Hammerlindl A. A computational method to extract macroscopic variables and their dynamics in multiscale systems. SIAM J. Appl. Dyn. Syst. 2014;13(4):1816–1846. doi: 10.1137/130943637. [DOI] [Google Scholar]
- Froyland G, Gottwald GA, Hammerlindl A. A trajectory-free framework for analysing multiscale systems. Phys. D Nonlinear Phenom. 2016;328:34–43. doi: 10.1016/j.physd.2016.04.010. [DOI] [Google Scholar]
- Huisinga W, Meyn S, Schütte C. Phase transitions and metastability in Markovian and molecular systems. Ann. Appl. Probab. 2004;14(1):419–458. doi: 10.1214/aoap/1075828057. [DOI] [Google Scholar]
- Hunt B, Kaloshin V. Regularity of embeddings of infinite-dimensional fractal sets into finite-dimensional spaces. Nonlinearity. 1999;12(5):1263–1275. doi: 10.1088/0951-7715/12/5/303. [DOI] [Google Scholar]
- Junge O, Koltai P. Discretization of the Frobenius–Perron operator using a sparse Haar tensor basis: the sparse Ulam method. SIAM J. Numer. Anal. 2009;47(5):3464–3485. doi: 10.1137/080716864. [DOI] [Google Scholar]
- Kevrekidis IG, Samaey G. Equation-free multiscale computation: algorithms and applications. Annu. Rev. Phys. Chem. 2009;60(1):321–344. doi: 10.1146/annurev.physchem.59.032607.093610. [DOI] [PubMed] [Google Scholar]
- Klus S, Koltai P, Schütte C. On the numerical approximation of the Perron–Frobenius and Koopman operator. J. Comput. Dyn. 2016;3(1):51–79. [Google Scholar]
- Klus, S., Nüske, F., Koltai, P., Wu, H., Kevrekidis, I., Schütte, C., Noé, F.: Data-driven model reduction and transfer operator approximation. ArXiv e-prints (2017)
- Kruskal JB. Multidimensional scaling by optimizing goodness of fit to a nonmetric hypothesis. Psychometrika. 1964;29(1):1–27. doi: 10.1007/BF02289565. [DOI] [Google Scholar]
- Kruskal JB. Nonmetric multidimensional scaling: a numerical method. Psychometrika. 1964;29(2):115–129. doi: 10.1007/BF02289694. [DOI] [Google Scholar]
- Kumar S, Rosenberg JM, Bouzida D, Swendsen RH, Kollman PA. The weighted histogram analysis method for free-energy calculations on biomolecules. I. The method. J. Comput. Chem. 1992;13(8):1011–1021. doi: 10.1002/jcc.540130812. [DOI] [Google Scholar]
- Laio A, Gervasio FL. Metadynamics: a method to simulate rare events and reconstruct the free energy in biophysics, chemistry and material science. Rep. Prog. Phys. 2008;71(12):126601. doi: 10.1088/0034-4885/71/12/126601. [DOI] [Google Scholar]
- Laio A, Parrinello M. Escaping free-energy minima. Proc. Natl. Acad. Sci. 2002;99(20):12562–12566. doi: 10.1073/pnas.202427399. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Legoll F, Lelièvre T. Effective dynamics using conditional expectations. Nonlinearity. 2010;23(9):2131. doi: 10.1088/0951-7715/23/9/006. [DOI] [Google Scholar]
- Li W, Ma A. Recent developments in methods for identifying reaction coordinates. Mol. Simul. 2014;40(10–11):784–793. doi: 10.1080/08927022.2014.907898. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Lu J, Vanden-Eijnden E. Exact dynamical coarse-graining without time-scale separation. J. Chem. Phys. 2014;141(4):07B619_1. doi: 10.1063/1.4890367. [DOI] [PubMed] [Google Scholar]
- Ma A, Dinner AR. Automatic method for identifying reaction coordinates in complex systems. J. Phys. Chem. B. 2005;109:6769–6779. doi: 10.1021/jp045546c. [DOI] [PubMed] [Google Scholar]
- Maragliano L, Vanden-Eijnden E. A temperature accelerated method for sampling free energy and determining reaction pathways in rare events simulations. Chem. Phys. Lett. 2006;426(1):168–175. doi: 10.1016/j.cplett.2006.05.062. [DOI] [Google Scholar]
- Mattingly JC, Stuart AM. Geometric ergodicity of some hypo-elliptic diffusions for particle motions. Markov Process. Relat. Fields. 2002;8(2):199–214. [Google Scholar]
- Mattingly JC, Stuart AM, Higham DJ. Ergodicity for SDEs and approximations: locally Lipschitz vector fields and degenerate noise. Stoch. Process. Appl. 2002;101(2):185–232. doi: 10.1016/S0304-4149(02)00150-3. [DOI] [Google Scholar]
- McGibbon RT, Husic BE, Pande VS. Identification of simple reaction coordinates from complex dynamics. J. Chem. Phys. 2017;146(4):044109. doi: 10.1063/1.4974306. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Metzner P, Schütte C, Vanden-Eijnden E. Transition path theory for Markov jump processes. Multiscale Model. Simul. 2009;7(3):1192–1219. doi: 10.1137/070699500. [DOI] [Google Scholar]
- Moroni D, van Erp TS, Bolhuis PG. Investigating rare events by transition interface sampling. Phys. A Stat. Mech. Appl. 2004;340(1):395–401. doi: 10.1016/j.physa.2004.04.033. [DOI] [Google Scholar]
- Nadler B, Lafon S, Coifman RR, Kevrekidis IG. Diffusion maps, spectral clustering and reaction coordinates of dynamical systems. Appl. Comput. Harmon. Anal. 2006;21(1):113–127. doi: 10.1016/j.acha.2005.07.004. [DOI] [Google Scholar]
- Noé F, Nüske F. A variational approach to modeling slow processes in stochastic dynamical systems. Multiscale Model. Simul. 2013;11(2):635–655. doi: 10.1137/110858616. [DOI] [Google Scholar]
- Noé F, Schütte C, Vanden-Eijnden E, Reich L, Weikl TR. Constructing the full ensemble of folding pathways from short off-equilibrium simulations. Proc. Natl. Acad. Sci. 2009;106:19011–19016. doi: 10.1073/pnas.0905466106. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Pande VS, Beauchamp K, Bowman GR. Everything you wanted to know about Markov state models but were afraid to ask. Methods. 2010;52(1):99–105. doi: 10.1016/j.ymeth.2010.06.002. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Pavliotis G, Stuart A. Multiscale Methods: Averaging and Homogenization. Berlin: Springer; 2008. [Google Scholar]
- Pérez-Hernández G, Paul F, Giorgino T, De Fabritiis G, Noé F. Identification of slow molecular order parameters for Markov model construction. J. Chem. Phys. 2013;139(1):015102. doi: 10.1063/1.4811489. [DOI] [PubMed] [Google Scholar]
- Pozun ZD, Hansen K, Sheppard D, Rupp M, Mller K-R, Henkelman G. Optimizing transition states via kernel-based machine learning. J. Chem. Phys. 2012;136(17):174101. doi: 10.1063/1.4707167. [DOI] [PubMed] [Google Scholar]
- Ren W, Vanden-Eijnden E, Maragakis P, E W. Transition pathways in complex systems: application of the finite-temperature string method to the alanine dipeptide. J. Chem. Phys. 2005;123(13):134109. doi: 10.1063/1.2013256. [DOI] [PubMed] [Google Scholar]
- Robinson JC. A topological delay embedding theorem for infinite-dimensional dynamical systems. Nonlinearity. 2005;18(5):2135–2143. doi: 10.1088/0951-7715/18/5/013. [DOI] [Google Scholar]
- Sarich M, Noé F, Schütte C. On the approximation quality of Markov state models. Multiscale Model. Simul. 2010;8(4):1154–1177. doi: 10.1137/090764049. [DOI] [Google Scholar]
- Sauer T, Yorke JA, Casdagli M. Embedology. J. Stat. Phys. 1991;65(3):579–616. doi: 10.1007/BF01053745. [DOI] [Google Scholar]
- Schervish MJ, Carlin BP. On the convergence of successive substitution sampling. J. Comput. Graph. Stat. 1992;1(2):111–127. [Google Scholar]
- Schütte, C., Sarich, M.: Metastability and Markov State Models in Molecular Dynamics: Modeling, Analysis. Algorithmic Approaches. Courant Lecture Notes in Mathematics, vol. 24. American Mathematical Society (2013)
- Schütte C, Fischer A, Huisinga W, Deuflhard P. A direct approach to conformational dynamics based on hybrid Monte Carlo. J. Comput. Phys. 1999;151(1):146–168. doi: 10.1006/jcph.1999.6231. [DOI] [Google Scholar]
- Schütte, C., Noé, F., Lu, J., Sarich, M., Vanden-Eijnden, E.: Markov state models based on milestoning. J. Chem. Phys. 134(20), 204105 (2011). doi:10.1063/1.3590108 [DOI] [PubMed]
- Singer A, Erban R, Kevrekidis IG, Coifman RR. Detecting intrinsic slow variables in stochastic dynamical systems by anisotropic diffusion maps. Proc. Natl. Acad. Sci. 2009;106(38):16090–160955. doi: 10.1073/pnas.0905547106. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Socci N, Onuchic JN, Wolynes PG. Diffusive dynamics of the reaction coordinate for protein folding funnels. J. Chem. Phys. 1996;104(15):5860–5868. doi: 10.1063/1.471317. [DOI] [Google Scholar]
- Takens, F.: Detecting strange attractors in turbulence. In: Rand, D., Young, L.-S. (eds.) Springer Lecture Notes in Mathematics, vol. 898, pp. 366–381. Springer, Berlin (1981)
- Tenenbaum JB, Silva VD, Langford JC. A global geometric framework for nonlinear dimensionality reduction. Science. 2000;290(5500):2319–2323. doi: 10.1126/science.290.5500.2319. [DOI] [PubMed] [Google Scholar]
- Torrie GM, Valleau JP. Nonphysical sampling distributions in Monte Carlo free-energy estimation: umbrella sampling. J. Comput. Phys. 1977;23(2):187–199. doi: 10.1016/0021-9991(77)90121-8. [DOI] [Google Scholar]
- Trefethen LN, Embree M. Spectra and Pseudospectra: The Behavior of Nonnormal Matrices and Operators. Princeton: Princeton University Press; 2005. [Google Scholar]
- Vanden-Eijnden, E.: Transition path theory. In: Ferrario, M., Ciccotti, G., Binder, K. (eds.) Computer Simulations in Condensed Matter Systems: From Materials to Chemical Biology, vol. 1, pp. 453–493 (2006)
- Vanden-Eijnden E. On HMM-like integrators and projective integration methods for systems with multiple time scales. Commun. Math. Sci. 2007;5(2):495–505. doi: 10.4310/CMS.2007.v5.n2.a14. [DOI] [Google Scholar]
- Weber, M.: Meshless methods in conformation dynamics. Ph.D. thesis, FU Berlin (2006)
- Weber, M.: A subspace approach to molecular Markov state models via a new infinitesimal generator. Habilitation thesis (2012)
- Weber M, Fackeldey K, Schütte C. Set-free Markov state model building. J. Chem. Phys. 2017;146:124133. doi: 10.1063/1.4978501. [DOI] [PubMed] [Google Scholar]
- Whitney H. Differentiable manifolds. Ann. Math. 1936;37(3):645–680. doi: 10.2307/1968482. [DOI] [Google Scholar]
- Zhang W, Schuette C. Reliable approximation of long relaxation timescales in molecular dynamics. Entropy. 2017;19:367. doi: 10.3390/e19070367. [DOI] [Google Scholar]
- Zhang W, Hartmann C, Schütte C. Effective dynamics along given reaction coordinates, and reaction rate theory. Faraday Discuss. 2016;195:365–394. doi: 10.1039/C6FD00147E. [DOI] [PubMed] [Google Scholar]



















