Skip to main content
Chaos logoLink to Chaos
. 2023 Dec 4;33(12):123108. doi: 10.1063/5.0178947

Locating saddle points using gradient extremals on manifolds adaptively revealed as point clouds

A Georgiou 1, H Vandecasteele 2, J M Bello-Rivas 3, I Kevrekidis 1,3,1,3,a)
PMCID: PMC10697725  PMID: 38048255

Abstract

Steady states are invaluable in the study of dynamical systems. High-dimensional dynamical systems, due to separation of time scales, often evolve toward a lower dimensional manifold M. We introduce an approach to locate saddle points (and other fixed points) that utilizes gradient extremals on such a priori unknown (Riemannian) manifolds, defined by adaptively sampled point clouds, with local coordinates discovered on-the-fly through manifold learning. The technique, which efficiently biases the dynamical system along a curve (as opposed to exhaustively exploring the state space), requires knowledge of a single minimum and the ability to sample around an arbitrary point. We demonstrate the effectiveness of the technique on the Müller–Brown potential mapped onto an unknown surface (namely, a sphere). Previous work employed a similar algorithmic framework to find saddle points using Newton trajectories and gentlest ascent dynamics; we, therefore, also offer a brief comparison with these methods.


The energy landscapes for molecular dynamics, protein folding, glassy materials, etc., may exhibit large numbers of minima and transition states. Locating these critical points—especially in high dimensions—is challenging. Current approaches (e.g., the nudged elastic band and the string method) start with two minima and a priori knowledge of collective variables, which represent the system well in a lower dimensional setting; other methods attempt to exhaustively explore the state space (metadynamics, adaptive biasing force, iMapD). In this paper, we present a technique to systematically bias dynamical systems from a single minimum to a saddle point (transition state) without a priori knowledge of good collective variables.

I. INTRODUCTION

Locating saddle points of dynamical systems is of primary importance to a variety of applications, including the location of transition states of chemical systems described at the atomic level1 or the study of transitions between molecular configurations in molecular dynamics.2 Unfortunately, finding these transition states simply via simulation is time-prohibitive: the trajectories spend the majority of their time in a basin of attraction around a stable equilibrium, rarely traveling to a different basin of attraction (potential well). Often, trajectories of the high-dimensional system quickly evolve to, and then move slowly along, a low-dimensional manifold (for example, due to a time-scale separation). The coarse (collective) variables that parameterize the slow manifold are either known a priori or can be determined through manifold learning techniques.

Many techniques exist to locate saddle points of these systems. The metadynamics3 and adaptive biasing force4 methods attempt to exhaustively explore the whole space; yet, they require a priori knowledge of collective variables; iMapD5 alleviates this need by exploiting manifold learning techniques. Another class of methods, namely, a nudged elastic band or string method variants,6–8 trace a curve between two known minima: specifically, they minimize the energy at each point along a path connecting two minima (reactant and product) while attempting to maintain equidistant spacing between the points. These methods require a global set of collective variables. A third class of methods require only a single minimum, utilizing local information to follow a continuous curve to a nearby fixed point (hopefully a saddle). These one-dimensional curves include Newton trajectories,9,10 gentlest ascent dynamics11 or dimer method12 trajectories, and gradient extremals,13 which we study here. These methods still require a global set of collective variables and were originally formulated for Euclidean space.

Gradient extremals have been extended to smooth Riemannian manifolds to locate fixed points.14,15 In this work, we forego a priori knowledge of good collective variables as well as an explicit formulation of the manifold. Instead, we follow gradient extremals on initially unexplored Riemannian manifolds gradually revealed by adaptively sampled point clouds. We utilize local sets of collective variables that are discovered on-the-fly through manifold learning (here, diffusion maps16). By sampling along the manifold, discovering a good local chart, resolving the path locally on the chart, and repeating this process at the newly found location, the system is driven from an initial conformation (typically a stable equilibrium) to a saddle point. Our methodology successfully locates new fixed points using a single initial point and without need for a priori knowledge of collective variables. The method does not exhaustively explore the state space, but rather follows a single path at a time. In prior work, we utilized a similar algorithmic approach, but with the resolved paths being those of Newton trajectories17 and gentlest ascent dynamics.18

II. GRADIENT EXTREMALS

Gradient extremals are a type of curve that connects fixed points of a dynamical system. We will assume, because of our motivating examples, that the dynamical system is a gradient system; i.e., the vector field X(x)=DE(x) for some twice-differentiable potential E. This assumption can be relaxed to arbitrary vector fields by working with the squared magnitude of the vector field.19 An example of this can be found toward the end of Sec. II A. Gradient extremals can be thought of as following the “valleys, ridges, cirques, or cliffs” of the potential E.20

A. Gradient extremals in Euclidean space

Gradient extremal curves13,19–21 are defined as the locus of points that minimize (or maximize) the gradient norm along each level set of the potential E,

argminxRn(max){DE(x)2:E(x)=L}. (1)

The points on the level set highlighted in Fig. 1 that minimize (respectively, maximize) the gradient are the farthest (respectively, closest) from the fixed point. Equivalently, one may formulate gradient extremals as the locus of points where the gradient of the potential is an eigenvector of the Hessian D2E(x),

D2E(x)DE(x)=λDE(x), (2)

along each level set E(x)=L. Here, λ is the associated eigenvalue. One can verify that both methods are equivalent by applying the method of Lagrange multipliers on (1). Gradient extremals can be followed through continuation methods22 where the level set value, L, acts as the parameter of interest and (1) is optimized for each L.23 Note that a specific gradient extremal curve may not traverse all fixed points—it might instead connect just a subset of them.24 See Fig. 2 for an example of gradient extremals on the Müller–Brown potential [defined in (8)]. We also note that a gradient extremal can exhibit turning points (or “meander”), as seen in Fig. 3, which depicts a gradient extremal on the potential defined by

E(x1,x2)=i=14Aiexp(ai(xi1x0i1)2+bi(xi2x0i2)2+ci(xi1xi2)) (3)

with the coefficients listed in Table I. The superscripts here, and throughout the paper, refer to the component indices (Einstein notation) unless otherwise noted.

FIG. 1.

FIG. 1.

(a) A harmonic potential E intersected by a plane at E=L. The elliptical contour line is colored by the length of the corresponding force, with red being a maximum and white a minimum. The points of the gradient extremal corresponding to the min and max of DE at E=L are marked with a ×. (b) A 2D schematic of (a).

FIG. 2.

FIG. 2.

Gradient extremals (following the eigenvector corresponding to the smallest eigenvalue of the Hessian of the potential) on the Müller–Brown potential, which is defined in (8). Minima are depicted by yellow circles and saddle points by squares. Note that the top minimum is not directly connected to the bottom two. They are instead connected through a long detour25 that extends beyond the boundaries of the figure.

FIG. 3.

FIG. 3.

The gradient extremal following the eigenvector corresponding to the smallest eigenvalue of the Hessian of (3). Note how the gradient extremal “meanders” or exhibits turning points. (a) A 2D depiction of the potential energy surface. (b) A 3D depiction.

TABLE I.

Coefficients of the potential defined by (3).

i A i a i b i c i x0i1 x0i2
1 10 13000 11000 0 0 0
2 −0.4 1300 130 1200 15 10
3 0.8 11500 11500 12000 25 100
4 6 13000 13000 320000 40 30

We previously assumed that the dynamical system is of a gradient type. However, this assumption can be removed by replacing the potential by the squared magnitude of any given vector field X.19 The gradient extremal then becomes the locus of points that solve the constrained optimization problem,

argminxRn(max){D(XTX)2:XTX=L}. (4)

Notice that E in (1) is replaced by the squared magnitude XTX. We illustrate this technique with the vector field describing a first order, exothermic, irreversible reaction AB taking place in a continuous-stirred tank reactor (CSTR). The governing mass and energy balances are

dx1dt=x1+Da(1x1)exp(x2),dx2dt=x2+BDa(1x1)exp(x2)β(x2xc2), (5)

where x1 is the conversion of species A and x2 is the dimensionless temperature as detailed in Ref. 26. (Please remember that the superscripts in x1 and x2 are indices, not powers, in the Einstein notation.) With parameters set to Da=0.085, B=22, β=3, and xc2=0.04, the dynamical system exhibits two unstable steady states: one stable steady state and one stable limit cycle. Figure 4 illustrates gradient extremals on the squared magnitude of the vector field (5).

FIG. 4.

FIG. 4.

Gradient extremal (orange curve) on the graph of the squared magnitude, XTX, for the CSTR. The steady states are shown in yellow.

B. Gradient extremals on Riemannian manifolds

We now turn our attention to resolving gradient extremals for vector fields defined on a manifold. Recall that the purpose of the proposed algorithm is to connect fixed points (hopefully a local minimum to a kinetically relevant nearby saddle): if one could sample and map a chart (defined as a small portion of the manifold represented with reduced coordinates) that includes both minimum and saddle point, one could directly solve for a gradient extremal by treating the chart as Euclidean space. The resulting path would not be a true gradient extremal but would still properly link the two critical points, which was our original intent. However, chances are that multiple local charts will be needed, whether because (a) the curvature of the manifold makes the projection onto a single chart impossible or (b) a single chart does not contain both the initial and final fixed points. In the former case, increased curvature makes an accurate, invertible mapping between the manifold and local chart difficult: choosing multiple smaller areas to work with alleviates this issue. In the latter case, unless both connected fixed points lie on the chart, a gradient extremal would terminate at the boundary of the chart without connecting to the next fixed point of interest.

To ensure the path transitions properly from one chart to the next, we use differential geometry to resolve the gradient extremal in the Riemannian setting of a d-dimensional smooth manifold, MRn, with Riemannian metric g. The Riemannian metric g defines the notions of distances and angles on the manifold by assigning inner products on the tangent spaces of the manifold. Let E:MR be the potential energy function on the manifold. We create a mapping φ:UMRd that defines a system of coordinates of an open subset UM on a local chart in Rd. Let ψ:RdU, uψ(u) be the inverse map of the chart φ. We define a Hessian matrix in terms of the covariant derivative on (M,g) with respect to the Levi–Civita connection. A connection describes how a vector changes as it moves from one point on a manifold to the next. We use the following notation when working on (M,g), with Z=ψE:RdR now representing the pullback energy:

Z=ψE=E°ψ.

Thus, the Euclidean formulation of gradient extremals (2) now becomes, on a local chart,

Hess(Z)grad(Z)=λgrad(Z). (6)

First, we define our Riemannian metric, g, as

g=i,j=1dgijdφidφj, (7)

where gij=(DψDψ)ij for i,j=1,,d. This metric allows us to define an inner product between two tangent vectors S and T defined in local coordinates: g(S,T)=i,j=1dgijSiTi. We can now compute the gradient and Hessian on the manifold. Each component of the gradient becomes

grad(Z)i=j=1dgijZφj,

where gij refers to element (i,j) of the inverse metric tensor. The Hessian is defined in terms of the covariant derivative,

Hess(Z)T=Tgrad(Z),

where the covariant derivative, which allows us to specify the parallel transport along tangent vectors of a manifold, is described as

TY=k=1d{i=1d(Ykφi+j=1dΓijkYj)Ti}φk.

The above expression constitutes the covariant derivative of Y in the direction of T. The coefficients Γijk act as a correction to otherwise straight trajectories on the tangent plane so that they remain on the manifold. These coefficients are unique to a given affine connection. Here, we use the Levi–Civita connection, the unique symmetric connection compatible with the metric, for which the coefficients, known as Christoffel symbols, are given by

Γjk=12i=1dgi(gijφk+gikφjgjkφi).

Beginning with an initial point p0, its energy level L=E(p0), and the smallest (or largest) eigenvalue, λ, of the Hessian of the pullback energy evaluated at the initial point, we can now use continuation to solve the following system of d+1 equations:

{ZL=0,Hess(Z)grad(Z)λgrad(Z)=0.

Note that naïvely using the Euclidean formulation of gradient extremals as defined in (2) on the local chart would be erroneous because it fails to take into account the geometry of the manifold. Indeed, we would obtain curves that would not, in general, coincide with the true gradient extremals. This is akin to how a straight line joining New York and Madrid drawn on a (Mercator) map is not actually a “straight-line” on the globe (i.e., a segment of a great circle—also known as a geodesic).

Example 1

We illustrate a gradient extremal resolved on a smooth Riemannian manifold defined by the unit sphere S2={(x,y,z)R3x2+y2+z2=1} with potential energy E(x,y,z)=xyz in Fig. 5. We define the mapping φ as the stereographic projection from the North pole to the tangent plane at the South pole such that φ(x,y,z)=(x1z,y1z)(u,v) and its inverse as ψ(u,v)=1u2+v2+1(2u,2v,u2+v21) and the pullback energy as Z(u,v)=4uv(u2+v21)(u2+v2+1)3. All fixed points except the North pole are mapped onto a single chart; in our actual numerical scheme, we relax this assumption to allow for sampling of only a small portion of the manifold at a time. We refer the reader to Ref. 18 for a full derivation of the gradient and Hessian used in this example.

FIG. 5.

FIG. 5.

(a) Two gradient extremal curves shown on the sphere S2={(x,y,z)x2+y2+z2=1R3x2+y2+z2=1} with the potential E(x,y,z)=xyz. (b) The stereographic projection from the North Pole to the tangent plane at the South pole is shown along with the associated pullback potential energy ψE. Full calculation is detailed in Example 1.

As we did in Sec. II A, we can relax the assumption that E corresponds to the potential energy of a gradient vector field if we work instead with the squared magnitude of a general vector field X on the manifold. If we let (u) be the pullback of the squared magnitude evaluated at point u, then the counterpart to (4) on a Riemannian manifold (M,g) becomes

argminu1({L})(max){gu(grad(u),grad(u))},

where

:MR0u(u)=gu(ψX,ψX)=ψ(XTX).

Here, grad is again the Riemannian gradient and (u) corresponds to the pullback of the term XX in the Euclidean case.

As an example of a non-gradient system, let us study the van der Pol oscillator on a Riemannian manifold. First, let R2 be the Euclidean plane with coordinates (x1,x2) and the usual Euclidean metric. The van der Pol oscillator is given by the vector field X,

dx1dt=x2,dx2dt=μ(1(x1)2)x2x1

for some constant μR. Next, the Poincaré disk model of the hyperbolic plane is the Riemannian manifold given by

H2={(y1,y2)R2:(y1)2+(y2)2<1}

and the metric

g=4(dy1dy1+dy2dy2)(1(y1)2(y2)2).

We define the van der Pol vector field X on H2 simply by restricting the Euclidean vector field to the subset H2. In other words, we retain only the vector field X as defined in the disk H2, and this yields, in the new hyperbolic metric g above, a different squared length function than the Euclidean one. Figure 6 shows the difference between the squared length function of the van der Pol oscillator X in R2 and its restriction to H2.

FIG. 6.

FIG. 6.

Squared length function of the van der Pol vector field ( μ=2) on two manifolds. (a) Euclidean plane. (b) Poincaré disk model of the hyperbolic plane.

III. ALGORITHM

We consider a high-dimensional dynamical system X(x)Rn whose trajectories become quickly attracted to a slow d-dimensional manifold MRn(nd). Suppose we have an initial point p0M near a minimum, obtained by running a single long-enough gradient descent trajectory, such that the fast dynamics have subsided and movement on the manifold dominates. Next, we draw N samples in a neighborhood UM around p0. Any sampling technique appropriate to the application can be used; in the subsequent example, we randomly perturb p0 in ambient space and run the N trajectories for a short time τ such that they now effectively lie on the manifold. Another option would be to use umbrella sampling27–29 based on the reduced local coordinates.17

Next, we apply a dimensionality reduction technique, such as diffusion maps,16 to uncover a set of reduced, latent coordinates that describe the manifold in the neighborhood of the initial point. It is important to emphasize that we use global manifold learning techniques on a neighborhood on the manifold: as a result, we obtain a good set of coordinates describing the single, local chart. A different set of collective variables may be considered for each subsequent chart, which could be better suited for the associated local neighborhood (which could correspond to a different operating regime, stage, conformation, etc. of the system). We then use a Gaussian process30 to learn the local mappings between the ambient space and discovered chart φ:URd and its inverse ψ:RdU, as well as mapping of the reduced coordinates to the potential energy, the pullback energy Z:RdR. In the simple examples of this paper, the potential function in ambient space is given. In more complex applications (e.g., effective descriptions of molecular dynamics simulations in terms of collective variables), where this potential function is not explicitly known, some additional local computations (e.g., through umbrella sampling) would be required to estimate its gradient (mean force) at each step. Furthermore, in the case of high-dimensional ambient space (of higher dimensions than the examples shown here), a Gaussian process may inadequately lift from latent to ambient space. Rather than fitting a function for ψ in these more complex applications, Ref. 17 presents instead a method using biased sampling by simulating a stochastic differential equation to estimate ψ(u).

After learning the mappings between the manifold and the local chart, we can now compute the gradient extremal in the Riemannian setting as explained in Sec. II B. Here, we use a pseudo-arclength continuation method22 with adaptive step sizing, stopping when an equilibrium point or a chart boundary is reached. Detection of a fixed point can be done by monitoring the norm of the vector field (either using the vector field itself in the ambient space or by inspecting the gradient of the pullback energy on the chart). If using a Gaussian process to obtain the diffeomorphisms, detection of a chart boundary can be accomplished by monitoring that the norm of the covariance matrix at a given point remains below a chosen threshold. If a boundary is reached, it is necessary to switch charts—that is, to resample around the point at the boundary of the previous chart and relearn the mappings and pullback energy function. Care has to be taken to continue the curve in the same direction between local charts, potentially oriented in opposing directions; this can be accomplished by mapping the last few ambient points of the curve onto the new local chart and ensuring continuation occurs in the same direction (e.g., the angle between the previous curve mapped onto the new chart and the continuation direction is acute). The full methodology is summarized in Algorithm 1.

ALGORITHM 1.

Continuation of gradient extremal curves on manifolds defined by point clouds.

Input: initial point p=p0 near minimum, samples per iteration N, continuation parameters, threshold ρ>0 used in the convergence
  criterion.
Output: equilibrium p of X.
   for n=1,2,3, do
            Sample N points from a neighborhood UM of pn-1.
            Use manifold learning to obtain local coordinates φ:URd
            Obtain a parameterization ψ:VU, where V=φ(U)Rd.
            Approximate Z=ψE via Gaussian process regression.
             qφ(pn1)
            if n=1 then
                   v Random direction in Rd
            else
                   vφ(q)wn1
            end if
             qn,vnresolve_extremal_curve(q,v) See Algorithm 2.
             pn,wnψ(qn),ψ(qn)vn
            if gradZ(qn)<ρ and n>1 then
                   return ppn
            end if
end for

ALGORITHM 2.

Resolve_extremal_curve.

Input: Initial point q0Rd and velocity v0Rd.
Output: Final point q and velocity v of the extremal curve γ.
Construct numerical continuation curve γ for
{Hess(Z)grad(Z)=λgrad(Z)Z(γ(t))=L
such that γ(0)=q0 and γ˙(0)=v0. L is the value of the level set at the current time step.
Prolong γ until the first instant t>0 such that γ(t)V but γ(t+s)V for any s>0.
return γ(t), γ˙(t)

While we believe our algorithm is as parsimonious as possible (i.e., starting from a single minimum, using a set of reduced collective variables discovered on-the-fly, and efficiently exploring in a single direction), it is not without limitations. Here, we detail those nuances and limitations:

  • Exploration of all directions: Given a particular minimum, there exist multiple different directions along which one can initialize a gradient extremal curve. For example, from a minimum on a 2D manifold, one could follow both sides (positive and negative) of both eigenvectors of the Hessian. Algorithm 1 assumes that we always follow a single, pre-chosen direction. A bookkeeping strategy would be required to systematically explore all gradient extremals of a dynamical system. The global terrain algorithm19 and the navigation algorithm15 utilize a graph structure to conduct such bookkeeping. These techniques could be integrated into Algorithm 1 to provide an orderly exploration of the fixed points of the dynamical system.

  • Disconnected minima: Unfortunately, a single gradient extremal curve is not guaranteed to connect all fixed points of a potential energy surface. Two minima may occasionally lie on separate gradient extremal curves24 or may be connected by a very long detour.25 This is the case in the Müller–Brown potential, where the topmost minimum follows a long detour to the bottom two (see Fig. 2). An alternative to following the long detour would be to perform a steepest descent trajectory from both sides of the newly located leftmost saddle. Discovery of all saddle points given a single starting minimum is not guaranteed.

  • Continuation parameters: Our implementation hinges on a predictor–corrector scheme for numerical continuation. Care must be taken to set reasonable tolerance values and step size maxima. This is especially true near fixed points, which give rise to numerical difficulties. Continuation sometimes becomes numerically challenging, e.g., in the neighborhood of a new fixed point in the current chart, because the potential energy surface is locally flat there. To locate the fixed point once we approach its neighborhood, Ref. 19 suggests switching to an equation-solving technique, such as quadratic acceleration or conjugate gradients, to accurately locate the saddle. These techniques can be employed using the learned mappings in the chart of interest.

  • Sampling and manifold learning: An adequate approximation of the manifold is critical. Poor sampling or failed manifold learning will undermine the success of the algorithm. The more curvature a manifold exhibits, the more challenging it is to create an accurate, invertible mapping between the manifold and local chart. Working instead with multiple, smaller charts will alleviate this issue.

IV. COMPARISON OF POTENTIAL CURVES CONNECTING FIXED POINTS

The briefly aforementioned paths (gradient extremals, Newton trajectories, gentlest ascent dynamics, and that of the nudged elastic band or string variants) are all examples of paths that connect a minimum to a saddle. Sometimes, they coincide with reaction paths: continuous curves that monotonically ascend toward a saddle and then monotonically decrease to the next minimum.31 (While we use the definition of the reaction path as described in Ref. 31, we note that there is disagreement in the literature surrounding the formal definition of a reaction path. For example, Ref. 32 uses the term reaction path as a synonym for a curve in which all points of the reaction path sit in a valley. Further complicating matters, the path described by Ref. 32 is sometimes referred to as the minimum energy path. We do neither use this definition for a reaction path nor a minimum energy path in this work.) Curves that display turning points in L and, thus, switch between increasing and decreasing in energy (i.e., curves that “meander”) are not reaction paths. Here, we briefly compare the paths, which differ in mathematical formulation and how they ascend to the saddle. Our comparison is done in Euclidean space.

The path of the nudged elastic band (NEB)7 and the string variants8,33,34 converge to the steepest descent path from each side of a saddle point.35 This curve is often—but not always—referred to in the literature as the minimum energy path (MEP) as it marks the “path of least resistance” or the most probable reaction path between a chemical reactant and product.36 (An alternative definition for the MEP is a reaction path in which all points sit in a valley. This alternative definition has been well studied by Quapp and Bofill in Refs. 24, 31, 37, and 38. We do not use this definition in this work.) The tangent of this curve is parallel to the gradient except at fixed points. The discovery of the MEP using NEB requires an estimated path initialization and knowledge of two minimum,39,40 which prevent it from being used with our algorithmic framework. Nevertheless, we present it here for comparison due to its significance.

A Newton trajectory (NT), also known as an isocline or reduced gradient following curve, is a curve in which the normalized vector field at each point along the curve is equal to a chosen, fixed vector.9,10 In the framework of mechanochemistry, the movement of fixed points under the application of a quasi-static external force that results in a linear perturbation of the potential is also described by a NT of the original potential.41,42 While in this work and in Ref. 17, our interpretation of a NT focuses on its role as a mathematical connection between fixed points, extending this mechanochemical interpretation to a Riemannian setting would indeed be a worthy exercise. A gradient extremal, as previously described, is a curve comprising points in which the gradient is always an eigenvector of the Hessian of the potential. A gradient extremal does, however, bear general physical significance as it usually follows a valley floor or ridge. Gentlest ascent dynamics forms a new dynamical system that is designed to search for and follow the eigenvector corresponding to the smallest eigenvalue of the Hessian of the potential such that a trajectory ascends toward a saddle point in the least steep manner possible.11 Note that this trajectory is not guaranteed to converge to a saddle. A GAD curve is different from a gradient extremal due to the presence of the ascent term unless the gradient extremal is confluent to an MEP. Further comparisons of gradient extremals, GAD, and NTs can be found in Refs. 10, 24, 43, and 44.

Figure 7 illustrates the four curves on the Müller–Brown potential and (3). We emphasize that the different techniques do produce four distinct curves and that the presence of “meandering” depends on the potential energy surface. See Ref. 31 for a mathematical description of when NT, GAD, and GE exhibit turning points. The gradient extremal curve is the most direct route (aside from that of the string method) to the saddle in the Müller–Brown potential but is the most indirect on (3) due to the presence of turning points. We have previously demonstrated the use of Newton trajectories and GAD within the algorithmic framework presented here to locate saddle points on Riemannian manifolds adaptively revealed in Refs. 17 and 18 respectively. The selection of which path is optimal for a given problem hinges on the shape of the potential energy surface, which, in the context of the algorithm, is likely unknown a priori.

FIG. 7.

FIG. 7.

(a) Gradient extremal (GE), the path of NEB (calculated via the String method), GAD (initialized with a velocity equivalent to the smallest eigenvector of the Hessian of the potential), and NT (with a horizontal slope) between the rightmost minimum and the nearby saddle on the Müller–Brown potential, which is further defined in (8). (b) The four curves starting at the bottom minimum of the potential defined by (3). Circles mark minima, and squares mark saddles. All four curves are unique.

V. NUMERICAL EXAMPLE

We demonstrate the algorithm on the 2D Müller–Brown potential surface,

U(w1,w2)=i=14Aiexp(ai(w1w0i1)2+bi(w1w0i1)(w2w0i2)+ci(w2w0i2)2), (8)

mapped onto the unit sphere using the transformation

V(x,y,z)=(U°κ)(arctan(y/x),arctan(z/x2+y2),

where κ(k1,k2)=(1.973521294k11.85,1.750704373k2+0.875) is an affine mapping. The constants in (8) are listed in Table II.

TABLE II.

Coefficients of the planar Müller–Brown potential.

i A i a i b i c i y0i1 y0i2
1 −200 −1 0 −10 1 0
2 −100 −1 0 −10 0 0.5
3 −170 −6.5 11 −6.5 −0.5 −1.5
4 15.0 0.7 0.6 0.7 −1 1

Beginning at the rightmost minimum, we sample 500 points in the neighborhood of an initial point as in Fig. 8, learn a local chart using diffusion maps, and learn the mappings between ambient space and charts using Gaussian processes as explained in Sec. III. Figure 9 shows the second iteration of a gradient extremal in the learned latent space and its equivalence on the sphere in ambient space; Fig. 10 shows the fifth. We reached the saddle point after iterating through a sequence of 11 successive charts; the full gradient extremal can be seen in Fig. 11. This example and a generic implementation of our algorithm can be found at https://github.com/tasiag/gradient-extremals-on-manifolds.

FIG. 8.

FIG. 8.

500 samples drawn around the rightmost minimum of the Müller-Brown potential mapped onto a unit sphere. (a) and (b) Samples are first drawn by randomly perturbing the known minima in ambient space. (c) and (d) The trajectories are then run for a short time such that they have fallen onto the manifold. These final samples, now on the manifold, are used in the subsequent steps.

FIG. 9.

FIG. 9.

The second iteration of the algorithm beginning from the rightmost minimum of the Müller–Brown potential mapped onto a unit sphere. (a) depicts the gradient extremal resolved in the learned latent chart. (b) shows the same path mapped back onto the sphere. The red outline depicts the boundary of our learned latent chart.

FIG. 10.

FIG. 10.

The fifth iteration of the algorithm beginning from the rightmost minimum of the Müller–Brown potential mapped onto a unit sphere. (a) depicts the gradient extremal resolved in the learned latent chart. (b) shows the same path mapped back onto the sphere. The red outline depicts the boundary of our learned latent chart.

FIG. 11.

FIG. 11.

The resolved gradient extremal after 11 iterations successfully reaches the saddle point.

VI. CONCLUSION

We have presented an algorithmic framework to locate saddle points (and other fixed points) of vector fields on manifolds gradually revealed by point clouds. The methodology assumes that a low-dimensional manifold, which can be uncovered on-the-fly by manifold learning techniques, quickly attracts trajectories of the high-dimensional dynamical system. Gradient extremals, curves that are known to connect fixed points, can then be followed on the manifold to the next point of interest. The technique does neither require a priori knowledge of the manifold nor of good collective variables; rather, it requires only an initial starting point close to a minimum and the ability to sample the dynamical system. The technique does not exhaustively search the whole space but predicts a single new location to sample at so that the simulator is efficiently and effectively biased toward the next point of interest.

We demonstrate the technique on the Müller–Brown potential mapped to a sphere, without prior knowledge of the manifold. The technique is applicable both to (effective) gradient problems, where it helps discover transition states, as well as to general vector fields, where it helps connect isolated solution branches. We also compare how a gradient extremal curve compares to our prior work, which shares algorithmic elements but uses Newton trajectories17 and GAD curves18 to locate saddle points. We expect the collection of the three techniques to act as a robust toolkit that can scale to systems of higher complexity.

ACKNOWLEDGMENTS

We thank Yannick De Decker (Université Libre de Bruxelles, Belgium) for providing (3). The work of A.G., J.M.B.-R., and I.K. has been partially supported by the US Air Force Office of Scientific Research (AFOSR MURI) and the US Department of Energy. The work of A.G. has been partially supported by T32CA153952 from NCI. The work of H.V. has been supported by grant 1179820N from the FWO (Belgium).

Contributor Information

A. Georgiou, Email: mailto:ageorgi3@jhu.edu.

I. Kevrekidis, Email: mailto:yannisk@jhu.edu.

AUTHOR DECLARATIONS

Conflict of Interest

The authors have no conflicts to disclose.

Author Contributions

A. Georgiou: Formal analysis (equal); Investigation (equal); Methodology (equal); Software (equal); Validation (equal); Visualization (equal); Writing – original draft (equal); Writing – review & editing (equal). H. Vandecasteele: Methodology (equal); Validation (equal); Visualization (equal); Writing – original draft (equal). J. M. Bello-Rivas: Conceptualization (equal); Formal analysis (equal); Methodology (equal); Validation (equal); Writing – original draft (equal); Writing – review & editing (equal). I. Kevrekidis: Conceptualization (equal); Formal analysis (equal); Funding acquisition (lead); Methodology (equal); Project administration (lead); Supervision (lead); Writing – review & editing (equal).

DATA AVAILABILITY

The data that support the findings of this study are openly available in GitHub at https://doi.org/10.5281/zenodo.8384160.

REFERENCES

  • 1.Bochevarov A. D., Harder E., Hughes T. F., Greenwood J. R., Braden D. A., Philipp D. M., Rinaldo D., Halls M. D., Zhang J., and Friesner R. A., “Jaguar: A high-performance quantum chemistry software program with strengths in life and materials sciences,” Int. J. Quantum Chem. 113, 2110–2142 (2013). 10.1002/qua.24481 [DOI] [Google Scholar]
  • 2.Leimkuhler B. and Matthews C., Molecular Dynamics: With Deterministic and Stochastic Numerical Methods,” Interdisciplinary Applied Mathematics (Springer, 2015).
  • 3.Laio A. and Parrinello M., “Escaping free-energy minima,” Proc. Natl. Acad. Sci. U.S.A. 99, 12562–12566 (2002). 10.1073/pnas.202427399 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Darve E., Rodríguez-Gómez D., and Pohorille A., “Adaptive biasing force method for scalar and vector free energy calculations,” J. Chem. Phys. 128, 144120 (2008). 10.1063/1.2829861 [DOI] [PubMed] [Google Scholar]
  • 5.Chiavazzo E., Covino R., Coifman R. R., Gear C. W., Georgiou A. S., Hummer G., and Kevrekidis I. G., “Intrinsic map dynamics exploration for uncharted effective free-energy landscapes,” Proc. Natl. Acad. Sci. U.S.A. 114, E5494–E5503 (2017). 10.1073/pnas.1621481114 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Olender R. and Elber R., “Calculation of classical trajectories with a very large time step: Formalism and numerical examples,” J. Chem. Phys. 105, 9299–9315 (1996). 10.1063/1.472727 [DOI] [Google Scholar]
  • 7.Jónsson H., Mills G., and Jacobsen K. W., “Nudged elastic band method for finding minimum energy paths of transitions,” in Classical and Quantum Dynamics in Condensed Phase Simulations (World Scientific, 1998), pp. 385–404.
  • 8.W. Ren W. E. and Vanden-Eijnden E., “String method for the study of rare events,” Phys. Rev. B 66, 052301 (2002). 10.1103/PhysRevB.66.052301 [DOI] [PubMed] [Google Scholar]
  • 9.Quapp W., Hirsch M., Imig O., and Heidrich D., “Searching for saddle points of potential energy surfaces by following a reduced gradient,” J. Comput. Chem. 19, 1087–1100 (1998). [DOI] [Google Scholar]
  • 10.Hirsch M. and Quapp W., “Reaction pathways and convexity of the potential energy surface: Application of Newton trajectories,” J. Math. Chem. 36, 307–340 (2004). 10.1023/B:JOMC.0000044520.03226.5f [DOI] [Google Scholar]
  • 11.E W. and Zhou X., “The gentlest ascent dynamics,” Nonlinearity 24, 1831–1842 (2010). 10.1088/0951-7715/24/6/008 [DOI] [Google Scholar]
  • 12.Henkelman G. and Jónsson H., “A dimer method for finding saddle points on high dimensional potential surfaces using only first derivatives,” J. Chem. Phys. 111, 7010 (1999). 10.1063/1.480097 [DOI] [Google Scholar]
  • 13.Basilevsky M., “The topography of potential energy surfaces,” Chem. Phys. 67, 337–346 (1982). 10.1016/0301-0104(82)85194-X [DOI] [Google Scholar]
  • 14.Rowe D. J. and Ryman A., “Valleys and fall lines on a Riemannian manifold,” J. Math. Phys. 23, 732–735 (1982). 10.1063/1.525427 [DOI] [Google Scholar]
  • 15.Filippidis I. and Kyriakopoulos K. J., “Roadmaps using gradient extremal paths,” in 2013 IEEE International Conference on Robotics and Automation (IEEE, 2013), pp. 370–375.
  • 16.Coifman R. R. and Lafon S., “Diffusion maps,” Appl. Comput. Harmon. Anal. 21, 5–30 (2006). 10.1016/j.acha.2006.04.006 [DOI] [Google Scholar]
  • 17.Bello-Rivas J. M., Georgiou A., Guckenheimer J., and Kevrekidis I. G., “Staying the course: Iteratively locating equilibria of dynamical systems on Riemannian manifolds defined by point-clouds,” J. Math. Chem. 61, 600–629 (2023). 10.1007/s10910-022-01425-9 [DOI] [Google Scholar]
  • 18.Bello-Rivas J. M., Georgiou A., Vandecasteele H., and Kevrekidis I. G., “Gentlest ascent dynamics on manifolds defined by adaptively sampled point-clouds,” J. Phys. Chem. B 127, 5178–5189 (2023). 10.1021/acs.jpcb.3c00897 [DOI] [PubMed] [Google Scholar]
  • 19.Lucia A. and Feng Y., “Global terrain methods,” Comput. Chem. Eng. 26, 529–546 (2002). 10.1016/S0098-1354(01)00777-3 [DOI] [Google Scholar]
  • 20.Hoffman D. K., Nord R. S., and Ruedenberg K., “Gradient extremals,” Theor. Chim. Acta 69, 265–279 (1986). 10.1007/BF00527704 [DOI] [Google Scholar]
  • 21.Quapp W., “Gradient extremals and valley floor bifurcations on potential energy surfaces,” Theor. Chim. Acta 75, 447–460 (1989). 10.1007/BF00527676 [DOI] [Google Scholar]
  • 22.Allgower E. L. and Georg K., Introduction to Numerical Continuation Methods (Society for Industrial and Applied Mathematics, 2003). [Google Scholar]
  • 23.Lucia A., Dimaggio P. A., and Depa P., “A geometric terrain methodology for global optimization,” J. Glob. Optim. 29, 297–314 (2004). 10.1023/B:JOGO.0000044771.25100.2d [DOI] [Google Scholar]
  • 24.Hirsch M. and Quapp W., “The reaction pathway of a potential energy surface as curve with induced tangent,” Chem. Phys. Lett. 395, 150–156 (2004). 10.1016/j.cplett.2004.07.079 [DOI] [Google Scholar]
  • 25.Ohno K. and Maeda S., “A scaled hypersphere search method for the topography of reaction pathways on the potential energy surface,” Chem. Phys. Lett. 384, 277–282 (2004). 10.1016/j.cplett.2003.12.030 [DOI] [Google Scholar]
  • 26.Uppal A., Ray W., and Poore A., “On the dynamic behavior of continuous stirred tank reactors,” Chem. Eng. Sci. 29, 967–985 (1974). 10.1016/0009-2509(74)80089-8 [DOI] [Google Scholar]
  • 27.Torrie G. and Valleau J., “Nonphysical sampling distributions in Monte Carlo free-energy estimation: Umbrella sampling,” J. Comput. Phys. 23, 187–199 (1977). 10.1016/0021-9991(77)90121-8 [DOI] [Google Scholar]
  • 28.Giacomo Fiorin M. L. K. and Hénin J., “Using collective variables to drive molecular dynamics simulations,” Mol. Phys. 111, 3345–3362 (2013). 10.1080/00268976.2013.813594 [DOI] [Google Scholar]
  • 29.Ferguson A. L., Panagiotopoulos A. Z., Debenedetti P. G., and Kevrekidis I. G., “Integrating diffusion maps with umbrella sampling: Application to alanine dipeptide,” J. Chem. Phys. 134, 135103 (2011). 10.1063/1.3574394 [DOI] [PubMed] [Google Scholar]
  • 30.Rasmussen C. E. and Williams C. K. I., “Gaussian processes for machine learning,” in Adaptive Computation and Machine Learning (MIT Press, 2006), pp. I–XVIII, 1–248.
  • 31.Bofill J. M. and Quapp W., “Calculus of variations as a basic tool for modelling of reaction paths and localisation of stationary points on potential energy surfaces,” Mol. Phys. 118, e1667035 (2020). 10.1080/00268976.2019.1667035 [DOI] [Google Scholar]
  • 32.Heidrich D., “An introduction to the nomenclature and usage of the reaction path concept,” in The Reaction Path in Chemistry: Current Approaches and Perspectives, edited by Heidrich D. (Kluwer Academic Publishers, 1995). 10.1007/978-94-015-8539-2 [DOI] [Google Scholar]
  • 33.Behn A., Zimmerman P. M., Bell A. T., and Head-Gordon M., “Incorporating linear synchronous transit interpolation into the growing string method: Algorithm and applications,” J. Chem. Theory Comput. 7, 4019–4025 (2011). 10.1021/ct200654u [DOI] [PubMed] [Google Scholar]
  • 34.Chaffey-Millar H., Nikodem A., Matveev A. V., Krüger S., and Rösch N., “Improving upon string methods for transition state discovery,” J. Chem. Theory Comput. 8, 777–786 (2012). 10.1021/ct200639w [DOI] [PubMed] [Google Scholar]
  • 35.Sheppard D. and Henkelman G., “Paths to which the nudged elastic band converges,” J. Comput. Chem. 32, 1769–1771 (2011). 10.1002/jcc.21748 [DOI] [PubMed] [Google Scholar]
  • 36.Dunitz J. D., “Chemical reaction paths,” Phil. Trans. R. Soc. Lond. B 272, 99–108 (1975). 10.1098/rstb.1975.0073 [DOI] [Google Scholar]
  • 37.Bofill J. M., Quapp W., and Caballero M., “The variational structure of gradient extremals,” J. Chem. Theory Comput. 8, 927–935 (2012). 10.1021/ct200805d [DOI] [PubMed] [Google Scholar]
  • 38.Bofill J. M. and Quapp W., “The variational nature of the gentlest ascent dynamics and the relation of a variational minimum of a curve and the minimum energy path,” Theor. Chem. Acc. 135, 11 (2015). 10.1007/s00214-015-1767-7 [DOI] [Google Scholar]
  • 39.Palenik M. C., “Initial estimate for minimum energy pathways and transition states using velocities in internal coordinates,” Chem. Phys. 542, 111046 (2021). 10.1016/j.chemphys.2020.111046 [DOI] [Google Scholar]
  • 40.Halgren T. A. and Lipscomb W. N., “The synchronous-transit method for determining reaction pathways and locating molecular transition states,” Chem. Phys. Lett. 49, 225–232 (1977). 10.1016/0009-2614(77)80574-5 [DOI] [Google Scholar]
  • 41.Quapp W. and Bofill J. M., “Reaction rates in a theory of mechanochemical pathways,” J. Comput. Chem. 37, 2467–2478 (2016). 10.1002/jcc.24470 [DOI] [PubMed] [Google Scholar]
  • 42.Quapp W. and Bofill J. M., “A contribution to a theory of mechanochemical pathways by means of Newton trajectories,” Theor. Chem. Acc. 135, 113 (2016). 10.1007/s00214-016-1880-2 [DOI] [Google Scholar]
  • 43.Quapp W., “Reduced gradient methods and their relation to reaction paths,” J. Theor. Comput. Chem. 02, 385–417 (2003). 10.1142/S0219633603000604 [DOI] [Google Scholar]
  • 44.Bofill J. M., Quapp W., and Caballero M., “Locating transition states on potential energy surfaces by the gentlest ascent dynamics,” Chem. Phys. Lett. 583, 203–208 (2013). 10.1016/j.cplett.2013.07.074 [DOI] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

The data that support the findings of this study are openly available in GitHub at https://doi.org/10.5281/zenodo.8384160.


Articles from Chaos are provided here courtesy of American Institute of Physics

RESOURCES