Abstract
We study the first-passage dynamics of a non-Markovian stochastic process with time-averaged feedback, which we model as a one-dimensional Ornstein–Uhlenbeck process wherein the particle drift is modified by the empirical mean of its trajectory. This process maps onto a class of self-interacting diffusions. Using weak-noise large deviation theory, we calculate the leading order asymptotics of the time-dependent distribution of the particle position, derive the most probable paths that reach the specified position at a given time and quantify their likelihood via the action functional. We compute the feedback-modified Kramers rate and its inverse, which approximates the mean first-passage time, and show that the feedback accelerates dynamics by storing finite-time fluctuations, thereby lowering the effective energy barrier and shifting the optimal first-passage time from infinite to finite. Although we identify alternative mechanisms, such as slingshot and ballistic trajectories, we find that they remain sub-optimal and hence do not accelerate the dynamics. These results show how memory feedback reshapes rare event statistics, thereby offering a mechanism to potentially control first-passage dynamics.
Introduction
Many real-world systems exhibit memory-feedback dynamics [1–4], where the evolution of a process depends on its past trajectory. Unlike standard Markovian systems, these processes are influenced by accumulated memory. Such feedback mechanisms introduce rich dynamical features, altering long-time behavior and fluctuation properties in ways not captured by memoryless models [5–13]. In particular, memory-feedback mechanisms are relevant for designing control strategies to achieve specific tasks [14–18] with specific thermodynamic properties [19–25].
Self-interacting systems [26–32] constitute a notable class of memory-feedback processes in which past states, encoded in the empirical occupation measure, or in a time-integrated functional of the process, influence future evolution. Examples include models in statistical physics relevant to biological [33–37] and colloidal systems [38–40], where interactions with the occupation measure affect the dynamics.
Although the stationary properties of self-interacting processes have been studied extensively, the role of feedback in shaping rare event statistics, and in particular first-passage dynamics, remains poorly understood. However, understanding the finite-time stochastic dynamics of feedback-driven systems has many important implications in, among other problems, reinforcement learning algorithms [14].
Only recently have large-deviation results for self-interacting systems begun to emerge, both in the long-time regime for discrete Markov chains [41, 42] and for diffusions [43]. For example, first-passage dynamics has been addressed in self-interacting discrete-time random walks [44], where universal scaling laws for first-passage time distributions were identified, and in diffusive systems [45], where by analyzing the asymptotic behavior of the first-passage time in the weak-noise limit a Kramers-type law was derived.
Here we analyze a representative case: a one-dimensional Ornstein–Uhlenbeck process where the drift term is influenced by its own time-averaged position. This feedback mechanism breaks Markovianity and defines a self-interacting diffusion process, fundamentally altering the first-passage dynamics of the system.
By using weak-noise large deviation theory, we show that self-interaction can lead to accelerated first-passage dynamics, and thus fluctuation-driven mechanisms induced by memory. In particular, we identify and characterize different first-passage scenarios and quantify their likelihood. Our large deviation approach allows us to derive a Kramers-type law for the mean first-passage time in the feedback Ornstein–Uhlenbeck process, which complements and extends the results in [44, 45]. We show that for a positive feedback parameter (see Eq. 1), the leading-order noise asymptotics of the first-passage time distribution develops a new global minimum in time. As a result, the effective energy barrier is reduced, causing exponential speed-up and an exponentially smaller mean first-passage time. Furthermore, our findings generalize recent results on first-passage dynamics in other classes of memory-driven systems, such as generalized Langevin equations [46, 47].
The paper is organized as follows. In Section 2, we introduce the Ornstein–Uhlenbeck process with empirical average position feedback, establish its mapping to self-interacting diffusion, and discuss its deterministic and long-time dynamics. In Section 3, we present the weak-noise large deviation framework used to derive our results. We apply it to both the standard and feedback Ornstein–Uhlenbeck processes, and derive the fundamental equation (61) governing finite-time dynamics in the weak-noise regime, which we have not found in the literature. In Section 4, we detail the consequences of solving (61). We describe the finite-time mechanisms driving first-passage dynamics in the feedback process, compare them to the standard Ornstein–Uhlenbeck case, and analyze their relative and absolute likelihoods. We characterize the quasi-potential (i.e., the large deviation rate function associated with the Arrhenius law) and the optimal first-passage times. Finally, in Section 5 we summarize our findings, highlight open questions, and discuss potential applications, particularly in the context of reinforcement learning.
Stochastic Feedback Ornstein–Uhlenbeck Processes and Deterministic Dynamics
Ornstein–Uhlenbeck Process with Empirical Average Position Feedback
We consider a one-dimensional process governed by the following stochastic differential equation (SDE):
| 1 |
with initial condition , noise intensity , and feedback parameter , the range of which we will justify in detail below. Here, denotes a standard Wiener process, or Brownian motion, in , characterized by independent Gaussian increments: and .
In the absence of the additional feedback term involving the time-averaged position , Eq. (1) is the classical Ornstein–Uhlenbeck (OU) process, describing a Brownian agent moving in a parabolic potential well centered on the origin. As we will show, the inclusion of the empirical time-averaged position introduces a non-Markovian feedback that can give rise to unique dynamical behavior, effectively modifying the potential in which the particle evolves. This results in an effective potential whose minimum shifts along depending on the cumulative history of the process. We discuss this in more detail in Section 2.2.
Our objective is to analyze how this non-Markovian feedback influences the first-passage dynamics of through a threshold , with a particular focus on the role of noise and memory. The feedback term can be expressed as
| 2 |
| 3 |
| 4 |
is the empirical occupation measure of the process. This measure captures the fraction of time the system has spent in the vicinity of each position x up to time t.
Although because its evolution depends on its entire history, the process is non-Markovian, the joint two-dimensional process is Markovian and satisfies:
| 5 |
Despite the fact that the process is time-inhomogeneous, the Markovian formulation (5) allows us to study the first-passage dynamics of the feedback OU process semi-analytically in the weak-noise regime, as shown in Section 3.3.
Mapping to Self-Interacting Diffusion Processes
The structure of the feedback embodied in Eqs. (2)–(4), allows the process in Eq. (1) to be interpreted as a specific case of a one-dimensional self-interacting diffusion (SID). A general SID in is typically expressed as
| 6 |
where is a background potential and is an interaction potential whose gradient is taken with respect to . The interaction term describes how the current position is influenced by the history of past positions, functionally through the empirical occupation measure defined in Eq. (4).
The long-time behavior of self-interacting processes, including random walks and diffusions, has been an active area of research in probability theory for several decades. Self-interacting random walks were first introduced in the discrete-time setting by Norris et al. [48], and continuous-time counterparts appeared soon after as models for polymer growth [49]. The normalized SID considered here (see Eq. (6)) was introduced by Benaïm et al. [27] as an example of a continuous-time path-interaction (or reinforcement) process. It extends earlier classes of self-interacting processes based on non-normalized occupation measures [50, 51] and discrete-time reinforced random walks. A thorough historical overview is given in [28].
The asymptotic properties of SIDs were first studied in [27, 52–54], with results established for compact spaces and symmetric interaction potentials. These were later extended to open domains, in particular to under confining potentials [30, 55, 56], which is directly relevant for our feedback OU process. Stochastic approximation methods were used in [27, 30] to describe the asymptotic behavior of the occupation measure via a limiting non-autonomous differential equation. A key result of [53] shows that for symmetric potentials converges almost surely to a local minimum of a nonlinear free-energy functional, which typically admits multiple critical points, each selected at random with some probability. In the special case of symmetric self-repelling potentials this functional is strictly convex, so converges almost surely to a unique global minimum in the long-time limit.
More recently, refinements of Kramers’ law for SIDs have been obtained [45, 57] (see also [46, 47] for generalized overdamped Langevin systems), and these results extend to systems with general additive noise [45, 57, 58]. A recent contribution by one of us establishes large-deviation results for SIDs on the ring [43]. We will detail below (Sec. 2.4) those results that are most relevant for the present work.
For the case of Eq. (1), the potentials in Eq. (6) take the form
| 7 |
| 8 |
Therefore, a quadratic interaction potential K describes the attraction of the current position to the past positions , so that the general behavior of the process is determined by the combination of this interaction and the background potential V.
By rewriting Eq. (1), we can define an effective potential
| 9 |
whose shape evolves according to the past trajectory of the process. For example, the minimum of this effective potential may shift to the left or right of the origin depending on whether the integral contribution in Eq. (9) is negative or positive, respectively.
Eqs. (7) and (8) make clear that for (), the background potential remains quadratic and attractive (repulsive) around the origin. Although the process is still attracted to its past positions through the interaction potential K, the repulsive background potential drives first-passage dynamics towards infinity, which justifies the choice described when we introduced Eq. (1).
More generally, Eq. (1) can be extended to incorporate alternative forms of feedback, such as
| 10 |
with , which ensures a decaying contribution, with slower decay as decreases, and f is an arbitrary function on . Although the methods developed here could be generalized to such circumstances, we focus on the simplest and most natural case: feedback via the empirical average position, corresponding to and . Importantly, this choice enables an exact mapping to SID processes. Alternative values of , or more general time-dependent functions, would typically not allow this mapping and thus preclude comparison with well-established convergence properties of SIDs, which we discuss in Section 2.4.
Exact Results for the Deterministic Dynamics
Before analyzing the stochastic dynamics of the feedback OU process, we examine some essential features of the deterministic dynamics, obtained by setting in Eqs. (1) or (5). This analysis provides insight into the behavior of the system and helps clarify the choice of the parameter range for .
We set in Eq. (5), which becomes the following first-order, time-inhomogeneous linear system of differential equations;
| 11 |
where denotes the time derivative, and the time-dependent drift matrix is
| 12 |
For , the only fixed point of the system is the origin (0, 0). However, when , all points satisfying are fixed points.
The stability of the origin depends on both and time as follows (see Fig. 1 for comparison):
- :
The origin is stable, but exhibits a spiral-like behavior in the finite time window defined below.
- :
The origin is a stable node.
- :
The origin becomes a saddle point, and an instability develops along a specific direction. At the bifurcation point , all points are neutrally stable.
Fig. 1.
Phase portraits of the deterministic dynamical systems in Eq. (11) with (12) for three illustrative values of (left), (middle), and (right)
Due to the linearity of Eq. (11), stability is fully characterized by the eigenvalues of the matrix , which are always real and given by:
| 13 |
For , the square root term in Eq. (13) satisfies:
| 14 |
implying that both eigenvalues are negative for all , and the origin is a stable node.
For , the discriminant in Eq. (13) can become negative, leading to complex eigenvalues. This occurs when , where
| 15 |
Since the real parts of both eigenvalues remain negative, the origin remains stable, but the dynamics exhibit a spiral behavior during this time interval.
For , the situation is reversed: one eigenvalue , while the other . An unstable direction thus emerges, associated with the eigenvector
| 16 |
Here, the fixed point at the origin becomes a saddle. When , the direction is neutrally stable with . Thus, any small perturbation to (0, 0) not orthogonal to will drive the system away from the fixed point, growing asymptotically along the direction of , with X and increasing proportionally, as in Eq. (16).
The key features of the deterministic dynamics that have important implications for the stochastic system are as follows:
If one conditions on , then when even a small perturbation due to noise will cause the process to escape the origin deterministically towards . In such a regime, noise merely initiates the escape, but deterministic dynamics dominate the evolution. We will not focus on this regime in our work. However, it would be interesting to study the fluctuations of the escape time from the origin, similarly to that done in [59].
The interesting regime for our work is when . For , the feedback in Eq. (1) is positive, so the most likely first-passage to occurs through the accumulation of positive fluctuations. However, for , this intuitive picture breaks down: as positive fluctuations accumulate, the feedback becomes increasingly negative, pulling the process back toward the origin and making first-passage through highly unlikely. Therefore, two alternative first-passage mechanisms for the stochastic dynamics emerge:
(a) A one-stage mechanism, whereby large fluctuations are sufficient to overcome the negative feedback. This is an extremely rare event in the weak-noise limit. (b) A two-stage mechanism, whereby the stochastic process first accumulates negative fluctuations, generating a strong positive feedback, which then drives a “slingshot” first-passage to the right; .
These -dependent qualitatively distinct first-passage pathways raise a natural question: In a symmetric domain , with , is the agent more likely to escape to the right via the accumulation of positive fluctuations, or to the left by exploiting a slingshot mechanism? We return to these questions in Section 4.
In the next section (2.4), we characterize the long-time behavior of general SIDs and feedback processes, which will serve as a useful reference for our analysis of finite-time first-passage dynamics.
Long-Time Behavior of the Feedback Ornstein–Uhlenbeck Processes
Having established in Section 2.2 that the feedback OU process in Eq. (1) belongs to the class of SIDs, we now draw on results from SID theory to understand the important asymptotic behavior. This will prove useful later, particularly in Sections 4.2 and 4.3, when interpreting our results.
Assuming stationary behavior at long times (as in [27, 43, 57]), one can formally derive the stationary occupation measure, or invariant density, of the general SID in Eq. (6) from the stationary Fokker–Planck equation (see [43]):
| 17 |
where Z is a normalization constant. The symmetry of the potential V in Eq. (7) is inherited by the invariant density [45, 57, 60], and in the case of the feedback OU process Eq. (1), Eq. (17) becomes
| 18 |
which coincides with the stationary distribution of the standard OU process. In consequence, the memory feedback term in Eq. (2) converges in probability to
| 19 |
Therefore, the classical OU process is recovered as the long-time limit of the feedback OU process. However, it is important to note that the convergence to this fixed point occurs very-slowly; slower than any algebraic rate. This has been shown in [45, 56, 57]: see Theorems 1.6 and 1.12 in [56], Theorem 2.7 in [57], and Theorem 2.10 and Propositions 2.11–2.12 in [45]. These long-time results serve as a valuable benchmark for understanding the finite-time dynamics studied in the remainder of this paper, where the memory feedback has a controlling influence.
We conclude this section by highlighting Theorem 2.13 from Aleksian et al. [45], which is relevant to our study. In the weak-noise limit , they established a Kramers-type law for the first-passage time OF the process in Eq. (6) THAT exits a domain , through the boundary , defined as . They assumed that the potentials V and W are uniformly convex with at most polynomial growth and V such that (this last condition does not hold for the V in Eq. (7)). They established that
| 20 |
where
| 21 |
and m is the point at which attains its maximum. For the feedback OU process in Eq. (1), we have and we typically consider a single first-passage boundary . Therefore, we find:
| 22 |
We note that this result was derived under the further assumption that the empirical occupation measure stabilizes prior to first-passage dynamics, which we believe is the reason the right hand side of Eq. (20) depends neither on the starting point x nor on . As we show in Section 3.1 using a weak-noise large deviation approach, our analysis does not require this assumption, which leads to a strong dependence of the first-passage dynamics on the value of , as we will explore in detail.
Finally, we emphasize that Eq. (20) holds in probability. As such, it does not constitute an Arrhenius-type law, which would describe the expected time to reach a specified position (this is also noted in the Appendix of [45]). The quantification of an Arrhenius law for the feedback OU process, and its relation to the underlying large deviation principles, is one of the main results of our work.
First-passage Dynamics and Weak-Noise Large Deviations
Given the Markovian structure of the dynamics of the process in Eq. (5), the associated time-dependent infinitesimal generator takes the form:
| 23 |
where and denotes the gradient with respect to the variables . In principle, this generator can be used to extract information on the first-passage dynamics, in particular: (i) the first-passage time probability density, and (ii) the mean first-passage time.
For the non-homogeneous Markov process with time-dependent infinitesimal generator given in (23), the first-passage time is defined as
| 24 |
with initial condition at time s. Its average, the mean first-passage time, is given by
| 25 |
which is the expected time (from time s) it takes for the process to hit . Notice that both the first-passage time and its average depend on the starting time s and starting position (due to the non-homogeneity of the process).
In order to describe a general procedure to calculate the first-passage time probability density and its first moment , we introduce the transition probability density of the process , with absorbing boundary at , which is written as
| 26 |
for . Given the pointwise initial condition, the transition probability density above is already normalized. We remark that is shorthand for the probability that lies in a small interval conditioned on having started the process at at time s, so that the probability density satisfies , where is the probability distribution.
Returning to the transition probability density function in (26), we remark that it satisfies the following Fokker–Planck boundary value problem:
| 27 |
with and where denotes the adjoint of the generator in Eq. (23). We also have that the first-passage time probability density is given by
| 28 |
where
| 29 |
is the survival probability density function, which is the probability that the particle starting at at time s has not hit by time (notice that the variable is also integrated out).
Therefore, the mean first-passage time satisfies the following boundary value problem:
| 30 |
whose derivation is given in Appendix 6.1. Because the infinitesimal generator is explicitly time-dependent due to the drift term, and the diffusion is degenerate since no noise acts directly on the dynamics of , exact analytical solutions to Eqs. (27) and (30) are generally not available.
However, asymptotic approaches are possible. For instance, approximate analytical solutions for both the autonomous and non-autonomous one-dimensional Ornstein–Uhlenbeck processes and for stochastic resonance have recently been obtained using matched asymptotic expansions [61–64]. While this approach could in principle be extended to the two-dimensional setting considered here, identifying the appropriate boundary layers in two dimensions is highly non-trivial and would require a separate in-depth study.
Instead, here we use the framework of weak-noise large deviation theory to derive asymptotic results for the first-passage dynamics of the feedback OU process. In the weak-noise regime, we can analyze the asymptotic behavior of the probability density by solving an effective approximation of Eq. (27) that is the associated Hamilton–Jacobi equation.
In Section 3.1, we briefly review the general weak-noise large deviation framework for Markov processes—see [65–67] for more comprehensive treatments. We then illustrate how this framework can be used to extract relevant information about first-passage dynamics and apply it to both the standard OU process (Section 3.2) and the feedback OU process (Section 3.3).
Weak-Noise Large Deviations
For simplicity, we introduce the weak-noise large deviation framework by considering a general two-dimensional Markov process, denoted, with slight abuse of notation, . We initially keep the notation as general as possible, since the framework extends to any d-dimensional Markov process, and then specialize to the case of interest in Sec. 3.3. The general process is governed by the stochastic differential equation,
| 31 |
where is a (possibly time-dependent) 2-dimensional drift vector field, is a constant noise matrix, is the noise covariance (or diffusion) matrix, and denotes a 2-dimensional Brownian motion. For simplicity, we assume additive noise, i.e., does not depend on position, and we consider a pointwise initial condition . We are interested in events wherein the process exhibits a specific outcome—for instance, when the trajectory reaches a prescribed set at time T, that is . While such events are impossible in the deterministic case (), they typically occur in the presence of noise (), though of course with decreasing frequency as the noise intensity vanishes, .
Similarly to the previous treatment, we introduce the normalized transition probability density function as in Eq. (26) with absorbing boundary . This function satisfies a Fokker–Planck boundary value problem equivalent to (27) and written below for thoroughness
| 32 |
with initial condition and absorbing boundary for and .
Now, by assuming an exponential scaling in we make the following change of variable,
| 33 |
where, for parsimony of notation, on the right hand side the dependence on the initial time and point is implicit. Taking the weak-noise limit, , and retaining terms up to order , gives the following Hamilton–Jacobi equation [67]:
| 34 |
Here, the time-dependent Hamiltonian is given by
| 35 |
where p denotes a vector of adjoint variables, or momentum coordinates. The approximation introduced through the change of variable (33) with is commonly referred to as the WKBJ (after Wentzel, Kramers, Brillouin and Jeffreys) ansatz in many areas of physics, including stochastic processes [61, 68, 69].
The Hamilton–Jacobi equation (34) is a first-order partial differential equation whose solution can be obtained from the characteristic curves of the associated Hamiltonian dynamics, which are
| 36 |
subject to appropriate boundary conditions. By integrating along these characteristic trajectories—often referred to as instantons in the large deviation literature [67, 68]—the action functional in Eq. (34) takes the following form;
| 37 |
This expression reflects the fact that the action is the generating function of a canonical transformation, linking the first-order PDE (34) with the second-order Hamiltonian dynamics of Eq. (36) [70, 71].
In the weak-noise regime, the action so obtained represents the dominant contribution to the probability density function in Eq. (33), wherein, with a slight abuse of notation, we denote both the full action and its leading-order contribution at with the same symbol. This action determines the asymptotic behavior in of the probability of finding the process at for the first time T, viz.
| 38 |
At leading order in , this characterizes the survival probability density function. Therefore, upon taking the time derivative of Eq. (38), we find that in the weak-noise limit the first-passage time distribution in Eq. (28) is also approximated by of Eq. (38).
The optimal trajectory is that which minimizes the action and thereby defines the instanton. Thus, the instanton corresponds to the most likely path connecting the initial point to the final point as , thereby providing the dominant path through which the system reaches at time T. By integrating over time, Eq. (38) leads to the Kramers-type, or Arrhenius-type, formula as
| 39 |
where is the probability density on the state space induced by time integration and
| 40 |
is the quasi-potential, or the large deviation rate function. We interpret as an effective energy barrier that the system must overcome to reach , and it governs the leading-order behavior of the first-passage probability. Additionally, the optimal first-passage time can be determined as
| 41 |
We stress that the equalities in Eq. (38) and Eq. (39) are exact only at the logarithmic scale. Consequently, the right hand side of both equations are not normalized, since terms of order are neglected. This explains the apparent divergence obtained when integrating Eq. (38) over time.
We note that in general this optimal first-passage time is not equivalent to the mean first-passage time defined in Eq. (25) and approximated by the inverse Kramers rate; . Although the two quantities are related, they have distinct interpretations. The optimal first-passage time is defined in the limit , where it is given by the value of T at which the action reaches its minimum. It is therefore independent of and should not be confused with either the mean or the typical first-passage time with finite noise. Instead, it captures the intrinsic tendency of the process in the absence of noise. For example, in the Ornstein–Uhlenbeck process of Sec. 3.2 below, the minimum of the action is attained only as , reflecting the fact that the noiseless dynamics keeps the process at the fixed point indefinitely (before escaping). In contrast, real first-passage events are noise-induced and occur asymptotically on the finite timescale set by the inverse Kramers rate in Eq. (39).
The Weak-Noise Regime of the Ornstein–Uhlenbeck Process
As noted above, a canonical analytically tractable stochastic process in statistical physics is the Ornstein–Uhlenbeck process, which is obtained by removing the feedback term, , from Eq. (1). The result derived in the following can be found in the standard literature of stochastic processes, see for instance [72, 73]. The transition probability density function is a Gaussian that can be computed explicitly by solving the Fokker–Planck equation
| 42 |
with initial condition , where L is the infinitesimal Markov generator given by
| 43 |
Equation (42) can be solved using Fourier transform or heat kernel methods giving the transition probability density as
| 44 |
which is the probability of finding the process at position x at time t, starting at at .
Because Eq. (44) takes a large deviation form, we identify the action in Eq. (38) as
| 45 |
As discussed in Section 3.1, this gives an asymptotic characterization of the probability of finding the OU process at for the first time T; it provides the leading-order asymptotic form of the first-passage time distribution in Eq. (28). By solving Eqs. (36) for the Hamiltonian in Eq. (35), which in this case is , we obtain the instanton equation
| 46 |
which gives the most probable path from to . Integrating Eq. (44) in time yields the probability of being at x as an implicit function of time;
| 47 |
and its leading-order behavior in , which can also be obtained from Eq. (40), is given by the following quasi-potential
| 48 |
which, as expected, is quadratic in the final position. Moreover, we note that the quasi-potential for the OU process coincides with in Eq. (22).
Finally, the first-passage time of the process generally depends on the noise intensity . Although the mean first-passage time is finite and given by the inverse Kramers rate, in the weak-noise limit inspection of Eq. (41) shows that the optimal first-passage time diverges; . Therefore, the weak-noise OU process tends to remain near the bottom of the potential well for as long as possible in order to minimize the action, eventually escaping the quadratic trap through a single large fluctuation.
The Weak-Noise Regime of Feedback Ornstein–Uhlenbeck Processes
We now turn our attention to the feedback OU process. Rather than analyzing the non-Markovian process defined by Eq. (1), we consider the two-dimensional Markovian process defined by Eq. (5). Using the notation of Eq. (31), the deterministic time-dependent drift is given by
| 49 |
and a noise matrix
| 50 |
showing that noise acts only on the component. (We discuss the consequences of the structure of the noise matrix below.) We consider the initial condition , which also regularizes the singularity at , and we aim to study the first-passage dynamics through .
The time-dependent Hamiltonian is given by
| 51 |
where and . The solutions to the Hamilton–Jacobi equation (34) are obtained by solving the associated Hamilton equations (36) and are:
| 52 |
In Eqs. (52), is a parameter implicitly determined. by the final condition through the Hamiltonian dynamics. As we highlight in the derivation below, it is more convenient to impose final conditions on the adjoint variables rather than initial conditions on the adjoint variables or final conditions on the coordinate variables. We also refer the reader to Section 2.4.1 of [74] for an explanation of how to transform final conditions on position variables into final conditions on adjoint variables within a Lagrangian framework.
We note that, in principle, for non-autonomous driving, the choice of boundary conditions in the propagator (see Eq. (32)) can influence even the leading-order behavior in the large-deviation regime (), because first-passage events may depend on whether earlier crossings are excluded. This would in turn modify the boundary conditions in (52). However, in the feedback OU process considered here the boundary conditions do not affect the rate function. The process starts at the stable fixed point , and we ask for a crossing at a distant value . In the weak-noise limit the deterministic drift drives trajectories into the fixed point where they vanish, so waiting before departure incurs no cost in the action. As a result, the minimization problem leading to the instanton is the same with or without a “no earlier crossing” constraint: the optimal path leaves the fixed point once, at the time that minimizes the action, and reaches at .
Starting from Eqs. (52), we derive a simplified differential equation the solution to which gives the instanton. Importantly, the adjoint equations are decoupled from the coordinate variables, which allows us to solve them independently and substitute their solution into the coordinate equations. We begin with the equation for . The formal solution is obtained by treating as a source term, and using separation of variables and variation of constants, and is given by
| 53 |
where the constant of integration is fixed by the boundary condition . Note that, because the momentum grows exponentially in time, this is why it is preferable to impose a final rather than an initial condition on the momentum, which naturally constrains the unbounded growth of the solution.
Substituting Eq. (53) into the equation for and differentiating both sides with respect to t, we obtain the following second-order differential equation:
| 54 |
This equation must be solved subject to the boundary conditions that , as in Eqs. (52), and , which follows directly from the evolution equation for . The general solution of Eq. (54) is given in terms of special functions as
| 55 |
where and are constants, the first term on the right-hand side is the Tricomi confluent hypergeometric function,
| 56 |
and is the Laguerre function, defined as the solution of Laguerre’s differential equation,
| 57 |
WHICH is non-singular at . It can be shown that Eq. (55) satisfies , ensuring regularity in the limit in Eqs. (52). The same regularity is obtained by formally considering the asymptotic limit of or for or both. In both cases, becomes , so that and hence .
Imposing the two boundary conditions above determines the constants in Eq. (55) as
| 58 |
| 59 |
This completes the solution Eq. (55) of Eq. (54), thereby providing an explicit expression for in Eq. (53).
Differentiating the first equation in (52), substituting the expression for , and using the first equation again to eliminate , we obtain the following second-order differential equation for ,
| 60 |
which can be rewritten, using Eq. (53), as
| 61 |
This equation must be solved using from Eq. (53), from Eq. (55), and boundary conditions and . The latter condition follows directly from the initial condition and the Hamiltonian equation for . The dynamics of can then be recovered by inverting the equation for once the solution to Eq. (61) is known.
In summary, we have reduced the full system Eqs. (52) to a single second-order differential equation (61), whose solution provides the instanton connecting to , which indirectly determines . The solution to Eq. (61) must be computed numerically. We compare the instantons of the OU process and the feedback OU process in Fig. 2 of Section 4.1.
Fig. 2.
Comparison between the instanton solving Eq. (61) for and numerical simulations of the feedback OU process governed by Eqs. (5), for three values of the feedback parameter: (a), (b), and (c). In the small-noise regime , the average simulated trajectory (solid orange line), computed over realizations and post-selected on reaching at time , closely follows the instanton. The shaded area shows one standard deviation around the average. The corresponding memory trajectory , obtained by inverting the equation for in Eq. (52), is also shown. These results support the accuracy of the analytical approximation
Once the instanton solution is obtained, we can compute the action using Eq. (37). Interestingly, by substituting the Hamiltonian (51) along with Hamilton’s equations (52) into Eq. (37), we obtain the following simplified form for the action:
| 62 |
with , since the process starts at the minimum of the potentials V and W given by Eqs. (7) and (8). The weak-noise approximation of the first-passage probability through is then given by Eq. (39), with:
| 63 |
where is prescribed, and is determined from the solution of Hamilton’s equations. The quasi-potential and optimal first-passage time are finally obtained using Eqs. (40) and (41), respectively.
We conclude this section by highlighting the following. A natural approach to deriving the instanton equation would have been to consider the Lagrangian , where angle brackets denote the Euclidean scalar product. This dual Lagrangian derivation would have ensured that we would have to grapple with consequences of the singular nature of the matrix D. Indeed, the kernel of D is one-dimensional and may encode important information about the dynamics. For instance, using the Moore–Penrose inverse1 (pseudo-inverse), one would obtain an equation of the form , the solution of which is the instanton. This equation for , in contrast to Eq. (61), entirely misses the contribution of the adjoint dynamics. Thus, in such a formulation, it would be necessary to supplement the Euler–Lagrange equations with an additional kernel-dependent term, the analysis of which would require alternative methods. On the contrary, within the Hamiltonian framework no such issue arises: the degeneracy of D is not a problem and the adjoint dynamics are fully incorporated into the solution.
Numerical Results and Discussion
Instanton Dynamics
We now compare and contrast the classical and feedback OU processes allowing us to characterize the physical mechanisms responsible for first-passage dynamics. In particular, we present our results on the instanton dynamics leading the feedback OU process to first-passage through , and compare them with the classical OU case.
To solve Eq. (61) numerically we use the solve_ivp solver from the scipy.integrate Python package, employing RK45, which corresponds to the explicit fourth order Runge–Kutta method of integration. In Fig. 2 we compare the numerical solution for the instanton , obtained from the analytical approximation of Eq. (61), and the memory component , with simulations for three representative cases: in Fig. 2(a), (OU process) in Fig. 2(b), and in Fig. 2(c). Clearly, the average behavior of sample-path simulations conducted at finite but small noise compares well with the instanton , thereby supporting our analytical derivation of Eq. (61).
Firstly, the case shown Fig. 2(b) exhibits the classical OU instanton given by Eq. (46). For completeness, we also show the memory , defined in Eq. (2), which does not influence the dynamics in this case. As expected, the classical OU instanton corresponds to a monotonic exponential approach to , as discussed in Section 3.2.
Secondly, Fig. 2 reveals qualitatively distinct first-passage mechanisms for negative and positive values of , although both resemble the classical OU first-passage at late times. For , Fig. 2(a) shows that since it is detrimental to accumulate positive fluctuations toward , the process instead accumulates fluctuations in the opposite direction. Because , this effectively moves the potential minimum toward , storing a form of ‘potential energy’ that is then released through a monotonic exponential approach to . We refer to this behavior as the ‘slingshot’ effect.
For , Fig. 2(c) shows that the process escapes optimally by initially accumulating small fluctuations toward , which are stored in the memory , effectively shifting the minimum of the potential closer to . As time progresses, the memory feedback becomes increasingly rigid due to averaging, which leads to a plateau in the intermediate time regime. However, since the effective potential minimum moves closer to , first-passage becomes less costly in terms of the action (see Fig. 3) and we see the final exponential rise, again akin to the late time behavior of the classical OU case in Fig. 2(b).
Fig. 3.
(a) The action (Eq. 37) as a function of time T, at a fixed , for several values of . The classical OU action is the black dashed line (), with the feedback OU process curves above () and below (). (b) The action plotted as a function of , at fixed , for several values of T. (c) The instanton and memory for and , showing nearly ballistic dynamics associated with the local minimum in (a) and the non-monotonic time behavior in (b). (d) A heatmap of the action plotted as a function of first-passage time T and feedback parameter , at fixed .
As previously noted, for simplicity we restrict our analysis to the case with , but note that the case can be treated analogously. Indeed, changing the sign of simply reflects the instanton trajectory, leaving the qualitative dynamics unchanged. Having described the first-passage mechanisms of the feedback OU process, next we analyze the action function in order to assess the likelihood of these mechanisms.
Time-Dependent Action
As defined in Eq. (37), the action depends both on the first-passage point in Eq. (63) and on the first-passage time T. Moreover, as shown in Eq. (38), the action determines the probability of finding the process at for the first time T. In Fig. 3, we collect a series of plots illustrating the important behavior of this action.
In Fig. 3(a), we plot the action at a fixed first-passage point as a function of time T, for five values of . When () the action lies above (below) the analytical action of the classical OU process given in Eq. (45) and the black dashed line for . This implies that, for a given , the finite-time slingshot mechanism employed by the feedback OU process with is less probable than the accumulation of positive fluctuations occurring for .
For , Fig. 3(a) shows the emergence of a global minimum as , as seen by the convergence of the colored curves towards the OU action from above. Thus, at large times, all first-passage mechanisms asymptote to the standard OU dynamics, confirming the theoretical results on the long-time behavior of self-interacting diffusions discussed in Section 2.4. On the other hand, at short times, solving Hamilton’s equations for reveals that the instanton dynamics is nearly ballistic, as seen in Fig. 3(c).
In contrast, when the global minimum of the action moves from to a finite time, the value of which decreases as decreases. Therefore, accumulating and storing positive fluctuations via memory feedback is always more favorable than escaping through classical OU dynamics. This shift in the global minimum significantly influences the computation of the quasi-potential and optimal first-passage time, as discussed in Section 4.3.
As seen in Fig. 3(b), for , classical OU first-passage dynamics are more probable as T increases, since the action decreases and approaches . This monotonic behavior persists for all negative values of . However, for , the behavior changes: reflecting the emergence of the global minimum shown in Fig. 3(a), wherein the action is minimized at an intermediate time.
A unified means of visualizing the information in Figs. 3(a) and (b) is obtained by constructing a heatmap of the action as a function of and first-passage time T, at fixed , as shown in Fig. 3(d). Here we see that across all times the first-passage dynamics for , be they ballistic or slingshot, are less probable than for , where the accumulation of positive fluctuations dominates. Similarly, in Fig. 4 we plot heatmaps of the action as a function of T and first-passage point , for three fixed values of the feedback parameter . For all values of the figures show that escaping through boundary points farther from the origin is unlikely, and least likely for .
Fig. 4.
Heatmaps of the action as a function of and T for three feedback parameters: (a), (b), and (c)
In Section 4.3, we turn to the time-independent properties of the first-passage dynamics, captured by the quasi-potential in Eq. (40). Additionally, we examine the Kramers first-passage rate and the optimal first-passage time introduced in Eq. (41).
Quasi-Potential and Optimal First-Passage Time
The quasi-potential is the large deviation rate function governing the probability of finding the process at a point in the small-noise limit . It characterizes the effective ‘energy barrier’ that the process must overcome to first-passage through . The larger the quasi-potential, the greater the barrier required for first-passage and the longer the mean first-passage time, which is approximated by the inverse of the probability in the Kramers rate formula in Eq. (39).
Fig. 5(a) shows as a function of the first-passage point for five values of straddling the origin. Clearly, for the quasi-potential lies below the black dashed line corresponding to the OU result given in Eq. (48). This is a consequence of the global minimum of the action occurring at finite time shown in Fig. 3(a). Thus, the most likely mechanism of reaching a given first-passage point must have , and involves the accumulation of positive fluctuations in the memory. Moreover, as increases, this mechanism becomes more likely.
Fig. 5.
(a) Quasi-potential as a function of . Curves for lie below the OU result , indicating that the most likely mechanism is the accumulation of positive fluctuations. For , curves collapse over . (b) Quasi-potential as a function of for several values of . For , the curves decrease monotonically, reinforcing the enhanced likelihood of first-passage via memory accumulation. For , the quasi-potential plateaus at . (c) Heatmap of quasi-potential as a function of and
Importantly, this behavior is significantly different from the scenario in which the empirical occupation measure stabilizes before first-passage, as discussed at the end of Section 2.4 and reflected in the -independent rate function in Eq. (22). We argue that this difference arises because the quasi-potential in Eq. (40), and shown in Fig. 5(a), is the infimum of the action, rather than its value in the infinite-time limit, which would be a sufficient condition for the stabilization of the occupation measure.
As expected, the quasipotential coincides with the OU result when , since the action reaches a global minimum as , as discussed in Fig. 3. In Fig. 5(b), we plot the quasi-potential as a function of for five first-passage points . The plateaus for correspond to the OU value of Eq. (48). For , decreases monotonically with , reflecting the shift of the global minimum of the action to finite times. This reduction in the quasi-potential effectively lowers the escape barrier, thereby accelerating first-passage dynamics by reducing the mean first-passage time, which is approximated by taking the inverse of the Kramers rate in Eq. (39).
As previously, a heatmap perspective combines the behavior in a single figure, in this case the quasi-potential as a function of both and , as shown in Fig. 5(c). Key here is the enhanced first-passage through a fixed with increasing . In summary, Fig. 5 demonstrates that, rather than classical OU dynamics or slingshot effects, memory-accumulated positive fluctuations towards the target is the most likely mechanism underlying the escape process through . The transition between classical OU first-passage dynamics and that due to memory-accumulated positive fluctuations is delineated by .
Finally, in Fig. 6 we plot the optimal first-passage time as a function of for three points . This provides further evidence that the global minimum of the action shifts from infinite time for (with the proviso of the cut off at ) to a finite time for . Near the transition appears abrupt, however, further investigation is needed to fully characterize the nature of this transition. The emergence of a finite optimal first-passage time for reflects the role that feedback plays in accelerating the first-passage dynamics through memory-accumulated fluctuations.
Fig. 6.

Optimal first-passage time as a function of the feedback parameter for several values of first-passage point . For , diverges and saturates at the numerical cutoff . For , the optimal time becomes finite, confirming the acceleration of first-passage dynamics. As , increases again and eventually reaches the cutoff value
Furthermore, for all , the optimal first-passage time initially decreases with , reaches a minimum, and then grows again for larger , eventually reaching the numerical cutoff . This suggests that, although feedback initially accelerates first-passage dynamics, the global minimum of the action gradually shifts to longer times as increases. Recall from Section 2.3 that denotes a bifurcation in the deterministic dynamics at which the origin becomes an unstable fixed point, and small fluctuations can rapidly drive the process away from it, in either direction along the x-axis. Thus, it is difficult to predict the behavior of the optimal time as , but the saturation at in Fig. 6 suggests a divergence in the optimal time, possibly due to the contribution of trajectories escaping to the left.
Conclusion
We have investigated the first-passage dynamics of an exemplary non-Markovian stochastic process. The non-Markovianity is treated by modifying a one-dimensional Ornstein–Uhlenbeck process using a time-averaged feedback wherein the drift is influenced by the empirical mean of its own trajectory. Importantly, this model belongs to the broader class of self-interacting diffusions and provides a simple, yet semi-analytically tractable, framework to study the effect of memory feedback on rare events. Beyond its theoretical appeal, such dynamics are relevant in many contexts, including feedback control systems [18] and reinforcement learning algorithms [14], where history-dependent feedback plays a central role in optimizing task execution and controlling stochastic agents.
We used a weak-noise large deviation approach to derive the equations governing the most probable paths that reach a specified position at a given time, or instantons, and characterized the associated action and quasi-potential. We found that the feedback parameter delineates three qualitatively distinct first-passage mechanisms. When , first-passage is driven by the gradual accumulation of positive fluctuations that shift the effective potential and lead to accelerated dynamics through a reduction of the effective energy barrier at leading order in the noise. The optimal first-passage time, which is the time that minimizes the action, also transitions from being infinite for to finite for , providing further evidence of accelerated first-passage dynamics. When , first-passage occurs either through a slingshot-like mechanism or through an almost-ballistic trajectory, both of which are suboptimal compared to classical OU behavior. Our results establish a clear link between feedback strength and first-passage likelihood, showing that memory can be harnessed to fundamentally reshape the statistics of rare events.
In addition to its basic theoretical relevance, our study suggests more broadly that empirical occupation measures could act as versatile memory devices in feedback-controlled systems. In particular, we suggest that one could encode information on transient rare events into a feedback mechanism that influences future dynamics. This feedback, naturally expressible in terms of the empirical occupation measure, is especially responsive on short time scales, where it adapts quickly to new information. On longer time scales, it relaxes towards the invariant distribution, becoming progressively harder to manipulate. This feature makes it particularly suitable for capturing finite-time rare events and using that memory to modulate a system response in a closed-loop control setting, effectively guiding the process to behave in a desired manner.
Our work suggests a number of immediate questions for future research. The model studied here is a continuous-space, continuous-time self-interacting process. It is instructive, however, to draw a parallel with the Elephant Random Walk (ERW) and the Gaussian ERW (GERW) [5, 11], defined in discrete space–time and in continuous space with discrete time respectively. Despite these differences in construction, their fluctuation paths—featuring giant leaps and long excursions—are reminiscent of the first-passage dynamics we observe for , where beneficial fluctuations are retained through memory feedback. The analogy, however, cannot be pushed too far: the ERW and GERW couple to the empirical current, whereas our model couples to the empirical density; in particular its mean. This raises the intriguing possibility of a broader “universality class” of fluctuation mechanisms in self-interacting processes, extending beyond the specific interaction considered here.
In designing feedback-controlled systems that use self-interaction as a versatile memory device, both theoretical questions and practical implementations are of interest. A compelling direction is to examine the role of finite-time window memory as a feedback mechanism. Rather than considering a full-time average of the form , we suggest introducing a feedback based on a time-dependent finite-time window, such as . This formulation is truly non-Markovian and, in contrast to the model considered here, does not appear to admit a mapping to a finite-dimensional Markov system. Therefore, within the framework developed in this paper, one could also explore alternative forms of feedback, including for example those governed by the empirical variance or standard deviation of the process, rather than the empirical mean. Finally, it is of interest to analyze the effect of a changing environment by incorporating multiplicative noise, explicitly time-dependent noise or a non-autonomous deterministic backbone.
Acknowledgements
FC gratefully acknowledges stimulating discussions with Cristobal Arratia, Ralf Eichhorn, and Supriya Krishnamurthy at Nordita and with Baruch Meerson at Les Houches during the School ‘Theory of Large Deviations and Applications’ held in July 2024. FC is also thankful to Tobias Grafke for insightful discussions and for pointing out the correct final conditions to be used when solving Hamilton’s equations in (52). The authors are grateful to Martin Rosinberg for stimulating discussions and for pointing out, after the arXiv submission, a numerical error in the code used to compute the action. This has since been corrected without qualitatively affecting the main results. The authors are also grateful to an anonymous referee for pointing out how to obtain Eq. (62). RD was supported by a scholarship from ENS Paris-Saclay (FR) during his Master 1 internship at Nordita. We acknowledge partial support from the Swedish Research Council under Grant No. 638-2013-9243 and from EPSRC under Grant No. EP/V031201/1.
Appendices
Derivation of boundary value problem for mean first-passage time in Eq. (30)
The boundary value problem in (30) can be obtained by adapting the derivation of result 7.1 in [73] by considering a time-dependent generator, such as that in (23). Given the definition of mean first-passage time in (25) and in terms of the survival probability (28), we can write
| 64 |
| 65 |
| 66 |
which, by first integrating by parts and by then using (29), can be re-written as
| 67 |
| 68 |
Then, by introducing the semigroup associated to the time-dependent infinitesimal generator (23) and using the initial condition in (27) we find
| 69 |
| 70 |
| 71 |
Applying on both sides, we obtain
| 72 |
which, after integrating by parts, yields
| 73 |
and , for .
Footnotes
Given any matrix , its Moore–Penrose inverse is the unique matrix satisfying the following four conditions: (1) , (2) , (3) and (4) . In the special case that A is square and invertible: . In the case discussed in the main text, and .
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
References
- 1.Beran, J., Feng, Y., Ghosh, S., Kulik, R.: Long-memory processes: Probabilistic properties and statistical methods. Springer-Verlag, Berlin Heidelberg (2013) [Google Scholar]
- 2.Scalliet, C., Berthier, L.: Rejuvenation and memory effects in a structural glass. Phys. Rev. Lett. 122, 255502 (2019) [DOI] [PubMed] [Google Scholar]
- 3.Zhang, J., Zhou, T.: Markovian approaches to modeling intracellular reaction processes with molecular memory. Proc. Natl. Acad. Sci. U.S.A. 116, 23542–23550 (2019) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.Song, Y.M., Campbell, S., Shiau, L., Kim, J.K., Ott, W.: Noisy delay denoises biochemical oscillators. Phys. Rev. Lett. 132, 078402 (2024) [DOI] [PubMed] [Google Scholar]
- 5.Schütz, G.M., Trimper, S.: Elephants can always remember: Exact long-range memory effects in a non-markovian random walk. Phys. Rev. E 70(4), 045101 (2004)
- 6.Rebenshtok, A., Barkai, E.: Distribution of time-averaged observables for weak ergodicity breaking. Phys. Rev. Lett. 99(21), 210601 (2007) [DOI] [PubMed] [Google Scholar]
- 7.Maes, C., Netočný, K., Wynants, B.: Dynamical fluctuations for semi-markov processes. J. Phys. A: Math. Theor. 42, 365002 (2009) [Google Scholar]
- 8.Harris, R.J., Touchette, H.: Current fluctuations in stochastic systems with long-range memory. J. Phys. A: Math. Theor. 42(34), 342001 (2009) [Google Scholar]
- 9.Mieghem, P.V., Bovenkamp, R.V.D.: Non-markovian infection spread dramatically alters the susceptible- infected-susceptible epidemic threshold in networks. Phys. Rev. Lett. 110, 108701 (2013) [DOI] [PubMed] [Google Scholar]
- 10.Harris, R.J.: Fluctuations in interacting particle systems with memory. J. Stat. Mech: Theory Exp. 2015(7), P07021 (2015) [Google Scholar]
- 11.Jack, R.L., Harris, R.J.: Giant leaps and long excursions: Fluctuation mechanisms in systems with long-range memory. Phys. Rev. E 102(1), 012154 (2020) [DOI] [PubMed] [Google Scholar]
- 12.Moran, J., Fosset, A., Luzzati, D., Bouchaud, J.-P., Benzaquen, M.: “By force of habit: Self-trapping in a dynamical utility landscape,” SSRN Electronic Journal, (2020)
- 13.Hartich, D., Godec, A.: Emergent memory and kinetic hysteresis in strongly driven networks. Phys. Rev. X 11, 041047 (2021) [Google Scholar]
- 14.Sutton, R.S., Barto, A.G.: Reinforcement Learning: An Introduction, 2nd edn. A Bradford Book (1998) [Google Scholar]
- 15.Bechhoefer, J.: Feedback for physicists: a tutorial essay on control. Rev. Mod. Phys. 77, 783–836 (2005) [Google Scholar]
- 16.Puterman, M.L.: Markov decision processes: Discrete stochastic dynamic programming. Wiley, Hoboken (2008) [Google Scholar]
- 17.Khadem, S.M., Klapp, S.H.: Delayed feedback control of active particles: A controlled journey towards the destination. Phys. Chem. Chem. Phys. 21, 13776–13787 (2019) [DOI] [PubMed] [Google Scholar]
- 18.Bechhoefer, J.: Control Theory for Physicists. Cambridge University Press, Cambridge (2021) [Google Scholar]
- 19.Horowitz, J.M., Vaikuntanathan, S.: Nonequilibrium detailed fluctuation theorem for repeated discrete feedback. Phys. Rev. E Stat. Nonlinear Soft Matter Phys. 82, 061120 (2010) [Google Scholar]
- 20.Sagawa, T., Ueda, M.: Nonequilibrium thermodynamics of feedback control. Phys. Rev. E Stat. Nonlinear Soft Matter Phys. 85, 021104 (2012) [Google Scholar]
- 21.Jun, Y., Gavrilov, M., Bechhoefer, J.: High-precision test of landauer’s principle in a feedback trap. Phys. Rev. Lett. 113, 190601 (2014) [DOI] [PubMed] [Google Scholar]
- 22.Loos, S.A., Klapp, S.H.: Heat flow due to time-delayed feedback. Sci. Rep. 9, 1–11 (2019) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23.Rico-Pasto, M., Schmitt, R.K., Ribezzi-Crivellari, M., Parrondo, J.M., Linke, H., Johansson, J., Ritort, F.: Dissipation reduction and information-to-measurement conversion in dna pulling experiments with feedback protocols. Phys. Rev. X 11, 031052 (2021) [Google Scholar]
- 24.Debiossac, M., Rosinberg, M.L., Lutz, E., Kiesel, N.: Non-markovian feedback control and acausality: an experimental study. Phys. Rev. Lett. 128, 200601 (2022) [DOI] [PubMed] [Google Scholar]
- 25.Kopp, R.A., Klapp, S.H.L.: Heat production in a stochastic system with nonlinear time-delayed feedback. Phys. Rev. E 110, 054126 (2024) [DOI] [PubMed] [Google Scholar]
- 26.Tóth, B.: Self-Interacting Random Motions, In: European Congress of Mathematics, pp. 555–564, (2001)
- 27.Benaïm, M., Ledoux, M., Raimond, O.: Self-interacting diffusions. Probab. Theory Relat. Fields 122(1), 1–41 (2002) [Google Scholar]
- 28.Pemantle, R.: A survey of random processes with reinforcement. Probab. Surv. 4(1), 1–79 (2007) [Google Scholar]
- 29.Del Moral, P., Miclo, L.: Self-interacting markov chains. Stoch. Anal. Appl. 24(3), 615–660 (2007) [Google Scholar]
-
30.Kurtzmann, A.: The ode method for some self-interacting diffusions on
. Annales de l’institut Henri Poincare (B) Probability and Statistics 46(3), 618–643 (2010) [Google Scholar] - 31.Brémont, J., Bénichou, O., Voituriez, R.: Exact propagators of one-dimensional self-interacting random walks. Phys. Rev. Lett. 133, 157101 (2024) [DOI] [PubMed] [Google Scholar]
- 32.Coghi, F., Garrahan, J. P.: Self-interacting processes via Doob conditioning, arXiv: 2503.01574, (2025)
- 33.Budrene, E.O., Berg, H.C.: Dynamics of formation of symmetrical patterns by chemotactic bacteria. Nature 376(6535), 49–53 (1995) [DOI] [PubMed] [Google Scholar]
- 34.Tsori, Y., de Gennes, P.-G.: Self-trapping of a single bacterium in its own chemoattractant. Europhys. Lett. 66(4), 599 (2004) [Google Scholar]
- 35.Reid, C.R., Latty, T., Dussutour, A., Beekman, M.: Slime mold uses an externalized spatial “memory’’ to navigate in complex environments. Proc. Natl. Acad. Sci. U.S.A. 109(43), 17490–17494 (2012) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 36.Zhao, K., Tseng, B.S., Beckerman, B., Jin, F., Gibiansky, M.L., Harrison, J.J., Luijten, E., Parsek, M.R., Wong, G.C.: Psl trails guide exploration and microcolony formation in pseudomonas aeruginosa biofilms. Nature 497(7449), 388–391 (2013) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37.Kranz, W.T., Gelimson, A., Zhao, K., Wong, G.C., Golestanian, R.: Effective dynamics of microorganisms that interact with their own trail. Phys. Rev. Lett. 117(3), 038101 (2016) [DOI] [PubMed] [Google Scholar]
- 38.Nakayama, B., Nagase, H., Takahashi, H., Saito, Y., Hatayama, S., Makino, K., Yamamoto, E., Saiki, T.: Tunable pheromone interactions among microswimmers. Proc. Natl. Acad. Sci. U.S.A. 120(9), e2213713120 (2023) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39.Kumar, M., Murali, A., Subramaniam, A.G., Singh, R., Thutupalli, S.: Emergent dynamics due to chemo-hydrodynamic self-interactions in active polymers. Nat. Commun. 15, 1–12 (2024) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40.Chen, W., Izzet, A., Zakine, R., Clément, E., Vanden-Eijnden, E., Brujic, J.: Evolving motility of active droplets is captured by a self-repelling random walk model. Phys. Rev. Lett. 134, 018301 (2025) [DOI] [PubMed] [Google Scholar]
- 41.Budhiraja, A., Waterbury, A., Zoubouloglou, P.: Large Deviations for Empirical Measures of Self Interacting Markov Chains, Stochastic Processes and their Applications, Volume 186, August 2025, 104640
- 42.Budhiraja, A., Waterbury, A., Zoubouloglou, P.: Large Deviations for Empirical Measures of Self-Interacting Markov Chains, arXiv:2304.01384, (2023)
- 43.Coghi, F.: Current fluctuations of a self-interacting diffusion on a ring. J. Phys. A: Math. Theor. 58, 015002 (2024) [Google Scholar]
- 44.Barbier-Chebbah, A., Bénichou, O., Voituriez, R.: Self-interacting random walks: aging, exploration, and first-passage times. Phys. Rev. X 12, 011052 (2022) [Google Scholar]
- 45.Aleksian, A., Moral, P., Kurtzmann, A., Tugaut, J.: Self-interacting diffusions: Long-time behaviour and exit-problem in the uniformly convex case. ESAIM: Probability and Statistics 28, 46–61 (2024) [Google Scholar]
- 46.Kappler, J., Daldrop, J. O., Brünig, F. N., Boehle, M. D., Netz, R. R.: Memory-induced acceleration and slowdown of barrier crossing, Journal of Chemical Physics, 148, (2018)
- 47.Barbier-Chebbah, A., Bénichou, O., Voituriez, R., Guérin, T.: Long-term memory induced correction to 858 arrhenius law. Nat. Commun. 15, 7408 (2024) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 48.Norris, J.R., Rogers, L.C., Williams, D.: Self-avoiding random walk: A brownian motion model with local time drift. Probab. Theory Relat. Fields 74(2), 271–287 (1987) [Google Scholar]
- 49.Durrett, R.T., Rogers, L.C.: Asymptotic behavior of brownian polymers. Probab. Theory Relat. Fields 92(3), 337–349 (1992) [Google Scholar]
- 50.Cranston, M., Jan, Y.: Self attracting diffusions: Two case studies. Math. Ann. 303(1), 87–93 (1995) [Google Scholar]
- 51.Raimond, O.: Self-attracting diffusions: Case of the constant interaction. Probab. Theory Relat. Fields 107(2), 177–196 (1997) [Google Scholar]
- 52.Benaïm, M., Raimond, O.: Self-interacting diffusions II: Convergence in law. Annales de l’Institut Henri Poincare (B) Probability and Statistics 39(6), 1043–1055 (2003) [Google Scholar]
- 53.Benaïm, M., Raimond, O.: Self-interacting diffusions. III. Symmetric interactions. Ann. Probab. 33(5), 1716–1759 (2005) [Google Scholar]
- 54.Benaïm, M., Raimond, O.: Self-interacting diffusions IV: Rate of convergence. Electron. J. Probab. 16(66), 1815–1843 (2011) [Google Scholar]
- 55.Chambeu, S., Kurtzmann, A.: Some particular self-interacting diffusions: Ergodic behaviour and almost sure convergence. Bernoulli 17(4), 1248–1267 (2011) [Google Scholar]
- 56.Kleptsyn, V., Kurtzmann, A.: Ergodicity of self-attracting motion. Electron. J. Probab. 17, 1–37 (2012) [Google Scholar]
- 57.Aleksian, A., Del Moral, P., Kurtzmann, A., Tugaut, J.: On the exit-problem for self-interacting diffusions, arXiv:2201.10428v1, (2022)
- 58.Aleksian, A.: Problème de Temps de Sortie pour les Diffusions Auto-interagissantes et Auto-stabilisantes. PhD thesis, (2024)
- 59.Krapivsky, P.L., Meerson, B.: Finite-time blowup of a brownian particle in a repulsive potential, Phys. Rev. E 112, 024128
- 60.Carrillo, J.A., McCann, R.J., Villani, C.: Kinetic equilibration rates for granular media and related equations: entropy dissipation and mass transportation estimates. Revista Matemática Iberoamericana 19, 971–1018 (2003) [Google Scholar]
- 61.Moon, W., Balmforth, N.J., Wettlaufer, J.S.: Nonadiabatic escape and stochastic resonance. J. Phys. A: Math. Theor. 53, 095001 (2020) [Google Scholar]
- 62.Moon, W., Giorgini, L.T., Wettlaufer, J.S.: Analytical solution of stochastic resonance in the nonadiabatic regime. Phys. Rev. E 104, 044130 (2021) [DOI] [PubMed] [Google Scholar]
- 63.Giorgini, L.T., Moon, W., Wettlaufer, J.S.: Analytical survival analysis of the Ornstein-Uhlenbeck process. J. Stat. Phys. 181, 2404–2414 (2020) [Google Scholar]
- 64.Giorgini, L.T., Moon, W., Wettlaufer, J.S.: Analytical survival analysis of the non-autonomous Ornstein-Uhlenbeck process. J. Stat. Phys. 191, 1–18 (2024) [Google Scholar]
- 65.Freidlin, M.I., Wentzell, A.D.: Random Perturbations of Dynamical Systems. Springer, Berlin (1984) [Google Scholar]
- 66.Touchette, H.: The large deviation approach to statistical mechanics. Phys. Rep. 478(1–3), 1–69 (2009) [Google Scholar]
- 67.Grafke, T., Vanden-Eijnden, E.: Numerical computation of rare events via large deviation theory. Chaos 29, 63118 (2019) [Google Scholar]
- 68.Assaf, M., Meerson, B.: Wkb theory of large deviations in stochastic populations. J. Phys. A: Math. Theor. 50, 263001 (2017) [Google Scholar]
- 69.Weber, M.F., Frey, E.: Master equations and the theory of stochastic path integrals. Rep. Prog. Phys. 80, 046601 (2017) [DOI] [PubMed] [Google Scholar]
- 70.Arnold, V.I.: Mathematical Methods of Classical Mechanics, vol. 60. Springer, New York (1989) [Google Scholar]
- 71.Goldstein, H., Poole, C., Safko, J.: Classical Mechanics. (2001)
- 72.Risken, H.: The Fokker-Planck Equation, 2nd edn. Springer, Berlin Heidelberg (1996) [Google Scholar]
- 73.Pavliotis, G.A.: Stochastic Processes and Applications, vol. 60. Springer, New York (2014) [Google Scholar]
- 74.Schulz, M.: Control Theory in Physics and other Fields of Science. Springer-Verlag, Berlin (2006) [Google Scholar]





