Skip to main content
Springer logoLink to Springer
. 2018 Feb 15;170(6):1019–1050. doi: 10.1007/s10955-018-1986-0

Canonical Structure and Orthogonality of Forces and Currents in Irreversible Markov Chains

Marcus Kaiser 1,, Robert L Jack 2,3,4, Johannes Zimmer 1
PMCID: PMC6566214  PMID: 31258181

Abstract

We discuss a canonical structure that provides a unifying description of dynamical large deviations for irreversible finite state Markov chains (continuous time), Onsager theory, and Macroscopic Fluctuation Theory (MFT). For Markov chains, this theory involves a non-linear relation between probability currents and their conjugate forces. Within this framework, we show how the forces can be split into two components, which are orthogonal to each other, in a generalised sense. This splitting allows a decomposition of the pathwise rate function into three terms, which have physical interpretations in terms of dissipation and convergence to equilibrium. Similar decompositions hold for rate functions at level 2 and level 2.5. These results clarify how bounds on entropy production and fluctuation theorems emerge from the underlying dynamical rules. We discuss how these results for Markov chains are related to similar structures within MFT, which describes hydrodynamic limits of such microscopic models.

Keywords: Nonequilibrium dynamical fluctuations, Large deviations, Microscopic fluctuation theory, Irreversible Markov chains

Introduction

We consider dynamical fluctuations in systems described by Markov chains. The nature of such fluctuations in physical systems constrains the mathematical models that can be used to describe them. For example, there are well-known relationships between equilibrium physical systems and detailed balance in Markov models [20, Sect. 5.3.4]. Away from equilibrium, fluctuation theorems [12, 19, 25, 32, 37] and associated ideas of local detailed balance [32, 39] have shown how the entropy production of a system must be accounted for correctly when modelling physical systems. However, the mathematical structures that determine the probabilities of non-equilibrium fluctuations are still only partially understood.

We characterise dynamical fluctuations using an approach based on the Onsager–Machlup (OM) theory [36], which is concerned with fluctuations of macroscopic properties of physical systems (for example, density or energy). Associated to these fluctuations is a large-deviation principle (LDP), which encodes the probability of rare dynamical trajectories. The classical ideas of OM theory have been extended in recent years, through the Macroscopic Fluctuation Theory (MFT) of Bertini et al. [7]. This theory uses an LDP to describe path probabilities for the density and current in diffusive systems, on the hydrodynamic scale. At the centre of MFT is a decomposition of the current into two orthogonal terms, one of which is symmetric under time-reversal, and another which is anti-symmetric. The resulting theory is a general framework for the analysis of dynamical fluctuations in a large class of non-equilibrium systems. It also connects dynamical fluctuations with thermodynamic quantities like free energy and entropy production, and with associated non-equilibrium objects like the quasi-potential (which extends the thermodynamic free energy to non-equilibrium settings).

Here, we show how several features that appear in MFT can be attributed to a general structure that characterises dynamical fluctuations in microscopic Markov models. That is, the properties of the hydrodynamic (MFT) theory can be traced back to the properties of the underlying stochastic processes. Our approach builds on recent work by Mielke, Renger and M. A. Peletier, in which the analogue of the OM theory for reversible Markov chains has been described in terms of a generalised gradient-flow structure [43]. To describe non-equilibrium processes, that theory must be generalised to include irreversible Markov chains. This can be achieved using the canonical structure of fluctuations discovered by Maes and Netočný [38]. Extending their approach, we decompose currents in the system into two parts, and we identify a kind of orthogonality relationship associated with this decomposition. However, in contrast to the classical OM theory and to MFT, the large deviation principles that appear in our approach have non-quadratic rate functions, which means that fluxes have non-linear dependence on their conjugate forces. Thus, the idea of orthogonality between currents needs to be generalised, just as the notion of gradient flows in macroscopic equilibrium systems can be extended to generalised gradient flows.

The central players in our analysis are the probability density ρ and the probability current j. For a given Markov chain, the relation between these quantities is fully encoded in the master equation, which also fully specifies the dynamical fluctuations in that model. However, thermodynamic aspects of the system—the roles of heat, free energy, and entropy production—are not apparent in the master equation. Within the Onsager–Machlup theory, these thermodynamic quantities appear in the action functional for paths, and solutions of the master equation appear as paths of minimal action. Hence, the structure that we discuss here, and particularly the decomposition of the current into two components, links the dynamical properties of the system to thermodynamic concepts, both for equilibrium and non-equilibrium systems.

Summary

We now sketch the setting considered in this article (precise definitions of the systems of interest and the relevant currents, densities and forces will be given in Sect. 2).

We introduce a large parameter N, which might be the size of the system (as in MFT) or a large number of copies of the system (an ensemble), as considered for Markov chains in [39]. Then let (ρ^tN,ȷ^tN)t[0,T] be the (random) path followed by the system’s density and current, in the time interval [0, T]. Consider a random initial condition such that Prob(ρ^0Nρ)exp[-NI0(ρ)], asymptotically as N, for some rate functional I0. Paths that in addition satisfy a continuity equation ρ˙+divj=0 have the asymptotic probability

Probρ^tN,ȷ^tNt[0,T](ρt,jt)t[0,T]exp-NI[0,T](ρt,jt)t[0,T] 1

with the rate functional

I[0,T]((ρt,jt)t[0,T])=I0(ρ0)+120TΦ(ρt,jt,F(ρt))dt; 2

here F(ρt) is a force (see (12) below for the precise definition) and Φ is what we call the generalised OM functional, which has the general form

Φ(ρ,j,f):=Ψ(ρ,j)-j·f+Ψ(ρ,f), 3

where j·f is a dual pairing between a current j and a force f, while Ψ and Ψ are a pair of functions which satisfy

Ψ(ρ,f)=supj[j·f-Ψ(ρ,j)],andΨ(ρ,j)=supf[j·f-Ψ(ρ,f)], 4

as well as Ψ(ρ,f)=Ψ(ρ,-f) and Ψ(ρ,j)=Ψ(ρ,-j). Note that (4) means that the two functions satisfy a Legendre duality. Moreover, these two functions Ψ and Ψ are strictly convex in their second arguments. Here and throughout, f indicates a force, while F is a function whose (density-dependent) value is a force.

The large deviation principle stated in (1) is somewhat abstract: for example, ρ^tN might be defined as a density on a discrete space or on Rd, depending on the system of interest. Specific examples will be given below. In addition, all microscopic parameters of the system (particle hopping rates, diffusion constants, etc.) will enter the (system-dependent) functions Ψ, Ψ and F.

As a preliminary example, we recall the classical Onsager theory [36], in which one considers n currents j=(jα)α=1n and a set of conjugate applied forces F=(Fα)α=1n. Examples of currents might be particle flow or heat flow, and the relevant forces might be pressure or temperature gradients. The large parameter N corresponds to the size of a macroscopic system. The theory aims to to describe the typical (average) response of the current j to the force F, and also the fluctuations of j. In this (simplest) case, the density ρ plays no role, so the force F has a fixed value in Rn. The dual pairing is simply j·f=αjαfα and Ψ is given by Ψ(ρ,j)=12α,βjαRαβjβ, where R is a symmetric n×n matrix with elements Rαβ. The Legendre dual of Ψ is Ψ(ρ,f)=12α,βfαLαβfβ, where L=R-1 is the Onsager matrix, whose elements are the linear response coefficients of the system. One sees that Ψ and Ψ can be interpreted as squared norms for currents and forces respectively. Denoting this norm by jL-12:=Ψ(ρ,j), one has

Φ(ρ,j,f)=j-LfL-12. 5

On applying an external force F, the response of the current j is obtained as the minimum of Φ, so j=LF (that is, jα=βLαβFβ). One sees that Φ measures the deviation of the current j from its expected value LF, within an appropriate norm. From the LDP (1), one sees that the size of this deviation determines the probability of observing a current fluctuation of this size.

In this article, we show in Sect. 2 that finite Markov chains have an LDP rate functional of the form (3), where Φ (and thus Ψ) are not quadratic. In that case, ρ and j correspond to probability densities and probability currents, while the transition rates of the Markov chain determine the functions F, Ψ and Ψ. Since Ψ and Ψ measure respectively the sizes of the currents and forces, we interpret them as generalisations of the squared norms that appear in the classical case. The resulting Φ is not a squared norm, but it is still a non-negative function that measures the deviation of j from its most likely value. This leads to nonlinear relations between forces and currents. The MFT theory [7] also fits in this framework, as we show in Sect. 4: in that case ρ,j are a particle density and a particle current. However, there are relationships between the functions Φ for MFT and for general Markov chains, as we discuss in Sect. 4.5.

Hence, the general structure of Eqs. (1)–(4) describes classical OM theory [36], MFT, and finite Markov chains. A benefit is that the terms have a physical interpretation. For a path (ρ,j), the time-reversed path is (ρt,jt):=(ρT-t,-jT-t). Since both Ψ and Ψ are symmetric in their second argument and thus invariant under time reversal, it holds that Φ(ρ,j,f)-Φ(ρ,j,f)=-2j·f. This allows us to identify j·F(ρ) as a rate of entropy production. In contrast, the term Ψ(ρ,j)+Ψ(ρ,F(ρ)) is symmetric under time reversal and encodes the frenesy (see [3]). Thus, within this general structure, the physical significance of Eqs. (1)–(4) is that they connect path probabilities to physical notions such as force, current, entropy production and breaking of time-reversal symmetry. Furthermore, we introduce in Sect. 3 decompositions of forces and the (path-wise) rate functional. Sect. 4 shows that some results of MFT originate from generalised orthogonalities of the underlying Markov chains derived in Sect. 3. Similar results hold for time-average large deviation principles, as shown in Sect. 5. In Sect. 6, we show how some properties of MFT can be derived directly from the canonical structure (1)–(4), independent of the specific models of interest. Hence these results of MFT have analogues in Markov chains. Finally we briefly summarise our conclusions in Sect. 7.

Onsager–Machlup Theory for Markov Chains

In this section, we collect results on forces and currents in Markov chains and on associated LDPs. In particular, we recall the setting of [38, 39]; other references for this section are for example [49] (for the definition of forces and currents in Markov chains) and [43] for LDPs.

Setting

We consider an irreducible continuous time Markov chain Xt on a finite state space V with a unique stationary distribution π that satisfies π(x)>0 for all xV. The transition rate from state x to state y is denoted with rxy. We assume that rxy>0 if and only if ryx>0.

We restrict to finite Markov chains for simplicity: the theory can be extended to countable state Markov chains, but this requires some additional assumptions. Briefly, one requires that the Markov chain should be positively recurrent and ergodic (see for instance [9]), for which it is sufficient that (i) the transition rates are not degenerate: yVrxy< for all xV, and (ii) for each xV, the Markov chain started in x almost all trajectories of the Markov chain do not exhibit infinitely many jumps in finite time (“no explosion”). Second, one has to invoke a summability condition for the currents considered below (see, e.g., Eqs. 9 and 10), such that in particular the discrete integration by parts (or summation by parts) formula (15) holds. Finally, note that the cited result for existence and uniqueness of the optimal control potential (the solution to (70)) is only valid for finite state Markov chains.

As usual, we can interpret the state space of the Markov chain as a directed graph with vertices V and edges E=xy|x,yV,rxy>0, such that xyE if and only if yxE. Let ρ be a probability measure on V. We define rescaled transition rates with respect to π as

qxy:=π(x)rxy, 6

so that ρ(x)rxy=ρ(x)π(x)qxy. With this notation, the detailed balance condition π(x)rxy=π(y)ryx reads qxy=qyx, so this equality holds precisely if the Markov chain is reversible (i.e. satisfies detailed balance). In general (not assuming reversibility), since π is the invariant measure for the Markov chain, one has (for all x) that

y(qxy-qyx)=0. 7

We further define the free energy F on V to be the relative entropy (or Kullback–Leibler divergence) with respect to π,

F(ρ):=xρ(x)log(ρ(x)π(x)). 8

The probability current J(ρ) is defined as [49, Eq. (7.4)]

Jxy(ρ):=ρ(x)rxy-ρ(y)ryx. 9

Moreover, for a general current j such that jxy=-jyx, we define the divergence as

divj(x):=yVjxy. 10

We say that j is divergence free if divj(x)=0 for every xV. The time evolution of the probability density ρ is then given by the master equation

ρ˙t=-divJ(ρt) 11

(which is often stated as ρ˙t=Lρt, with the (forward) generator L).

Non-linear Flux–Force Relation and the Associated Functionals Ψ and Ψ

To apply the theory outlined in Sect. 1.1, the next step is to identify the appropriate forces F(ρ) and also a set of mobilities a(ρ). In this section we define these forces, following [38, 39, 49]. This amounts to a reparameterisation of the rates of the Markov process in terms of physically-relevant variables: an example is given in Sect. 3.5.

To each edge in E we assign a force F and a mobility a, as

Fxy(ρ):=logρ(x)rxyρ(y)ryxandaxy(ρ):=2ρ(x)rxyρ(y)ryx. 12

Note that Fxy=-Fyx, while axy=ayx: forces have a direction but the mobility is a symmetric property of each edge. The fact that Fxy depends on the density ρ means that these forces act in the space of probability distributions. This definition of the force is sometimes also called affinity [49, Eq. (7.5)]; see also [1]. With this definition, the probability current (9) is

Jxy(ρ)=axy(ρ)sinh(12Fxy(ρ)), 13

which may be verified directly from the definition sinh(x)=(ex-e-x)/2. In contrast to the classical OM theory, this is a non-linear relation between forces and fluxes, although one recovers a linear structure for small forces (recall the classical theory in Sect. 1.1, for which j=Lf).

Now consider a current j defined on E, with jxy=-jyx, and a general force f that satisfies fxy=-fyx (which is not in general given by (12)). Define a dual pair on E as

j·f:=12xyjxyfxy, 14

where the summation is over all xyE (the normalisation 1 / 2 appears because each connected pair of states should be counted only once, but E is a set of directed edges, so it contains both xy and yx, which have the same contribution to j·f).

We define the discrete gradient g by x,yg:=g(y)-g(x). The discrete gradient and the divergence defined in (10) satisfy a discrete integration by parts formula: for any function g:VR, since jxy=-jyx, we have

-xVg(x)divj(x)=12xyjxyx,yg=j·g. 15

We will show in Sect. 2.3 that there is an OM functional associated with these forces and currents, which is of the form (3). Since Ψ and Ψ are convex and related by a Legendre transformation, it is sufficient to specify only one of them. The appropriate choice turns out to be

Ψ(ρ,f):=xyaxy(ρ)(cosh(12fxy)-1). 16

This means that Φ(ρ,j,f) defined in (3) is uniquely minimised for the current jxy=jxyf(ρ) with

jxyf(ρ)=2(δΨ/δf)xy=axy(ρ)sinh(fxy/2), 17

as required for consistency with (13). From (4) and (14), one has also

Ψ(ρ,j)=12xyjxyfxyj(ρ)-xyaxy(ρ)(cosh(12fxyj(ρ))-1), 18

where

fxyj(ρ):=2arcsinhjxy/axy(ρ) 19

is the force required to induce the current j.

Physically, Ψ(ρ,f) is a measure of the strength of the force f and Ψ(ρ,j) is a measure of the magnitude of the current j. Consistent with this interpretation, note that Ψ and Ψ are symmetric in their second arguments. Moreover, for small forces and currents, Ψ and Ψ are quadratic in their second arguments, and can be interpreted as generalisations of squared norms of the force and current respectively. Note that Eqs. (16) and (18) can alternatively be represented as

Ψ(ρ,j)=xy[12jxyfxyj(ρ)-jxy2+axy(ρ)2+axy(ρ)] 20

and

Ψ(ρ,f):=xy[jxyf(ρ)2+axy(ρ)2-axy(ρ)]. 21

Large Deviations and the Onsager–Machlup Functional

As anticipated in Sect. 1.1, the motivation for the definitions of Ψ, Ψ, and F is that there is a large deviation principle for these Markov chains, whose rate function is of the form given in (2). This large deviation principle appears when one considers N independent copies of the Markov chain.

We denote the ith copy of the Markov chain by Xti and define the empirical density for this copy as ρ^ti(x)=δXti,x, where δ is a Kronecker delta function. Let the times at which the Markov chain Xti has jumps in [0, T] be t1i,t2i,,tKii. Further denote the state just before the kth jump with xk-1i (such that the state after the kth jump is xki). With this, the empirical current is given by

ȷ^tixy=k=1Ki(δx,xk-1iδy,xki-δy,xk-1iδx,xki)δt-tki,

where δ(t-tk) denotes a Dirac delta. Note that (ȷ^ti)xy=-(ȷ^ti)yx and the total probability is conserved, as xdivȷ^ti(x)=0 (which holds for any discrete vector field with (ȷ^ti)xy=-(ȷ^ti)yx). With a slight abuse of notation we define a similar empirical density and current for the full set of copies as

ρ^tN:=1Ni=1Nρ^ti,andȷ^tN:=1Ni=1Nȷ^ti. 22

Next, we state the large deviation principle where the OM functional appears. For this, we fix a time interval [0, T] and consider the large N limit. We assume that the N copies at time t=0 have initial conditions drawn from the invariant measure of the process (the generalisation to other initial conditions is straightforward). Then, the probability to observe a joint density and current (ρt,jt)t[0,T] over the time interval [0, T] is in the limit as N given by (1). That is,

Prob((ρ^tN,ȷ^tN)t[0,T](ρt,jt)t[0,T])exp{-NI[0,T]((ρt,jt)t[0,T])} 23

with

I[0,T]((ρt,jt)t[0,T])=F(ρ0)+120TΦ(ρt,jt,F(ρt))dtifρ˙t+divjt=0+otherwise 24

Here, F(ρ) is the force defined in (12) and the condition ρ˙t+divjt=0 has to hold for almost all t[0,T]. Moreover, Φ is of the form Φ(ρ,j,f)=Ψ(ρ,j)-j·f+Ψ(ρ,f) stated in (3), and the relevant functions Ψ, Ψ and F are those of (16), (18) and (8). This LDP was formally derived in [38, 39]. Since the quantities defined in (22) are simple averages over independent copies of the same Markov chain, this LDP may also be proven by direct application of Sanov’s theorem, which provides an interpretation of I[0,T] as a relative entropy between path measures; we sketch the derivation in Appendix A. For finite-state Markov chains, (23) and (24) also follow (by contraction) from [48, Theorem 4.2], which provides a rigorous proof.

We emphasise that the arguments ρ and j of the function Φ correspond to the random variables that appear in the LDP, while the functions F, Ψ and Ψ that appear in Φ encapsulate the transition rates of the Markov chain. Thus, by reparameterising the rates rxy in terms of forces F and mobilities a, we arrive at a representation of the rate function which helps to make its properties transparent (convexity, positivity, symmetries such as (25)).

We note that for reversible Markov chains, the force F(ρ) is a pure gradient F=G for some potential G (see Sect. 3), in which case one may write j·F=xρ˙(x)G(x), which follows from an integration by parts and application of the continuity equation. In this case, Mielke, M. A. Peletier, and Renger [43] also identified a slightly different canonical structure to the one presented here, in which the dual pairing is xv(x)G(x), for a velocity v(x)=ρ˙(x) and a potential G. The analogues of Ψ and Ψ in that setting depend on v and G respectively, instead of j and F. The setting of (3) and (4) is more general, in that the functions Ψ,Ψ for the velocity/potential setting are fully determined by those for the current/force setting. Also, focusing on the velocity v prevents any analysis of the divergence-free part of the current, and restricting to potential forces does not generalise in a simple way to irreversible Markov chains. For this reason, we use the current/force setting in this work.

In a separate development, Maas [35] identified a quadratic cost function for paths (in fact a metric structure) for which the master equation (11) is the minimiser in the case of reversible dynamics. This metric corresponds to the solution of an optimal mass transfer problem which seems to have no straightforward extension to irreversible systems. Of course, in the reversible case, the pathwise rate function (24) has the same minimiser, but is non-quadratic and therefore does not correspond to a metric structure, so there is no simple geometrical interpretation of (24). It seems that the non-quadratic structure in the rate function is essential in order capture the large deviations encoded by (23).

Time-Reversal Symmetry, Entropy Production, and the Gallavotti–Cohen Theorem

The rate function for the large-deviation principle (23) is given by (24), which has been written in terms of forces F, currents j, and densities ρ. To explain why it is useful to write the rate function in this way, we compare the probability of a path (ρt,jt)t[0,T] with that of its time-reversed counterpart (ρt,jt)t[0,T], where (ρt,jt)=(ρT-t,-jT-t) as before.

In this case, the fact that Ψ and Ψ are both even in their second argument means that

-1NlogProb((ρ^tN,ȷ^tN)t[0,T](ρt,jt)t[0,T])Prob((ρ^tN,ȷ^tN)t[0,T](ρt,jt)t[0,T])I[0,T]((ρt,jt)t[0,T])-I[0,T]((ρt,jt)t[0,T])=F(ρ0)-F(ρT)-0Tjt·F(ρt)dt. 25

This formula is a (finite-time) statement of the Gallavotti–Cohen fluctuation theorem [19, 32]: see also [12, 37]. It also provides a connection to physical properties of the system being modelled, via the theory of stochastic thermodynamics [50]. The terms involving the free energy F come from the initial conditions of the forward and reverse paths, while the integral of j·F corresponds to the heat transferred from the system to its environment during the trajectory [50, Eqs. (18), (20)]. This latter quantity—which is the time-reversal antisymmetric part of the pathwise rate function—is related (by a factor of the environmental temperature) to the entropy production in the environment [37]. The definition of the force F in (12) has been chosen so that the dual pairing j·F is equal to this rate of heat flow: this means that the forces and currents are conjugate variables, just as (for example) pressure and volume are conjugate in equilibrium thermodynamics. See also the example in Sect. 3.5.

Decomposition of Forces and Rate Functional

We now introduce a splitting of the force F(ρ) into two parts FS(ρ) and FA, which are related to the behaviour of the system under time-reversal, as well as to the splitting of the heat current into “excess” and “housekeeping” contributions [50]. We use this splitting to decompose the function Φ into three pieces, which allows us to compare (for example) the behaviour of reversible and irreversible Markov chains. This splitting also mirrors a similar construction within MFT [7], and this link will be discussed in Sect. 4. Related splittings have been introduced elsewhere; see [30] and [47] for decompositions of forces in stochastic differential equations, and [13] for decompositions of the instantaneous current in interacting particle systems.

Splitting of the Force According to Time-Reversal Symmetry

We define the adjoint process associated with the original Markov chain of interest. The transition rates of the adjoint process are rxy:=π(y)ryxπ(x)-1. It is easily verified that the adjoint process has invariant measure π, so qxy:=π(x)rxy=qyx. Under the assumption that the initial distribution is sampled from the steady state, the probability to observe a trajectory for the adjoint process coincides with the probability to observe the time-reversed trajectory for the original process.

From the definition of F(ρ) in (12), we can decompose this force as

Fxy(ρ)=FxyS(ρ)+FxyA 26

with

FxyS(ρ):=-x,ylogρπ,FxyA:=logqxyqyx. 27

With this choice, we note that the equivalent force for the adjoint process

F(ρ)=logρ(x)rxyρ(y)ryx,

satisfies F(ρ)=FS(ρ)-FA. So taking the adjoint inverts the sign of FA (the “antisymmetric” force) but leaves FS(ρ) unchanged (the “symmetric” force). For a reversible Markov chain, the adjoint process coincides with the original one, and FA=0.

Lemma 1

Given ρ, with the mobility a(ρ) of (12), the forces FS(ρ) and FA satisfy

xysinh(FxyS(ρ)/2)axy(ρ)sinh(FxyA/2)=0. 28

Proof

From the definitions of FS(ρ), FA, axy and sinh, one has

axy(ρ)sinh(FxyS(ρ)/2)=ρ(x)π(x)-ρ(y)π(y)qxyqyx

and sinh(FxyA/2)=(qxyqyx)-1/2(qxy-qyx)/2. Hence

xysinh(FxyS(ρ)/2)ax,y(ρ)sinh(FxyA/2)=12xyρ(x)π(x)-ρ(y)π(y)(qxy-qyx)=xρ(x)π(x)y(qxy-qyx)=0,

where the last equality uses (7). This establishes (28).

In Sect. 4.4, we will reformulate the so-called Hamilton–Jacobi relation of MFT in terms of forces, and show that this yields an equation analogous to (28).

Physical Interpretation of FS and FA

In stochastic thermodynamics, one may identify FxyA as the housekeeping heat (or adiabatic entropy production) associated with a single transition from state x to state y, see [16, 50]. (Within the Markov chain formalism, there is some mixing of the notions of force and energy: usually an energy would be a product of a force and a distance but there is no notion of a distance between states of the Markov chain, so forces and energies have the same units in our analysis.) Hence j·FA is the rate of flow of housekeeping heat into the environment. The meaning of the housekeeping heat is that for irreversible systems, transitions between states involve unavoidable dissipated heat which cannot be transformed into work (this dissipation is required in order to “do the housekeeping”).

To obtain the physical interpretation of FS, we also define

D(ρ,j):=12xyjxylogρ(y)π(x)ρ(x)π(y). 29

For a general path (ρt,jt)t[0,T] that satisfies ρ˙t=-divjt, we also identify

ddtF(ρt)=xρ˙t(x)logρt(x)π(x)=12xy(jt)xyx,ylogρπ=D(ρt,jt), 30

where we used (8), (15). That is, D(ρ,j) is the change in free energy induced by the current j. Moreover it is easy to see that

FxyS(ρ)=-x,yδFδρ, 31

where δFδρ denotes the functional derivative of the free energy F given in (8). (Note that the functional derivative δF/δρ is simply F/ρ in this case, since ρ is defined on a discrete space. We retain the functional notation to emphasise the connection to the general setting of Sect. 1.1.) Also, the last identity in (30) can be phrased as

j·FS(ρ)=-D(ρ,j). 32

The same identity, with an integration by parts, shows that

D(ρ,j)=0ifjis divergence free. 33

Equation (31) shows that the symmetric force FS is minus the gradient of the free energy, so the heat flow associated with the dual pairing of j and FS is equal to (the negative of) the rate of change of the free energy. It follows that the right hand side of (25) can alternatively be written as -j·FAdt.

We also recall from Sect. 2.2 that the force F acts in the space of probability densities: Fxy depends not only on the states xy but also on the density ρ. (Physical forces acting on individual copies of the system should not depend on ρ since each copy evolves independently, but F includes entropic terms associated with the ensemble of copies.) To understand this dependence, it is useful to write F(ρ)=-xρ(x)logπ(x)+xρ(x)logρ(x). We also write the invariant measure in a Gibbs-Boltzmann form: π(x)=exp(-U(x))/Z, where U(x) is the internal energy of state x and Z=xexp(-U(x)) is a normalisation constant. Then -xρ(x)logπ(x)=Eρ(U)+logZ depends on the mean energy of the system, while xρ(x)logρ(x) is (the negative of) the mixing entropy, which comes from the many possible permutations of the copies of the system among the states of the Markov chain. From (31) one then sees that FS has two contributions: one term (independent of ρ) that comes from the gradient of the energy U and the other (which depends on ρ) comes from the gradient of the entropy. These entropic forces account for the fact that a given empirical density ρN can be achieved in many different ways, since individual copies of the system can be permuted among the different states of the system.

Generalised Orthogonality for Forces

Recalling the definitions of Sect. 3.1, one sees that the current in the adjoint process satisfies an analogue of (13):

Jxy(ρ):=axy(ρ)sinh(12Fxy(ρ)),withFxy(ρ):=FxyS(ρ)-FxyA. 34

Comparing with (27), one sees that the adjoint process may also be obtained by inverting FA (while keeping FS(ρ) as it is). For axyS(ρ):=axy(ρ)cosh(FxyA/2) the symmetric current is defined as

JxyS(ρ):=axyS(ρ)sinh(FxyS(ρ)/2), 35

which satisfies JxyS(ρ)=(Jxy(ρ)+Jxy(ρ))/2. It is the same for the process and the adjoint process, and also coincides with the current for reversible processes (where qxy=qyx, or equivalently FA=0). An analogous formula can also be obtained for the anti-symmetric current. With axyA(ρ):=axy(ρ)cosh(FxyS(ρ)/2)=axy(π)(ρ(x)π(x)+ρ(y)π(y))/2, the anti-symmetric current is defined as

JxyA(ρ):=axyA(ρ)sinh(FxyA/2). 36

It satisfies JxyA(ρ)=(Jxy(ρ)-Jxy(ρ))/2.

Let ΨS be the symmetric version of Ψ obtained from (16) with axy(ρ) replaced by axyS(ρ). (The Legendre transform of ΨS is similarly denoted ΨS). This leads to a separation of Ψ(ρ,F(ρ)) in a term corresponding to FS(ρ) and a term corresponding to FA.

Lemma 2

The two forces FS(ρ) and FA defined in (27) satisfy

Ψ(ρ,F(ρ))=ΨS(ρ,FS(ρ))+Ψ(ρ,FA), 37

Proof

Using cosh(x+y)=cosh(x)cosh(y)+sinh(x)sinh(y), Lemma 1 and the definition of axyS(ρ), we obtain that the left hand side of (37) is given by

xyaxy(ρ)(cosh(Fxy(ρ)/2)-1)=xyaxy(ρ)(cosh(FxyS(ρ)/2)cosh(FxyA(ρ)/2)-1)=xyaxyS(ρ)(cosh(FxyS(ρ)/2)-1)+xyaxy(ρ)(cosh(FxyA(ρ)/2)-1), 38

which coincides with the right hand side of (37).

The physical interpretation of Lemma 2 is that the strength of the force F(ρ) can be written as separate contributions from FS(ρ) and FA. The following corollary allows us to think of a generalised orthogonality of the forces FS(ρ) and FA.

Proposition 3

(Generalised orthogonality) The forces FS(ρ) and FA satisfy

Ψ(ρ,FS(ρ)+FA)=Ψ(ρ,FS(ρ)-FA). 39

Proof

This follows directly from Lemma 2 and the symmetry of Ψ(ρ,·).

We refer to Proposition 3 as a generalised orthogonality between FS and FA because Ψ is acting as generalisation of a squared norm (see Sect. 1.1), so (39) can be viewed as a nonlinear generalisation of FS+FA2=FS-FA2, which would be a standard orthogonality between forces.

Moreover, Lemma 2 can be used to decompose the OM functional as a sum of three terms.

Corollary 4

Let ΦS be defined as in (3) with (Ψ,Ψ) replaced by (ΨS,ΨS), and D(ρ,j) as defined in (29). Then

Φ(ρ,j,F(ρ))=D(ρ,j)+ΦS(ρ,0,FS(ρ))+Φ(ρ,j,FA). 40

Proof

We use the definition of Φ in (3) and (32) together with Lemma 2 to decompose Φ(ρ,j,F(ρ)) as

Φ(ρ,j,F(ρ))=D(ρ,j)+ΨS(ρ,FS(ρ))+[Ψ(ρ,j)-j·FA+Ψ(ρ,FA)]=D(ρ,j)+ΦS(ρ,0,FS(ρ))+Φ(ρ,j,FA), 41

which proves the claim.

Recall from Sect. 1.1 that Φ measures how much the current j deviates from the typical (or most likely) current J(ρ). One sees from (40) that it can be large for three reasons. The first term is large if the current is pushing the system up in free energy (because D is the rate of change of free energy induced by the current j). The second term comes from the time-reversal symmetric (gradient) force FS(ρ), which is pushing the system towards equilibrium. The third term comes from the time-reversal anti-symmetric force FA; namely, it measures how far the current j is from the value induced by the force FA.

Corollary 4 also makes it apparent that the free energy F is monotonically decreasing for solutions of (11), which are minimisers of I[0,T].

Corollary 5

The free energy F is monotonically decreasing along minimisers of the rate function I[0,T]. Its rate of change is given by

ddtF(ρt)=-ΨS(ρt,FS(ρt))-Φ(ρt,J(ρt),FA(ρt)). 42

Proof

For minimisers of the rate function one has Φ=0. Hence (30) and Corollary 4 imply that

ddtF(ρt)=D(ρ,j)=-ΨS(ρt,FS(ρt))-Φ(ρt,J(ρt),FA(ρt)). 43

Both Ψ and Φ are non-negative, so F is indeed monotonically decreasing.

Hamilton–Jacobi Like Equation for Markov Chains

It is also useful to note at this point an additional aspect of the orthogonality relationships presented here, which has connections to MFT (see Sect. 4). We formulate an analogue of the Hamilton–Jacobi equation of MFT, as follows. Define

H(ρ,ξ)=12Ψ(ρ,F(ρ)+2ξ)-Ψ(ρ,F(ρ)), 44

which we refer to as an extended Hamiltonian, for reasons discussed in Sect. 6.3 (see also Sect. IV.G of [7]).

The extended Hamilton–Jacobi equation for a functional S is then (cf. equation (100) in Sect. 6.3) given by

Hρ,δSδρ=0. 45

Note that the free energy F defined in (8) solves (45), which follows from Proposition 3 (using (31) and that Ψ is symmetric in its second argument). In fact (see Proposition 13), the free energy is the maximal solution to this equation. In MFT, the analogous variational principle can be useful, as a characterisation of the invariant measure of the process. Here, one has a similar characterisation of the (non-equilibrium) free energy.

Since (45) with S=F provides a characterisation of the free energy F, which is uniquely determined by the invariant measure π of the process, it follows that (45) must be equivalent to the condition that π satisfies divJ(π)=0: recall (11). Writing everything in terms of the rates of the Markov chain and its adjoint, (45) becomes

xρ(x)y[rxy-rxy]=0,

which must hold for all ρ: from the definition of r one then has yπ(x)rxy=yπ(y)ryx, which is indeed satisfied if and only if π is invariant (cf. Eq. (7)).

Example: Simple Ring Network

To illustrate these abstract ideas, we consider a very simple Markov chain, in which n states are arranged in a circle, see Fig. 1. So V={1,2,,n} and the only allowed transitions take place between state x and states x±1 (to incorporate the circular geometry we interpret n+1=1 and 1-1=n). In physics, such Markov chains arise (for example) as simple models of nano-machines or motors, where an external energy source might be used to drive circular motion [29, 53]. Alternatively, such a Markov chain might describe a protein molecule that goes through a cyclic sequence of conformations, as it catalyses a chemical reaction [31]. In both cases, the systems evolve stochastically because the relevant objects have sizes on the nano-scale, so thermal fluctuations play an important role.

Fig. 1.

Fig. 1

Illustration of a simple Markov chain with n=5 states arranged in a circle. The transition rates between states are ri,i±1. If the Markov chain is not reversible, there will be a steady-state probability current J corresponding to a net drift of the system around the circle

To apply the analysis presented here, the first step is to identify forces and mobilities, as in (12). Let Rx=rx,x+1rx+1,x. The invariant measure may be identified by solving yπ(x)rxy=yπ(y)ryx subject to yπ(y)=1. Finally, one computes the steady state current J=π(x)rx,x+1-π(x+1)rx+1,x, where the right hand side is independent of x (this follows from the steady-state condition on π). The original Markov process has 2n parameters, which are the rates rx,x±1: these are completely determined by the n-1 independent elements of π, the n mobilities (Rx)x=1n and the current J. The idea is that this reparameterisation allows access to the physically important quantities in the system.

From the definitions of J and R, it may be verified that

2π(x)rx,x+1=J2+4Rx2π(x)π(x+1)+J,

and similarly 2π(x+1)rx+1,x=J2+4Rx2π(x)π(x+1)-J. Then write

ρ(x)rx,x+1=Rxρ(x)ρ(x+1)×ρ(x)π(x+1)ρ(x+1)π(x)·×J2+4Rx2π(x)π(x+1)+JJ2+4Rx2π(x)π(x+1)-J1/2. 46

In this case, we can identify the three terms as

ρ(x)rx,x+1=12ax,x+1(ρ)×expFx,x+1S(ρ)/2×expFx,x+1A/2, 47

which allows us to read off the mobility a and the forces FS and FA. The physical meaning of these quantities may not be obvious from these definitions, but we show in the following that reparameterising the transition rates in this way reveals structure in the dynamical fluctuations.

For example, equilibrium models (with detailed balance) can be identified via Fx,x+1A=0 (for all x). In general Fx,x+1A is the (steady-state) entropy production associated with a transition from x to x+1, see Sect. 3.2. The steady state entropy production associated with going once round the circuit is xFx,x+1A=logx(rx,x+1/rx+1,x), as it must be [1].

Now consider the LDP in (23). We consider a large number (N) of identical nano-scale devices, each of which is described by an independent copy of the Markov chain. Typically, each device goes around the circle at random, and the average current is J (so each object performs J/n cycles per unit time). The LDP describes properties of the ensemble of devices. If N is large and the distribution of devices over states is ρ, then the (overwhelmingly likely) time evolution of this distribution is ρ˙=-divJ(ρ), where the current J obeys the simple formula

Jx,x+1(ρ)=ax,x+1(ρ)sinh12Fx,x+1S(ρ)+Fx,x+1A, 48

which is (13), applied to this system. The simplicity of this expression motivates the parametrisation of the transition rates in terms of forces and mobilities. In addition, if one observes some current j [not necessarily equal to J(ρ)] then the rate of change of free energy of the ensemble can be written compactly as D(ρ,j)=-j·FS(ρ), from (32). The quantity j·FA is the rate of dissipation via housekeeping heat (see Sect. 3.2). This (physically-motivated) splitting of j·F=j·(FS+FA) motivates our introduction of the two forces FS and FA. Note that j·F is the rate of heat flow from the system to its environment, and appears in the fluctuation theorem (25).

Finally we turn to the large deviations of this ensemble of nano-scale objects. There is an LDP (23), whose rate function can be decomposed into three pieces (Corollary 4), because of the generalised orthogonality of the forces FS and FA (Lemma 2). This splitting of the rate function is useful because the symmetry properties of the various terms yields bounds on rate functions for some other LDPs obtained from Φ by contraction, see Sect. 5.

Connections to MFT

MFT is a field theory which describes the mass evolution of particle systems in the drift-diffusive regime, on the level of hydrodynamics. In this setting, it can be seen as generalisation of Onsager–Machlup theory [36]. For a comprehensive review, we refer to [7]. This section gives an overview of the theory, focussing on the connections to the results presented in Sects. 2 and 3.

We seek to emphasise two points: first, while the particle currents in MFT and the probability current in Markov chains are very different objects, they both obey large-deviation principles of the form presented in Sect. 1.1. This illustrates the broad applicability of this general setting. Second, we note that many of the particle models for which MFT gives a macroscopic description are Markov chains on discrete spaces. Starting from this observation, we argue in Sect. 4.5 that some results that are well-known in MFT originate from properties of these underlying Markov chains, particularly Proposition 3 and Corollary 4.

Setting

We consider a large number N of indistinguishable particles, moving on a lattice ΛL (indexed by LN, such that the number of sites |ΛL| is strictly increasing with L). These particles are described by a Markov chain, so the relevant forces and currents satisfy the equations derived in Sects. 2 and 3. The hydrodynamic limit is obtained by letting L such that the total density N/|ΛL| converges to a fixed number ρ¯. In this limit, the lattice ΛL is rescaled into a domain ΛRd and one can characterise the system by a local (mass) density ρ:Λ[0,) together with a local current j:ΛRd, which evolve deterministically as a function of time [7, 28]. This time evolution depends on some (density-dependent) applied forces F(ρ):ΛRd. The force at xΛ can be written as

F(ρ)(x)=f^(ρ(x))ρ(x)+E(x), 49

where the gradient denotes a spatial derivative, the function f^:[0,)R is a free energy density and E:ΛRd is a drift. (The free energy f^ is conventionally denoted by f [7]; here we use a different notation since f indicates a force in this work.) With these definitions, the deterministic currents satisfy the linear relation [41]

J(ρ)=χ(ρ)F(ρ), 50

which is the hydrodynamic analogue of (13). Here, χ(ρ)Rd×d is a (density-dependent) mobility matrix.

Onsager–Machlup Functional

Within MFT, the system is fully specified once the functions f,χ,E are given. These three quantities are sufficient to specify both the deterministic evolution of the most likely path ρ, and the fluctuations away from it. We can again define an OM functional given by

ΦMFT(ρ,j,f):=12Λ(j-χf)·χ-1(j-χf)dx. 51

To cast this functional in the form (3), we define the dual pair Λ(j·f)dx, together with the Legendre duals

ΨMFT(ρ,j):=12Λj·χ-1jdxandΨMFT(ρ,f):=12Λf·χfdx. 52

Given ρ and f, we have that ΦMFT is uniquely minimised (and equal to zero) for the current j=χ(ρ)f.

Large Deviation Principle

Within MFT, one considers an empirical density and an empirical current. We emphasise that these refer to particles, which are interacting and move on the lattice ΛL; this is in contrast to the case of Markov chains, where the copies of the system were non-interacting and one considers a density and current of probability. The averaged number of particles at site iΛL is denoted with ρ^tL(xi), where xi is the image in the rescaled domain Λ of site iΛL, and the corresponding particle current is given by ȷ^tL (cf. Sect. VIII.F in [7] for details). Note that both the particle density ρ^tL and the particle current ȷ^tL are random quantities (see also Sect. 4.5).

In keeping with the setting of Sect. 1.1, we focus on paths (ρ^tL,ȷ^tL)t[0,T] in the limit as L, where the probability is, analogous to (1), given by

Prob((ρ^tL,ȷ^tL)t[0,T](ρt,jt)t[0,T])exp{-|ΛL|I[0,T]MFT((ρt,jt)t[0,T])}. 53

Note that the parameter N in (1), which is the speed of the LDP, corresponds to the lattice size |ΛL|. For the force F(ρ) defined in (49), the rate functional in (53) is given by

I[0,T]MFT((ρt,jt)t[0,T])=V(ρ0)+120TΦMFT(ρt,jt,F(ρt))dtifρ˙t+divjt=0+otherwise. 54

Here V is the quasipotential, which plays the role of a non-equilibrium free energy. We may think of V as the macroscopic analogue of the free energy F defined in (8). It is the rate functional for the process sampled from the invariant measure, which is consistent with the case for Markov chains in (24). We assume that V has a unique minimiser π, which is the steady-state density profile (so V(π)=0).

An important difference between the Markov chain setting and MFT is that the OM functional for Markov chains is non-quadratic, which is equivalent to a non-linear flux force relation, whereas MFT is restricted to quadratic OM functionals.

Equation (53) is the basic assumption in MFT [7], in the sense that all systems considered by MFT are assumed to satisfy this pathwise LDP. In fact, both the process and its adjoint are assumed to satisfy such LDPs (with similar rate functionals, but different forces) [7].

Decomposition of the Force F

The force F in (49) can be written as the sum of a symmetric and an anti-symmetric part, F(ρ)=FS(ρ)+FA(ρ), just as in Sect. 3.1. The force for the adjoint process is given by F(ρ)=FS(ρ)-FA(ρ). Note that, unlike in the case of Markov chains, FA(ρ) can here depend on ρ. More precisely, FS(ρ)=-δVδρ and FA(ρ) is given implicitly by FA(ρ)=F(ρ)-FS(ρ).

The symmetric and anti-symmetric currents are defined in terms of the forces FS(ρ) and FA(ρ) as JS(ρ):=χ(ρ)FS(ρ) and JA(ρ):=χ(ρ)FA(ρ). An important result in MFT is the so-called Hamilton–Jacobi orthogonality, which states that

ΛJS(ρ)·χ(ρ)-1JA(ρ)dx=0. 55

In terms of the forces FS(ρ) and FA(ρ), we can restate (55) as

ΛFS(ρ)·χ(ρ)FA(ρ)dx=0. 56

The latter is the quadratic version of the orthogonality (28) of Lemma 1; it is equivalent to

Λ(FS(ρ)+FA(ρ))·χ(ρ)(FS(ρ)+FA(ρ))dx=Λ(FS(ρ)-FA(ρ))·χ(ρ)(FS(ρ)-FA(ρ))dx, 57

or in other words, from (52),

ΨMFT(ρ,FS(ρ)+FA(ρ))=ΨMFT(ρ,FS(ρ)-FA(ρ)), 58

which is the result of Proposition 3 in the context of MFT. One can see (39), and hence Proposition 3, as the natural generalisation to the Hamilton–Jacobi orthogonality (55). Again, the MFT describes systems on the macroscopic scale, but the result (58) originates from the result (39), on the microscopic level.

Relating Markov Chains to MFT: Hydrodynamic Limits

We have discussed a formal analogy between current/density fluctuations in Markov chains and in MFT: the large deviation principles (23) and (53) refer to different objects and different limits, but they both fall within the general setting described in Sect. 1.1. We argue here that the similarities between these two large deviation principles are not coincidental—they arise naturally when MFT is interpreted as a theory for hydrodynamic limits of interacting particle systems.

To avoid confusion between particle densities and probability densities, we introduce (only for this section) a different notation for some properties of discrete Markov chains, which is standard for interacting particle systems. Let η represent a state of the Markov chain (in place of the notation x of Sect. 2), and let μ be a probability distribution over these states (in place of the notation ρ of Sect. 2). Let ȷ be the probability current.

We illustrate our argument using the weakly asymmetric simple exclusion process (WASEP) in one dimension, so the lattice is ΛL={1,2,,L}, and each lattice site contains at most one particle, so V={0,1}L. The lattice has periodic boundary conditions and the occupancy of site i is η(i). Particles hop to the right with rate L2 and to the left with rate L2(1-(E/L)), but in either case only if the destination site is empty. Here E is a fixed parameter (an external field); the dependence of the hop rates on L is chosen to ensure a diffusive hydrodynamic limit (as required for MFT).

The spatial domain relevant for MFT is Λ=[0,1]: site iΛL corresponds to position i/LΛ. For any probability measure μ on V, one can write a corresponding smoothed particle density ρϵ on Λ, as

ρϵ(x)=1LηVi=1Lμ(η)η(i/L)δϵ(x-(i/L)), 59

where δϵ is a smoothed delta function (for example a Gaussian with unit weight and width ϵ, or—more classically—a top-hat function of width ϵ, cf. [28]). Similarly if there is a probability current ȷ in the Markov chain, one can write a smoothed particle current as

jϵ(x)=1LηVi=1Lȷη,ηi,i+1δϵx-2i+12L, 60

where ηi,i+1 is the configuration obtained from η by moving a particle from site i to site i+1; if there is no particle on site i then define ηi,i+1=η so that ȷη,ηi,i+1=0. Physically, ρϵ is the average particle density associated to μ, and jϵ is the particle current associated to ȷ.

As noted above, MFT is concerned with the limit L. The LDP (23) is not relevant for that limit (it applies when one considers many (N) independent copies of the Markov chain, with L being finite for each copy). However, the rate function I[0,T] that appears in (23) has an alternative physical interpretation, as the relative entropy between two path measures: see Appendix A. This relative entropy can be seen as a property of the WASEP; there is no requirement to invoke many copies of the system. Physically, the relative entropy measures how different is the WASEP from an alternative Markov process with a given probability and current (μt,ȷt)t[0,T].

The key point is that in cases where MFT applies, one expects that the rate function I[0,T]MFT can be related to this relative entropy. In fact, there is a deeper relation between relative entropies and rate functionals: it can be shown that Large Deviation Principles are equivalent to Γ-convergence of relative entropy functionals (see [42] for details).

Returning to the WASEP, we consider a particle density (ρt,jt)t[0,T] that satisfies ρ˙t=-divjt. One then can find (for each L) a time-dependent probability and current (μtL,ȷtL)t[0,T], with μ˙tL=-divȷtL, such on taking the limit ϵ0 after L, the associated particle densities (ρtϵ,jtϵ)(ρt,jt) and moreover

limL1|ΛL|I[0,T]((μtL,ȷtL)t[0,T])=I[0,T]MFT((ρt,jt)t[0,T]). 61

In order to find (μtL,ȷtL)t[0,T], one defines a “controlled” WASEP (similar to (69) in Sect. 5.3), in which the particle hop rates depend on position and time, such that the particle density in the hydrodynamic limit obeys ρ˙t=-divjt.

For interacting particle systems, this “controlled” process is usually obtained by adding a time dependent external field to the system that acts on the individual particles. This was first derived for the symmetric SEP in [27] (see also [4] for a treatment of the zero-range process). For the WASEP (in a slightly different situation with open boundaries) a proof of (61) can e.g. be found in [6], Lemma 3.7.

Moreover, on decomposing I[0,T]MFT and I[0,T] as in (3), the separate functions Ψ and Ψ obey formulae analogous to (61): this is the sense in which the structure of the MFT rate function is inherited from the relative entropy of the Markov chains. The quadratic functions Ψ and Ψ in MFT arise because the forces that appear in the underlying Markov chains are small (compared to unity), so second order Taylor expansions of Ψ and Ψ give in the limit the accurate description, similar to [2]. We will return to this discussion in a later publication.

LDPs for Time-Averaged Quantities

So far we have considered large deviation principles for hydrodynamic limits, and for systems consisting of many independent copies of a single Markov chain. We now show how some of the results derived in Sects. 2 and 3 also have analogues for large deviations for a single Markov chain, in the large-time limit.

Large Deviations at Level 2.5

Analogous to (22), we define the time averaged empirical measure of a single copy of the Markov chain ρ^[0,T] and the time averaged empirical current ȷ^[0,T] as

ρ^[0,T]:=1T0Tρ^tdtandȷ^[0,T]:=1T0Tȷ^tdt 62

(where we choose ρ^t=ρ^t1 and ȷ^t=ȷ^t1 for the empirical density and current of the single Markov chain, as defined above in Sect. 2.3). For countable state Markov chains, the quantity (ρ^[0,T],ȷ^[0,T]) satisfies a LDP as T:

Prob((ρ^[0,T],ȷ^[0,T])(ρ,j))exp{-TI2.5(ρ,j)}. 63

We refer to such principles as level 2.5 LDPs. For countable state Markov chains the rate functional I2.5(ρ,j) was derived in [39], and was proven rigorously in [8, 9] for Markov chains in the setting of Sect. 2.1 under some additional conditions (see [8, 9] for the details). We can recast the rate functional (see [8, Theorem 6.1]) as

I2.5(ρ,j)=12Φ(ρ,j,F(ρ))ifdivj=0+otherwise, 64

with Φ again given by (3), together with (14), (16) and (18).

We have stated this LDP for joint fluctuations of the density and the current. For Markov chains, the LDP for the density and the flow is also known as a level-2.5 LDP [9], so our general use of the name level-2.5 for (63) may be non-standard, but it seems reasonable. The rate functional for the density and the current in (63) can be obtained by contraction from the rate functional for the density and the flow (see Theorem 6.1 in [8]).

Using the splitting obtained in Sect. 3.3, we obtain the following representation for the rate functional on level-2.5.

Proposition 6

Let j be divergence free. Then the level-2.5 rate functional (64) is given by

I2.5(ρ,j)=12[ΦS(ρ,0,FS(ρ))+Φ(ρ,j,FA)]. 65

Proof

We note from (33) that D(ρ,j) vanishes for divergence free currents j. The result then directly follows from Corollary 4.

Large Deviations for Currents

Proposition 6 is connected to recently-derived bounds on rate functions for currents, see [22, 23, 45, 46]. Indeed, the rate function for current fluctuations can be obtained by contraction from level-2.5, as

Icurrent(j):=infρI2.5(ρ,j). 66

Then, following [23, 46], it may be shown that for any ρ,j,f one has for Φ as in (3) with (14), (16)–(18) that

Φ(ρ,j,f)xyjxy-jxyf(ρ)2bxy(ρ,f) 67

with bxy(ρ,f)=fxy/(4jxyf(ρ)) if fxy0; otherwise bxy is continuously extended by taking bxy(ρ,f)=1/(2axy(ρ)). Hence one has the result of [22], that the curvature of the rate function is controlled by the housekeeping heat FA, as

Icurrent(j)I2.5(π,j)=12Φ(π,j,FA)12xy(jxy-Jxyss)24(Jxyss)2JxyssFxyA, 68

where Jss:=J(π) is the steady state current (recall (9)), and the ratio FxyA/Jxyss must again be interpreted as 2/axy(ρ) in the case where FxyA (and hence Jxyss) vanish. The first step in (68) comes from (66), the second step uses (65) as well as Φ(π,0,FS)=0, and the third uses (67).

The significance of the splitting (65) for this result is that JxyssFxyA is the rate of flow of housekeeping heat associated with edge xy: the appearance of the housekeeping heat is natural since the bound comes from the second term in (65), which is independent of FS and depends only on FA.

Optimal Control Theory

It will be useful to introduce ideas of optimal control theory, whose relationship with large deviation theory is discussed in [10, 11, 18, 24]. In parallel with our given transition rates rxy we introduce a new process, the controlled process, where the rates are modified by a control potential φ, as

r~xy:=rxyexp((φ(y)-φ(x))/2). 69

For a given probability distribution ρ, we seek a potential φ such that the controlled process has invariant measure π~:=ρ. For this we need

yρxrxyexp((φ(y)-φ(x))/2)-ρyryxexp((φ(x)-φ(y))/2)=0,

or equivalently

divjF+φ(ρ)=yaxy(ρ)sinh(Fxy(ρ)+x,yφ)/2=0. 70

We stress that, for any fixed ρ, (70) is equivalent to solving the minimisation problem

infdivj=0Φ(ρ,j,F(ρ)), 71

which is also equivalent to maximisation of the Donsker–Varadhan functional, see for example Chapter IV.4 in [15]. A proof for the existence and uniqueness of φ can, e.g., be found in [40]. Now assume that φ solves (70). The resulting controlled process depends on ρ and has rates r~ given by (69). Throughout this section, we use tildes to indicate properties of the controlled process: all these quantities depend implicitly on the fixed probability ρ. Hence the (time-dependent) measure of the controlled process is ρ~.

Repeating the analysis of Sect. 2.1 and noting that r~xyr~yx=rxyryx, we find that a~xy(ρ~):=2ρ~(x)r~xyρ~(y)r~yx=axy(ρ~). Also, the force for the controlled process is

F~(ρ~)=F(ρ~)+φ, 72

which may be decomposed as

F~S(ρ~):=FS(ρ~)+logρπ=-logρ~ρ,F~A:=F(ρ)+φ=FA-logρπ+φ. 73

Thus, the symmetric force in the controlled process vanishes when ρ~=ρ. The antisymmetric force F~A represents the force observed in the new non-equilibrium steady state ρ. If the original process is reversible, then φ=logρπ so F~A=FA=0.

It is useful to define J~xy(ρ~):=axy(ρ~)sinh(F~xy(ρ~)/2) and to identify the steady-state current for the controlled process as

J~ss:=J~(ρ). 74

Decomposition of Rate Functions

The ideas of optimal control theory are useful since they facilitate the further decomposition of the level-2.5 rate function into several contributions.

Lemma 7

Suppose that ρ and j are given and that divj=0. Then

I2.5(ρ,j)=12[Φ(ρ,J~ss,F(ρ))+Φ(ρ,j,F~A)], 75

where J~ss is given by (74), evaluated in the optimally controlled process whose steady state is ρ.

Proof

We write

2I2.5(ρ,j)=Ψ(ρ,j)-j·F(ρ)+Ψ(ρ,F(ρ))=[Ψ(ρ,j)-j·F~(ρ)+Ψ(ρ,F~(ρ))]+Ψ(ρ,F(ρ))-Ψ(ρ,F~(ρ))-j·(F(ρ)-F~(ρ))=Φ(ρ,j,F~(ρ))+Ψ(ρ,F(ρ))-Ψ(ρ,F~(ρ))+j·φ 76

where the first line is (3) and (64); the second line is simple rewriting; and the third uses the definition of Φ in (3) and also (72) with ρ~=ρ.

The current J~(ρ) satisfies Φ(ρ,J~(ρ),F~(ρ))=0 so one has (by definition of Φ) that Ψ(ρ,F~(ρ))=J~(ρ)·F~(ρ)-Ψ(ρ,J~(ρ)). Using this relation together with (72) and (76), one has

2I2.5(ρ,j)=Φ(ρ,j,F~(ρ))+Ψ(ρ,F(ρ))-J~(ρ)·F(ρ)+Ψ(ρ,J~(ρ))-J~(ρ)·φ+j·φ. 77

Finally we note that divJ~(ρ)=0 (since ρ is the invariant measure for the controlled process) and divj=0 (by assumption), so integration by parts yields J~(ρ)·φ=0=j·φ; using once more the definition of Φ yields (82).

The physical interpretation of (75) is as follows. The contribution 12Φ(ρ,j,F~A) is a rate functional for observing an empirical current j in the controlled process, while 12Φ(ρ,J~ss,F(ρ)) is the rate functional for observing an empirical current J~ss in the original process. Since J~ss is the (deterministic) probability current for the controlled process, one has that the more the controlled process differs from the original one, the larger will be Φ(ρ,J~ss,F(ρ)). Hence the level-2.5 rate functional is large if the controlled process is very different from the original one, as one might expect. The rate functional also takes larger values if the empirical current j is very different from the probability current of the controlled process.

We obtain our final representation for the level-2.5 rate functional, consisting of the sum of three different OM functionals.

Proposition 8

Let j be divergence free. We can represent the level-2.5 rate functional (64) as

I2.5(ρ,j)=12ΦS(ρ,0,FS(ρ))+Φ(ρ,J~ss,FA)+Φ(ρ,j,F~A). 78

Proof

This follows immediately from Lemma 7 followed by an application of Corollary 4 to Φ(ρ,J~ss,FA) and that D=0, from (33).

The three terms in (78) also appear in Lemma 7 and Corollary 4, and their interpretations have been discussed in the context of those results. Briefly, we recall that I2.5(ρ,j) sets the probability of fluctuations in which a non-typical density ρ and current j are sustained over a long time period. The first term in (78) reflects the fact that the free-energy gradient FS(ρ) tends to push ρ towards the steady state π, so maintaining any non-typical density is unlikely if FS(ρ) is large. Similarly, the second term in (78) reflects the fact that large non-gradient forces FA also tend to suppress the probability that ρ maintains its non-typical value. The final term is the only place in which the (divergence-free) current j appears: it vanishes if the current j is typical within the controlled process (see Corollary 9); otherwise it reflects the probability cost of maintaining a non-typical circulating current.

Large Deviations at Level 2

As well the LDP (63), we also consider an (apparently) simpler object, called a level-2 LDP, where one considers the density only. It is formally given by

Probρ^Tρexp(-TI2(ρ)). 79

The contraction principle for LDPs [52, Sect. 3.6] states that

I2(ρ)=infj:divj=0I2.5(ρ,j). 80

Equation (75) is uniquely minimised in its second argument for the divergence free current jF~A, such that the contraction over all divergence-free vector fields j yields the level-2 rate functional

I2(ρ)=12Φ(ρ,J~ss,F(ρ)). 81

The same splitting as above finally allows us to write the level 2 rate functional as follows.

Corollary 9

The level-2 rate functional can be written as the sum

I2(ρ)=12[ΦS(ρ,0,FS(ρ))+Φ(ρ,J~ss,FA)]. 82

Proof

This follows from (80) and (78), since Φ(ρ,j,F~A) has a minimal value of zero.

This last identity extends the results obtained in [26] on the accelerated convergence to equilibrium for irreversible processes using LDPs from the macroscopic scale (i.e. in the regime of MFT) to Markov chains. The level-2 rate function in (82) can be interpreted as a rate of convergence to the steady state. It was shown in [26] that the rate is higher for irreversible processes, as opposed to reversible ones (as the second term Φ(ρ,J~ss,FA)=0 for reversible processes). We remark that splitting techniques for irreversible jump processes have been used to devise efficient MCMC samplers; see for example [5, 34].

Connection to MFT

Under the assumption that no dynamical phase transition takes place, the time averaged density ρ^[0,T]L:=1T0Tρ^tLdt and current ȷ^[0,T]L:=1T0Tȷ^tLdt in MFT (recall Sect. 4.3 for definitions) also satisfy a joint LDP in the limit L,T: one takes first L and then T, see [26, Eq. (36)]. The LDP is similar to (63):

Probρ^[0,T]L,ȷ^[0,T]L(ρ,j)exp-T|ΛL|IjointMFT(ρ,j), 83

where the rate function is, for a density profile ρ and a current j with divj=0, given by

IjointMFT(ρ,j)=12ΦMFT(ρ,j,F(ρ)). 84

As for Markov chains (see Sect. 5.1) IjointMFT(ρ,j)= if j is not divergence free. If divj=0 then the rate function can be written in the form [26]

IjointMFT(ρ,j)=14ΛδVδρ·χδVδρdx+14Λφ·χφdx+14Λ(JF-j)·χ-1(JF-j)dx, 85

such that a contraction to to the density only yields

IdensityMFT(ρ)=14ΛδVδρ·χδVδρdx+14Λφ·χφdx. 86

The function φ in (85) and (86) is obtained by solving

divJF(ρ)=0,JF(ρ):=χφ+JA(ρ). 87

Clearly the solution φ depends on ρ. In essence, we have reduced the minimisation problem (80) to the solution of this PDE. Comparing with (78), we identify the terms JF=χF~A in the MFT setting, and also J~ss=χF~A, so (J~ss-χFA(ρ))=χφ. We obtain the following representations for (85) and (86) reminiscent of Proposition 8 and Corollary 9.

Proposition 10

The rate functional for the joint density and current in MFT, which is given by (85), can be written in terms of the OM functional (51) as

IjointMFT(ρ,j)=12[ΦMFT(ρ,0,FS(ρ))+ΦMFT(ρ,J~ss,FA(ρ))+ΦMFT(ρ,j,F~A)], 88

and (86), the rate functional for the density in MFT, is given by

IdensityMFT(ρ)=12[ΦMFT(ρ,0,FS(ρ))+ΦMFT(ρ,J~ss,FA(ρ))]. 89

This proposition is equivalent to Proposition 5 of [26], but has now been rewritten in the language of optimal control theory. As discussed in [26], Eq. (89) quantifies the extent to which breaking detailed balance accelerates convergence of systems to equilibrium, at the hydrodynamic level. For this work, the key point is that this result originates from Corollary 9, which is the equivalent statement for Markov chains (without taking any hydrodynamic limit).

Consequences of the Structure of the OM Functional Φ

We have shown that the rate functions for several LDPs in several different contexts depend on functionals Φ with the general structure presented in (3) and (4). In this section, we show how this structure alone is sufficient to establish some features that are well-known in MFT. This means that these results within MFT have analogues for Markov chains. Our derivations mostly follow the standard MFT routes [7], but we use a more abstract notation to emphasise the minimal assumptions that are required.

Assumptions

The following minimal assumptions are easily verified for Markov chains; they are also either assumed or easily proven for MFT. The results of this section are therefore valid in both settings.

We consider a process described by a time-dependent density ρ and current j, with an associated continuity equation ρ˙=-divj and unique steady state π. We are given a set of (ρ-dependent) forces denoted by F(ρ), a dual pairing j·f between forces and currents, and a function Ψ(ρ,j) which is convex in j and satisfies Ψ(ρ,j)=Ψ(ρ,-j). With these choices, the functions Ψ and Φ are fully specified via (3) and (4). We assume that for initial conditions chosen from the invariant measure, the system satisfies an LDP of the form (1) with rate function of the form (2).

We define an adjoint process for which the probability of a path (ρt,jt)t[0,T] is equal to the probability of the time-reversed path (ρt,jt)t[0,T] in the original process. As above, we define (ρt,jt)=(ρT-t,-jT-t). We assume that the adjoint process also satisfies an LDP of the form (1), with rate function I[0,T]. Hence we must have

I[0,T]((ρt,jt)t[0,T])=I[0,T]((ρt,jt)t[0,T]). 90

Moreover, we assume that I[0,T] may be obtained from I by replacing the force F(ρ) with some adjoint force F(ρ). That is,

I[0,T]((ρt,jt)t[0,T])=I0(ρ0)+120TΦ(ρt,jt,F(ρt))dt. 91

Here, I0 is the rate function associated with fluctuations of the density ρ, for a system in its steady state. That is, within the steady state, Prob(ρ^Nρ)exp(-NI0(ρ)). For Markov chains, I0=F, the free energy; for MFT we have I0=V, the quasipotential. In the following we refer to I0 as the free energy.

Symmetric and Anti-symmetric Forces

Define

FS(ρ):=12[F(ρ)+F(ρ)],FA(ρ):=12[F(ρ)-F(ρ)]. 92

As the following proposition shows, FS is connected to the gradient of the free energy (or quasipotential) I0, and the forces FA and FS satisfy a generalised orthogonality (in the sense of Proposition 3). The proof follows Section II.C of [7], but uses only the assumptions of Sect. 6.1, showing that the result applies also to Markov chains.

Proposition 11

The forces FS and FA satisfy

FS(ρ)=-δI0δρ, 93

and

Ψ(ρ,FS(ρ)+FA)=Ψ(ρ,FS(ρ)-FA). 94

Proof

Combining (90) and (91), we obtain (for any path (ρt,jt)t[0,T] that obeys the continuity equation ρ˙=-divj)

I0(ρ0)+120TΦ(ρt,jt,F(ρt))dt=I0(ρT)+120TΦ(ρT-t,-jT-t,F(ρT-t))dt. 95

Differentiating with respect to T and using (3) together with Ψ(ρ,j)=Ψ(ρ,-j) and (92), one has

I˙0(ρ)+j·FS(ρ)+12[Ψ(ρ,F(ρ))-Ψ(ρ,F(ρ))]=0.

Using the continuity equation and an integration by parts, one finds I˙0(ρ)=j·δI0δρ, so that

j·FS(ρ)+δI0δρ+12[Ψ(ρ,F(ρ))-Ψ(ρ,F(ρ))]=0.

This equation must hold for all (ρ,j), which means that the two terms in square parentheses both vanish separately. Combining the last equation with (92), we obtain (93) and (94).

Proposition 11 also yields a variational characterisation of I0. The following corollary is analogous to Eq. (4.8) of [7], as is its proof.

Corollary 12

The free energy I0 satisfies

I0(ρ^)=inf12-0Φ(ρt,jt,F(ρt))dt, 96

where the infimum is taken over all paths (ρt,jt)t(-,0] that satisfy ρ˙t+divjt=0, as well as limt-ρt=π and ρ0=ρ^. Moreover, the optimal path is given by the time reversal of the solution of the adjoint dynamics (ρt,-J(ρt))t(-,0].

Proof

We obtain from (95) (together with (2) and (90)) that

12-0Φ(ρt,jt,F(ρt))dt=I0(ρ^)+12-0Φ(ρt,jt,F(ρt))dt.

Taking the infimum on both sides yields (96); indeed the infimum of 12-0Φ(ρt,jt,F(ρt))dt is 0, and this infimum is attained uniquely for the optimal path for (96). To see this, we note that Φ(ρt,-jt,F(ρt)) is uniquely minimised for jt=-J(ρt), and (ρt,-J(ρt))t(-,0] satisfies the conditions above, so the optimal path is indeed the time-reversal of the solution of the adjoint dynamics.

Hamilton–Jacobi Like Equation for the Extended Hamiltonian

Another important relationship within MFT is the Hamilton–Jacobi equation [7, Eq. (4.13)]. This provides a characterisation of the quasipotential, as its maximal non-negative solution. The following formulation of that result uses only the assumptions of Sect. 6.1 and therefore applies also to Markov chains. The functional

L(ρ,j):=12Φ(ρ,j,F(ρ)) 97

can be interpreted as an extended Lagrangian. (Note that L(ρ,j) should not be interpreted as a Lagrangian in the classical sense, as it depends on density and current (ρ,j), rather than the pair consisting of density and associated velocity (ρ,ρ˙)). We follow Sect. IV.G of [7]: given a sample path (ρt,jt)t[0,T]), define a vector field At=A0-0tjsds. The initial condition A0 is chosen so that there is a bijection between the paths (ρt,jt)t[0,T] and (At)t[0,T]. For example, in finite Markov chains, define ρ¯ as a constant density, normalised to unity, and let A0=h, where h solves div(h)=(ρ0-ρ¯), see [13] for the relevant properties of these vector fields. With this choice, and using ρ˙=-divj, one has ρt=ρ¯+divAt for all t, and one may also write (formally) At=div-1(ρt-ρ¯). Comparing with [7, Sect. IV.G], we write ρ=ρ¯+divA instead of ρ=divA since for Markov chains one has (for any discrete vector field A) that xdivA(x)=0, so it is not possible to solve divA=ρ if ρ is normalised to unity (recall that discrete vector fields have by definition Axy=-Ayx [13]).

The fluctuations of A are therefore determined by the fluctuations of (ρ,j), so the LDP (1) implies a similar LDP for A, whose rate function is I[0,T]ex((At)t[0,T])=I0ex(A0)+0TLex(At,A˙t)dt, where Lex is a Lagrangian that depends on A and its time derivative (which we again refer to as extended Lagrangian, cf. [7]). The function L in (97) is then related to Lex via the bijection between (ρ,j) and A. Considering again the case of Markov chains, the time evolution of the system depends only on divA (which is ρ-ρ¯) and not on A itself, one sees that Lex(A,A˙) depends only on divA and A˙ (which is j). Hence we write, formally, L(ρ,j)=Lex(div-1(ρ-ρ¯),-j), and we recover (97).

Hence L is nothing but the extended Lagrangian Lex, written in different variables: for this reason we refer to L as an (extended) Lagrangian.

To arrive at the corresponding (extended) Hamiltonian, one should write Hex(A,ξ)=supA˙[ξ·A˙-Lex(At,A˙t)], or equivalently

H(ρ,ξ)=supj(j·ξ-L(ρ,j)), 98

where ξ is a conjugate field for the current j. We identify H as the scaled cumulant generating function associated with the rate function I2.5(ρ,j)=L(ρ,j) [52, Sect. 3.1]. Analysis of rare fluctuations in terms of the field ξ is often more convenient than direct analysis of the rate function [32, 33] and is the basis of the “s-ensemble” method that has recently been exploited in a number of physical applications (for example [21, 24]). Using (3) and (4), we obtain

H(ρ,ξ)=12Ψ(ρ,F(ρ)+2ξ)-12Ψ(ρ,F(ρ)). 99

(This generalises the definition (44), which was restricted to Markov chains.)

To relate this extended Hamiltonian to the free energy (quasipotential), one can define an extended Hamilton–Jacobi equation, which is for a functional S given by

Hρ,δSδρ=0. 100

The relation of this equation to the free energy is given by the following proposition, which mirrors equation (4.18) of [7], but now in our generalised setting, so that it applies also to Markov chains.

Proposition 13

The free energy I0 is the maximal non-negative solution to (100) which vanishes at the steady state π. In other words, any functional S that solves (100) and has S(π)=0 also satisfies SI0.

Proof

From (92), (93), (94) and Ψ(ρ,F)=Ψ(ρ,-F), one has

Ψρ,F(ρ)+2δI0δρ=Ψ(ρ,-FS(ρ)+FA(ρ))=Ψ(ρ,F(ρ)). 101

Thus (99) yields H(ρ,δI0δρ)=0, so I0 does indeed solve (100). In addition, (101) is valid also with I0 replaced by any S that solves (100); combining this result with (3) yields

Φ(ρ,j,F(ρ))=Φρ,j,F(ρ)+2δSδρ+2j·δSδρ2j·δSδρ, 102

where the second step uses Φ0. Moreover, for any path (ρt,jt)t(-,0] with ρ˙t+divjt=0 and limt-ρt=π, we have from (102) that

I(-,0]((ρ,j)t(-,0])=-0Φ(ρt,jt,F(ρt))dt-0j(x)·δSδρ(x)dt=S(ρ0),

where the final equality uses an integration by parts, together with the continuity equation. Finally, taking the infimum over all paths and using Corollary 12, one obtains S(ρ)I0(ρ), as claimed.

Generalisation of Lemma 2

Before ending, we note that (94) is analogous to Proposition 3 in the general setting of this section, but we have not yet proved any analogue of Lemma 2. Hence we have not obtained a generalisation of Corollary 4, nor any of its further consequences. To achieve this, one requires a further assumption within the general framework considered here, which amounts to a splitting of the Hamiltonian. This assumption holds for MFT and for Markov chains, and is a sufficient condition for a generalised Lemma 2.

To state the assumption, we consider a reversible process in which the forces are FS(ρ). (For Markov chains we should consider the process with rates rxyS=12(rxy+rxy); for MFT it is the process with J(ρ)=JS(ρ) and the same mobility χ as the original process.) We assume that such a process exists and that its Hamiltonian can be written as HS(ρ,ξ)=12[ΨS(ρ,FS(ρ)+2ξ)-ΨS(ρ,FS(ρ))] for some function ΨS (compare (99) and see Sect. 3.4 for the case of Markov chains). Also let the Hamiltonian for the adjoint process be H(ρ,ξ), which is constructed by replacing F by F in (99). Then, one assumes further that

HS(ρ,ξ)=12[H(ρ,ξ)+H(ρ,ξ)], 103

which may be verified to hold for Markov chains and for MFT. Writing ξ=-FS/2 and using (99) with (94) and Ψ(ρ,f)=Ψ(ρ,-f), one then obtains

ΨS(ρ,FS(ρ))=Ψ(F(ρ))-Ψ(FA(ρ)), 104

which is the promised generalisation of Lemma 2.

Conclusion

In this article, we have presented several results for dynamical fluctuations in Markov chains. The central object in our discussion has been the function Φ, which plays a number of different roles—it is the rate function for large deviations at level 2.5 (Eq. 64), and it also appears in the rate function for pathwise large deviation functions (Eq. 2). These results—derived originally by Maes et al. [38, 39]—originate from the relationship between Φ and the relative entropy between path measures (Appendix A). The canonical (Legendre transform) structure of Φ (Eq. 4) and its relation to time reversal (Eq. 25) have also been discussed before [38].

The function Φ depends on probability currents j and their conjugate forces f. Our Proposition 3 and Corollary 4 show how the rate functions in which Φ appears have another level of structure, based on the decomposition of the forces F in two pieces F=FS+FA, according to its behaviour under time-reversal. A similar decomposition is applied in MFT [7]: the discussion of Sects. 5 and 6 show how several results of that theory—which applies on macroscopic (hydrodynamic) scales—already have analogues for Markov chains, which provide microscopic descriptions of interacting particle systems. These results—which concern symmetries, gradient structures and (generalised) orthogonality relationships—show how properties of the rate functions are directly connected to physical ideas of free energy, dissipation, and time-reversal.

Looking forward, we hope that these structures can be exploited both in mathematics and physics. From a mathematical viewpoint, the canonical structure and generalised orthogonality relationships may provide new routes for scale-bridging calculations, just as the geometrical structure identified by Maas [35] has been used to develop new proofs of hydrodynamic limits [17]. In physics, a common technique is to propose macroscopic descriptions of physical systems based on symmetries and general principles—examples in non-equilibrium (active) systems include [51, 54]. However, this level of description leaves some ambiguity as to the best definitions of some physical quantities, such as the local entropy production [44]. We hope that the structures identified here can be useful in relating such macroscopic theories to underlying microscopic behaviour.

Acknowledgements

We thank Freddy Bouchet, Davide Gabrielli, Juan Garrahan, Jan Maas, Michiel Renger and Hugo Touchette for useful discussions. MK is supported by a scholarship from the EPSRC Centre for Doctoral Training in Statistical Applied Mathematics at Bath (SAMBa), under the project EP/L015684/1. JZ gratefully acknowledges funding by the EPSRC through project EP/K027743/1, the Leverhulme Trust (RPG-2013-261) and a Royal Society Wolfson Research Merit Award. The authors thank the anonymous referees for their careful reading of the manuscript and for many helpful comments and suggestions.

Appendix A: Relative Entropy on Path Space

Consider a Markov process with rates r(xy) and initial distribution Q0. We fix a time interval [0, T] for some T>0 and denote the distribution of the Markov process on this time interval with Q. For each path (xu)u[0,T] with jumps at times t1,,tn the density of Q can be found by solving the associated master equation (11); it is given by

Q((xu)u[0,T])=Q0(x0)exp{0T(i=1nlogrt(xt-,xt)δ(t-ti)-yrt(xt,y))dt},

where xt-:=limϵ0xt-ϵ is the state of the process just before time t.

Now consider a second Markov process with time-dependent rates r^t(x,y) and initial distribution P0. The distribution of this process is denoted by P. The logarithmic density of P with respect to Q is given by

logdPdQ((xu)u[0,T])=logdP0dQ0(x0)+0T(i=1nlog(r^t(xt-,xt)r(xt-,xt))δ(t-ti)-y[r^t(xt,y)-r(xt,y)])dt.

We further denote the distribution of P at time t with ρt, such that ρt=PXt-1 where Xt denotes the evaluation of the path at time t (such that in particular P0=ρ0). The relative entropy on path space

H(P|Q):=EP[log(dPdQ)]

is then equal to

EP0[log(dP0dQ0)]+0Tx,yρt(x)(r^t(x,y)log(r^t(x,y)r(x,y))-r^t(x,y)+r(x,y))dt.

Let (ρt,jt)t[0,T] be given, with ρt>0 for all times t[0,T]. We then can rewrite the relative entropy H(P|Q) in terms of the flow Ct(x,y):=ρt(x)r^t(x,y) as

H(ρ0|Q0)+0Tx,y(Ct(x,y)log(Ct(x,y)ρt(x)r(x,y))-Ct(x,y)+ρt(x)r(x,y))dt. 105

Note that the relative entropy H(P|Q) can (just as the Markov chain) be completely characterised by the probability distribution (ρt)t[0,T] and the flow (Ct)t[0,T].

We are interested in a special flow (Ct)t[0,T] which recovers a given current (jt)t[0,T] as (jt)xy=Ct(x,y)-Ct(y,x). The force associated to jt is by (13) given by fjt(ρt):=2arcsinh(jt/a(ρt)) and the flow of interest is defined as Ct(x,y)=12axy(ρt)exp(12fxyjt(ρt)). It can be interpreted as the optimal flow that creates the current (jt)t[0,T].

We define the rates r~t(x,y):=Ct(x,y)/ρt(x) and denote the law of the associated (time heterogeneous) Markov process on [0, T] with P~. The relative entropy of this new process P~ with respect to the reference process Q is

H(P~|Q)=H(ρ0|Q0)+120TΦ(ρt,jt,F(ρt))dt 106

with Φ given by (3); to see this, we argue as follows. Symmetrising (105) and considering each summand separately gives

12(Ct(x,y)logCt(x,y)CtQ(x,y)+Ct(y,x)logCt(y,x)CtQ(y,x))+12(CtQ(x,y)-Ct(x,y)+CtQ(y,x)-Ct(y,x)),

where the first summand coincides with

12(12axy(ρt)sinh(12fxyjt(ρt))fxyjt(ρt)-12axy(ρt)sinh(12fxyjt(ρt))Fxy(ρt))

and the second is given by

12(axy(ρt)cosh(12Fxy(ρt))-axy(ρt)cosh(12fxyjt(ρt))).

Combining this with (15) and (16) yields (106).

Pathwise Large Deviation Principle: Let x1,x2, be a sequence of iid copies of the Markov chains with law Q. By Sanov’s Theorem (see, e.g., Theorem 6.2.10 in [14]), the empirical average 1Ni=1Nδxi of the Markov chains satisfies a LDP with the rate functional H(·|Q). We can interpret H(·|Q) as the rate functional for the joint LDP of (ρt,Ct)t[0,T] by defining this rate functional I[0,T]((ρt,Ct)t[0,T]) as the right-hand side of (105).

We contract the above rate functional to obtain the rate functional for the joint empirical measure and current (ρt,jt)t[0,T]. It is given by

I[0,T]((ρt,jt)t[0,T]):=inf(Ct)t[0,T]I[0,T]((ρt,Ct)t[0,T]), 107

where the infimum is taken over the set of all flows which yield the current (jt)t[0,T], i.e. over the set {(Ct)t[0,T]| for all t[0,T]:Ct(x,y)0 and Ct(x,y)-Ct(y,x)=(jt)xy}. It was shown in [38] and [9] that the minimising flow is the current Ct(x,y)=12axy(ρt)exp(12fxyjt(ρt)) introduced above, such that I[0,T]((ρt,jt)t[0,T]) coincides with (106).

Contributor Information

Marcus Kaiser, Email: m.kaiser@bath.ac.uk.

Robert L. Jack, Email: rlj22@cam.ac.uk

Johannes Zimmer, Email: j.zimmer@bath.ac.uk.

References

  • 1.Andrieux D, Gaspard P. Fluctuation theorem for currents and Schnakenberg network theory. J. Stat. Phys. 2007;127(1):107–131. doi: 10.1007/s10955-006-9233-5. [DOI] [Google Scholar]
  • 2.Basile, G., Benedetto, D., Bertini, L.: A gradient flow approach to linear Boltzmann equations. arXiv preprint. arXiv:1707.09204 (2017)
  • 3.Basu U, Maes C. Nonequilibrium response and frenesy. J. Phys: Conf. Ser. 2015;638(1):012001. [Google Scholar]
  • 4.Benois O, Kipnis C, Landim C. Large deviations from the hydrodynamical limit of mean zero asymmetric zero range processes. Stoch. Process. Appl. 1995;55(1):65–89. doi: 10.1016/0304-4149(95)91543-A. [DOI] [Google Scholar]
  • 5.Bernard EP, Krauth W. Two-step melting in two dimensions: first-order liquid-hexatic transition. Phys. Rev. Lett. 2011;107:155704. doi: 10.1103/PhysRevLett.107.155704. [DOI] [PubMed] [Google Scholar]
  • 6.Bertini L, Landim C, Mourragui M, et al. Dynamical large deviations for the boundary driven weakly asymmetric exclusion process. Ann. Probab. 2009;37(6):2357–2403. doi: 10.1214/09-AOP472. [DOI] [Google Scholar]
  • 7.Bertini L, De Sole A, Gabrielli D, Jona-Lasinio G, Landim C. Macroscopic fluctuation theory. Rev. Mod. Phys. 2015;87(2):593–636. doi: 10.1103/RevModPhys.87.593. [DOI] [Google Scholar]
  • 8.Bertini L, Faggionato A, Gabrielli D. Flows, currents, and cycles for Markov chains: large deviation asymptotics. Stoch. Process. Appl. 2015;125(7):2786–2819. doi: 10.1016/j.spa.2015.02.001. [DOI] [Google Scholar]
  • 9.Bertini L, Faggionato A, Gabrielli D. Large deviations of the empirical flow for continuous time Markov chains. Ann. Inst. Henri Poincaré Probab. Stat. 2015;51(3):867–900. doi: 10.1214/14-AIHP601. [DOI] [Google Scholar]
  • 10.Chernyak VY, Chertkov M, Bierkens J, Kappen HJ. Stochastic optimal control as non-equilibrium statistical mechanics: calculus of variations over density and current. J. Phys. A. 2014;47(2):022001. doi: 10.1088/1751-8113/47/2/022001. [DOI] [Google Scholar]
  • 11.Chetrite, R., Touchette, H.: Variational and optimal control representations of conditioned and driven processes. J. Stat. Mech. Theory Exp. 2015(12), P12001, 42 (2015)
  • 12.Crooks GE. Path-ensemble averages in systems driven far from equilibrium. Phys. Rev. E. 2000;61:2361–2366. doi: 10.1103/PhysRevE.61.2361. [DOI] [Google Scholar]
  • 13.De Carlo L, Gabrielli D. Gibbsian stationary non-equilibrium states. J. Stat. Phys. 2017;168(6):1191–1222. doi: 10.1007/s10955-017-1852-5. [DOI] [Google Scholar]
  • 14.Dembo, A., Zeitouni, O.: Large Deviations Techniques and Applications. Stochastic Modelling and Applied Probability, vol. 38. Springer, Berlin (2010). Corrected Reprint of 2nd edn (1998)
  • 15.den Hollander, F.: Large Deviations. Fields Institute Monographs, vol. 14. American Mathematical Society, Providence (2000)
  • 16.Esposito M, Van den Broeck C. Three faces of the second law. I. Master equation formulation. Phys. Rev. E. 2010;82:011143. doi: 10.1103/PhysRevE.82.011143. [DOI] [PubMed] [Google Scholar]
  • 17.Fathi M, Simon M. The Gradient Flow Approach to Hydrodynamic Limits for the Simple Exclusion Process. Cham: Springer; 2016. pp. 167–184. [Google Scholar]
  • 18.Fleming, W.H., Soner, H.M.: Controlled Markov Processes and Viscosity Solutions. Stochastic Modelling and Applied Probability, vol. 25, 2nd edn. Springer, New York (2006)
  • 19.Gallavotti G, Cohen EGD. Dynamical ensembles in stationary states. J. Stat. Phys. 1995;80(5–6):931–970. doi: 10.1007/BF02179860. [DOI] [Google Scholar]
  • 20.Gardiner, C.: Stochastic Methods. A Handbook for the Natural and Social Sciences. Springer Series in Synergetics, 4th edn. Springer, Berlin (2009)
  • 21.Garrahan, J.P., Jack, R.L., Lecomte, V., Pitard, E., van Duijvendijk, K., van Wijland, F.: First-order dynamical phase transition in models of glasses: an approach based on ensembles of histories. J. Phys. A 42(7), 075007, 34 (2009)
  • 22.Gingrich TR, Horowitz JM, Perunov N, England JL. Dissipation bounds all steady-state current fluctuations. Phys. Rev. Lett. 2016;116:120601. doi: 10.1103/PhysRevLett.116.120601. [DOI] [PubMed] [Google Scholar]
  • 23.Gingrich TR, Rotskoff GM, Horowitz JM. Inferring dissipation from current fluctuations. J. Phys. A. 2017;50(18):184004. doi: 10.1088/1751-8121/aa672f. [DOI] [Google Scholar]
  • 24.Jack RL, Sollich P. Effective interactions and large deviations in stochastic processes. Eur. Phys. J. Spec. Top. 2015;224(12):2351–2367. doi: 10.1140/epjst/e2015-02416-9. [DOI] [Google Scholar]
  • 25.Jarzynski C. Nonequilibrium equality for free energy differences. Phys. Rev. Lett. 1997;78:2690–2693. doi: 10.1103/PhysRevLett.78.2690. [DOI] [Google Scholar]
  • 26.Kaiser M, Jack RL, Zimmer J. Acceleration of convergence to equilibrium in Markov chains by breaking detailed balance. J. Stat. Phys. 2017;168:259–287. doi: 10.1007/s10955-017-1805-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Kipnis C, Olla S, Varadhan SRS. Hydrodynamics and large deviation for simple exclusion processes. Commun. Pure Appl. Math. 1989;42(2):115–137. doi: 10.1002/cpa.3160420202. [DOI] [Google Scholar]
  • 28.Kipnis, C., Landim, C.: Scaling Limits of Interacting Particle Systems. Grundlehren der Mathematischen Wissenschaften [Fundamental Principles of Mathematical Sciences], vol. 320. Springer, Berlin (1999)
  • 29.Kolomeisky AB, Fisher ME. Molecular motors: a theorist’s perspective. Annu. Rev. Phys. Chem. 2007;58(1):675–695. doi: 10.1146/annurev.physchem.58.032806.104532. [DOI] [PubMed] [Google Scholar]
  • 30.Kwon C, Ao P, Thouless DJ. Structure of stochastic dynamics near fixed points. Proc. Natl Acad. Sci. U.S.A. 2005;102(37):13029–13033. doi: 10.1073/pnas.0506347102. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31.Lavorel J. Matrix analysis of the oxygen evolving system of photosynthesis. J. Theor. Biol. 1976;57(1):171–185. doi: 10.1016/S0022-5193(76)80011-2. [DOI] [PubMed] [Google Scholar]
  • 32.Lebowitz JL. A Gallavotti–Cohen-type symmetry in the large deviation functional for stochastic dynamics. J. Stat. Phys. 1999;95(1–2):333–365. doi: 10.1023/A:1004589714161. [DOI] [Google Scholar]
  • 33.Lecomte V, Appert-Rolland C, van Wijland F. Thermodynamic formalism for systems with Markov dynamics. J. Stat. Phys. 2007;127(1):51–106. doi: 10.1007/s10955-006-9254-0. [DOI] [Google Scholar]
  • 34.Ma, Y.-A., Fox, E.B., Chen, T., Wu, L.: A unifying framework for devising efficient and irreversible MCMC samplers. arXiv preprint. arXiv:1608.05973 (2016)
  • 35.Maas J. Gradient flows of the entropy for finite Markov chains. J. Funct. Anal. 2011;261(8):2250–2292. doi: 10.1016/j.jfa.2011.06.009. [DOI] [Google Scholar]
  • 36.Machlup S, Onsager L. Fluctuations and irreversible process. II. Systems with kinetic energy. Phys. Rev. 1953;91:1512–1515. doi: 10.1103/PhysRev.91.1512. [DOI] [Google Scholar]
  • 37.Maes C. The fluctuation theorem as a Gibbs property. J. Stat. Phys. 1999;95(1–2):367–392. doi: 10.1023/A:1004541830999. [DOI] [Google Scholar]
  • 38.Maes, C., Netočný, K.: Canonical structure of dynamical fluctuations in mesoscopic nonequilibrium steady states. Europhys. Lett. EPL 82(3), Art. 30003, 6 (2008)
  • 39.Maes C, Netočný K, Wynants B. On and beyond entropy production: the case of Markov jump processes. Markov Process. Relat. Fields. 2008;14(3):445–464. [Google Scholar]
  • 40.Maes, C., Netočný, K., Wynants, B.: Monotonicity of the dynamical activity. J. Phys. A 45(45), 455001, 13 (2012)
  • 41.Maes C. Netočný, Karel: Revisiting the Glansdorff-Prigogine criterion for stability within irreversible thermodynamics. J. Stat. Phys. 2015;159(6):1286–1299. doi: 10.1007/s10955-015-1239-4. [DOI] [Google Scholar]
  • 42.Mariani, M.: A gamma-convergence approach to large deviations. arXiv preprint. arXiv:1204.0640 (2012)
  • 43.Mielke A, Peletier MA, Renger DRM. On the relation between gradient flows and the large-deviation principle, with applications to Markov chains and diffusion. Potential Anal. 2014;41(4):1293–1327. doi: 10.1007/s11118-014-9418-5. [DOI] [Google Scholar]
  • 44.Nardini C, Fodor É, Tjhung E, van Wijland F, Tailleur J, Cates ME. Entropy production in field theories without time-reversal symmetry: quantifying the non-equilibrium character of active matter. Phys. Rev. X. 2017;7:021007. [Google Scholar]
  • 45.Pietzonka P, Barato AC, Seifert U. Universal bounds on current fluctuations. Phys. Rev. E. 2016;93:052145. doi: 10.1103/PhysRevE.93.052145. [DOI] [PubMed] [Google Scholar]
  • 46.Polettini M, Lazarescu A, Esposito M. Tightening the uncertainty principle for stochastic currents. Phys. Rev. E. 2016;94:052104. doi: 10.1103/PhysRevE.94.052104. [DOI] [PubMed] [Google Scholar]
  • 47.Qian H. A decomposition of irreversible diffusion processes without detailed balance. J. Math. Phys. 2013;54(5):053302. doi: 10.1063/1.4803847. [DOI] [Google Scholar]
  • 48.Renger, D.R.M.: Large deviations of specific empirical fluxes of independent Markov chains, with implications for macroscopic fluctuation theory. Weierstrass Institute. Preprint 2375 (2017)
  • 49.Schnakenberg J. Network theory of microscopic and macroscopic behavior of master equation systems. Rev. Mod. Phys. 1976;48(4):571–585. doi: 10.1103/RevModPhys.48.571. [DOI] [Google Scholar]
  • 50.Seifert U. Stochastic thermodynamics, fluctuation theorems and molecular machines. Rep. Prog. Phys. 2012;75(12):126001. doi: 10.1088/0034-4885/75/12/126001. [DOI] [PubMed] [Google Scholar]
  • 51.Toner J, Yuhai T. Long-range order in a two-dimensional dynamical XY model: how birds fly together. Phys. Rev. Lett. 1995;75:4326–4329. doi: 10.1103/PhysRevLett.75.4326. [DOI] [PubMed] [Google Scholar]
  • 52.Touchette H. The large deviation approach to statistical mechanics. Phys. Rep. 2009;478(1–3):1–69. doi: 10.1016/j.physrep.2009.05.002. [DOI] [Google Scholar]
  • 53.Vaikuntanathan S, Gingrich TR, Geissler PL. Dynamic phase transitions in simple driven kinetic networks. Phys. Rev. E. 2014;89:062108. doi: 10.1103/PhysRevE.89.062108. [DOI] [PubMed] [Google Scholar]
  • 54.Wittkowski R, Tiribocchi A, Stenhammar J, Allen RJ, Marenduzzo D, Cates ME. Scalar ϕ4 field theory for active-particle phase separation. Nat. Commun. 2014;5:4351. doi: 10.1038/ncomms5351. [DOI] [PubMed] [Google Scholar]

Articles from Journal of Statistical Physics are provided here courtesy of Springer

RESOURCES