Skip to main content
Springer logoLink to Springer
. 2019 Aug 13;79(5):1831–1883. doi: 10.1007/s00285-019-01412-w

Generalizations of the ‘Linear Chain Trick’: incorporating more flexible dwell time distributions into mean field ODE models

Paul J Hurtado 1,, Adam S Kirosingh 2
PMCID: PMC6800873  PMID: 31410551

Abstract

In this paper we generalize the Linear Chain Trick (LCT; aka the Gamma Chain Trick) to help provide modelers more flexibility to incorporate appropriate dwell time assumptions into mean field ODEs, and help clarify connections between individual-level stochastic model assumptions and the structure of corresponding mean field ODEs. The LCT is a technique used to construct mean field ODE models from continuous-time stochastic state transition models where the time an individual spends in a given state (i.e., the dwell time) is Erlang distributed (i.e., gamma distributed with integer shape parameter). Despite the LCT’s widespread use, we lack general theory to facilitate the easy application of this technique, especially for complex models. Modelers must therefore choose between constructing ODE models using heuristics with oversimplified dwell time assumptions, using time consuming derivations from first principles, or to instead use non-ODE models (like integro-differential or delay differential equations) which can be cumbersome to derive and analyze. Here, we provide analytical results that enable modelers to more efficiently construct ODE models using the LCT or related extensions. Specifically, we provide (1) novel LCT extensions for various scenarios found in applications, including conditional dwell time distributions; (2) formulations of these LCT extensions that bypass the need to derive ODEs from integral equations; and (3) a novel Generalized Linear Chain Trick (GLCT) framework that extends the LCT to a much broader set of possible dwell time distribution assumptions, including the flexible phase-type distributions which can approximate distributions on R+ and can be fit to data.

Keywords: Gamma chain trick, Linear Chain Trick, Distributed delay, Mean field model, Phase-type distributions, Time lag

Introduction

Many scientific applications involve systems that can be framed as continuous time state transition models (e.g., see Strogatz 2014), and these are often modeled using mean field ordinary differential equations (ODE) of the form

dxdt=f(x,θ,t), 1

where x(t)Rn, parameters θRp, and f:RnRn is smooth. The abundance of such applications, and the accessibility of ODE models in terms of the analytical techniques and computational tools for building, simulating, and analyzing ODE models [e.g., straightforward simulation methods and bifurcation analysis software like MatCont (Dhooge et al. 2003) and XPPAUT (Ermentrout 1987, 2019)], have made ODEs one of the most popular modeling frameworks in scientific applications.

Despite their widespread use, one shortcoming of ODE models is their inflexibility when it comes to specifying probability distributions that describe the duration of time individuals spend in a given state. The basic options available for assuming specific dwell time distributions within an ODE framework can really be considered as a single option: the 1st event time distribution for a (nonhomogeneous) Poisson process, which includes the exponential distribution as a special case.

To illustrate this, consider the following SIR model of infectious disease transmission by Kermack and McKendrick (1927),

ddtS(t)=-λ(t)S(t) 2a
ddtI(t)=λ(t)S(t)-γI(t) 2b
ddtR(t)=γI(t) 2c

where S(t), I(t), and R(t) correspond to the number of susceptible, infected, and recovered individuals in a closed population at time t and λ(t)βI(t) is the per-capita infection rate (also called the force of infection by Anderson and May (1992) and others). This model can be thought of as the mean field model for some underlying stochastic state transition model where a large but finite number of individuals transition from state S to I to R [see the Appendix, Kermack and McKendrick (1927) for a derivation, and see Armbruster and Beck (2017), Banks et al. (2013), and references therein for examples of the convergence of stochastic models to mean field ODEs].

Although multiple stochastic models can yield the same mean field deterministic model, it is common to consider a stochastic model based on Poisson processes. For the SIR model above, for example, this stochastic analog would assume that, over the time interval [t,t+Δt] (for very small Δt), an individual in state S or I at time t is assumed to transition from S to I with probability λ(t)Δt, or from I to R with probability γΔt, respectively. Taking Δt0 yields the desired continuous time stochastic model. Here, the linear rate of transitions from I to R (γI(t)) arises from assuming the dwell time for an individual in the infected state (I) follows an exponential distribution with rate γ (i.e., the 1st event time distribution for a homogeneous Poisson process with rate γ). Similarly, assuming the time spent in state S follows the 1st event time distribution under a nonhomogeneous (also called inhomogeneous) Poisson process with rate λ(t) yields a time-varying per capita transition rate λ(t). This association of a mean field ODE with a specific underlying stochastic model provides very valuable intuition in an applied context. For example, it allows modelers to ascribe application-specific (e.g., biological) interpretations to parameters and thus estimate parameter values (e.g., for γ above, the mean time spent infectious is 1/γ), and it provides intuition and a clear mathematical foundation from which to construct and evaluate mean field ODE models based on individual-level, stochastic assumptions.

To construct similar models using other dwell time distributions, a standard approach is to formulate a continuous time stochastic model and from it derive mean field distributed delay equations, typically represented as integro-differential equations (IDEs) or integral equations (IEs) (e.g., see Kermack and McKendrick 1927; Hethcote and Tudor 1980; Feng et al. 2007, 2016, or see the example derivation in the Appendix). Readers unfamiliar with IEs and IDEs are referred to Burton (2005) or similar texts. IEs and IDEs have proven to be quite useful models in biology, e.g., they have been used to model chemical kinetics (Roussel 1996), gene expression (Smolen et al. 2000; Takashima et al. 2011; Guan and Ling 2018), physiological processes such as glucose-insulin regulation (Makroglou et al. 2006, and references therein), cell proliferation and differentiation (Özbay et al. 2008; Clapp and Levy 2015; Yates et al. 2017), cancer biology and treatment (Piotrowska and Bodnar 2018; Krzyzanski et al. 2018; Câmara De Souza et al. 2018), pathogen and immune response dynamics (Fenton et al. 2006), infectious disease transmission (Anderson and Watson 1980; Lloyd 2001a, b; Feng and Thieme 2000; Wearing et al. 2005; Lloyd 2009; Feng et al. 2007; Ciaravino et al. 2018), and population dynamics (MacDonald 1978a; Blythe et al. 1984; Metz and Diekmann 1986; Boese 1989; Nisbet et al. 1989; Cushing 1994; Wolkowicz et al. 1997; Gyllenberg 2007; Wang and Han 2016; Lin et al. 2018; Robertson et al. 2018). See also Campbell and Jessop (2009) and the applications reviewed therein.

However, while distributed delay equations are very flexible, in that they can incorporate arbitrary dwell time distributions, when compared to ODEs they also can be more challenging to derive, to analyze mathematically, and to simulate (Cushing 1994; Burton 2005). Thus, many modelers face a trade-off between building appropriate dwell time assumptions into their mean field models (i.e., opting for an IE or IDE model) and constructing parsimonious models that are more easily analyzed both mathematically and computationally (i.e., opting for an ODE model). For example, the following system of integral equations generalizes the SIR example above by incorporating an arbitrary distribution for the duration of infectiousness (i.e., the dwell time in state I):

S(t)=S(0)(1-FS(t)) 3a
I(t)=I(0)(1-FI(t))+0tβI(u)S(u)(1-FI(u))du 3b
R(t)=N-S(t)-I(t) 3c

where N=S(0)+I(0)+R(0), 1-FS(t)=exp(-0tβI(u)du) is the survival function for the distribution of time spent in susceptible state S [i.e., the 1st event time under a Poisson process with rate λ(t)=βI(t)], and 1-FI(t)=exp(-γt) is the survival function for the time spent in the infected state I (related models can be found in, e.g., Feng and Thieme 2000; Ma and Earn 2006; Krylova and Earn 2013; Champredon et al. 2018). A different choice of the CDF FI allows us to generalize the SIR model to other dwell time distributions that describe the time individuals spend in the infected state. Integral equations like those above can also be differentiated (assuming the integrands are differentiable) and represented as integrodifferential equations (e.g., as in Hethcote and Tudor 1980).

There have been some efforts in the past to identify which categories of integral and integro-differential equations can be reduced to systems of ODEs (e.g., MacDonald 1989; Metz and Diekmann 1991; Ponosov et al. 2002; Jacquez and Simon 2002; Burton 2005; Goltser and Domoshnitsky 2013; Diekmann et al. 2017, and references therein), but in practice the most well known case is the reduction of IEs and IDEs that assume Erlang1 distributed dwell times. This is done using what has become known as the Linear Chain Trick (LCT, also referred to as the Gamma Chain Trick; MacDonald 1978b; Smith 2010) which dates at least back to Fargue (1973) and earlier work by Theodore Vogel (e.g., Vogel 1961, 1965, according to Câmara De Souza et al. 2018). However, for more complex models that exceed the level of complexity that can be handled by existing “rules of thumb” like the LCT, the current approach is to derive mean field ODEs from mean field integral equations that might themselves first need to be derived from system-specific stochastic state transition models (e.g., Kermack and McKendrick 1927; Feng et al. 2007; Banks et al. 2013; Feng et al. 2016, and see the Appendix for an example.). Unfortunately, modelers often avoid these extra (often laborious) steps in practice by assuming (sometimes only implicitly) very simplistic dwell time distributions based on Poisson process 1st event times as in the SIR example above.

In light of the widespread use of ODE models, these challenges and trade-offs underscore a need for a more rigorous theoretical foundation to more effectively and more efficiently construct mean field ODE models that include more flexible dwell time distribution assumptions (Wearing et al. 2005; Feng et al. 2016; Robertson et al. 2018). The goal of this paper is to address these needs by (1) providing a theoretical foundation for constructing the desired system of ODEs directly from “first principles” (i.e., stochastic model assumptions), without the need to derive ODEs from intermediate IDEs or explicit stochastic models, and by (2) providing similar analytical results for novel extensions of the LCT which allow more flexible dwell time distributions, and conditional relationships among dwell time distributions, to be incorporated into ODE models. We also aim to clarify how underlying (often implicit) stochastic model assumptions are reflected in the structure of corresponding mean field ODE model equations.

The remainder of this paper is organized as follows. An intuitive description of the Linear Chain Trick (LCT) is given in Sect. 1.1 as a foundation for the extensions that follow. In Sect. 2 we review key notation and properties of Poisson processes and certain probability distributions needed for the results that follow. In Sect. 3.1.1 we highlight the association between Poisson process intensity functions and per capita rates in mean field ODEs, and in Sect. 3.1.2 we introduce what we call the weak memorylessness property of (nonhomogeneous) Poisson process 1st event time distributions. In Sects. 3.2 and 3.3 we give a formal statement of the LCT and in Sect. 3.4 a generalization that allows time-varying rates in the underlying Poisson processes. We then provide similar LCT generalizations for more complex cases: In Sect. 3.5 we provide results for multiple ways to implement transitions from one state to multiple states (which arise from different stochastic model assumptions and lead to different systems of mean field ODEs), and we address dwell times that obey Erlang mixture distributions. In Sect. 3.6 we detail how to construct mean field ODEs in which a sub-state transition (e.g., from infected to quarantined) does or does not alter the overall dwell time distribution in that state (e.g., the duration of infection). Lastly, in Sect. 3.7 we present a Generalized Linear Chain Trick (GLCT) which describes how to construct mean field ODEs from first principles based on assuming a very flexible family of dwell time distributions that include the phase-type distributions, i.e., hitting time distributions for continuous time Markov chains (Reinecke et al. 2012a; Horváth et al. 2016) which we address more specifically in Sect. 3.7.1. Tools for fitting phase-type distributions to data, or using them to approximate other distributions, are also mentioned in Sect. 3.7.1.

Intuitive description of the Linear Chain Trick

To begin, an intuitive understanding of the Linear Chain Trick (LCT) based on properties of Poisson processes, is helpful for drawing connections between underlying stochastic model assumptions and the structure of corresponding mean field ODEs. Here we consider a very basic case (detailed in Sect. 3.2): the mean field ODE model for a stochastic process in which particles in state X remain there for an Erlang(rk) distributed amount of time before exiting to some other state.

In short, the LCT exploits a natural stage structure within state X imposed by assuming an Erlang(rk) distributed dwell time [i.e., a gamma(rk) distribution with integer shape k]. Recall that the time to the kth event under a homogeneous Poisson process with rate r>0 is Erlang(rk) distributed. In that context, each event is preceded by a length of time that is exponentially distributed with rate r, and thus the time to the kth event is the sum of k independent and identically distributed exponential random variables [i.e., the sum of kiid exponential random variables with rate r is Erlang(rk) distributed]. Particles in state X at a given time can therefore be classified by which event they are awaiting next, i.e., a particle is in state Xi if it is waiting for the ith event to occur, and X1, , Xk is a partition of X. The dwell time distribution for each sub-state Xi is exponential with rate r, and particles leave the last state Xk (and thus X) upon the occurrence of the kth event.

This sub-state partition is useful to impose on X because the corresponding mean field equations for these sub-states are linear (or nearly linear) ODEs. Specifically, if xi(t) is the amount in Xi at time t, these mean field ODEs are

ddtx1(t)=-rx1(t),ddtxi(t)=rxi-1(t)-rxi(t)fori=2,,k 4

where the total amount in X at time t is x(t)=i=1kxi(t), x1(0)=x0 and xi(0)=0 for i=2,,k.

As we show below, this Poisson process based perspective allows us to generalize the LCT to other more complex scenarios where we partition a focal state X in a similar fashion. Those results are further extended from exponential dwell times to 1st event time distributions under a nonhomogeneous Poisson processes with time varying rate r(t), allowing for time-varying (or state-dependent) dwell time distributions to be used in extensions of LCT.

Model framework

The context in which we consider applications of the Linear Chain Trick (LCT) is the derivation of continuous time mean field model equations for stochastic state transition models with a distributed dwell time in a focal state, X. Such mean field models might otherwise be modeled as integral equations (IEs) or integro-differential equations (IDEs), and we seek to identify generalizations of the LCT that allow us to replace mean field integral equations with equivalent systems of 1st order ODEs.

Distributions and notation

Below we will extend the LCT from Erlang(rk) distributions (i.e., kth event time distributions under homogeneous Poisson processes with rate r) to event time distributions under nonhomogeneous Poisson processes with time varying rate r(t), and related distributions like the minimum of multiple Erlang random variables. In this section we will first review some basic properties of these distributions.

Gamma distributions can be parameterized2 by two strictly positive quantities: rater and shapek (sometimes denoted α and β, respectively). The Erlang family of distributions are the subset of gamma distributions with integer-valued shape parameters kZ+, or equivalently, the distributions resulting from the sum of kiid exponential distributions. That is, if a random variable T=i=1kTi, where all Ti are independent exponential distributions with rate r, then T is Erlang(rk) distributed. Since the inter-event times under a homogeneous Poisson process are exponentially distributed, the time to the kth event is thus Erlang(rk). This construction is foundational to a proper intuitive understanding of the LCT and its extensions.

If random variable T is gamma(rk) distributed, then its mean μ, variance σ2, and coefficient of variation cv are given by

μ=kr,σ2=kr2, andcv=1k. 5

Note that gamma distributions can be parameterized by their mean μ and variance σ2 by rewriting rate r and shape k using Eq. (5):

r=μσ2, andk=μ2σ2. 6

However, to ensure this gamma distribution is also Erlang (i.e., to ensure the shape parameter k is an integer) one must adjust the assumed variance up or down by rounding the value of k in Eq. (6) down or up, respectively, to the nearest integer (see Appendix B for details, and alternatives).

The Erlang density function (g), CDF (G), and survival3 function (S=1-G; also called the complementary CDF) are given by

grk(t)=r(rt)k-1(k-1)!e-rt 7a
Grk(t)=1-j=1k(rt)j-1(j-1)!e-rt=1-j=1k1rgrj(t) 7b
Srk(t)=1-Grk(t)=j=1k1rgrj(t). 7c

The results in Sect. 3 use (and generalize) the property of Erlang distributions detailed in Lemma 1 (eqs. 7.11 in Smith 2010, restated here without proof), which is the linchpin of the LCT.

Lemma 1

Erlang density functions grj(t), with rate r and shape j, satisfy

ddtgr1(t)=-rgr1(t),wheregr1(0)=r, 8a
ddtgrj(t)=r[grj-1(t)-grj(t)],wheregrj(0)=0forj2. 8b

Since homogeneous Poisson processes are a special case of nonhomogeneous Poisson processes4 from here on we will use “Poisson process” or “Poisson process with rate r(t)” to refer to cases that apply to both homogeneous [i.e., r(t)=r constant] and nonhomogeneous Poisson processes.

The more general kth event time distribution under a Poisson process with rate r(t), starting from some time τ<t has a density function (hrk), survival function (Srk), and CDF (Hrk1-Srk) given by

hrk(t,τ)=r(t)m(t,τ)k-1(k-1)!e-m(t,τ)and 9a
Srk(t,τ)=j=1khrj(t,τ)r(t)where 9b
m(t,τ)τtr(s)dsandddtm(t,τ)=r(t). 9c

For an arbitrary survival function starting at time τ (i.e., over the period [τ,t] where tτ) we use the notation S(t,τ). In some instances, we also use the notation S(t)S(t,0).

Lastly, in the context of state transitions models, it is common to assume that, upon leaving a given state (e.g., state X) at time t, individuals are distributed across multiple recipient states according to a generalized Bernoulli distribution (also known as the categorical distribution or the multinomial distribution with 1 trials) defined on the integers 1 through k where the probability of a particle entering the jth of k recipient states (j1,,k) is pj(t), which define probability vector p(t)=[p1(t),,pk(t)]T and j=1kpj(t)=1.

Results

The results below focus on one or more states, within a potentially larger state transition model, for which we would like to assume a particular dwell time distribution and derive a corresponding system of mean field ODEs using the LCT or a generalization of the LCT. In particular, the results below describe how to construct those mean field ODEs directly from stochastic model assumptions without needing to derive them from equivalent mean field integral equations (which themselves might need to be derived from an explicit continuous-time stochastic model).

Preliminaries

Before presenting extensions of the LCT, we first illustrate in Sect. 3.1.1 how mean field ODEs include terms that reflect underlying Poisson process rates. In Sect. 3.1.2, we highlight a key property of these Poisson process 1st event time distributions that we refer to as a weak memorylessness property since it is a generalization of the well known memorylessness property of the exponential and geometric distributions.

Transition rates in ODEs reflect underlying Poisson process rates

To build upon the intuition spelled out in Sect. 1.1, where particles exit X following an exponentially distributed dwell time, we now assume that particles exit X following the 1st event time under nonhomogeneous Poisson processes with rate r(t) (recall the 1st event time distribution is exponential if r(t)=r is constant). As illustrated by the corresponding mean field equations given below, the rate function r(t) can be viewed as either the intensity function5 for the Poisson process governing when individuals leave state X, or as the (mean field) per-capita rate of loss from state X as shown in Eq. (11). This dual perspective provides valuable intuition for critically evaluating mean field ODEs.

Example 1

(Equivalence between Poisson process rates and per capita rates in mean field ODEs) Consider the scenario described above. The survival function for the dwell time distribution for a particle entering X at time τ is S(t,τ)=exp(-τtr(u)du), and it follows that the expected proportion of such particles remaining in X at time t>τ is given by S(t,τ). Let x(t) be the total amount in state X at time t, x(0)=x0, and assume that I(t) and r(t) are integrable, non-negative functions of t. Then the corresponding mean field integral equation for this scenario is

x(t)=x0S(t,0)+0tI(τ)S(t,τ)dτ 10

and the integral equation (10) above is equivalent to

ddtx(t)=I(t)-r(t)x(t),withx(0)=x0. 11
Proof

The Leibniz rule for differentiating integrals, Eq. (10), and Lemma 1 yield

ddtx(t)=x0ddtS(t,0)+ddt0tI(τ)S(t,τ)dτ=-r(t)x0e-0tr(u)du+I(t)-r(t)0tI(τ)e-τtr(u)dudτ=I(t)-r(t)[x0e-0tr(u)du+0tI(τ)e-τtr(u)dudτ]=I(t)-r(t)x(t). 12

Weak memoryless property of Poisson process 1st event time distributions

The intuition behind the LCT, and the history-independent nature of ODEs, are related to the memorylessness property of exponential distributions. For example, when particles accumulate in a state with an exponentially distributed dwell time distribution, then at any given time all particles currently in that state have iid exponentially distributed amounts of time left before they leave that state regardless of the duration of time already spent in that state. Each of these can be viewed as a reflection of the history-independent nature of Poisson processes.

Accordingly, the familiar memorylessness property of exponential and geometric distributions can, in a sense, be generalized to (nonhomogeneous) Poisson process 1st event time distributions. Recall that if an exponentially distributed (rate r) random variable T represents the time until some event, then if the event has not occurred by time s the remaining duration of time until the event occurs is also exponential with rate r. The analogous weak memorylessness property of nonhomogeneous Poisson process 1st event time distributions is detailed in the following definition.

Definition 1

(Weak memorylessness property of Poisson process 1st event times) Assume T is a Poisson process 1st event time starting at time τ, which has rate r(t) and CDF Hr1(t,τ)=1-exp(-m(t,τ)) [see Eqs. (9) and (9c)]. If the event has not occurred by time s>τ the distribution of the remaining time TsT-s|T>s follows a shifted but otherwise identical Poisson process 1st event time distribution with CDF P(Tst)=Hr1(t+s,s). If r(t)=r is a positive constant we recover the memorylessness property of the exponential distribution.

Proof

The CDF of Ts (for t>τ) is given by

P(Tst)=P(T-st|T>s)=P(s<Ts+t)P(s<T)=Hr1(t+s,τ)-Hr1(s,τ)1-Hr1(s,τ)=1-1-Hr1(t+s,τ)1-Hr1(s,τ)=1-e-m(t+s,τ)e-m(s,τ)=1-e-m(t+s,s)=Hr1(t+s,s). 13

In other words, Poisson process 1st event time distributions are memoryless up to a time shift in their rate functions. In the context of multiple particles entering a given state X at different times and leaving according to independent Poisson process 1st event times with identical rates r(t) (i.e., t is absolute time, not time since entry into X), then for all particles in state X at a given time the distribution of time remaining in state X is (1) independent of how much time each particle has already spent in X and (2) follows iid Poisson process 1st event time distributions with rate r(t).

Simple case of the LCT

To illustrate how the LCT follows from Lemma 1, consider the simple case of the LCT illustrated in Fig. 1, where a higher dimensional model includes a state transition into, then out of, a focal state X. Assume the time spent in that state (TX) is Erlang(rk) distributed [i.e., TX Erlang(rk)]. Then the LCT provides a system of ODEs equivalent to the mean field integral equations for this process as detailed in the following theorem:

Fig. 1.

Fig. 1

Example diagram where state X has an Erlang(rk) distributed dwell time, represented either as a a single state and corresponding integral equation, or b as a set of k sub-states each with exponential dwell time distributions whose mean field equations can be represented as either integral equations or a system of ODEs (see Theorem 1). Rate I(t) is an integrable non-negative function describing the mean field influx rate into state X

Theorem 1

(Simple LCT) Consider a continuous time state transition model with inflow rate I(t)0 (an integrable function of t) into state X which has an Erlang(rk) distributed dwell time [with survival function Srk from Eq. (7c)]. Let x(t) be the (mean field) amount in state X at time t and assume x(0)=x0.

The mean field integral equation for this scenario (see Fig. 1a) is

x(t)=x0Srk(t)+0tI(s)Srk(t-s)ds. 14

State X can be partitioned (see Fig. 1b) into k sub-states Xi, i=1,,k, where particles in Xi are those awaiting the ith event as the next event under a homogeneous Poisson process with rate r. Let xi(t) be the amount in Xi at time t, and x(t)=j=1kxj(t). Equation (14) is equivalent to the mean field ODEs

ddtx1(t)=I(t)-rx1(t) 15a
ddtxj(t)=rxj-1(t)-rxj(t),j=2,,k 15b

with initial conditions x1(0)=x0, xj(0)=0 for j2, and

xj(t)=x01rgrj(t)+0tI(s)1rgrj(t-s)ds. 16

Proof

Substituting Eq. (7c) into Eq. (14) and then substituting Eq. (16) yields

x(t)=x0Srk(t)+0tI(s)Srk(t-s)ds=x0j=1k1rgrj(t)+0tI(s)j=1k1rgrj(t-s)ds=j=1kx01rgrj(t)+0tI(s)1rgrj(t-s)ds=j=1kxj(t). 17

Differentiating Eq. (16) (for j=1,,k) yields Eq. (15) as follows.

For j=1, Eq. (16) reduces to

x1(t)=x0e-rt+0tI(s)e-r(t-s)ds. 18

Differentiating x1(t) using the Leibniz integral rule and substituting (18) yields

ddtx1(t)=-rx0e-rt-r0tI(s)e-r(t-s)ds+I(t)=I(t)-rx1(t). 19

Similarly, for j2, Lemma 1 yields

ddtxj(t)=x01rddtgrj(t)+0tI(s)ddt1rgrj(t-s)ds=x0grj-1(t)-grj(t)+0tI(s)grj-1(t-s)-grj(t-s)ds=r(x0rgrj-1(t)+0tI(s)1rgrj-1(t-s)ds)-r(x0rgrj(t)+0tI(s)1rgrj(t-s)ds)=rxj-1(t)-rxj(t). 20

The Xj dwell time distributions are exponential with rate r. To see why, let χi(t) (where 1ik) be the (mean field) number of particles in state X at time t that have not reached the ith event. Then

χi(t)=x0Sri(t)+0tI(s)Sri(t-s)ds 21

and by Eq. (7) we see from Eqs. (18) and (21) that xj(t)=χj(t)-χj-1(t). That is, particles in state Xj are those for which the (j-1)th event has occurred, but not the jth event. Thus, by properties of Poisson processes the dwell time in state Xj must exponential with rate r.

Standard LCT

The following theorem and corollary together provide a more general, formal statement of the standard Linear Chain Trick (LCT) as used in practice. These extend Theorem 1 (compare Figs. 1, 2) to explicitly include that particles leaving X enter state Y then remain in Y according to an arbitrary distribution with survival function S, and include transitions into Y from other sources.

Fig. 2.

Fig. 2

(Standard LCT results) This generic case assumes that the dwell times in state X (see panel a) are Erlang(rk) distributed with inflow rates IX(t)0 into state X and IY(t)0 into state Y. Panels b and c show sub-states resulting from applying the LCT and Corollary 1 assuming either (b) dwell times in state Y are determined by per-capita rate function μ(t), or (c) dwell times in Y follow and Erlang distribution with shape parameter κ

Theorem 2

(Standard LCT) Consider a continuous time dynamical system model of mass transitioning among various states, with inflow rate IX(t) to a state X and an Erlang(rk) distributed delay before entering state Y. Let x(t) and y(t) be the amount in each state, respectively, at time t. Further assume an inflow rate IY(t) into state Y from other non-X states, and that the underlying stochastic model assumes that the duration of time spent in state Y is determined by survival function SY(t,τ). Assume Ii(t) are integrable non-negative functions of t, and assume non-negative initial conditions x(0)=x0 and y(0)=y0.

The mean field integral equations for this scenario are

x(t)=x0Srk(t)+0tIX(s)Srk(t-s)ds 22a
y(t)=y0SY(t,0)+0t(IY(τ)+x0grk(τ)+0τIX(s)grk(τ-s)ds)SY(t,τ)dτ. 22b

Equations (22) are equivalent to

ddtx1(t)=IX(t)-rx1(t) 23a
ddtxj(t)=rxj-1(t)-rxj(t),j=2,,k 23b
y(t)=y0SY(t,0)+0tIY(τ)+rxk(τ)Net input rate at timeτSY(t,τ)dτ 23c

where x(t)=j=1kxj(t), x1(0)=x0, xj(0)=0 for j2, and

xj(t)=x01rgrj(t)+0tIX(s)1rgrj(t-s)ds. 24

Proof

Eqs. (23a), (23b) and (24) follow from Theorem 1. Equation (23c) follows from substituting (24) into (22b). The definition of xj and initial condition x(0)=x0 together imply x1(0)=x0 and xj(0)=0 for the remaining j2.

Corollary 1

Integral equations like Eq. (23c) may have ODE representations depending on the Y dwell time distribution, i.e., SY(t,τ), for example:

  1. If particles leave Y after a Poisson process 1st event time distributed dwell time [i.e., the per-capita loss rate from Y is μ(t)], then SY(t,τ)=exp(-τtμ(u)du), and letting I(t)=IY(t)+rxk(t), Theorem 2 yields
    ddty(t)=IY(t)+rxk(t)-μ(t)y(t). 25
  2. If particles leave Y after an Erlang(μ,κ) delay, then SY(t,τ)=Sμκ(t-τ) and letting I(t)=IY(t)+rxk(t) Theorem 2 gives that y=i=1κyi and
    ddty1(t)=IY(t)+rxk(t)-μy1(t) 26a
    ddtyi(t)=μyi-1(t)-μyi(t),i=2,,κ. 26b
  3. As implied by 1 and 2 above, if the per-capita loss rate μ(t)=μ is constant or time spent in Y is otherwise exponentially distributed, then
    ddty(t)=IY(t)+rxk(t)-μy(t). 27
  4. Any of the more general cases considered in the sections below.

Example 2

To illustrate how the Standard LCT can be used to substitute an implicit exponential dwell time distribution with an Erlang distribution, consider the SIR example discussed in the Introduction [Eqs. (2) and (3), see also Anderson and Watson 1980; Lloyd 2001a, b], but assume the dwell time distribution for the infected state I is Erlang (still with mean 1/γ) with variance6σ2, i.e., by Eq. (6), Erlang with a rate r=γk and shape k=σ2/γ2.

By Theorem 2 and Corollary 1, with II(t)=λ(t)S(t) and I(t)=j=1kIj(t), the corresponding mean field ODEs are

ddtS(t)=-λ(t)S(t) 28a
ddtI1(t)=λ(t)S(t)-γkI1(t) 28b
ddtIj(t)=γkIj-1(t)-γkIj(t),forj=2,,k 28c
ddtR(t)=γkIk(t). 28d

Notice that if σ2=γ2 (i.e., k=1), the dwell time in I is exponentially distributed with rate γ, I(t)=I1(t), and Eq. (28) reduce to Eq. (2).

This example nicely illustrates how using Theorem 2 to relax an exponential dwell time assumption implicit in a system of mean field ODEs is much more straightforward than constructing them after first deriving the integral equations, like Eq. (3), and then differentiating them using Lemma 1. In the sections below, we present similar theorems intended to be used for constructing mean field ODEs directly from stochastic model assumptions.

Extended LCT for Poisson process kth event time distributed dwell times

The Standard LCT assumes an Erlang(rk) distributed dwell time, i.e., a kth event time distribution under a homogeneous Poisson process with rate r. Here we generalize the Standard LCT by assuming the X dwell time follows a more general kth event time distribution under a Poisson process with rate r(t).

First, observe that Eq. (8) in Lemma 1 are more practical when written in terms of 1rgrj(t) (see the proof of Theorem 1),

ddt[1rgr1(t)]=-r[1rgr1(t)], 29a
ddt[1rgrj(t)]=r[1rgrj-1(t)-1rgrj(t)],j2 29b

where 1rgr1(0)=1 and 1rgrj(0)=0.

Lemma 2

A similar relationship to Eq. (29) above (i.e., to Lemma 1) holds true for the Poisson process jth event time distribution density functions hrj given by Eq. (9a). Specifically,

ddt[1r(t)hr1(t,τ)]=-r(t)[1r(t)hr1(t,τ)], 30a
ddt[1r(t)hrj(t,τ)]=r(t)[1r(t)hrj-1(t,τ)-1r(t)hrj(t,τ)], 30b

where 1r(τ)hr1(τ,τ)=1 and 1r(τ)hrj(τ,τ)=0 for j2. Note that, if for some tr(t)=0, this relationship can be written in terms of

urk(t,τ)m(t,τ)k-1(k-1)!e-m(t,τ), 31

as shown in the proof below, where hrk(t,τ)=r(t)urk(t,τ), ur1(τ,τ)=1, and urj(τ,τ)=0 for j2.

Proof

For j=1,

ddt[ur1(t,τ)]=ddte-m(t,τ)=-r(t)e-m(t,τ)=-r(t)ur1(t,τ). 32

Likewise, for j2, we have

ddt[urj(t,τ)]=ddtm(t,τ)k-1(k-1)!e-m(t,τ)=r(t)m(t,τ)k-2(k-2)!e-m(t,τ)-r(t)m(t,τ)k-1(k-1)!e-m(t,τ)=r(t)[urj-1(t,τ)-urj(t,τ)]. 33

Lemma 2 allows us to generalize Erlang-based results like Theorem 2 to their time-varying counterparts with a time-dependent (or state-dependent) rate r(t), as in the following generalization of the Standard LCT (Theorem 2).

Theorem 3

(Extended LCT for dwell times distributed as Poisson process kthevent times) Consider the Standard LCT in Theorem 2 but assume the X dwell time is a Poisson process kth event time, rate r(t). The corresponding mean field integral equations, where hrj and Srj are given in Eq. (9), are

x(t)=x0Srk(t,0)+0tIX(s)Srk(t,s)ds 34a
y(t)=y0SY(t,0)+0t(IY(τ)+x0hrk(τ,0)+0τIX(s)hrk(τ,s)ds)SY(t,τ)dτ. 34b

The above Eq. (34), with x(t)=j=1kxj(t), are equivalent to

ddtx1(t)=IX(t)-r(t)x1(t) 35a
ddtxj(t)=r(t)xj-1(t)-r(t)xj(t),j=2,,k 35b
y(t)=y0SY(t,0)+0tIY(τ)+r(τ)xk(τ)SY(t,τ)dτ 35c

with initial conditions x1(0)=x0, xj(0)=0 for j2 and

xj(t)=x01r(t)hrj(t,0)+0tIX(s)1r(t)hrj(t,s)ds. 36

Equation (35c) may be further reduced to ODEs, e.g., via Corollary 1.

Proof

Substituting Eq. (9b) into Eq. (34a) and substituting Eq. (36) yields x(t)=j=1kxj(t). Differentiating Eq. (36) with j=1 using the Liebniz integration rule as well as Eq. (30a) from Lemma 2 yields Eq. (35a). Likewise, for j2, differentiation of Eq. (36) and Lemma 2 yields

ddtxj(t)=x(0)r(t)[1r(t)hrj-1(t,0)-1r(t)hrj(t,0)]+0tIX(s)r(t)[1r(t)hrj-1(t,τ)-1r(t)hrj(t,τ)]ds=r(t)(xj-1(t)-xj(t)). 37

Equation (35c) follows from substituting (36) into (34b). The definition of xj and initial condition x(0)=x0 together imply x1(0)=x0 and xj(0)=0 for the remaining j2. If r(t)=0 for some t, Eq. (35) still hold, since Eqs. (36) and (37) can be rewritten using u as in the proof of Lemma 2.

Having generalized the Standard LCT (Lemma 1 and Theorem 2) to include Poisson process kth event time distributed dwell times, we may now address more complex stochastic model assumptions and how they are reflected in the structure of corresponding mean field ODEs.

Transitions to multiple states

Modeling the transition from one state to multiple states following a distributed delay (as illustrated in Fig. 3) can be done under different sets of assumptions about the underlying stochastic processes, particularly with respect to the rules governing how individuals are distributed across multiple recipient states upon exiting X, and how those rules depend on the dwell time distribution(s) for individuals in that state. Importantly, those different sets of assumptions can yield very different mean field models (e.g., see Feng et al. 2016) and so care must be taken to make those assumptions appropriately for a given application.

Fig. 3.

Fig. 3

Example diagram of transitions out of a given state (X) and into multiple states (Y1 and Y2). Different assumptions about (1) the dwell times in X, and (2) rules governing the subsequent transitions to Y1 and/or Y2 will lead to different sub-state partitions of X, and thus different mean field equations, as detailed in Sect. 3.5. Fortunately, different scenarios often encountered in applications can be reduced to ODEs by applying the results in Sect. 3.5 as detailed in Theorems 4, 5, and 6 and as illustrated in Figs. 4, 5, and 6

While modelers have some flexibility to choose appropriate assumptions, in practice modelers sometimes make unintended and undesirable implicit assumptions, especially when constructing ODE models using “rules of thumb” instead of deriving them from first principles. In this section we present results aimed at helping guide (a) the process of picking appropriate dwell time distribution assumptions, and (b) directly constructing corresponding systems of ODEs without deriving them from explicit stochastic models or intermediate integral equations.

Each of the three cases detailed below yield different mean field ODE models for the scenario depicted in Fig. 3.

First, in Sect. 3.5.1, we consider the extension of Theorem 3 where upon leaving X particles are distributed across m1 recipient states according to a generalized Bernoulli distribution with (potentially time varying) probabilities/proportions pj(t), j=1,,m. Here the outcome of which state a particle transitions to is independent of the time spent in the first state.

Second, in Sects. 3.5.2 and 3.5.3, particles entering the first state (X) do not all follow the same dwell time distribution in X. Instead, upon entering X they are distributed across n2 sub-states of X, Xi, according to a generalized Bernoulli distribution, and each sub-state Xi has a dwell time given by a Poisson process kith event time distribution with rate ri(t). That is, the X dwell time is a finite mixture of Poisson process event time distributions. Particles transition out of X into m subsequent states Yj according to the probabilities/proportions pij(t), the probability of going to Yj from Xi, i=1,,n and j=1,,m. Here the determination of which recipient state Y a particle transitions to depends on which sub-state of X the particle was assigned to upon entering X (see Fig. 5).

Fig. 5.

Fig. 5

The sub-state diagram (cf. Fig. 3) for Example 4 after applying Theorem 5 (Extended LCT for finite mixtures of Poisson process event time distributions and output to multiple states). Upon entering X particles have an Erlang(ri,ki) dwell time in X with probability ρi, i=1,2,3, i.e., the X dwell time follows an Erlang mixture distribution (see Sect. 3.5.3)

Third, in Sect. 3.5.4, the outcome of which recipient state a particle transitions to upon leaving X is determined by a “race” between multiple competing Poisson process kth event time distributions, and is therefore not independent of the time spent in the first state (as in Sect. 3.5.1), nor is it pre-determined upon entry into X (as in Sects. 3.5.2 and 3.5.3). This result is obtained using yet another novel extension of Lemma 1 in which the dwell time in state X is the minimum of n2 independent Poisson process event time distributions.

Lastly (Sect. 3.5.5), we describe an equivalence between (1) the more complex case addressed in Sect. 3.5.4 assuming a dwell time that obeys the minimum of Poisson process 1st event times, before being distributed across m recipient states, and (2) the conceptually simpler case in Sect. 3.5.1 where the dwell time follows a single Poisson process 1st event time distribution before being distributed among m recipient states. This is key to understanding the scope of the Generalized Linear Chain Trick in Sect. 3.7.

Transition to multiple states independent of the X dwell time distribution

Here we extend the case in the previous section and assume that, upon leaving state X, particles can transition to one of m states (call them Yi, i=1,,m), and that a particle leaving X at time t enters state Yi with probability pi(t), where i=1mpi(t)=1 [i.e., particles are distributed across all Yi following a generalized Bernoulli distribution with parameter vector p(t)=(p1(t),,pm(t))]. See Fig. 4 for a simple example with constant p and m=2. An important assumption in this case is that the determination about which state a particle goes to after leaving X is made once it leaves X, and thus the state it transitions to upon exiting X is determined independent of the dwell time in X. Examples from the literature include Model II in Feng et al. (2016), where infected individuals (state X) either recovered (Y0) or died (Y1) after an Erlang distributed time delay.

Fig. 4.

Fig. 4

The special case of Fig. 3 under the assumptions of Theorem 4 (Extended LCT with proportional outputs to multiple states; see Sect. 3.5.1). Specifically, this case assumes that (1) the dwell time distribution for X is Erlang(rk), and (2) upon exiting X particles are distributed to multiple recipient states, here Y and Z, with probabilities p and 1-p, respectively

Theorem 4

(Extended LCT with proportional output to multiple states) Consider the case addressed by Theorem 3, and further assume particles go to one of m states (call them Yj) with pj(t) being the probability of going to Yj. Let Sj be the survival functions for the dwell times in Yj.

The mean field integral equations for this case, with x(0)=x0 and yj(0)=yj0, are

x(t)=x0Srk(t,0)+0tIX(s)Srk(t,s)ds 38a
yj(t)=yj(0)Sj(t,0)+0t(Ij(τ)+pj(t)(x0hrk(τ,0)+0τIX(s)hrk(τ,s)ds))Sj(t,τ)dτ 38b

These integral equations are equivalent to the following system of equations:

ddtx1(t)=IX(t)-r(t)x1(t) 39a
ddtxi(t)=r(t)xi-1(t)-r(t)xi(t),i=2,,k 39b
yj(t)=yj(0)Sj(t,0)+0t(Ij(τ)+pj(t)r(t)xk(τ))Sj(t,τ)dτ 39c

where x(t)=i=1kxi(t), x1(0)=x0, xi(0)=0 for i2, and

xi(t)=x01r(t)hrk(t,0)+0tIX(s)1r(t)hrk(t,s)ds. 40

Equations (39c) may be further reduced to ODEs, e.g., via Corollary 1.

Proof

Equations (39a), (39b) and (40) follow from Theorem 3. Equation (39c) follows from substitution of Eq. (40) into (38b). The derivation of Eq. (38b) is similar to the derivation in Appendix A.1 but accounts for the expected proportion entering each Yj at time t being equal to pj(t).

Example 3

Consider the example shown in Fig. 4, where the dwell time distribution for X is Erlang(rk) and the dwell times in Y and Z follow 1st event times under nonhomogeneous Poisson processes with respective rates μY(t) and μZ(t). By Theorem 4 the corresponding mean field ODEs are

ddtx1(t)=IX(t)-rx1(t) 41a
ddtxi(t)=rxi-1(t)-rxi(t),i=2,,k 41b
ddty(t)=IY(t)+prxk(t)-μY(t)y(t) 41c
ddtz(t)=IZ(t)+(1-p)rxk(t)-μZ(t)z(t). 41d

Transition from sub-states of X with differing dwell time distributions and differing output distributions across states Yj

In this second case, particles in state X can be treated as belonging to a heterogeneous population, where each remains in that state according to one of N possible dwell time distributions, the ith of these being the kith event time distribution under a Poisson process with rate ri(t). Each particle is assigned one of these N dwell time distributions (i.e., it is assigned to sub-state Xi) upon entry into X according to a generalized Bernoulli distribution with a (potentially time varying) probability vector ρ(t)=(ρ1(t),,ρN(t)), i=1Nρi(t)=1. In contrast to the previous case, here the outcome of which recipient state a particle transitions to is not necessarily independent of the dwell time distribution.

Note that the dwell time distribution in this case is a finite mixture of N independent Poisson processes event time distributions. If a random variable T is a mixture of Erlang distributions, or more generally a mixture of N independent Poisson process event time distributions, then the corresponding density function (f) and survival function (Φ) are

fθ(t,τ)=i=1Nρi(τ)hriki(t,τ) 42a
Φθ(t,τ)=i=1Nρi(τ)Sriki(t,τ)=i=1Nρi(τ)j=1ki1ri(t)hrij(t,τ) 42b

where θ(t)=(ρ1(t),r1(t),k1,,ρN(t),rN(t),kN) is the potentially time varying parameter vector for the N distributions that constitute the mixture distribution. Note that if all ri(t)=ri are constant, this is a mixture of Erlang distributions, or if also all ki=1, a mixture of exponentials.

Theorem 5

(Extended LCT for dwell times given by mixtures of Poisson process event time distributions and outputs to multiple states) Consider a continuous time state transition model with inflow rate IX(t) into state X. Assume that the duration of time spent in state X follows a finite mixture of N independent Poisson process event time distributions. That is, X can be partitioned into N sub-states Xi, i=1,,N, each with dwell time distributions given by a Poisson process kith event time distributions with rates ri(t). Suppose the inflow to state X at time t is distributed among this partition according to a generalized Bernoulli distribution with probabilities ρi(t), where i=1Nρi(t)=1, so that the input rate to Xi is ρi(t)IX(t). Assume that particles leaving sub-state Xi then transition to state Y with probability pi(t), =1,,m, where the duration of time spent in state Y follows a delay distribution give by survival function Sj. Then we can partition each Xi into Xij, j=1,,ki, according to Theorem 3 and let x(t), xi(t), xij(t), and y(t) be the amounts in states X, Xi, Xij, and Y at time t, respectively. Assume non-negative initial conditions x(0)=x0, xi(0)=ρi(0)x0, xi1(0)=ρi(0)x0, xij(0)=0 for j2, and y(0)0.

The mean field integral equations for this scenario are

x(t)=x0Φθ(t,0)+0tIX(s)Φθ(t,s)ds 43a
y(t)=y(0)S(t,0)+0t(I(τ)+i=1Npij(τ)(x0ρi(τ)hriki(τ,0)+0τρi(s)IX(s)hriki(τ,s)ds))S(t,τ)dτ. 43b

The above system of equations (43) are equivalent to

ddtxi1(t)=ρi(t)IX(t)-ri(t)xi1(t),i=1,,N 44a
ddtxij(t)=ri(t)(xi,j-1(t)-xij(t)),i=1,,N;j=2,,ki 44b
y(t)=y(0)S(t,0)+0t(I(τ)+i=1Nri(t)xiki(τ)pi(τ))S(t,τ)dτ 44c

with initial conditions xi1(0)=ρi(0)x0, xij(0)=0 for j2, where x(t)=i=1Nxi(t), and xi(t)=j=1kixij(t). The amounts in Xi and Xij are

xi(t)=ρi(0)x0Sriki(t,0)+0tρi(s)IX(s)Sriki(t,s)ds 45
xij(t)=ρi(0)x0hrij(t,0)ri(t)+0tρi(s)IX(s)hrij(t,s)ri(t)ds. 46

Equations (44c) may be reduced to ODEs, e.g., via Corollary 1.

Proof

Substituting Eq. (42b) into Eq. (43a) and then substituting Eq. (45) yields x(t)=i=1Nxi(t). Applying Theorem 3 to each Xi [i.e., to each Eq. (45)] then yields Eqs. (46), (44a) and (44b). (Alternatively, one could prove this directly by differentiating Eq. (46) using Eq. (30) from Lemma 2). The y(t) equations (44c) are obtained from (43b) by substitution of Eq. (46).

Example 4

Consider the scenario in Fig. 5, where particles entering state X at rate IX(t) enter sub-state Xi with probability ρi, ρ1+ρ2+ρ3=1, and the Xi dwell time is Erlang(ri,ki) distributed. Particles exiting X1 and X2 transition to Y with probability 1, while particles exiting X3 transition either to state Y or Z with probabilities pY and pZ=1-pY. Assume particle may also enter states Y and Z from sources other than state X (at rates IY(t) and IZ(t), respectively), and the dwell times in those two states follow the 1st event times of independent nonhomogeneous Poisson processes with rates μY(t) and μZ(t). Theorem 5 yields the following mean field ODEs (see Fig. 5).

ddtxi,1(t)=ρiIX(t)-rixi,1(t),i=1,,3, 47a
ddtxi,j(t)=ri(xi,j-1(t)-xij(t)),j=2,,ki 47b
ddty(t)=IY(t)+r1x1,k1(t)+r2x2,k2(t)+r3pYx3,k3(t)-μY(t)y(t) 47c
ddtz(t)=IZ(t)+r3pZx3,k3(t)-μZ(t)z(t). 47d

Extended LCT for dwell times given by finite mixtures of Poisson process event time distributions

It’s worth noting that in some applied contexts one may want to approximate a non-Erlang delay distribution with a mixture of Erlang distributions (see Sect. 3.7.1 and Appendix B for more details on making such approximations). Theorem 5 above details how assuming such a mixture distribution would be reflected in the structure of the corresponding mean field ODEs. This case can also be addressed in the more general context provided in Sect. 3.7.1.

Transition to multiple states following “competing” Poisson processes

We now consider the case where T, the time a particle spends in a given state X, follows the distribution given by T=miniTi, the minimum of n2 independent random variables Ti, where Ti has either an Erlang(ri,ki) distribution or, more generally, a Poisson process kith event time distribution with rate ri(t). Upon leaving state X, particles have the possibility of transitioning to any of m recipient states Y, =1,,m, where the probability of transitioning to state Y depends on which of the n random variables Ti was the minimum. That is, if a particle leaves X at time T=Ti=t, then the probability of entering state Y is pi(t).

The distribution associated with T is not itself an Erlang distribution or a Poisson process event time distribution, however its survival function is the product7 of such survival functions, i.e.,

S(t,τ)i=1nSriki(t,τ). 48

As detailed below, we can further generalize the recursion relation in Lemma 1 for the distributions just described above, which can then be used to produce a mean field system of ODEs based on appropriately partitioning X into sub-states.

Before considering this case in general, it is helpful to first describe the sub-states of X imposed by assuming the dwell time distribution described above, particularly the case where the distribution for each Ti is based on 1st event times (i.e., all ki=1). Recall that the minimum of n exponential random variables (which we may think of as 1st event times under a homogeneous Poisson process) is exponential with a rate that is the sum of the individual rates r=i=1nri. More generally, it is true that the minimum of n 1st event times under independent Poisson processes with rates ri(t) is itself distributed as the 1st event time under a single Poisson processes with rate r(t)i=1nri(t), thus in this case S(t,τ)=i=0nSri1(t,τ)=Sr1(t,τ). Additionally, if particles leaving state X are then distributed across the recipient states Y as described above, then this scenario is equivalent to the proportional outputs case described in Theorem 4 with a dwell time that follows a Poisson process 1st event time distribution with rate r(t)i=1nri(t) and a probability vector p=i=1npi(t)ri(t)/r(t), since P(T=Ti)=ri(T)/r(T). (This mean field equivalence of these two cases is detailed in Sect. 3.5.5.) Thus, the natural partitioning of X in this case is into sub-states with dwell times that follow iid 1st event time distributions with rate r(t)i=1Nri(t).

We may now describe the mean field ODEs for the more general case using the following notation. To index the sub-states of X, consider the ith Poisson process and its kith event time distribution which defines the distribution of Ti. Let ai{1,,ki} denote the event number a particle is awaiting under the ith Poisson process. Then we can describe the particle’s progress through X according to its progress along each of these n Poisson processes using the index vector αK, where

K={(a1,a2,,an)|aj{1,,kj}}. 49

Let KiK denote the subset of indices where ai=ki (where we think of particles in these sub-states as being poised to reach the kith event related to the ith Poisson process, and thus poised to transition out of state X).

To extend Lemma 2 for these distributions, define

u(t,τ,α)i=1ne-mi(t,τ)mi(t,τ)ai-1(ai-1)! 50

where mi(t,τ)=exp(-τtri(s)ds), and u(τ,τ,α)=1 if α=(1,,1) and u(τ,τ,α)=0 otherwise. Note that i=1nhriai(t,τ)=u(t,τ,α)i=1nri(t) (c.f. Lemma 2). Applying Eq. (9b) to S(t,τ) in Eq. (48), the survival function given by Eq. (48) [c.f. Eq. (31) and (9b)] can be written

S(t,τ)=αKu(t,τ,α). 51

We will also refer to the quantities u and S with the jth element of each product (in u) removed using the notation

u\j(t,τ,α)i=1,ijne-mi(t,τ)mi(t,τ)ai-1(ai-1)! 52a
S\j(t,τ)αKju\j(t,τ,α). 52b

This brings us to the following lemma, which generalizes Lemma 1 and Lemma 2 to distributions that are the minimum of n different (independent) Poisson process event times. As with the above lemmas, Lemma 3 will allow one to partition X into sub-states corresponding to each of the event indices in K describing the various stages of progress along each Poisson process prior to the first of them reaching the target event number.

Lemma 3

For u as defined in Eq. (50), differentiation with respect to t yields

ddtu(t,τ,α)=j=1nrj(t)u(t,τ,αj,-1)1[aj>1](α)-j=1nrj(t)u(t,τ,α) 53

where the notation αj,-1 denotes the index vector generated by decrementing the jth element of α, aj [assuming aj>1; for example, α2,-1=(a1,a2-1,,an)], and the indicator function 1[aj>1](α) is 1 if aj>1 and 0 otherwise.

Proof

Using the definition of u in Eq. (50) above, it follows that

ddtu(t,τ,α)=ddti=1ne-mi(t,τ)mi(t,τ)ai-1(ai-1)!=j=1n(i=1ijne-mi(t,τ)mi(t,τ)ai-1(ai-1)!)[-rj(t)e-mj(t,τ)mj(t,τ)aj-1(aj-1)!+1[aj>1](α)rj(t)e-mj(t,τ)mj(t,τ)aj-2(aj-2)!]=j=1n-rj(t)i=1ne-mi(t,τ)mi(t,τ)ai-1(ai-1)!+j=1n1[aj>1](α)rj(t)e-mj(t,τ)mj(t,τ)aj-2(aj-2)!i=1ijne-mi(t,τ)mi(t,τ)ai-1(ai-1)!=j=1nrj(t)u(t,τ,αj,-1)1[aj>1](α)-j=1nrj(t)u(t,τ,α). 54

The next theorem details the LCT extension that follows from Lemma 3.

Theorem 6

(Extended LCT for dwell times given by competing Poisson processes) Consider a continuous time dynamical system model of mass transitioning among multiple states, with inflow rate IX(t) to a state X. The distribution of time spent in state X (call it T) is the minimum of n random variables, i.e., T=mini(Ti), i=1,,n, where Ti are either Erlang(ri,ki) distributed or follow the more general (nonhomogeneous) Poisson process kith event time distribution with rate ri(t). Assume particles leaving X can enter one of m states Y, =1,,m. If a particle leaves X at time Ti (i.e., Ti occurred first, so T=Ti), and then the particle transitions into state Y with probability pi(T). Let x(t), and y(t) be the amount in each state, respectively, at time t, and assume non-negative initial conditions.

The mean field integral equations for this scenario, for =1,,m and i=1,,n, are

x(t)=x0S(t,0)+0tIX(s)S(t,s)ds 55a
y(t)=y(0)S(t,0)+0t(I(τ)+i=1npi(x0S\i(τ,0)hriki(τ,0)+0τIX(s)S\i(τ,s)hriki(τ,s)ds))S(t,τ)dτ. 55b

Equations (55) above are equivalent to

ddtx(1,,1)(t)=IX(t)-r(t)x(1,,1)(t), 56a
ddtxα(t)=i=1nri(t)xαi,-1(t)1[ai>1](α)-r(t)xα(t) 56b
y(t)=y(0)S(t,0)+0t(I(τ)+i=1npi(τ)αKiri(t)xα(τ))S(t,τ)dτ 56c

for all αK\(1,,1), r(t)=i=1nri(t), x(t)=αKxα(t), and

xα(t)=x0u(t,0,α)+0tIX(s)u(t,s,α)ds. 57

The y(t) equations (56c) may be further reduced to a system of ODEs, e.g., via Corollary 1.

Proof

Substituting Eq. (51) into Eq. (55a) yields

x(t)=x0αKu(t,0,α)+0tIX(s)αKu(t,s,α)ds=αK(x0u(t,0,α)+0tIX(s)u(t,s,α)ds)=αKxα(t). 58

Differentiating (57) yields equations Eqs. (56a) and (56b) as follows. First, if α=(1,,1) then by Lemma 3

ddtx(1,,1)(t)=-x0i=1nri(t)u(t,0,α)-i=1nri(t)0tIX(s)u(t,s,α)ds+IX(t)=IX(t)-i=1nri(t)x(1,,1)(t). 59

Next, if α has any ai>1, differentiating Eq. (57) and applying Lemma 3 yields

ddtxα(t)=x0ddtu(t,0,α)+0tIX(s)ddtu(t,s,α)ds=x0(i=1nri(t)u(t,0,αi,-1)1[ai>1](α)-i=1nri(t)u(t,α))+0tIX(s)(i=1nri(t)u(t,s,αi,-1)1[ai>1](α)-i=1nri(t)u(t,s,α))ds=i=1nri(t)xαi,-1(t)1[ai>1](α)-i=1nri(t)xα(t) 60

Note that, by the definitions of xα and u that initial condition x(0)=x0 becomes x(1,,1)(0)=x0 and xα(0)=0 for the remaining αK.

Eqs. (55b) become (56c), where Ki={α|αK,ai=ki}, by substituting Eqs. (52), (57), and S\i(t,τ)hriki(t,τ)=αKiri(t)u(t,τ,α), which yields

x0S\i(τ,0)hriki(τ,0)+0τIX(s)S\i(τ,s)hriki(τ,s)ds=x0αKiri(t)u(τ,0,α)+0τIX(s)αKiri(t)u(τ,s,α)ds=ri(t)αKi(x0u(τ,0,α)+0τIX(s)u(τ,s,α)ds)=αKiri(t)xα(τ). 61

Example 5

See Fig. 6. Suppose the X dwell time is T=min(T1,T2) where T1 and T2 are the k1th and k2th event time distributions under independent Poisson processes (call these PP1 and PP2) with rates r1(t) and r2(t), respectively. Assume that, upon leaving X, particles transition to Y1 if T=T1 or to Y2 if T=T2. By Theorem 6, we can partition X into sub-states defined by which event (under each Poisson process) particles are awaiting next. Upon entry into X, all particles enter sub-state X1,1 where they each await the 1st events under PP1 or PP2. If the next event to occur for a given particle is from PP1, the particle transitions to X2,1 where it awaits a 2nd event from PP1 or 1st event from PP2 (hence the subscript notation). Likewise, if PP2’s 1st event occurs before PP1’s 1st event, the particle would transition to X1,2, and so on. Particles would leave these two states to either X2,2, Y1, or Y2 depending on which event occurs next. Under these assumptions, with k1=k2=2 and exponential Yi dwell times with rates μ, then the corresponding mean field equations (using r(t)=r1(t)+r2(t)) are

dx11dt=IX(t)-r(t)x11(t) 62a
dx21dt=r1(t)x11(t)-r(t)x21(t) 62b
dx12dt=r2(t)x11(t)-r(t)x12(t) 62c
dx22dt=r1(t)x12(t)+r2(t)x21(t)-r(t)x12(t) 62d
dy1dt=r1(t)x22(t)-μ(t)y1(t) 62e
dy2dt=r2(t)x22(t)-μ(t)y2(t). 62f
Fig. 6.

Fig. 6

The sub-state diagram (cf. Fig. 3) for the scenario detailed in Example 5 after the application of Theorem 6 (extended LCT for dwell times given by competing Poisson processes). Here the X dwell time distribution is the minimum of two Erlang random variables TiErlang(ri,ki), i=1,2, which can be thought of as event time distribution under two homogeneous Poisson processes as detailed in the main text. We here assume that whichever of these occurs first determines whether particles leaving X transition to Y1 or Y2, respectively

It’s worth pointing out that the dwell times for the above sub-states of X are all identically distributed Poisson process 1st event times (note the loss rates in Eqs. (62a)–(62d), and recall the weak memorylessness property from Sect. 3.1.2). All particles in a X sub-state at time τ will spend a remaining amount of time in that state that follows a 1st event time distributions under a Poisson process with rate r(t)=r1(t)+r2(t). This is a slight generalization of the familiar fact that the minimum of n independent exponentially distributed random variables (with respective rates ri) is itself an exponential random variable (with rate ri=1nri). The next section addresses the generality of this observation about the sub-states of X.

Mean field equivalence of proportional outputs and competing Poisson processes for 1st event time distributions

The scenarios described in Sect. 3.5.1 (proportional distribution across multiple states Y after an Erlang dwell time in X) and Sect. 3.5.4 (proportional distribution across multiple states based upon competing Poisson processes), can lead to equivalent mean field equations when the X dwell times follow Poisson process 1st event time distributions, as is Example 5. This equivalence is detailed in Theorem 7, and is an important aspect of the GLCT detailed in Sect. 3.7 because it helps to show how sub-states with dwell times distributed as Poisson process 1st event times are the fundamental buildings blocks of the GLCT.

Theorem 7

(Mean field equivalence of proportional outputs and competing Poisson processes for 1st event time distributions) Consider the special case of Theorem 6 (the Extended LCT for competing Poisson processes) where X has a dwell time given by T=miniTi, where each Ti is a Poisson process 1st event time with rate ri(t), i=1,,n and particles transition to Y with probability pi(T) when T=Ti. The corresponding mean field model is equivalent to the special case of Theorem 4 (the Extended LCT for multiple outputs) where the X dwell time is a Poisson process 1st event time distribution with rate r(t)=i=1nri(t), and the transition probability vector for leaving X and entering state Y is given by p(t)=i=1npi(t)ri(t)/r(t).

Proof

First, in this case S(t,τ)=i=0nSri1(t,τ)=Sr1(t,τ). Since all ki=1, the probability that T=Ti is ri(T)/r(T), thus the probability that a particle leaving X at t goes to Y is p(t)=i=1nri(t)r(t)pi(t). Substituting the above equalities into the mean field Eq. (56a) (where there’s only one possible index in K={(1,1,,1)}) and (56c) gives

ddtx(t)=IX(t)-r(t)x(t) 63a
yj(t)=yj(0)Sj(t,0)+0t(Ij(τ)+r(t)pj(τ)x(τ))Sj(t,τ)dτ 63b

which are the mean field equations for the aforementioned special case of Theorem 4.

Modeling intermediate state transitions: reset the clock, or not?

We next describe how to apply extensions of the LCT in two similar but distinctly different scenarios (see Fig. 7) where the transition to one or more intermediate sub-states either resets an individual’s overall dwell time in state X (by assuming the time spent in an intermediate sub-state XIi is independent of time already spent in X0; see Sect. 3.6.1), or instead leaves the overall dwell time distribution for X unchanged (by conditioning the time spent in intermediate state XIi on the time already spent in X0; see Sect. 3.6.2).

Fig. 7.

Fig. 7

Should the overall dwell time distribution for state X be “reset” by the transition from base sub-state X0 to intermediate sub-state XI (i.e., should the dwell time in state XI be independent of the time already spent in X0?), or should the XI dwell time be conditioned on time already spent in X0 so that the X0XI transition does not alter the overall dwell time in state X? How do these different assumptions alter the structure of the corresponding mean field ODEs? We answer these question in Sect. 3.6 where we describe how to apply the LCT in scenarios with intermediate states, assuming in Sect. 3.6.1 that the dwell time distribution for XI is independent of the amount of time spent in X0, and assuming in Sect. 3.6.2 that the overall dwell time for X is unaffected by transitions from X0 to XI

An example of these different assumptions leading to important differences in practice comes from Feng et al. (2016) where individuals infected with Ebola can either leave the infected state (X) directly (either to a recovery or death), or after first transitioning to an intermediate hospitalized state (XI) which needs to be explicitly modeled in order to incorporate a quarantine effect into the rates of disease transmission (i.e., the force of infection should depend on the number of non-quarantined individuals, i.e., X0). As shown in Feng et al. (2016), the epidemic model output depends strongly upon whether or not it is assumed that moving into the hospitalized sub-state impacts the distribution of time spent in the infected state X.

To most simply illustrate these two scenarios, consider the simple case in Fig. 7 where a single intermediate sub-state XI is being modeled, and particles enter X into sub-state X0 at rate IX(t). Let X=X0XI. Both cases assume particles transition out of X0 either to sub-state XI or they leave state X directly and enter state Y. Both cases also assume the distribution of time spent in X0 is T=min(T0,T1) where particles transition to XI if T1<T0 (i.e., if T=T1) or to Y if T0<T1 (where each Ti is the kith event time under Poisson processes with rates r0(t) and r1(t) (see Sects. 3.5.4 and 3.5.5). Let TI denote the distribution of time spent in intermediate state XI. The first case assumes TI is independent of time spent in X0 (i.e., the transition to XI ‘resets the clock’; see Sect. 3.6.1). The second case assumes TI is conditional on time already spent in X0 (call it t0), such that the total amount of time spent in X, t0+TI, is equivalent in distribution to T0 (i.e., the transition to XI does not change the overall distribution of time spent in X; see Sect. 3.6.2).

In the next two sections, we provide extensions of the LCT for generalizations of these two scenarios, extended to multiple possible intermediate states with eventual transitions out of X into multiple recipient states.

Intermediate states that reset dwell time distributions

First, we consider the case in which the time spent in the intermediate state XI is independent of the time already spent in X (i.e., in the base state X0). This is arguably the more commonly encountered (implicit) assumption found in ODE models that aren’t explicitly derived from a stochastic model and/or mean field integro-differential delay equations.

The construction of mean field ODEs for this case is a straightforward application of Theorem 6 from the previous section, combined with the extended LCT with output to multiple states (Theorem 4). Here we have extended this scenario to include MX intermediate sub-states XIj where the transition to those sub-states from base state X0 is based on the outcome of N competing Poisson process event time distributions (Ti), and upon leaving the intermediate states particles transition out of state X into one of MY possible recipient states Y.

Theorem 8

(Extended LCT with dwell time altering intermediate sub-state transitions) Suppose particles enter X at rate IX(t) into a base sub-state X0. Assume particles remain in X0 according to a dwell time distribution given by T, the minimum of N+1 independent Poisson process kith event time distributions with rates ri(t), i=0,,N, i.e., T=mini(Ti). Particles leaving X0 transition to one of MX1 intermediate sub-states XIi or to one of MY1 recipient states Y according to which Ti=T. If T0=T then the particle leaves X and the probability of transitioning to Y is p0(T), where =1MYp0(T)=1. If Ti=T for i1 then the particle transitions to XIj with probability pij(T), where j=1MXpij(T)=1. Particles in intermediate state XIj remain there according to the κith event times under a Poisson process with rate ϱi(t), and then transition to state Y with probability qj(t), where (for fixed t) =1MYqj(t)=1, and they remain in Y according to a dwell time with survival function S(t,τ).

In this case the corresponding mean field equations are

ddtx0(1,,1)(t)=IX(t)-i=0Nri(t)x(1,,1)(t) 64a
ddtx0α(t)=i=0Nri(t)(x0αi,-1(t)1[ai>1](α)-x0α(t)) 64b
ddtxIj1(t)=IXIj(t)+pij(t)(αKiri(t)x0α(t))-ϱj(t)xIj1(t) 64c
ddtxIjk(t)=ϱj(t)(xIj,k-1(t)-xIjk(t)),k=2,,κj 64d
y(t)=y(0)S(t,0)+0t(IY(τ)+p0(τ)αK0r0(τ)x0α(τ)+j=1MXϱj(τ)xIjκj(τ)qj(τ))SY(t,τ)dτ. 64e

where the amount in base sub-state X0 is x0(t)=αKx0α(t), the amount in the jth intermediate state XIj is xIj(t)=k=1κjxIjk(t) (see Theorem 6 for notation), K={(a0,a1,,aN)|aj{1,,kj}}, Ki={αK|ai=ki}, j=1,,N, =1,,MY, and in Eq. (64b) α=(a0,,aN)K\(1,,1). Note that the y(t) equations (64e) may be further reduced to a system of ODEs, e.g., via Corollary 1, and that more complicated distributions for dwell times in intermediate states XIi (e.g., an Erlang mixture) could be similarly modeled according to other cases addressed in this manuscript.

Proof

This follows from applying Theorem 6 to X0 and treating the intermediate states XIj as recipient states, then applying Theorem 4 to each intermediate state to partition each XIj into XIjk, k=1,,κj, yielding Eq. (64).

Example 6

Consider the scenario in Fig. 7. Let the X0 dwell time be the minimum of T0Erlang(r0,2) and T1Erlang(r1,2), with intermediate state dwell time TI1Erlang(ϱ1,κ1=3) and an exponential (rate μ) dwell time in Y. Assume the only inputs into X are into X0 at rate IX(t). By Theorem 8 the corresponding mean field ODEs (see Fig. 8) are Eq. (65), where x0(t)=x0(1,1)(t)+x0(2,1)(t)+x0(1,2)(t)+x0(2,2)(t) and xI1(t)=xI11(t)+xI12(t)+xI13(t).

ddtx0(1,1)(t)=IX(t)-(r0+r1)x0(1,1)(t) 65a
ddtx0(2,1)(t)=r0x0(1,1)(t)-(r0+r1)x0(2,1)(t) 65b
ddtx0(1,2)(t)=r1x0(1,1)(t)-(r0+r1)x0(1,2)(t) 65c
ddtx0(2,2)(t)=r0x0(1,2)(t)+r1x0(2,1)(t)-(r0+r1)x0(2,2)(t) 65d
ddtxI11(t)=r1x0(1,2)(t)+r1x0(2,2)(t)-ϱ1xI11(t) 65e
ddtxI12(t)=ϱ1xI11(t)-ϱ1xI12(t) 65f
ddtxI13(t)=ϱ1xI12(t)-ϱ1xI13(t) 65g
ddty(t)=IY(t)+r0x0(2,1)(t)+r0x0(2,2)(t)+ϱ1xI13(t)-μy(t). 65h
Fig. 8.

Fig. 8

When modeling sub-state transitions, e.g., from a base state X0 to intermediate state XI1 (state X=X0XI1), the intermediate sub-state dwell-time can be (a) independent of time already spent in X0 (Sect. 3.6.1, Theorem 8), or (b) conditioned on that time so that the dwell time in state X is unaffected by the sub-state transition (Sect. 3.6.2, Theorem 9). In both cases, the dwell time distribution for X0 is the minimum of independent Erlang distributions, and sub-states within X0 are as discussed in Sect. 3.5.4. Panel (a): The dwell time in XI1 is Erlang(ϱ1,κ1) and independent of time spent in X0. Thus, the sub-state transition alters the overall dwell time in state X. For the more general case, see Sect. 3.6.1 and Theorem 8. Panel (b): The overall X dwell time distribution T0Erlang(r0,k0) is preserved by conditioning the dwell time in the intermediate state on the prior progress through sub-states of X0, as detailed in Sect. 3.6.2. Compare the transitions from X0 to XI1 in these two cases, and recall Fig. 6 and the weak memoryless property of Poisson process first event times from Sect. 3.1.2

Intermediate states that preserve dwell time distributions

In this section we address how to modify the outcome in the previous section to instead construct mean field ODE models that incorporate ‘dwell time neutral’ sub-state transitions, i.e., where the distribution of time spent in X is the same regardless of whether or not particles transition (within X) from some base sub-state X0 to one or more intermediate sub-states XIj. This is done by conditioning the dwell time distributions in XIi on time spent in X0 in a way that leverages the weak memorylessness property discussed in Sect. 3.1.2.

In applications, this case (in contrast to the previous case) is perhaps the more commonly desired assumption, since modelers often seek to partition states into sub-states where key characteristics, like the overall X dwell time distribution, remain unchanged, but where the different sub-states have functional differences elsewhere in the model. For example, transmission rate reductions from quarantine in SIR type infectious disease models.

One approach to deriving such a model is to condition the dwell time distribution for an intermediate state XIi on the time already spent in X0 (as in Feng et al. 2016). We take a slightly different approach and exploit the weak memoryless property of Poisson process 1st event time distributions (see Theorem 1 in Sect. 3.1.2, and the notation used in the previous section) to instead condition the dwell time distribution for intermediate states XIj on how many of the k0 events have already occurred when a particle transitions from X0 to XIj (rather than conditioning on the exact elapsed time spent in X0). In this case, since each sub-state of X0 has iid dwell time distributions that are Poisson process 1st event times with rate r(t)=i=0Nri(t), if i of the k0 events had occurred prior to the transition out of X0, then the weak memoryless property of Poisson process 1st event time distributions implies that the remaining time spent in XIj should follow a (k0-i)th event time distribution under an Poisson process with rate r0(t), thus ensuring that the total time spent in X follows a k0th event time distribution with rate r0(t). With this realization in hand, one can then apply Theorem 6 and Theorem 4 as in the previous section to obtain the desired mean field ODEs, as detailed in the following Theorem, and as illustrated in Fig. 8.

Theorem 9

(Extended LCT with dwell time preserving intermediate states) Consider the mean field equations for a system of particles entering state X (into sub-state X0) at rate IX(t). As in the previous case, assume the time spent in X0 follows the minimum of N+1 independent Poisson process kith event time distributions with respective rates ri(t), i=0,,N (i.e., T=mini(Ti)). Particles leaving X0 at time T transition to recipient state Y with probability p0(T) if T=T0, or if T=Ti (i=1,,N) to the jth of MX intermediate sub-states, XIj, with probability pij(T). If T<T0, we may define a random variable K{0,,k0-1} indicating how many events had occurred under the Poisson process associated with T0 at the time of the transition out of X0 (at time T). In order to ensure the overall time spent in X follows a Poisson process k0th event time distribution with rate r0(t), it follows that particles entering state, XIj will remain there for a duration of time that is conditioned on K=k such that the conditional dwell time for that particle in XIj will be given by a Poisson process (k0-k)th event time with rate r0(t). Finally, assume that particles leaving X via intermediate sub-state XIj at time t transition to Y with probability qj, where they remain according to a dwell time determined by survival function S(t,τ).

The corresponding mean field equations are

ddtx0(1,,1)(t)=IX(t)-i=0Nri(t)x0(1,,1)(t) 66a
ddtx0α(t)=i=0Nri(t)x0αi,-1(t)1[ai>1](α)-i=0Nri(t)x0α(t) 66b
ddtxIjk(t)=r0(t)(xIj,k-1(t)1[k>1]-xIjk(t))+αKijri(t)xα(t)pij(t) 66c
y(t)=y(0)S(t,0)+0t(IY(τ)+αK0r0(τ)xα(τ)+j=1MXr0(τ)xIjk0(τ)qj(τ))S(t,τ)dτ 66d

where K={(a0,a1,,aN)|aj{1,,kj}}, αK\(1,,1), j=1,,MX, k=1,,k0, =1,,MY, KiK are the subset of indices where ai=ki, KijKi are the subset of indices where ai=ki and a0=j, x0(t)=αKx0α(t), xiI(t)=j=1k0xiIj(t), and x(t)=x0(t)+i=1nxiI(t). The y(t) equations (66d) may be reduced to ODEs, e.g., via Corollary 1.

Proof

The proof of Theorem 9 parallels the proof of Theorem 8, but with the following modifications. First, each sub-state of XIj (for all j) has the same dwell time distribution, namely, they are all 1st event time distributions under a Poisson process with rate r0(t). Second, upon leaving X0 where T=Ti and K(T)=k<k0 (i.e., when only k<k0 events have occurred under the 0th Poisson process; see the definition of K in the text above) particles will enter (with probability pij(T)) the jth intermediate state XIj by entering sub-state XIjk which (due to the weak memorylessness property described in Theorem 1) ensures that, upon leaving XIj particles will have spent a duration of time that follows the Poisson process k0th event time distribution with rate r0(t).

Example 7

Consider Example 6 in the previous section, but instead assume that the transition to the intermediate state does not impact the overall time spent in state X, as detailed in Theorem 9. The corresponding mean field ODEs for this case are given by Eq. (67) below [compare Eqs. (67e)–(67g) to Eqs. (65e)–(65h)].

ddtx0(1,1)(t)=IX(t)-(r0+r1)x0(1,1)(t) 67a
ddtx0(2,1)(t)=r0x0(1,1)(t)-(r0+r1)x0(2,1)(t) 67b
ddtx0(1,2)(t)=r1x0(1,1)(t)-(r0+r1)x0(1,2)(t) 67c
ddtx0(2,2)(t)=r0x0(1,2)(t)+r1x0(2,1)(t)-(r0+r1)x0(2,2)(t) 67d
ddtxI11(t)=r1x0(1,2)(t)-r0xI11(t) 67e
ddtxI12(t)=r1x0(2,2)(t)+r0xI11(t)-r0xI12(t) 67f
ddty(t)=IY(t)+r0x0(2,1)(t)+r0x0(2,2)(t)+r0xI12(t)-μy(t). 67g

Generalized Linear Chain Trick (GLCT)

In the preceding sections we have provided various extensions of the Linear Chain Trick (LCT) that describe how the structure of mean field ODE models reflects the assumptions that define corresponding continuous time stochastic state transition models. Each case above can be viewed as a special case of the following more general framework for constructing mean field ODEs, which we refer to as the Generalized Linear Chain Trick (GLCT).

The cases we have addressed thus far share the following stochastic model assumptions, which constitute the major assumptions of the GLCT.

  1. A focal state (which we call state X) can be partitioned into a finite number of sub-states (e.g., X1,,Xn), each with independent (across states and particles) dwell time distributions that are either exponentially distributed with rates ri or, more generally, are distributed as independent 1st event times under nonhomogeneous Poisson processes with rates ri(t), i=1,,n. Recall the equivalence relation in Sect. 3.5.5.

  2. Inflow rates into X can be described by non-negative, integrable inflow rates into each of these sub-states (e.g., IX1(t),,IXn(t)), some or all of which may be zero. This includes the case where particles enter X at rate Λ(t) and are distributed across sub-states Xi according to the probabilities ρ(t)=[ρ1(t),,ρn(t)]T (i.e., we let IXi(t)ρi(t)Λ(t)) where iρi=1.

  3. Particles that transition out of a sub-state Xi at time t transition either into sub-state Xj with probability pij(t), or into one of recipient states Y, =1,,m, with probability pi,n+. That is, let pij(t) denote the probability that a particle leaving state Xi at time t enters either Xj if jn or Yj-n if j>n, where i=1,,n, j=1,,n,n+1,,n+m.

  4. Recipient states Y, =1,,m, also have dwell time distributions defined by survival functions SY(t,τ) and integrable, non-negative inflow rates IY(t) that describe inputs from all other non-X sources.

The GLCT (Theorem 10) describes how to construct mean field ODEs for states X and Y for state transition models satisfying the above assumptions.

Theorem 10

(Generalized Linear Chain Trick) Consider a stochastic, continuous time state transition model of particles entering state X and transitioning to states Y, =1,,m, according to the above assumptions A1-A4. Then the corresponding mean field model is given by

ddtxi(t)=IXi(t)+j=1npji(t)rj(t)xj(t)-ri(t)xi(t),i=1,,n, 68a
y(t)=y(0)SY(t,0)+0t(IY(τ)+j=1nrj(t)xj(τ)pj,n+(t))SY(t,τ)dτ 68b

where x(t)=i=1nxi(t), and we assume non-negative initial conditions xi(0)=xi0, y(0)=y0. Note that the y(t) equations might be reducible to ODEs, e.g., via Corollary 1 or other results presented above.

Furthermore, Eq. (68a) may be written in vector form where PX(t)=(pij(t)) (i,j{1,,n}) is the n×n matrix of (potentially time-varying) probabilities describing which transitions out of Xi at time t go to Xj (likewise, one can define PY(t)=(pij(t)), i{1,,n}, j{n+1,,n+m}, which is the n×m matrix of probabilities describing which transitions from Xi at time t go to Yj-n), IX(t)=[IX1,,IXn]T, R(t)=[r1(t),,rn(t)]T, and x(t)=[x1(t),,xn(t)]T which yields

ddtx(t)=IX(t)+PX(t)T(R(t)x(t))-R(t)x(t). 69

where indicates the Hadamard (element-wise) product.

Proof

The proof of the theorem above follows directly from applying Theorem 4 to each sub-state.

Example 8

(Dwell time given by the maximum of independent Erlang random variables) We here illustrate how the GLCT can provide a conceptually simpler framework for deriving ODEs relative to derivation from mean field integral equations by assuming the X dwell time obeys the maximum of multiple Erlang distributions. While the survival function for this distribution is not straightforward to write down, it is fairly straightforward to construct a Markov Chain that yields such a dwell time distribution (see Fig. 9).

Fig. 9.

Fig. 9

The X sub-state structure for Example 8 where X dwell time distribution follows the maximum of two Erlang random variables with rates r1 and r2, respectively, and shape parameters ki=2. As in Sect. 3.5.4, where the minimum is assumed instead of the maximum, X can be partitioned using indices based on organizing particles by which events they are awaiting under each Poisson process associated with each Erlang distribution. Upon reaching the target event (here, the 2nd event) under any given Poisson process, particles transition to sub-states with an asterisk in the corresponding index position (e.g., see figure). In general, these sub-states all have a dwell times given by the 1st event time under a Poisson process, but with differing rates (see Example 8): here they follow exponential distributions with either rate r=r1+r2 (white backgrounds), rate r2 (gray backgrounds), or rate r1 (lined backgrounds)

Recall that, in Sect. 3.5.4, we considered a dwell time given by the minimum of N Erlang distributions. Here we instead consider the case where the dwell time distribution is given by the maximum of multiple Erlang distributions, T=max(T1,T2) where TiErlang(ri,2). For simplicity, assume the dwell time in a single recipient state Y is exponential with rate μ. We again partition X according to which events (under the two independent homogeneous Poisson processes associated with each of T1 and T2) particles are awaiting, and index those sub-states accordingly (see Fig. 9). These sub-states are X11, X21, X12, X1, X22, X1, X2, and X2, where a ‘’ in the ith index position indicates that particles in that sub-state have already had the ith Poisson process reach the kith event (in this case, the 2nd event). Each such sub-state has exponentially distributed dwell times, but rates for these dwell time distributions differ (unlike the cases in Sect. 3.5.4 where all sub-states had the same rate): the Poisson process rates for sub-states X11, X21, X12, and X22 are r=r1+r2 (see Fig. 9 and compare to Theorem 6 and Fig. 6), but the rate for the states X1 and X2 (striped circles in Fig. 9) are r1 , and for X1 and X2 (shaded circles in Fig. 9) are r2.

In the context of the GLCT, let x(t)=[x11(t), x21(t), x12(t), x1(t), x22(t), x1(t), x2(t), x2(t)]T then by the assumptions above R(t)=[r, r, r, r2, r, r1, r2, r1]T, ρ(t)=[1,0,,0]T and hence IX(t)=[IX(t),0,,0]T. Denote p1r1/r and p2r2/r (à la Theorem 7 in Sect. 3.5.5). Then the first eight rows of 9×9 matrix P are given by

graphic file with name 285_2019_1412_Equ136_HTML.gif 70

Thus, by the GLCT (Theorem 10), the corresponding mean field ODEs are

ddtx11(t)=IX(t)-rx11(t) 71a
ddtx21(t)=r1x11(t)-rx21(t) 71b
ddtx12(t)=r2x11(t)-rx12(t) 71c
ddtx1(t)=r1x21(t)-r2x1(t) 71d
ddtx22(t)=r2x21(t)+r1x12(t)-rx22(t) 71e
ddtx1(t)=r2x12(t)-r1x1(t) 71f
ddtx2(t)=r2x1(t)+r1x22(t)-r2x2(t) 71g
ddtx2(t)=r1x1(t)+r2x22(t)-r1x2(t) 71h
ddty(t)=r1x2(t)+r2x2(t)-μy(t). 71i

GLCT for phase-type distributions

The GLCT above extends the LCT to a very flexible family of dwell time distributions known as (continuous) phase-type distributions (Asmussen et al. 1996; Pérez and Riaño 2006; Osogami and Harchol-Balter 2006; Thummler et al. 2006; Reinecke et al. 2012a; Horváth et al. 2012; Komárková 2012; Horváth et al. 2016; Okamura and Dohi 2015; Horváth and Telek 2017), i.e., the hitting time distributions for Continuous Time Markov Chains (CTMC). These CTMC hitting time distributions include the hypoexponential distributions, hyper-exponential and hyper-Erlang distributions, and generalized Coxian distributions (Reinecke et al. 2012a; Horváth et al. 2016). Importantly, these distributions can be fit to data or can be used to approximate other named distributions (e.g., see Horváth and Telek 2017; Osogami and Harchol-Balter 2006; Altiok 1985; Pérez and Riaño 2006; Komárková 2012; Reinecke et al. 2012b, and related publications). As detailed below, this enables modelers to incorporate a much broader set of dwell time distributions into ODEs than is afforded by the standard LCT.

Consider the assumptions of Theorem 10 above. Assume vectors ρ(t)=ρ and R(t)=R, and matrices PX(t)=PX, and PY(t)=PY are all constant. As above, assume the probability of entering states in Y is zero, thus our initial distribution vector for this CTMC (with n+m states) is fully determined by the (length n) vector ρ. Also assume—just to define the CTMC that describes transitions among transient states X up to (but not after) entering states Y—that each state in Y is absorbing (i.e., pii=1, i>n). Then the X dwell time distribution follows the hitting time distribution for a CTMC with initial distribution vector ρ and (n+m)×(n+m) transition probability matrix

P=PXPY0I. 72

To clearly state the GLCT for phase-type distributions, we must reparameterize the above CTMC. First, there is an equivalent parameterization of this CTMC which corresponds to thinning the underlying Poisson processes so that we only track transitions between distinct states, and ignore when an individual leaves and instantly returns to it’s current state (this thinned process is often called the embedded jump process).

The rate for the thinned process that determines a particle’s dwell time in transient state i goes from ri to

λiri(1-pii) 73

If 0pii<1, the transition probabilities out of state i then get normalized to

p~ijpij(1-pii),forji,andp~ii=0. 74

The rows for absorbing states (Y) remain unchanged, i.e., p~ij=pij for i>n. The resulting transition probability matrix P~ and rate vector λ define the embedded jump process description of the CTMC with transition probability matrix P and rate vector R (initial probability vector ρ is the same for both representations of this CTMC).

Lastly, this CTMC can again be reparameterized by combining the jump process transition probability matrix P~ and rate vector λ to yield this CTMC’s transition rate matrix (also sometimes called the infinitesimal generator matrix or simply the generator matrix) which is defined as follows. Let matrix G be the same dimension as P (and thus, P~) and let the first n terms in the diagonal of G be the negative of the jump process rates, -λ (i.e., Gii=-λi, in). Let the off diagonal entries of the first n rows of G be the jump process transition probabilities p~ij multiplied by the ith rate λi (i.e., Gij=λip~ij, ji). Thus, the first row of G is

[-λ1p~12λ1p~1nλ1p~1,n+mλ1]

and so on. Since the transition rates out of absorbing states (e.g., the last m rows of G) are 0, G has the form

G=GXGY00. 75

Note that P~ and λ can be recovered from G using the definitions above.

This third parameterization of the given CTMC, determined solely by initial distribution ρ and transition rate matrix G, can now to used to formally describe the phase-type distribution associated with this CTMC, i.e., the distribution of time spent in the transient states X before hitting absorbing states Y. Specifically, the phase-type distribution density function and CDF are

f(t)=ρTetGX(-GX1) 76a
F(t)=1-ρTetGX1 76b

where 1 is a n×1 vector of ones. Importantly, this distribution depends only on the n×n matrix GX and length n initial distribution vector ρ.

The GLCT for phase-type distributions can now be stated as follows.

Corollary 2

(GLCT for phase-type distributions) Assume particles enter state X at rate Λ(t) and that the dwell time distribution for a state X follows a continuous phase-type distribution given by the n×1 initial probability vector ρ and n×n matrix GX. Let IXi(t)=ρiΛ(t) and IX(t)=[IX1,,IXn]T. Then Eq. (69) in Theorem 10 becomes

ddtx(t)=IX(t)+GXTx(t). 77
Example 9

(Serial LCT and hypoexponential distributions) Assume the dwell time in state X is given by the sum of independent (not identically distributed) Erlang distributions or, more generally, Poisson process kith event time distributions with rates ri(t), i.e., T=iTi, i=1,,N (note the special case where all ki=1 and ri(t)=ri are constant, which yields that T follows a hypoexponential distribution). Let n=iki and further assume particles go to Y with probability p upon leaving X, =1,,m. Using the GLCT framework above, this corresponds to partitioning X into sub-states Xj, where j=1,,n, and

R(t)=[r1(t),r1(t),,r2(t),,rn(t)]T 78

where the first k1 elements of R(t) are r1(t), the next k2 are r2(t), etc., and

PX=0100000100000100000100000n×n,PY=000000p1p2pmn×. 79

By the GLCT (Theorem 10), using r(j)(t) to denote the jth element of R(t) in Eq. (78), the corresponding mean field equations are

ddtx1(t)=IX(t)-r1(t)x1(t) 80a
ddtxj(t)=r(j-1)(t)xj-1(t)-r(j)(t)xj(t),forj1, 80b
y(t)=y(0)SY(t,0)+0t(IY(τ)+j=1mr(j)(t)xj(τ)pj(t))SY(t,τ)dτ. 80c

Note, the phase-type distribution form of the hypoexponential distribution could be used with Corollary 2 to derive the xi equations in Eq. (9).

Example 10

(SIR model with Phase-Type Duration of Infectiousness.) Consider the simple SIR model given by Eq. (2) with mass action transmission λ(t)=βI(t). Suppose the assumed exponentially distributed dwell times in state I were to be replaced by a phase-type distribution with initial distribution vector α and matrix A (note that, in some cases, it is possible to match the first three or more moments using only a 2×2 or 3×3 matrix A and note that Matlab and Python routines for making such estimates are freely available in Horváth and Telek 2017). Then, letting x be the vector of substates of I and I(t)=x1(t)++xn(t), by the GLCT for phase-type distributions (Corollary 2) the corresponding mean field ODE model, with R(t)=N0-S(t)-I(t), is

ddtS(t)=-βI(t)S(t) 81a
ddtx(t)=αβI(t)S(t)+ATx(t). 81b

Discussion

ODEs are arguably the most popular framework for modeling continuous time dynamical systems in applications. This is in part due to the relative ease of building, simulating, and analyzing ODE models, which makes them very accessible to a broad range of scientists and mathematicians. The above results generalize the Linear Chain Trick (LCT), and detail how to construct mean field ODE models for a broad range of scenarios found in applications based on explicit individual-level stochastic model assumptions. Our hope is that these contributions improve the speed and efficiency of constructing mean field ODE models, increase the flexibility to make more appropriate dwell time assumptions, and help clarify (for both modelers and those reading the results of their work) how individual-level stochastic assumptions are reflected in the structure of mean field ODE model equations. We have provided multiple novel theorems that describe how to construct such ODEs directly from underlying stochastic model assumptions, without formally deriving them from an explicit stochastic model or from intermediate integral equations. The Erlang distribution recursion relation (Lemma 1) that drives the LCT has been generalized to include the time-varying analogues of Erlang distributions—i.e., kth event time distributions under nonhomogeneous Poisson processes (Lemma 2)—and distributions that reflect “competing Poisson process event times” defined as the minimum of a finite number of independent Poisson process event times (Lemma 3). These new lemmas, and our generalization of the memorylessness property of the exponential distribution (which we refer to in Sect. 3.1.2 as the weak memorylessness property of nonhomogeneous Poisson process 1st event time distributions) together allow more flexible dwell time assumptions to be incorporated into mean field ODE models. We have also introduced a novel Generalized Linear Chain Trick (GLCT; Theorem 10 in Sect. 3.7) which complements previous extensions of the LCT (e.g., Jacquez and Simon 2002; Diekmann et al. 2017) and allows modelers to construct mean field ODE models for a broad range of dwell time assumptions and associated sub-state configurations (e.g., conditional dwell time distributions for intermediate sub-state transitions), including the phase-type family of distributions and their time-varying analogues as detailed in Sect. 3.7.1. The GLCT also provides a framework for considering other scenarios not specifically addressed by the preceding results, as illustrated by Example 8 in which the dwell time distribution follows the maximum of multiple Erlang distributions.

These results not only provide a framework to incorporate more accurate dwell time distributions into ODE models, but also hopefully encourage more comparative studies, such as Feng et al. (2016), that explore the dynamic and application-specific consequences of incorporating non-Erlang dwell time distributions, and conditional dwell time distributions, into ODE models. More study is needed, for example, comparing integral equation models with non-exponential dwell time distributions and their ODE counterparts based on approximating those distributions with Erlang and phase-type distributions.

The GLCT also permits modelers to incorporate the flexible phase-type distributions (i.e., the hitting-time distributions for Continuous Time Markov Chains) into ODE models. This family of probability distributions includes mixtures of Erlang distributions (a.k.a. hyper-Erlang distributions), the minimum or maximum of multiple Erlang distributions, the hypoexponential distributions, generalized Coxian distributions, and others (Reinecke et al. 2012a; Horváth et al. 2016). While the phase-type distributions are currently mostly unknown to mathematical biologists, they have received some attention in other fields and modelers can take advantage of existing methods that have been developed to fit phase-type distributions to other distributions on R+ and to data (see Sect. 3.7.1 for references). These results provide a flexible framework for approximating dwell time distributions, and incorporating those empirically or analytically derived dwell time distributions into ODE models.

There are some additional considerations, and potential challenges to implementing these results in applications, that are worth addressing. First, the increase in the number of state variables may lead to both computational and analytical challenges, however we have a growing number of tools at our disposal for tackling higher dimensional systems (e.g., van den Driessche and Watmough 2002; Guo et al. 2008; Li and Shuai 2010; Du and Li 2014; Bindel et al. 2014). We intend to explore the costs associated with this increase in dimensionality in future publications. Second, it is tempting to assume that the sub-states resulting from the above theorems correspond to some sort of sub-state structure in the actual system being modeled. This is not necessarily the case, and we should be cautious about interpreting these sub-states as evidence of, e.g., cryptic population structure. Third, some of the above theorems make a simplifying assumption that, either at time t=0 or upon entry into X, the initial distribution of particles is only into the first sub-state. This may not be the appropriate assumption to make in some applications, but it is fairly straight forward to modify these these initial condition assumptions within the context of the GLCT. Fourth, it may be appropriate in some applications to avoid mean field models, and instead analyze the stochastic model dynamics directly [e.g., see Allen (2010, 2017) and references therein]. Lastly, the history of successful attempts to garner scientific insights from mean field ODE models seems to suggest that such distributional refinements are unnecessary. However, this is clearly not always the case, as evidenced by the many instances in which modelers have abandoned ODEs and instead opted to use integral equations to model systems with non-Erlang dwell time distributions. We hope these results will facilitate more rigorous comparisons between such models and their simplified and/or ODE counterparts, like Feng et al. (2016) and Piotrowska and Bodnar (2018), to help clarify when the use of these ODE models is warranted.

In closing, these novel extensions of the LCT help clarify how underlying assumptions are reflected in the structure of mean field ODE models, and provide a means for incorporating more flexible dwell time assumptions into mean field ODE models directly from first principles without a need to derive ODEs from stochastic models or intermediate mean field integral equations. The Generalized Linear Chain Trick (GLCT) provides both a conceptual framework for understanding how individual-level stochastic assumptions are reflected in the structure of mean field model equations, and a practical framework for incorporating exact, approximate, or empirically derived dwell time distributions and related assumptions into mean field ODE models.

Acknowledgements

The authors thank Michael H. Cortez, Jim Cushing, Marisa Eisenberg, Jace Gilbert, Zoe Haskell, Tomasz Kozubowski, Miles Moran, Catalina Medina, Amy Robards, Deena R. Schmidt, Joe Tien, Narae Wadsworth, and an anonymous referee for comments and suggestions that improved this work.

Appendix A Deterministic models as mean field equations

We here give a brief description of the process of deriving deterministic mean field equations from stochastic first principles.

Consider a stochastic model of discrete particles (e.g., individual organisms in a population, molecules in a solution, etc.) that transition among a finite number of n states in continuous time. Let the state variables of interest be the amount of particles in each state. This set of counts in each state can be thought of as a state vector in the state spaceNnRn, and our model describes the (stochastic) rules governing transitions from one state vector to the next (i.e., from a given state there is some probability distribution across the state space describing how the system will proceed).

Mean field models essentially average that distribution, and thus describe the mean state transitions from any given state of the system. That is, for a given state in Rn, we can think of a probability distribution that describes where the system would move from that point in Rn, and find the mean transition direction in state space, which then defines a deterministic dynamical system on Rn which we refer to as a mean field model for the given stochastic process. To derive a continuous time mean field model, we do this for an arbitrary discrete step size Δt so we can then take the limit as Δt0.

A.1 Deriving mean field integral equations and ODEs

Here we illustrate one approach to derive mean field equations starting from an explicit stochastic model using a simple toy example. To do this, we begin by describing a discrete time approximation (with arbitrarily small time step Δt) of a stochastic system of particles transitioning among multiple states, then derive the corresponding discrete-time mean field model which then yields the desired mean field model by taking the limit as Δt0. Other more systematic approaches exist, e.g., see Kurtz (1970, 1971).

Consider a large number of particles that begin (at time t=0) in state W and then each transitions to state X after an exponentially distributed amount of time (with rate a). Particles remain in state X according to a Erlang(rk) distributed delay before entering state Y. They then go to state Z after an exponentially distributed time delay with rate μ. Let w~(t), x~(t), y~(t), and z~(t) be the amount in each of the corresponding states at time t0, with w(0)=N0, and x(0)=y(0)=z(0)=0. Then, as we will derive below, the corresponding continuous time mean field model is given by Eq. (A1) where the state variables w, x, y, and z correspond to the amount in each of the corresponding states, and we assume the initial conditions w(0)=w01 and x(0)=y(0)=z(0)=0.

ddtw(t)=-aw(t) A1a
x(t)=0taw(s)I(s)Srk(t-s)proportion remainingds A1b
y(t)=0t0τaw(s)grk(τ-s)dsNet input rate at timeτSμ1(t-τ)dτ A1c
ddtz(t)=μy(t) A1d

First, to derive the linear ODEs (A1a) and (A1d), note that the number of particles that transition from state W to state X in a short time interval (t,t+Δt) is binomially distributed: if we think of a transition from W to X as a “success” then the number of “trials” n=w~(t) and the probability of success p is given by the exponential CDF value p=1-exp(-aΔt). For sufficiently small Δt this implies p=aΔt+O(Δt2). Let w(t+s)=E(w~(t+s)|w~(t)) for s0. Since the expected value of a binomial random variable is np it follows that

w(t+Δt)-w(t)E(w~(t+Δt)-w~(t)|w~(t))=-aw(t)Δt+O(Δt2). A2

Dividing both sides of (A2) by Δt and then taking the limit as Δt0 yields

ddtw(t)=-aw(t). A3

Similarly, define z(t) in terms of z~(t) then it follows that

ddtz(t)=μy(t). A4

Next, we derive the integral equations (A1b) and (A1c) by similarly deriving a discrete time mean field model and then taking its limit as bin width Δt0 (i.e., as the number of bins M).

Partition the interval [0, t] into M1 equally wide intervals of width Δtt/M (see Fig. 10). Let Ii=((i-1)Δt,iΔt] denote the ith such time interval (i=1,,M) and let ti=(i-1)Δt denote the start time of the ith interval. We may now account for the number transitioning into and out of state X during time interval Ii, and sum these values to compute x(t).

Fig. 10.

Fig. 10

Illustration of discretizing the time interval [0, t] into M bins (each Δt=t/M wide) in order to derive the integral equations (A1b) and (A1c)

The number in state X at time t (x~(t)) is the number that entered state X between time 0 and t, less the number that transitioned out of X before time t (for now, assume x(0)=0). A particle that enters state X at time s(0,t) will still be in state X according to a Bernoulli random variable with p=Srk(t-s) (the expected proportion under the given gamma distribution). Therefore, to compute x~(t) we can sum over our M intervals and add up the number that entered state X during interval Ii and, from each of those M cohorts, count how many remain in X at time t. Specifically, the number entering X during the ith interval [t,t+Δt] is given by Niw~(ti+Δt)-w~(ti) (see Fig. 10), and thus the number remaining in X at time t is the sum of the number remaining at time t from each such cohort (i.e., the sum over i=1 to M) where the number remaining in X at t from each cohort follows a compound binomial distribution given by the sum of Ni Bernoulli random variables BX each with probability p=Srk(t-ti)+O(Δt). This defines our stochastic state transition model, which yields a mean field model as follows.

The expected amount entering X during [ti,ti+Δt] is E(Ni)=E(w~(ti)-w~(ti+Δt))=aw(ti)Δt, and the expected proportion of the ith cohort remaining at time t is E(BX)=Srk(t-ti)+O(Δt). Thus, the expected number from the ith cohort remaining in X at time t is E(Ni)E(BX)=aw(ti)Srk(t-ti)Δt+O(Δt2). Summing these expected values over all intervals, and letting Δt0, yields

x(t)=E(x~(t))=limMi=1Maw(ti)Srk(t-ti)Δt+O(Δt2)=0taw(s)Srk(t-s)ds. A5

To calculate y(t), let Kj be the number entering state Y during interval Ij that are still in state Y at time t. As above, Kj can be calculated by summing (over I1 to Ij-1) the number in each cohort that entered state X during Ii then transitioned to state Y during time interval Ij and are still in state Y at t. Therefore Kj can be written as the sum of j-1 compound distributions given by counting how many of the Ni particles that entered state X during Ii then transitioned to state Y during interval Ij and then persisted until time t without transitioning to Z. To count these, notice that each such particle entering state X during Ii, state Y during Ij and persisting in Y at time t follows a Bernoulli random variable Bij with probability

pij=(Srk(tj-ti)-Srk(tj+1-ti)+O(Δt))P(XY inIj|WX inIi)(Sμ1(t-tj)+O(Δt))P(still in Y att). A6

Therefore, the number of particles that entered Y at Ij and remain in state Y at time t, Kj, can be written as a compound random variable Kj=i=1j-1k=1NiB(pij). The expected Kj is thus

E(Kj)=i=1j-1E(Ni)E(Bij)=i=1j-1(aw(ti)Δt)[(Srk(tj-ti)-Srk(tj+1-ti))Sμ1(t-tj)+O(Δt)]=i=1j-1aw(ti)(Srk(tj-ti)-Srk(tj-ti+Δt))ΔtSμ1(t-tj)Δt2+O(Δt2)=i=1j-1aw(ti)(Grk(tj-ti+Δt)-Grk(tj-ti))ΔtSμ1(t-tj)Δt2+O(Δt2). A7

Summing over all intervals and letting Δt0 (M) gives

y(t)=limME(y~(t))=limME(j=1MKj)=limMj=1ME(Kj)=limMj=1Mi=1j-1aw(ti)(Grk(tj-ti+Δt)-Grk(tj-ti))ΔtΔt·Sμ1(t-tj)Δt+O(Δt2)=0t0τaw(s)grk(τ-s)dsSμ1(t-τ)dτ. A8

Applying Theorem 2, the mean field ODEs equivalent to Eq. (A1), where x(t)=j=1kxj(t), are given by

ddtw(t)=-aw(t) A9a
ddtx1(t)=aw(t)-rx1(t) A9b
ddtxj(t)=rxj-1(t)-rxj(t),forj=2,,k A9c
ddty(t)=rxk(t)-μy(t) A9d
ddtz(t)=μy(t). A9e

Appendix B Erlang mixture approximation of Gamma(α,β)

Here we give a simple example of analytically approximating a gamma distribution with a mixture of two Erlang distributions by moment matching.

Suppose random variable T follows a gamma(α,β) distribution, which has mean μ=β/α and variance σ2=μ/α, and shape β is not an integer. One can approximate this gamma distribution with a mixture of two Erlang distributions that yields the same mean and variance.

These Erlang distributions are TErlang(r,k) and TErlang(r,k) where the shape parameters are obtained by rounding β down and up, respectively, to the nearest integer (kβ and kβ) and the rate parameters are given by r=αββ (r=kμ) and r=αββ (rkμ). This ensures that T and T have mean μ.

To calculate their variance, let pβ-β and qβ-β (note p+q=1). By rounding shape β down(up) the resulting Erlang distribution has higher(lower) variance, i.e.,

σ2μr=σ2(1+qk)andσ2μr=σ2(1-pk). B1

To calculate the mixing proportion, let the mixture distribution Tρ=BρT+(1-Bρ)T, where Bρ is a Bernoulli random variable with P(Bp=1)=ρ and

ρ=p/kp/k+q/k. B2

This Erlang mixture has the desired mean E(Tρ)=ρμ+(1-ρ)μ=μ and variance σ2, since

Var(Tρ)=E((BρT+(1-Bρ)T)2))-μ2=E((BρT)2)+E(((1-Bρ)T)2)-μ2=E(Bρ2)E(T2)+E((1-Bρ)2)E(T2)-μ2=E(Bρ)E(T2)+E(1-Bρ)E(T2)-μ2=ρE(T2)+(1-ρ)E(T2)-μ2=ρσ2+(1-ρ)σ2=ρσ2(1+qk)+(1-ρ)σ2(1-pk)=σ2+σ2(ρqk-(1-ρ)pk)=σ2+σ2(p/k·q/kp/k+q/k-q/k·p/kp/k+q/k)=σ2. B3

Numerical comparisons suggest this mixture is a very good approximation of the target gamma(α,β) distribution for shape β values larger than roughly 3 to 5, depending (to a lesser extent) on α.

Alternatively, to approximate a gamma(α,β) distribution with a mixture of Erlang distributions as described above, one could also select the mixing probabilities by, for example, using an alternative metric such as a distance in probability space (e.g., see Rachev 1991), e.g., the L-norm on their CDFs, or information-theoretic quantities such as KL or Jensen-Shannon divergence.

Funding

This work was supported by funds provided to PJH by the University of Nevada, Reno (UNR) Office of Research and Innovation, and by a grant from the Sloan Scholars Mentoring Network of the Social Science Research Council with funds provided by the Alfred P. Sloan Foundation.

Compliance with ethical standards

Conflict of interest

The authors declare that they have no conflict of interest.

Footnotes

1

Erlang distributions are Gamma distributions with integer-valued shape parameters.

2

They can also be parameterized in terms of their mean and variance (see Appendix B), or with a shape and scale parameters, where the scale parameter is the inverse of the rate.

3

A useful interpretation of survival functions, which is used below, is that they give the expected proportion remaining after a give amount time.

4

...despite the implied exclusivity of the adjective nonhomogeneous.

5

That is, the probability of a given individual exiting state X during a brief time period [t,t+Δt] is approximately r(t)Δt.

6

Here the variance is assumed to have been chosen so that the resulting shape parameter is integer valued. See Appendix B for related details.

7

It is generally true that the survival function for a minimum of multiple independent random variables is the product of their survival functions.

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

References

  1. Allen LJS. An introduction to Stochastic processes with applications to biology. 2. Boca Raton: Chapman and Hall/CRC; 2010. [Google Scholar]
  2. Allen LJ. A primer on stochastic epidemic models: formulation, numerical simulation, and analysis. Infect Dis Model. 2017;2(2):128–142. doi: 10.1016/j.idm.2017.03.001. [DOI] [PMC free article] [PubMed] [Google Scholar]
  3. Altiok T. On the phase-type approximations of general distributions. IIE Trans. 1985;17(2):110–116. doi: 10.1080/07408178508975280. [DOI] [Google Scholar]
  4. Anderson RM, May RM. Infectious diseases of humans: dynamics and control. Oxford: Oxford University Press; 1992. [Google Scholar]
  5. Anderson D, Watson R. On the spread of a disease with gamma distributed latent and infectious periods. Biometrika. 1980;67(1):191–198. doi: 10.1093/biomet/67.1.191. [DOI] [Google Scholar]
  6. Armbruster B, Beck E. Elementary proof of convergence to the mean-field model for the SIR process. J Math Biol. 2017;75(2):327–339. doi: 10.1007/s00285-016-1086-1. [DOI] [PubMed] [Google Scholar]
  7. Asmussen S, Nerman O, Olsson M. Fitting phase-type distributions via the EM algorithm. Scand J Stat. 1996;23(4):419–441. [Google Scholar]
  8. Banks HT, Catenacci J, Hu S. A comparison of stochastic systems with different types of delays. Stoch Anal Appl. 2013;31(6):913–955. doi: 10.1080/07362994.2013.806217. [DOI] [Google Scholar]
  9. Bindel D, Friedman M, Govaerts W, Hughes J, Kuznetsov YA. Numerical computation of bifurcations in large equilibrium systems in matlab. J Comput Appl Math. 2014;261:232–248. doi: 10.1016/j.cam.2013.10.034. [DOI] [Google Scholar]
  10. Blythe S, Nisbet R, Gurney W. The dynamics of population models with distributed maturation periods. Theor Popul Biol. 1984;25(3):289–311. doi: 10.1016/0040-5809(84)90011-X. [DOI] [PubMed] [Google Scholar]
  11. Boese F. The stability chart for the linearized Cushing equation with a discrete delay and with gamma-distributed delays. J Math Anal Appl. 1989;140(2):510–536. doi: 10.1016/0022-247X(89)90081-4. [DOI] [Google Scholar]
  12. Burton TA. Volterra integral and differential equations, mathematics in science and engineering. 2. Amsterdam: Elsevier; 2005. [Google Scholar]
  13. Câmara De Souza D, Craig M, Cassidy T, Li J, Nekka F, Bélair J, Humphries AR. Transit and lifespan in neutrophil production: implications for drug intervention. J Pharmacokinet Pharmacodyn. 2018;45(1):59–77. doi: 10.1007/s10928-017-9560-y. [DOI] [PubMed] [Google Scholar]
  14. Campbell SA, Jessop R. Approximating the stability region for a differential equation with a distributed delay. Math Model Nat Phenom. 2009;4(2):1–27. doi: 10.1051/mmnp/20094201. [DOI] [Google Scholar]
  15. Champredon D, Dushoff J, Earn D (2018) Equivalence of the Erlang SEIR epidemic model and the renewal equation. bioRxiv 10.1101/319574
  16. Ciaravino G, García-Saenz A, Cabras S, Allepuz A, Casal J, García-Bocanegra I, Koeijer AD, Gubbins S, Sáez J, Cano-Terriza D, Napp S. Assessing the variability in transmission of bovine tuberculosis within Spanish cattle herds. Epidemics. 2018;23:110–120. doi: 10.1016/j.epidem.2018.01.003. [DOI] [PubMed] [Google Scholar]
  17. Clapp G, Levy D. A review of mathematical models for leukemia and lymphoma. Drug Discov Today Dis Models. 2015;16:1–6. doi: 10.1016/j.ddmod.2014.10.002. [DOI] [PMC free article] [PubMed] [Google Scholar]
  18. Cushing JM. The dynamics of hierarchical age-structured populations. J Math Biol. 1994;32(7):705–729. doi: 10.1007/BF00163023. [DOI] [Google Scholar]
  19. Dhooge A, Govaerts W, Kuznetsov YA. MATCONT: a MATLAB package for numerical bifurcation analysis of ODEs. ACM Trans Math Softw. 2003;29(2):141–164. doi: 10.1145/779359.779362. [DOI] [Google Scholar]
  20. Diekmann O, Gyllenberg M, Metz JAJ. Finite dimensional state representation of linear and nonlinear delay systems. J Dyn Differ Equ. 2017 doi: 10.1007/s10884-017-9611-5. [DOI] [Google Scholar]
  21. Du P, Li MY. Impact of network connectivity on the synchronization and global dynamics of coupled systems of differential equations. Phys D Nonlinear Phenom. 2014;286–287:32–42. doi: 10.1016/j.physd.2014.07.008. [DOI] [Google Scholar]
  22. Ermentrout B (1987) Simulating, analyzing, and animating dynamical systems: a guide to XPPAUT for researchers and students. Software, environments and tools (Book 14), Society for Industrial and Applied Mathematics
  23. Ermentrout B (2019) XPP/XPPAUT homepage. http://www.math.pitt.edu/~bard/xpp/xpp.html. Accessed Apr 2019
  24. Fargue D. Réductibilité des systèmes héréditaires à des systèmes dynamiques (régis par des équations différentielles ou aux dérivées partielles) C R Acad Sci Paris Sér A-B. 1973;277:B471–B473. [Google Scholar]
  25. Feng Z, Thieme H. Endemic models with arbitrarily distributed periods of infection I: fundamental properties of the model. SIAM J Appl Math. 2000;61(3):803–833. doi: 10.1137/S0036139998347834. [DOI] [Google Scholar]
  26. Feng Z, Xu D, Zhao H. Epidemiological models with non-exponentially distributed disease stages and applications to disease control. Bullet Math Biol. 2007;69(5):1511–1536. doi: 10.1007/s11538-006-9174-9. [DOI] [PubMed] [Google Scholar]
  27. Feng Z, Zheng Y, Hernandez-Ceron N, Zhao H, Glasser JW, Hill AN. Mathematical models of Ebola-Consequences of underlying assumptions. Math Biosci. 2016;277:89–107. doi: 10.1016/j.mbs.2016.04.002. [DOI] [PubMed] [Google Scholar]
  28. Fenton A, Lello J, Bonsall M. Pathogen responses to host immunity: the impact of time delays and memory on the evolution of virulence. Proc Biol Sci. 2006;273(1597):2083–2090. doi: 10.1098/rspb.2006.3552. [DOI] [PMC free article] [PubMed] [Google Scholar]
  29. Goltser Y, Domoshnitsky A. About reducing integro-differential equations with infinite limits of integration to systems of ordinary differential equations. Adv Differ Equ 3013. 2013;1:187. doi: 10.1186/1687-1847-2013-187. [DOI] [Google Scholar]
  30. Guan Zhi-Hong, Ling Guang. Emergence, Complexity and Computation. Berlin, Heidelberg: Springer Berlin Heidelberg; 2017. Dynamic Analysis of Genetic Regulatory Networks with Delays; pp. 285–309. [Google Scholar]
  31. Guo H, Li MY, Shuai Z. A graph-theoretic approach to the method of global lyapunov functions. Proc Am Math Soc. 2008;136(08):2793–2802. doi: 10.1090/S0002-9939-08-09341-6. [DOI] [Google Scholar]
  32. Gyllenberg M. Mathematical aspects of physiologically structured populations: the contributions of J. A. J. Metz. J Biol Dyn. 2007;1(1):3–44. doi: 10.1080/17513750601032737. [DOI] [PubMed] [Google Scholar]
  33. Hethcote HW, Tudor DW. Integral equation models for endemic infectious diseases. J Math Biol. 1980;9(1):37–47. doi: 10.1007/BF00276034. [DOI] [PubMed] [Google Scholar]
  34. Horváth G, Telek M (2017) BuTools 2: a rich toolbox for Markovian performance evaluation. In: ValueTools 2016—10th EAI international conference on performance evaluation methodologies and tools, association for computing machinery, pp 137–142. 10.4108/eai.25-10-2016.2266400
  35. Horváth Gábor, Reinecke Philipp, Telek Miklós, Wolter Katinka. Analytical and Stochastic Modeling Techniques and Applications. Berlin, Heidelberg: Springer Berlin Heidelberg; 2012. Efficient Generation of PH-Distributed Random Variates; pp. 271–285. [Google Scholar]
  36. Horvath Andras, Scarpa Marco, Telek Miklos. Springer Series in Reliability Engineering. Cham: Springer International Publishing; 2016. Phase Type and Matrix Exponential Distributions in Stochastic Modeling; pp. 3–25. [Google Scholar]
  37. Jacquez JA, Simon CP. Qualitative theory of compartmental systems with lags. Math Biosci. 2002;180(1):329–362. doi: 10.1016/S0025-5564(02)00131-1. [DOI] [PubMed] [Google Scholar]
  38. Kermack WO, McKendrick AG. A contribution to the mathematical theory of epidemics. Proc R Soc Lond Ser A Contain Pap Math Phys Character. 1927;115(772):700–721. doi: 10.1098/rspa.1927.0118. [DOI] [Google Scholar]
  39. Komárková Z (2012) Phase-type approximation techniques. Bachelor’s Thesis, Masaryk University, https://is.muni.cz/th/ysfsq
  40. Krylova O, Earn DJD. Effects of the infectious period distribution on predicted transitions in childhood disease dynamics. J R Soc Interface. 2013 doi: 10.1098/rsif.2013.0098. [DOI] [PMC free article] [PubMed] [Google Scholar]
  41. Krzyzanski W, Hu S, Dunlavey M. Evaluation of performance of distributed delay model for chemotherapy-induced myelosuppression. J Pharmacokinet Pharmacodyn. 2018;45(2):329–337. doi: 10.1007/s10928-018-9575-z. [DOI] [PubMed] [Google Scholar]
  42. Kurtz TG. Solutions of ordinary differential equations as limits of pure jump Markov processes. J Appl Probab. 1970;7(1):49–58. doi: 10.2307/3212147. [DOI] [Google Scholar]
  43. Kurtz TG. Limit theorems for sequences of jump Markov processes approximating ordinary differential processes. J Appl Probab. 1971;8(2):344–356. doi: 10.2307/3211904. [DOI] [Google Scholar]
  44. Li MY, Shuai Z. Global-stability problem for coupled systems of differential equations on networks. J Differ Equ. 2010;248(1):1–20. doi: 10.1016/j.jde.2009.09.003. [DOI] [Google Scholar]
  45. Lin CJ, Wang L, Wolkowicz GSK. An alternative formulation for a distributed delayed logistic equation. Bull Math Biol. 2018;80(7):1713–1735. doi: 10.1007/s11538-018-0432-4. [DOI] [PubMed] [Google Scholar]
  46. Lloyd AL. Destabilization of epidemic models with the inclusion of realistic distributions of infectious periods. Proc R Soc Lond B Biol Sci. 2001;268(1470):985–993. doi: 10.1098/rspb.2001.1599. [DOI] [PMC free article] [PubMed] [Google Scholar]
  47. Lloyd AL. Realistic distributions of infectious periods in epidemic models: changing patterns of persistence and dynamics. Theor Popul Biol. 2001;60(1):59–71. doi: 10.1006/tpbi.2001.1525. [DOI] [PubMed] [Google Scholar]
  48. Lloyd A.L. Mathematical and Statistical Estimation Approaches in Epidemiology. Dordrecht: Springer Netherlands; 2009. Sensitivity of Model-Based Epidemiological Parameter Estimation to Model Assumptions; pp. 123–141. [Google Scholar]
  49. Ma J, Earn DJD. Generality of the final size formula for an epidemic of a newly invading infectious disease. Bull Math Biol. 2006;68(3):679–702. doi: 10.1007/s11538-005-9047-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
  50. MacDonald Norman. Time Lags in Biological Models. Berlin, Heidelberg: Springer Berlin Heidelberg; 1978. [Google Scholar]
  51. MacDonald Norman. Time Lags in Biological Models. Berlin, Heidelberg: Springer Berlin Heidelberg; 1978. Stability Analysis; pp. 13–38. [Google Scholar]
  52. MacDonald N. Biological delay systems: linear stability theory, cambridge studies in mathematical biology. Cambridge: Cambridge University Press; 1989. [Google Scholar]
  53. Makroglou A, Li J, Kuang Y. Mathematical models and software tools for the glucose-insulin regulatory system and diabetes: an overview. Appl Numer Math. 2006;56(3):559–573. doi: 10.1016/j.apnum.2005.04.023. [DOI] [Google Scholar]
  54. Metz J. A. J., Diekmann O., editors. The Dynamics of Physiologically Structured Populations. Berlin, Heidelberg: Springer Berlin Heidelberg; 1986. [Google Scholar]
  55. Metz J, Diekmann O (1991) Exact finite dimensional representations of models for physiologically structured populations. I: The abstract formulation of linear chain trickery. In: Goldstein JA, Kappel F, Schappacher W (eds) Proceedings of differential equations with applications in biology, physics, and engineering 1989, vol 133, pp 269–289
  56. Nisbet R. M., Gurney W. S. C., Metz J. A. J. Applied Mathematical Ecology. Berlin, Heidelberg: Springer Berlin Heidelberg; 1989. Stage Structure Models Applied in Evolutionary Ecology; pp. 428–449. [Google Scholar]
  57. Okamura Hiroyuki, Dohi Tadashi. Quantitative Evaluation of Systems. Cham: Springer International Publishing; 2015. mapfit: An R-Based Tool for PH/MAP Parameter Estimation; pp. 105–112. [Google Scholar]
  58. Osogami T, Harchol-Balter M. Closed form solutions for mapping general distributions to quasi-minimal PH distributions. Perform Eval. 2006;63(6):524–552. doi: 10.1016/j.peva.2005.06.002. [DOI] [Google Scholar]
  59. Özbay H, Bonnet C, Clairambault J (2008) Stability analysis of systems with distributed delays and application to hematopoietic cell maturation dynamics. In: 2008 47th IEEE conference on decision and control, pp 2050–2055. 10.1109/CDC.2008.4738654
  60. Pérez JF, Riaño G (2006) jPhase: an object-oriented tool for modeling phase-type distributions. In: Proceeding from the 2006 Workshop on tools for solving structured Markov chains, ACM, New York, NY, USA, SMCtools ’06. 10.1145/1190366.1190370
  61. Piotrowska M, Bodnar M. Influence of distributed delays on the dynamics of a generalized immune system cancerous cells interactions model. Commun Nonlinear Sci Numer Simul. 2018;54:389–415. doi: 10.1016/j.cnsns.2017.06.003. [DOI] [Google Scholar]
  62. Ponosov A, Shindiapin A, Miguel JJ. The W-transform links delay and ordinary differential equations. Funct Differ Equ. 2002;9(3–4):437–469. [Google Scholar]
  63. Rachev ST (1991) Probability metrics and the stability of stochastic models. Wiley Series in Probability and Mathematical Statistics, Wiley, New York
  64. Reinecke Philipp, Bodrog Levente, Danilkina Alexandra. Resilience Assessment and Evaluation of Computing Systems. Berlin, Heidelberg: Springer Berlin Heidelberg; 2012. Phase-Type Distributions; pp. 85–113. [Google Scholar]
  65. Reinecke P, Krauß T, Wolter K. Cluster-based fitting of phase-type distributions to empirical data. Comput Math Appl. 2012;64(12):3840–3851. doi: 10.1016/j.camwa.2012.03.016. [DOI] [Google Scholar]
  66. Robertson SL, Henson SM, Robertson T, Cushing JM. A matter of maturity: to delay or not to delay? Continuous-time compartmental models of structured populations in the literature 2000–2016. Nat Resour Model. 2018;31(1):e12160. doi: 10.1111/nrm.12160. [DOI] [Google Scholar]
  67. Roussel MR. The use of delay differential equations in chemical kinetics. J Phys Chem. 1996;100(20):8323–8330. doi: 10.1021/jp9600672. [DOI] [Google Scholar]
  68. Smith H. An introduction to delay differential equations with applications to the life sciences. Berlin: Springer; 2010. [Google Scholar]
  69. Smolen P, Baxter DA, Byrne JH. Modeling transcriptional control in gene networks—methods, recent results, and future directions. Bull Math Biol. 2000;62(2):247–292. doi: 10.1006/bulm.1999.0155. [DOI] [PubMed] [Google Scholar]
  70. Strogatz SH (2014) Nonlinear dynamics and chaos: with applications to physics, biology, chemistry, and engineering, 2nd edn. Studies in Nonlinearity, Westview Press
  71. Takashima Y, Ohtsuka T, González A, Miyachi H, Kageyama R. Intronic delay is essential for oscillatory expression in the segmentation clock. Proc Natl Acad Sci. 2011;108(8):3300–3305. doi: 10.1073/pnas.1014418108. [DOI] [PMC free article] [PubMed] [Google Scholar]
  72. Thummler A, Buchholz P, Telek M. A novel approach for phase-type fitting with the EM algorithm. IEEE Trans Dependable Secure Comput. 2006;3(3):245–258. doi: 10.1109/TDSC.2006.27. [DOI] [Google Scholar]
  73. van den Driessche P, Watmough J. Reproduction numbers and sub-threshold endemic equilibria for compartmental models of disease transmission. Math Biosci. 2002;180(1–2):29–48. doi: 10.1016/S0025-5564(02)00108-6. [DOI] [PubMed] [Google Scholar]
  74. Vogel T (1961) Systèmes déferlants, systèmes héréditaires, systèmes dynamiques. In: Proceedings of the international symposium nonlinear vibrations, IUTAM, Kiev, pp 123–130
  75. Vogel T (1965) Théorie des Systèmes Évolutifs. No. 22 in Traité de physique théorique et de physique mathématique, Gauthier-Villars, Paris
  76. Wang N, Han M. Slow-fast dynamics of Hopfield spruce-budworm model with memory effects. Adv Differ Equ. 2016;1:73. doi: 10.1186/s13662-016-0804-8. [DOI] [Google Scholar]
  77. Wearing HJ, Rohani P, Keeling MJ. Appropriate models for the management of infectious diseases. PLOS Med. 2005 doi: 10.1371/journal.pmed.0020174. [DOI] [PMC free article] [PubMed] [Google Scholar]
  78. Wolkowicz G, Xia H, Ruan S. Competition in the chemostat: a distributed delay model and its global asymptotic behavior. SIAM J Appl Math. 1997;57(5):1281–1310. doi: 10.1137/S0036139995289842. [DOI] [Google Scholar]
  79. Yates CA, Ford MJ, Mort RL. A multi-stage representation of cell proliferation as a Markov process. Bull Math Biol. 2017;79(12):2905–2928. doi: 10.1007/s11538-017-0356-4. [DOI] [PMC free article] [PubMed] [Google Scholar]

Articles from Journal of Mathematical Biology are provided here courtesy of Springer

RESOURCES