Skip to main content
Springer logoLink to Springer
. 2017 Dec 15;77(2):261–312. doi: 10.1007/s00285-017-1196-4

Time-dependent propagators for stochastic models of gene expression: an analytical method

Frits Veerman 1,, Carsten Marr 2, Nikola Popović 1
PMCID: PMC6061071  PMID: 29247320

Abstract

The inherent stochasticity of gene expression in the context of regulatory networks profoundly influences the dynamics of the involved species. Mathematically speaking, the propagators which describe the evolution of such networks in time are typically defined as solutions of the corresponding chemical master equation (CME). However, it is not possible in general to obtain exact solutions to the CME in closed form, which is due largely to its high dimensionality. In the present article, we propose an analytical method for the efficient approximation of these propagators. We illustrate our method on the basis of two categories of stochastic models for gene expression that have been discussed in the literature. The requisite procedure consists of three steps: a probability-generating function is introduced which transforms the CME into (a system of) partial differential equations (PDEs); application of the method of characteristics then yields (a system of) ordinary differential equations (ODEs) which can be solved using dynamical systems techniques, giving closed-form expressions for the generating function; finally, propagator probabilities can be reconstructed numerically from these expressions via the Cauchy integral formula. The resulting ‘library’ of propagators lends itself naturally to implementation in a Bayesian parameter inference scheme, and can be generalised systematically to related categories of stochastic models beyond the ones considered here.

Keywords: Stochastic gene expression, Probability generating function, Propagator, Dynamical systems, Perturbation techniques, Asymptotic analysis

Introduction

Motivation

Understanding the process of gene expression in the context of gene regulatory networks is indispensable for gaining insight into the fundamentals of numerous biological processes. However, gene expression can be highly stochastic in nature, both in prokaryotic and in eukaryotic organisms; see e.g. the work by Elowitz et al. (2002), Raj and Oudenaarden (2008), Shahrezaei and Swain (2008b), and the references therein. This inherent stochasticity has a profound influence on the dynamics of the involved species, in particular when their abundance is low. Therefore, gene expression is often appropriately described by stochastic models (Bressloff 2014; Karlebach and Shamir 2008; Thattai and Oudenaarden 2001; Wilkinson 2009). A schematic of the canonical model for gene expression is depicted in Fig. 1. Here, the processes of transcription, translation, and degradation are approximated by single rates.

Fig. 1.

Fig. 1

The canonical model of gene expression. Transcription of mRNA occurs with rate ν0; mRNA is translated to protein with rate ν1. Both mRNA and protein decay, with rates d0 and d1, respectively.

Figure courtesy of Shahrezaei and Swain (2008a) (Copyright (2008) National Academy of Sciences, U.S.A.)

To test the validity of such stochastic models, a comparison with experimental data needs to be performed. The development of experimental techniques, such as time-lapse fluorescence microscopy (Coutu and Schroeder 2013; Elowitz et al. 2002; Larson et al. 2011; Muzzey and Oudenaarden 2009; Raj et al. 2006; Young et al. 2011), allows for real-time tracking of gene expression dynamics in single cells, providing mRNA or protein abundance time series, such as those depicted in Fig. 2. To select between competing hypotheses on the underlying regulatory networks given measurement data, as well as to infer the values of the corresponding model parameters, we can apply Bayesian inference theory to calculate the likelihood of a given model, which is constructed as follows.

Fig. 2.

Fig. 2

Sketch of a potential time series of protein abundance n, sampled at times ti, with regular, fixed, sampling interval Δt

The abundance of a protein, denoted by n, is sampled at times ti, see Fig. 2, yielding a list of measurement pairs (ti,ni) from which transitions (Δt,nini+1) between states can be extracted; here, Δt=ti+1-ti is the regular, fixed, sampling interval. Next, the underlying stochastic model with parameter set Θ is used to calculate the probabilities of these transitions, which are denoted by Pni+1|ni(Δt). These so-called propagators give the probability of ni+1 protein being present after time Δt, given an initial protein abundance of ni. The log-likelihood L(Θ|D) of the parameter set Θ, given the observed data D, is now defined in terms of the propagators as

L(Θ|D)=ilogPni+1|ni(Δt). 1.1

To infer the values of parameters in the model, propagators are calculated for a wide range of parameter combinations, resulting in a ‘log-likelihood landscape’; the maximal value of the log-likelihood as a function of the model parameters yields the most likely parameter values, given the experimental data. An example realisation of the above procedure can e.g. be found in the work by Feigelman et al. (2015).

To calculate accurately the log-likelihood in (1.1), it is imperative that the values of the propagators can be extracted from the underlying stochastic model for any desired combination of parameters in Θ. In particular, we need to be able to calculate the propagator Pn|n0(t) as a function of time t for any initial protein number n0. Unfortunately, the highly complex nature of the stochastic models involved makes it very difficult to obtain explicit expressions for these probabilities. Some analytical progress can be made when a steady-state approximation is performed, i.e. when it is assumed that the system is allowed to evolve for a sufficiently long time, such that it converges to a time-independent state. However, the sampling interval Δt used for obtaining experimental data, as seen in Fig. 2, is often short with respect to the protein life time. As that life time represents a natural time scale for the system dynamics, it follows that the evolution of the probabilities Pn|n0(t) should be studied over short times, in contradiction with the steady-state or long-evolution-time approximations which have previously been employed to derive analytical results (Bokes et al. 2012a; Hornos et al. 2005; Iyer-Biswas and Jayaprakash 2014; Shahrezaei and Swain 2008a).

The complex nature of stochastic models for gene expression has led to the widespread use of stochastic simulation techniques, such as Gillespie’s algorithm (Gillespie 1977), with the aim of predicting values for the associated propagators from these models; see Feigelman et al. (2016) for recent work combining stochastic simulation with a particle filtering approach. However, these approaches can still be very time-consuming, due to the (relatively) high dimensionality of the model parameter space, combined with the fact that, for each combination of parameter values, the stochastic model has to be simulated sufficiently many times to yield a probability distribution that can be used to infer the corresponding propagator. For that reason, it is desirable to be able to obtain explicit expressions for the propagator Pn|n0(t) directly in terms of the model parameters, if necessary in an appropriate approximation.

Analytical method

In the present article, we develop an analytical method for the efficient evaluation of time-dependent propagators in stochastic gene expression models, for arbitrary values of the model parameters. The results of our analysis can be implemented in a straightforward fashion in a Bayesian parameter inference framework, as outlined above.

To demonstrate our approach, we analyse two different stochastic models for gene expression. The first model, henceforth referred to as ‘model A’, is a model that incorporates autoregulation, where transcription and translation are approximated by single rates and protein can either stimulate or inhibit its own production by influencing the activity of DNA; see Fig. 3. That model was first studied by Iyer-Biswas and Jayaprakash (2014) via a steady-state approximation. The second model, henceforth referred to as ‘model B’, models both mRNA and protein explicitly and again incorporates DNA switching between an active and an inactive state; see Fig. 4. That model was first studied by Shahrezaei and Swain (2008a) in a long-evolution-time approximation.

Fig. 3.

Fig. 3

Schematic of model A, a gene expression model with autoregulation.

Base figure courtesy of Shahrezaei and Swain (2008a). (Copyright (2008) National Academy of Sciences, U.S.A.)

Fig. 4.

Fig. 4

Schematic of model B, a gene expression model that explicitly incorporates transcription (Shahrezaei and Swain 2008a).

Figure courtesy of Shahrezaei and Swain (2008a). (Copyright (2008) National Academy of Sciences, U.S.A.)

Both model A and model B are formulated in terms of the chemical master equation (CME), which is the accepted mathematical representation of stochastic gene expression in the context of the model categories considered here; cf. Iyer-Biswas and Jayaprakash (2014) and Shahrezaei and Swain (2008a), respectively. Mathematically speaking, the CME is an infinite-dimensional system of linear ordinary differential equations (ODEs) that describes the evolution in time of the probabilities of observing a specific state in the system, given some initial state. Numerous approaches have been suggested for the (approximate) solution of the CME; see e.g. Popović et al. (2016) and the references therein for details. Our method relies on a combination of various techniques from the theory of differential equations and dynamical systems; specifically, we perform three consecutive steps, as follows.

  1. CME system PDE system: We introduce a probability-generating function to convert the CME into a (system of) partial differential equations (PDEs).

  2. PDE system ODE system: Applying the method of characteristics—combined, if necessary, with perturbation techniques—we transform the system of PDEs obtained in step 1 into a dynamical system, that is, a system of ODEs.

  3. ODE system Explicit solution: Making use of either special functions (model A) or multiple-time-scale analysis (model B), we obtain explicit solutions to the dynamical system found in step 2.

We emphasise that the ‘characteristic system’ of ODEs which is obtained in step 2 is low-dimensional, in contrast to the underlying CME system, as well as that it exhibits additional structure, allowing for the derivation of a closed-form analytical approximation for the associated generating function.

To convert the results of the above procedure into solutions to the original stochastic model, the three steps involved in our analysis have to be reverted. To that end, we require the following three ingredients:

  1. Initial conditions are originally stated in terms of the CME, and first have to be reformulated in terms of the corresponding system of PDEs to ensure well-posedness; then, initial conditions can be extracted for the dynamical system that was obtained via the method of characteristics, reverting step 3.

  2. To transform solutions to the characteristic system into solutions of the underlying PDE system, the associated ‘characteristic transformation’ has to be inverted, reverting step 2.

  3. Lastly, solutions of the CME have to be extracted from solutions to the resulting PDE system, reverting step 1. Although the correspondence between the two sets of solutions is exact, theoretically speaking, the complexity of the expressions involved precludes the efficient analytical reconstruction of propagators from their generating functions. Therefore, we propose a novel hybrid analytical-numerical approach which relies on the Cauchy integral formula.

The various steps in our analytical method, as indicated above, are represented in Fig. 5. It is important to mention that the implementation of Bayesian parameter inference, as outlined in Sect. 1.1, is not a topic for the present article; rather, the aim here is to describe our method, and to present analytical results which can readily be implemented in the context of parameter inference. The article hence realises the first stage of our research programme; the natural next stage, which is precisely that implementation, will be the subject of a follow-up article by the same authors.

Fig. 5.

Fig. 5

Schematic overview of the analytical method

Outline

The present article is organised as follows. In Sect. 2, we apply the analytical method outlined in Sect. 1.2 to model A, the gene expression model with autoregulation. Here, we use a perturbative approach to incorporate the autoregulatory aspects of the model; the resulting dynamical system can be solved in terms of confluent hypergeometric functions, see §13 in NIST Digital Library of Mathematical Functions . In Sect. 3, the same method is applied to model B, the model that explicitly incorporates transcription. We also indicate how autoregulation can be added to that model, and how the resulting extended model can be analysed on the basis of our treatment of model A. The analysis carried out in Sects. 2 and 3 yields a ‘library’ of explicit asymptotic expressions for the probability-generating functions associated to the underlying stochastic models. To obtain quantifiable expressions for their propagators, we introduce a novel hybrid analytical–numerical approach in Sect. 4, which can be readily implemented in the Bayesian parameter inference framework that provided the motivation for our analysis; see Sect. 1.1. We conclude with a discussion of our results, and an outlook to future work, in Sect. 5.

Model A: gene expression with autoregulation

We first demonstrate our analytical method in the context of an autoregulatory stochastic gene expression model, as presented by Iyer-Biswas and Jayaprakash (2014); see also Fig. 3. In the original article (Iyer-Biswas and Jayaprakash 2014), a Poisson representation was used to obtain analytical descriptions for time-independent solutions to the model. For a visual guide to the upcoming analysis, the reader is referred to Fig. 5.

Stochastic model and CME

The basic stochastic model for gene expression is represented by the reaction scheme

graphic file with name 285_2017_1196_Equ2_HTML.gif 2.1

The gene can hence switch between the inactive state D and the active state D, with switching rates cf and cb, respectively. The active gene produces protein (P) with rate pb, while protein decays with rate pd.

The autoregulatory part of the model is implemented as either positive or negative feedback:

D+PaD+P(autoactivation), 2.2a
D+PrD+P(autorepression). 2.2b

In the case of autoactivation, viz. (2.2a), protein induces activation of the gene with activation rate a, thereby accelerating its own production; in the case of autorepression, viz. (2.2b), protein deactivates the active gene with repression rate r, impeding its own production.

The CME system that is associated to the reaction scheme in (2.1), with autoactivation as in (2.2a), is given by

dPn(0)dt=-κf+apdnPn(0)+κbPn(1)+(n+1)Pn+1(0)-nPn(0), 2.3a
dPn(1)dt=κf+apdnPn(0)-κbPn(1)+(n+1)Pn+1(1)-nPn(1)+λPn-1(1)-Pn(1). 2.3b

Here, Pn(j)(t) (j=0,1) represents the probability of n protein being present at time t while the gene is either inactive (0) or active (1). The time variable is nondimensionalised by the protein decay rate pd; other model parameters are scaled as

κf=cfpd,κb=cbpd,andλ=pbpd. 2.4

Analogously, the CME system for the case of autorepression, as defined in (2.2b), is given by

dPn(0)dt=-κfPn(0)+κb+rpdnPn(1)+(n+1)Pn+1(0)-nPn(0), 2.5a
dPn(1)dt=κfPn(0)-κb+rpdnPn(1)+(n+1)Pn+1(1)-nPn(1)+λPn-1(1)-Pn(1). 2.5b

Remark 2.1

A priori, it is possible to incorporate both autoactivation and autorepression in a single model, by merging systems (2.3) and (2.5). However, since autoactivation and autorepression precisely counteract each other, a partial cancellation would ensue, resulting in effective activation or repression. It can hence be argued that the simultaneous inclusion of both effects would introduce superfluous terms and parameters, which could be considered as poor modelling practice. Therefore, we choose to model the two autoregulation mechanisms separately.

Generating function PDE

Rather than investigating the dynamics of (2.3) and (2.5) numerically, using stochastic simulation, we aim to employ an analytical approach. To that end, we define the probability-generating functions F(j)(z,t) (j=0,1) as follows; see e.g. Gardiner (2009):

F(j)(z,t)=n=0znPn(j)(t). 2.6

In the case of autoactivation, the generating functions F(j)(z,t) can be seen to satisfy

tF(0)+(z-1)zF(0)+apdzzF(0)=-κfF(0)+κbF(1), 2.7a
tF(1)+(z-1)zF(1)-apdzzF(0)=κfF(0)-κbF(1)+λ(z-1)F(1) 2.7b

if the coefficients Pn(j)(t) in (2.6) obey the CME system (2.3); likewise, in the case of autorepression, (2.3) gives rise to

tF(0)+(z-1)zF(0)-rpdzzF(1)=-κfF(0)+κbF(1), 2.8a
tF(1)+(z-1)zF(1)+rpdzzF(1)=κfF(0)-κbF(1)+λ(z-1)F(1). 2.8b

Both (2.7) and (2.8) are systems of coupled, linear, first-order, hyperbolic partial differential equations (PDEs). Systems of this type are typically difficult to analyse; existing techniques only provide general results (Courant and Hilbert 1962; Taylor 2011).

To allow for an explicit analysis of systems (2.7) and (2.8), we make the following assumption:

Assumption 2.2

We assume that the autoactivation rate a in (2.2) is small in comparison with the other model parameters; specifically, we write

a=αpdδ, 2.9

where 0<δ<1 is sufficiently small. Likewise, we assume that the autorepression rate r is small in comparison with the other model parameters, writing

r=ρpdδ. 2.10

Previous work on the inclusion of autoregulatory effects in model selection by Feigelman et al. (2016) suggests that, in the context of Nanog expression in mouse embryonic stem cells, autoregulation rates are indeed small compared to other model parameters.

Based on Assumption 2.2, we can expand the generating functions F(j) (j=0,1) as power series in δ:

F(j)(z,t)=m=0δmFm(j)(z,t). 2.11

Substitution of (2.11) into (2.7) yields

tFm(0)+(z-1)zFm(0)=-κfFm(0)+κbFm(1)-αzzFm-1(0), 2.12a
tFm(1)+(z-1)zFm(1)=κfFm(0)-κbFm(1)+λ(z-1)Fm(1)+αzzFm-1(0); 2.12b

analogously, we substitute (2.11) into (2.8) to find

tFm(0)+(z-1)zFm(0)=-κfFm(0)+κbFm(1)+ρzzFm-1(1), 2.13a
tFm(1)+(z-1)zFm(1)=κfFm(0)-κbFm(1)+λ(z-1)Fm(1)-ρzzFm-1(1). 2.13b

We observe that, in (2.12) and (2.13), the same leading-order differential operator acts on both Fm(0) and Fm(1), which allows us to apply the method of characteristics to solve the equations for Fm(j) (j=0,1) simultaneously. In particular, we emphasise that, mathematically speaking, the resulting perturbation is regular in the perturbation parameter δ.

Dynamical systems analysis

In this section, we apply the method of characteristics to derive the ‘characteristic equations’ that are associated to the PDE systems (2.12) and (2.13), respectively; the former are systems of ODEs, which are naturally analysed in the language of dynamical systems.

Autoactivation

We first consider the case of autoactivation; to that end, we rewrite system (2.12) as

(t+vv)Fm(0)+κfFm(0)-κbFm(1)=-α(v+1)vFm-1(0), 2.14a
(t+vv)Fm(1)-κfFm(0)+κbFm(1)-λvFm(1)=α(v+1)vFm-1(0), 2.14b

where we have introduced the new variable

v=z-1. 2.15

The differential operator t+vv in Eq. (2.14) gives rise to characteristics ξ(s;v0) that obey the characteristic equation

vs=v; 2.16

these characteristics can thus be expressed as

ξ(s;v0)=s,v0es. 2.17

Since the partial differential operators in (2.14) transform into

t+vv=sand 2.18a
α(v+1)v=αv0+e-sv0, 2.18b

we arrive at the following system:

sFm(0)+κfFm(0)-κbFm(1)=-αv0+e-sv0Fm-1(0), 2.19a
sFm(1)-κfFm(0)+κbFm(1)-λv0esFm(1)=αv0+e-sv0Fm-1(0). 2.19b

Note that Eq. (2.19) is a recursive (nonhomogeneous) system of ordinary differential equations for Fm(j) (j=0,1). Henceforth, we will therefore refer to (2.19) as such, while retaining the use of partial derivatives s due to the presence of v0 in the corresponding right-hand sides.

To solve system (2.19), we rewrite it as a second-order ODE for Fm(0): we hence obtain

[s2+(κf+κb-λv0es)s-κfλv0es]Fm(0)=[λv0es-s]α(v0+e-s)v0Fm-1(0), 2.20

which can be solved recursively to determine Fm(0) for any m0. To simplify (2.20), we introduce the variable

w=λv0es, 2.21

which transforms the partial derivatives s and v0 into

swwandv0wv0w+v0; 2.22

Equation (2.20) hence reads

[(ww)2+(κf+κb-w)(ww)-κfw]Fm(0)=(w-ww)αw+λw+v0wv0Fm-1(0). 2.23

Using (2.19a), we can express the second component Fm(1) in terms of Fm(0) as

Fm(1)=1κbsFm(0)+κfFm(0)+αv0+e-sv0Fm-1(0)=1κbwwFm(0)+κfFm(0)+α(w+λ)w+v0wv0Fm-1(0). 2.24

At leading order, i.e. for m=0, (2.23) reduces to

[(ww)2+(κf+κb-w)(ww)-κfw]F0(0)=0, 2.25

the solutions of which can be expressed in terms of the confluent hypergeometric function 1F1, see §13 of NIST Digital Library of Mathematical Functions , to yield

F0(0)(w)=c11F1(κf,1+κf+κb,w)+c2w-κf-κb1F1(-κb,1-κf-κb,w). 2.26

Using (2.24), we can determine

F0(1)(w)=1κbwwF0(0)+κfF0(0), 2.27

with F0(0) as given in (2.26); for an explicit expression, see Eq. (A.1) in Appendix A.

The expression for F0(0)(w) in (2.26) allows us to determine the first-order correction F1(0)(w): substituting m=1 in (2.23), we obtain

[(ww)2+(κf+κb-w)(ww)-κfw]F1(0)=(w-ww)αw+λw+v0wv0F0(0). 2.28

Next, we apply the method of variation of constants to express the solution to (2.28) as

F1(0)(w)=1κf+κb[1F1(κf,1+κf+κb,w)c3w1F1(-κb,1-κf-κb,w^)g(w^)dw^-w-κf-κb1F1(-κb,1-κf-κb,w)c4w1F1(κf,1+κf+κb,w^)g(w^)w^-κf-κbdw^], 2.29

where

g(w)=e-ww(w-ww)αw+λw+v0wv0F0(0), 2.30

with F0(0) as given in (2.26). Finally, we may again use (2.24) to determine

F1(1)(w)=1κbwwF1(0)+κfF1(0)+α(w+λ)w+v0wv0F0(0), 2.31

with F0(0) and F1(0) as given in (2.26) and (2.29), respectively.

At this point in the analysis, the constants c1 and c2 in (2.26) and the integration limits c3 and c4 in (2.29) remain undetermined. To fix these constants, and thereby determine a unique solution to (2.19), we have to prescribe appropriate initial conditions.

Autorepression

Given the analysis of autoactivation in the previous subsection, the case of autorepression can be analysed in an analogous manner. Employing the same characteristics as before, recall (2.17), we obtain

sFm(0)+κfFm(0)-κbFm(1)=ρv0+e-sv0Fm-1(1), 2.32a
sFm(1)-κfFm(0)+κbFm(1)-λv0esFm(1)=-ρv0+e-sv0Fm-1(1) 2.32b

from system (2.13); cf. Eq. (2.19). Next, we rewrite (2.32) as a second-order ODE for Fm(1), using again the variable transformation in (2.21):

ww2+κf+κb-www-(1+κf)wFm(1)=-wwρw+λw+v0wv0Fm-1(1), 2.33

which can be solved recursively to obtain Fm(1) for any m1. The first component Fm(0) can be expressed in terms of Fm(1) as

Fm(0)=1κfwwFm(1)+(κb-w)Fm(1)+ρw+λw+v0wv0Fm-1(1); 2.34

to leading order, we thus obtain

F0(1)(w)=c^11F1(1+κf,1+κf+κb,w)+c^2w-κf-κb1F1(1-κb,1-κf-κb,w) 2.35

and

F0(0)=1κfwwF0(1)+(κb-w)F0(1). 2.36

The corresponding equation for the first-order correction F1(1) reads

[(ww)2+(κf+κb-w)(ww)-(1+κf)w]F1(1)=-(ww)ρ(w+λ)w+v0wv0F0(1), 2.37

which can be solved via the method of variation of constants to give

F1(1)(w)=1κf+κb[1F1(1+κf,1+κf+κb,w)c^3w1F1(1-κb,1-κf-κb,w^)h(w^)dw^-w-κf-κb1F1(1-κb,1-κf-κb,w)c^4w1F1(1+κf,1+κf+κb,w^)h(w^)w^-κf-κbdw^]; 2.38

here,

h(w)=-e-wwwwρw+λw+v0wv0F0(1), 2.39

with F0(1) as in (2.35). The first-order correction to the first component F0(0) is hence given by

F1(0)(w)=1κfwwF1(1)+(κb-w)F1(1)+ρw+λw+v0wv0F0(1), 2.40

with F0(1) and F1(1) as in (2.35) and (2.38), respectively. As in the case of autoactivation, the constants c^1 and c^2 in (2.35) and the integration limits c^3 and c^4 in (2.38) remain undetermined, and have to be fixed through suitable initial conditions.

Initial conditions

To determine appropriate initial conditions for the dynamical systems (2.19) and (2.32), we consider the original CME systems (2.3) and (2.5), respectively.

At time t=0, we impose an initial protein number n=n0, which implies

Pn(0)(0)+Pn(1)(0)=δn,n0 2.41

for the probabilities Pn(j)(t) (j=0,1); here, δn,n0 denotes the standard Kronecker symbol, with δn,n0=1 for n=n0 and δn,n0=0 otherwise. Using the definition of the generating functions F(j)(z,t) in (2.6), we find that

F(0)(z,0)+F(1)(z,0)=zn0=(v+1)n0, 2.42

taking into account the change of variables in (2.15). Thus, (2.42) provides an initial condition for the PDE systems (2.7) and (2.8). Given the power series expansion in (2.11), we infer that the coefficients Fm(j)(v,t) (j=0,1) satisfy

F0(0)(v,0)+F0(1)(v,0)=(v+1)n0and 2.43a
Fm(0)(v,0)+Fm(1)(v,0)=0for allm1, 2.43b

which holds for both (2.12) and (2.13).

To be able to interpret the initial conditions in (2.43) in the context of the dynamical systems (2.19) and (2.32), we revisit the method of characteristics, which was used to map the PDE systems (2.12) and (2.13) to the former, respectively.

The characteristics of the differential operator t+vv(=s) in (2.19) and (2.32) are the integral curves of the vector field (1, v). Geometrically speaking, these characteristic curves foliate the (tv)-plane over the v-axis. Therefore, each characteristic can be identified by its base point, which is the point where the characteristic curve intersects the v-axis, at v=v0; see Eq. (2.17) and Fig. 6.

Fig. 6.

Fig. 6

The (tv)-coordinate plane, on which the PDE systems (2.12) and (2.13) are solved. The characteristics are integral curves of the vector field (1, v), indicated in blue. The black characteristic curve intersects the v-axis at v=v0 (colour figure online)

Equivalently, each characteristic can be represented as a graph over the t-axis. Indeed, the differential equation for the v-component of a characteristic curve (2.16) can be solved to obtain the natural parametric description of a characteristic ‘fibre’ with base point v=v0, which is given by (s,v0es). Here, the parameter s along the characteristic is chosen such that its point of intersection with the v-axis lies at s=0. Given that choice, it is natural to identify the parameter along the characteristic (s) with the time variable (t). Hence, the initial conditions in (2.43), which determine a relation between Fm(0) and Fm(1) on the v-axis, can be interpreted on every characteristic as

F0(0)+F0(1)s=0=(v0+1)n0and 2.44a
Fm(0)+Fm(1)s=0=0for allm1, 2.44b

which again holds for both (2.19) and (2.32).

Remark 2.3

For CME systems such as (2.3) and (2.5), it is customary to impose a ‘normalisation’ condition of the form

m=0Pm(0)(t)+Pm(1)(t)=1, 2.45

as Pm(0)(t) and Pm(1)(t) represent probabilities. Recast in the framework of generating functions, recall (2.6), the above normalisation condition yields the boundary condition

F(0)(z,t)+F(1)(z,t)z=1=F(0)(v,t)+F(1)(v,t)v=0=1. 2.46

It is worthwhile to note that (2.46) is automatically satisfied whenever (2.43) is imposed: by adding the two equations in system (2.12)—or, equivalently, in (2.13)—one sees that Fm(0)+Fm(1) satisfies

sFm(0)+sFm(1)=λvFm(1). 2.47

The line {v=0} is represented by the ‘trivial’ characteristic, which is identified by v0=0; therefore, on that characteristic, we impose (2.43) to find

sFm(0)+sFm(1)=0, 2.48a
F0(0)+F0(1)s=0=1,and 2.48b
Fm(0)+Fm(1)s=0=0for allm1, 2.48c

which implies that F0(0)+F0(1)=1 for all s, as well as that Fm(0)+Fm(1) vanishes identically for all m1. Substituting these results into the power series representation for F(j)(z,t) (j=0,1) in (2.11), we obtain (2.46).

At this point, it is important to observe that Eq. (2.43) determines a line of initial conditions in the phase spaces of the dynamical systems (2.19) and (2.32). Therefore, at every order in δ, we can only fix one of the two free parameters that arise in the solution of the corresponding differential equations. In particular, (2.43) only fixes either c1 or c2 in (2.26), and either c3 or c4 in (2.29), and so forth. That indeterminacy motivates us to introduce a new parameter χ, which is defined as follows:

Definition 2.4

Consider the CME systems (2.3) and (2.5). We define χ(n0) to be the probability that the gene modelled by the reaction scheme in (2.1) is switched off at time t=0, given an initial protein number n0.

Definition 2.4 immediately specifies initial conditions for systems (2.3) and (2.5) via

Pn(0)(0)=χ(n0)δn,n0andPn(1)(0)=(1-χ(n0))δn,n0; 2.49

the above expression, in turn, provides us with a complete set of initial conditions for the PDE systems (2.7) and (2.8), to wit

F(0)(v,0)=χ(n0)(v+1)n0and 2.50a
F(1)(v,0)=(1-χ(n0))(v+1)n0. 2.50b

Here, we allow for the fact that χ(n0) may depend on other model parameters, and in particular on the autoregulation rates a and r; for a discussion of alternative options, see Sect. 4.1. We therefore expand χ(n0) as a power series in δ,

χ(n0)=m=0δmχm(n0). 2.51

The above expansion can be used to infer a complete set of initial conditions for the PDE systems (2.12) and (2.13), yielding

Fm(0)(v,0)=χm(v+1)n0for allm0, 2.52a
F0(1)(v,0)=(1-χ0)(v+1)n0,and 2.52b
Fm(1)(v,0)=-χm(v+1)n0for allm1. 2.52c

By the same reasoning that inferred (2.44) from (2.43), we can conclude that the complete set of initial conditions for the dynamical systems (2.19) and (2.32) is given by

Fm(0)s=0=χm(v0+1)n0for allm0, 2.53a
F0(1)s=0=(1-χ0)(v0+1)n0,and 2.53b
Fm(1)s=0=-χm(v0+1)n0for allm1. 2.53c

We can now use the conditions in (2.53) to determine the free constants c1 and c2 in (2.26), which yields

c1=(1+v0)n0e-λv0κbκf+κb(1F1(-κb,1-κf-κb,λv0)+χ0λv01-κf-κb1F1(1-κb,2-κf-κb,λv0))and 2.54a
c2=(1+v0)n0e-λv0(λv0)κf+κbκf+κb{[χ0(κf+κb)-κb]1F1(κf,1+κf+κb,λv0)+κfχ0λv01+κf+κb1F1(1+κf,2+κf+κb,λv0)}. 2.54b

Analogously, for c^1 and c^2 in (2.35), we obtain

c^1=c1κfκbandc^2=c2, 2.55

with c1 and c2 as in (2.54). Using conversion formulas found in §13.3(i) of NIST Digital Library of Mathematical Functions , one can show that the expressions resulting from (2.35) and (2.36) match those in (2.27) and (2.26), respectively, as expected; see also Eq. (A.1). Explicit expressions for the integration limits c3 and c4 in (2.29), as well as for the corresponding limits c^3 and c^4 in (2.38), can be found in Appendix A.

Inverse transformation

The final step towards providing explicit solutions to the PDE systems (2.7) and (2.8) consists in interpreting the solutions to the dynamical systems (2.19) and (2.32), with initial conditions as in (2.53), as solutions to the PDE systems (2.12) and (2.13), respectively. To that end, we again consider the corresponding characteristics from a geometric viewpoint.

As mentioned in Sect. 2.4, the (tv)-coordinate plane is foliated by characteristics, which are the integral curves of the vector field (1, v); recall Fig. 6. Hence, any point (tv) lies on a unique characteristic; flowing backward along that characteristic to its intersection with the v-axis, we can determine the corresponding base point v0 by the inverse transformation

(t,v)v0(t,v)=ve-t, 2.56

since we have identified the parameter along the characteristic (s) with the time variable (t).

To determine the value of Fm(j)(v,t) (j=0,1), interpreted as a solution to the PDE system (2.12) or (2.13), we proceed as follows. We first apply the inverse transformation in (2.56) to establish on which characteristic the coordinate pair (tv) lies. For that characteristic, identified by its base point v0, we then find the solution to the dynamical system (2.19) or (2.32), which is a function of s and v0. Next, we substitute s=t and v0=ve-t into that solution to obtain an explicit expression for the solution to the PDE system (2.12) or (2.13):

Fm(j)(v,t)[as solution to(2.12)or(2.13)]=Fm(j)(s,v0)=(t,ve-t), 2.57

where Fm(j) (j=0,1) on the right-hand side denotes the solution to (2.19) or (2.32), with initial conditions as in (2.53). Lastly, we substitute Fm(j)(v,t) into the power series in (2.11) to obtain an explicit solution to (2.7) or (2.8), to satisfactory order in δ.

Remark 2.5

The geometric interpretation of characteristics that was used to motivate the inverse transformation in (2.56) also shows that the introduction of the new system parameter χ in Definition 2.4 is necessary for the generating functions F(j) to be determined uniquely as solutions to (2.7) or (2.8), even if we are only interested in their sum F(0)(v,t)+F(1)(v,t). The crucial point is that any free constants obtained in the process of solving the dynamical systems (2.19) and (2.32)—or, equivalently, their second-order one-dimensional counterparts (2.20) and (2.33), respectively—are constant in s. In other words, they are constant along the particular characteristic on which the dynamical system is solved. These constants—such as e.g. c1 and c2 in (2.26)—can, and generally will, depend on the base point v0 of the characteristic; see for example (2.54). The inverse transformation in (2.56) that is used to reconstruct the solution to the original PDE from that of the corresponding dynamical system would then yield undetermined functions c(ve-t) in the resulting solutions to (2.7) and (2.8), respectively.

Summary of main result

To summarise Sect. 2, we combine the analysis of the previous subsections to state our main result.

Main result: The PDE system (2.7) can be solved for sufficiently small autoactivation rates a; see Assumption 2.2. Its solutions F(j)(z,t) (j=0,1) are expressed as power series in the small parameter δ; recall (2.11). The coefficients Fm(j)(z,t) in these series, written in terms of the shifted variable v defined in (2.15), can be found by

  1. solving recursively the second-order ODE (2.20) and using the identity in (2.24), incorporating the initial conditions in (2.53);

  2. and, subsequently, applying the inverse transformation in (2.56) to the resulting solutions.

Likewise, we can solve the PDE system (2.8) for sufficiently small autorepression rates r; see Assumption 2.2. Its solutions F(j)(z,t) are again expressed as power series in the small parameter δ; cf. (2.11). The coefficients Fm(j)(z,t) in these series, written in terms of the shifted variable v defined in (2.15), can be found by

  1. solving recursively the second-order ODE (2.33) and using the identity in (2.34), incorporating the initial conditions in (2.53);

  2. and, subsequently, applying the inverse transformation in (2.56) to the resulting solutions.

To illustrate the procedure described above, we state the resulting explicit expressions for the leading-order solution to (2.7)—or, equivalently, to (2.8)—in terms of the original variables z and t:

F0(0)(z,t)=[1+(z-1)e-t]n0e-λ(z-1)e-tκbκf+κb[1F1(-κb,1-κf-κb,λ(z-1)e-t)+χ0λ(z-1)e-t1-κf-κb1F1(1-κb,2-κf-κb,λ(z-1)e-t)]×1F1(κf,1+κf+κb,λ(z-1))-[1+(z-1)e-t]n0e-λ(z-1)e-te-(κf+κb)tκf+κb×{[χ0(κf+κb)-κb]1F1(κf,1+κf+κb,λ(z-1)e-t)+κfχ0λ(z-1)e-t1+κf+κb1F1(1+κf,2+κf+κb,λ(z-1)e-t)}×1F1(-κb,1-κf-κb,λ(z-1))and 2.58a
F0(1)(z,t)=[1+(z-1)e-t]n0e-λ(z-1)e-tκfκf+κb×[1F1(-κb,1-κf-κb,λ(z-1)e-t)+χ0λ(z-1)e-t1-κf-κb1F1(1-κb,2-κf-κb,λ(z-1)e-t)]×1F1(1+κf,1+κf+κb,λ(z-1))+[1+(z-1)e-t]n0e-λ(z-1)e-te-(κf+κb)tκf+κb×{[χ0(κf+κb)-κb]1F1(κf,1+κf+κb,λ(z-1)e-t)+κfχ0λ(z-1)e-t1+κf+κb1F1(1+κf,2+κf+κb,λ(z-1)e-t)}×1F1(1-κb,1-κf-κb,λ(z-1)). 2.58b

Note that similar expressions were derived by Iyer-Biswas et al. (2009), where an analogous generating function approach was applied under the assumption that the gene is initially inactive—i.e. that χ(n0)=1, see Definition 2.4—as well as that the initial protein number n0 is zero. With these choices, the expression for F0(0)+F0(1) from (2.58) can be seen to coincide with that found in Equations (5) through (7) of Iyer-Biswas et al. (2009), using §13.3.4 and §13.2.39 of NIST Digital Library of Mathematical Functions .

Model B: gene expression with explicit transcription

In this section, we apply our analytical method to model B, a stochastic gene expression model presented by Shahrezaei and Swain (2008a) which explicitly incorporates the transcription stage in the expression process, as well as DNA switching; see also Fig. 4. In the original article by Shahrezaei and Swain (2008a), a generating function approach was used to obtain analytical expressions for the time-independent (‘stationary’) solution to the model. For a visual guide to the upcoming analysis, the reader is again referred to Fig. 5.

Stochastic model and CME

The model for stochastic gene expression considered here is given by the reaction scheme

graphic file with name 285_2017_1196_Equ82_HTML.gif 3.1

The modelled gene can hence switch between inactive and active states which are denoted by D and D, respectively, with corresponding switching rates k0 and k1. The active gene is transcribed to mRNA (M) with rate ν0; mRNA is translated to protein (P) with rate ν1. Finally, mRNA decays with rate d0, while protein decays with rate d1.

As in model A, autoregulatory terms can be added to the core reaction scheme in (3.1). Since both mRNA and protein are modelled explicitly, we can identify four distinct autoregulatory mechanisms, in analogy to those in (2.2a) and (2.2b):

D+MaMD+M(autoactivation through mRNA), 3.2a
D+MrMD+M(autorepression through mRNA), 3.2b
D+PaPD+P(autoactivation through protein), 3.2c
D+PrPD+P(autorepression through protein). 3.2d

Autoactivation can be achieved either by mRNA or by protein, with rates aM or aP, respectively; similarly, autorepression can occur either through mRNA or through protein, with respective rates rM or rP.

The CME system associated to the reaction scheme in (3.1) is given by

dPm,n(0)dt=-κ0Pm,n(0)+κ1Pm,n(1)+γ(m+1)Pm+1,n(0)-mPm,n(0)+(n+1)Pm,n+1(0)-nPm,n(0)+γμmPm,n-1(0)-mPm,n(0), 3.3a
dPm,n(1)dt=κ0Pm,n(0)-κ1Pm,n(1)+γ(m+1)Pm+1,n(1)-mPm,n(1)+(n+1)Pm,n+1(1)-nPm,n(1)+γμmPm,n-1(1)-mPm,n(1)+λPm-1,n(1)-Pm,n(1). 3.3b

Here, Pm,n(j)(t) (j=0,1) represents the probability of m mRNA and n protein being present at time t while the gene is either inactive (0) or active (1). As in (2.3) and (2.5), the time variable is nondimensionalised by the protein decay rate d1; other model parameters are scaled as

κ0=k0d1,κ1=k1d1,γ=d0d1,λ=ν0d1,andμ=ν1d0. 3.4

We note that the above scaling was also used by Shahrezaei and Swain (2008a). The effects of incorporating the autoregulatory mechanisms in (3.2) into the CME system (3.3) are specified in Table 1.

Table 1.

Contribution to the right-hand sides of (3.3) that is due to incorporation of the autoregulatory mechanisms in (3.2)

Autoregulation type Contribution to (3.3a) Contribution to (3.3b)
mRNA autoactivation -aMd1mPm,n(0) +aMd1mPm,n(0)
mRNA autorepression +rMd1mPm,n(1) -rMd1mPm,n(1)
Protein autoactivation -aPd1nPm,n(0) +aPd1nPm,n(0)
Protein autorepression +rPd1nPm,n(1) -rPd1nPm,n(1)

Generating function PDE

Since the probabilities Pm,n(j)(t) (j=0,1) in (3.3) depend on both the mRNA number m and the protein number n, we introduce probability-generating functions that are defined by double asymptotic series:

F(j)(w,z,t)=m=0n=0wmznPm,n(j)(t)forj=0,1. 3.5

For coefficients Pm,n(j)(t) that obey the CME system (3.3), the associated generating functions F(j)(w,z,t) satisfy

tF(0)+(z-1)zF(0)+γ(w-1)wF(0)-γμ(z-1)wwF(0)=-κ0F(0)+κ1F(1), 3.6a
tF(1)+(z-1)zF(1)+γ(w-1)wF(1)-γμ(z-1)wwF(1)=κ0F(0)-κ1F(1)+λ(w-1)F(1). 3.6b

The effects of incorporating the autoregulatory mechanisms in (3.2) into system (3.6) are specified in Table 2.

Table 2.

Contribution to the right-hand sides of (3.6) that is due to incorporation of the autoregulatory mechanisms in (3.2)

Autoregulation type Contribution to (3.6a) Contribution to (3.6b)
mRNA autoactivation -aMd1wwF(0) +aMd1wwF(0)
mRNA autorepression +rMd1wwF(1) -rMd1wwF(1)
Protein autoactivation -aPd1zzF(0) +aPd1zzF(0)
Protein autorepression +rPd1zzF(1) -rPd1zzF(1)

Dynamical systems analysis

As before, the PDE system (3.6) for the generating function can be reformulated as a system of ODEs via the method of characteristics. The differential operator t+(z-1)z+γ(w-1)w-γμ(z-1)ww now gives rise to the characteristic system

u˙=γ[u-μv(u+1)], 3.7a
v˙=v, 3.7b

where u˙=duds is the derivative along the characteristic, which is parametrised by s. For simplicity, we have introduced the new variables u and v, which are defined as

u=w-1andv=z-1, 3.8

respectively. On the resulting characteristics, Eq. (3.6) yields the system of ODEs

F˙(0)=-κ0F(0)+κ1F(1), 3.9a
F˙(1)=κ0F(0)-κ1F(1)+λuF(1). 3.9b

First, we note that the characteristic system (3.7) can be solved explicitly in terms of the incomplete Gamma function Γ(a,z), see §8.2(i) of NIST Digital Library of Mathematical Functions ), yielding

u(s)=eγ(s-μv0es){u0eγμv0+-γμv0γ[Γ(1-γ,-γμv0)-Γ(1-γ,-γμv0es)]}, 3.10a
v(s)=v0es. 3.10b

However, due to its complex nature, the expression for u(s) given in (3.10a) cannot be used to obtain explicit expressions for the generating functions F(j) that solve system (3.9). Therefore, inspired by the analysis by Shahrezaei and Swain (2008a) and Popović et al. (2016), we make the following assumption:

Assumption 3.1

We assume that the decay rate of protein (d1) is smaller than the decay rate of mRNA (d0); specifically, we write

d0d1=γ=1ε, 3.11

where 0<ε<1 is sufficiently small.

The resulting scale separation between mRNA and protein decay rates is well-documented in many microbial organisms, including in bacteria and yeast (Shahrezaei and Swain 2008a; Yu et al. 2006).

For clarity of presentation, we make an additional assumption here.

Assumption 3.2

We assume that all other model parameters κ0, κ1, λ, and μ, as defined in (3.4), are O(1) in ε.

Remark 3.3

Although Assumption 3.2 is not strictly necessary for the upcoming analysis, it is beneficial. It is worthwhile to note that the analytical scheme presented in this section can be applied in a straightforward fashion in cases where Assumption 3.2 fails, which is particularly relevant in relation to previous work (Shahrezaei and Swain 2008a; Feigelman et al. 2015), where the CME system (3.3) is studied for parameter values far beyond the range implied by Assumption 3.2.

Using Assumption 3.1, we can write the characteristic system (3.7) as

εu˙=u-μv(u+1), 3.12a
v˙=v. 3.12b

Since ε is assumed to be small, we can classify (3.12) as a singularly perturbed slow-fast system in standard form; see Kuehn (2015). A comprehensive slow-fast analysis of Eq. (3.12) was carried out by Popović et al. (2016); we highlight some relevant aspects of that analysis here.

System (3.12) gives rise to a critical manifold C0=(u,v)|u=μv1-μv. If ε is asymptotically small, orbits of (3.12) can be separated into slow and fast segments, using Fenichel’s geometric singular perturbation theory (Kuehn 2015). The critical manifold C0 is normally repelling; in other words, orbits converge to C0 in backward time at an exponential rate. For initial conditions asymptotically close to C0, orbits initially follow C0 closely for some time, after which they move away from C0 under the fast dynamics; see also Fig. 7.

Fig. 7.

Fig. 7

Phase space dynamics of systems (3.12) and (3.13). The slow flow along C0 is indicated by single arrows; the fast dynamics transverse to C0 are denoted by double arrows

To analyse the fast dynamics of system (3.12), we introduce the fast variable σ=sε; in terms of σ, (3.12) is hence expressed as

u=u-μv(u+1), 3.13a
v=εv, 3.13b

where u=dudσ. We can solve (3.13b) explicitly and write the result as a power series in ε, which yields

v(σ)=v0eεσ=v0n=0εnσnn!. 3.14

Expressing the solution to (3.13a) as a power series in ε, as well, i.e. writing

u(σ)=n=0εnu^n(σ), 3.15

substituting (3.15) into (3.13a), and making use of (3.14), we find

du^ndσ=u^n-μv0σnn!1+l=0nσ-ln!(n-l)!u^l, 3.16

with initial conditions

u^0(0)=u0andu^n(0)=0forn1. 3.17

The first two terms in (3.15) are thus given by

u^0(σ)=μv01-μv0+u0-μv01-μv0e(1-μv0)σand 3.18a
u^1(σ)=μv0(1-μv0)21-e(1-μv0)σ1-μv0+σ-u0-μv01-μv0μv0σ22e(1-μv0)σ. 3.18b

We now employ the expansion in (3.15) to obtain explicit expressions for the generating functions F(j) (j=0,1). In the fast variable σ=sε, system (3.9) takes the form

F(0)=-εκ0F(0)+εκ1F(1), 3.19a
F(1)=εκ0F(0)-εκ1F(1)+ελuF(1). 3.19b

As in the analysis of model A, recall Sect. 2.3, we rewrite system (3.19) as a second-order ODE for F(0) to find

d2dσ2F(0)+ε(κ0+κ1-λu)ddσF(0)-ε2λκ0uF(0)=0. 3.20

Next, we use (3.19a) to express F(1) in terms of F(0) as

F(1)=1εκ1F(0)+εκ0F(0). 3.21

To incorporate the expansion for u in (3.15), we also expand F(j) (j=0,1) in powers of ε, writing

F(j)(σ)=n=0εnFn(j)(σ). 3.22

Substitution of (3.22) into (3.20) then yields

d2dσ2F0(0)=0, 3.23a
d2dσ2F1(0)+(κ0+κ1-λu^0)ddσF0(0)=0, 3.23b

and

d2dσ2Fn(0)+(κ0+κ1-λu^0)ddσFn+1(0)-λk=0nu^n+1-kddσ+κ0u^n-kFk(0)=0 3.24

for n2. By combining (3.21) with (3.22), we obtain

ddσF0(0)=0 3.25

and

Fn(1)=κ0κ1Fn(0)+1κ1ddσFn+1(0) 3.26

for n0. We can solve Eqs. (3.23) and (3.24) iteratively—taking into account the additional condition on F0(0) in (3.25)—to find

F0(0)=f0, 3.27a
F1(0)=f1+g1σ,and 3.27b
F2(0)=f2+g2σ+λ(f0κ0+g1)1-μv0e(1-μv0)σ1-μv0u0-μv01-μv0+μv0σ22-g1(κ0+κ1)σ22 3.27c

for the first three terms which, upon substitution into (3.26), yields

F0(1)=f0κ0+g1κ1and 3.28a
F1(1)=f1κ0+g2κ1+λ(f0κ0+g1)κ1(1-μv0)e(1-μv0)σu0-μv01-μv0+σμv0-σg1; 3.28b

here, fi and gi are free constants, to be determined by initial conditions; see Sect. 3.4.

Contrary to common practice in the study of slow-fast systems such as (3.19), we forego a detailed analysis of the slow system (3.9), continuing our discussion with the determination of appropriate initial conditions; cf. Sect. 2.4. For details on why the slow dynamics is disregarded, the reader is referred to Remark 3.5.

Initial conditions and inverse transformation

To complete our analytical method, we discuss the determination of initial conditions and the reconstruction of the solution to the original PDE system (3.6), as was done for model A in Sects. 2.4 and 2.5, respectively.

Initial conditions

We follow the reasoning of Sect. 2.4, and determine appropriate initial conditions by considering the original CME system (3.3).

At time t=0, we prescribe initial mRNA and protein numbers m0 and n0, respectively. As in Sect. 2.4, we again introduce the parameter χ, which is defined as follows; compare Definition 2.4:

Definition 3.4

Consider the CME system (3.3). We define χ(m0,n0) to be the probability that the gene modelled by the reaction scheme in (3.1) is switched off at time t=0, given initial mRNA and protein numbers m0 and n0, respectively.

For the probabilities Pm,n(j)(t) (j=0,1), Definition 3.4 implies that

Pm,n(0)(0)=χδm,m0δn,n0and 3.29a
Pm,n(1)(0)=(1-χ)δm,m0δn,n0. 3.29b

Using the definition of the generating functions F(j)(w,z,t) in (3.5), we find

F(0)(w,z,0)=χwm0zn0=χ(1+u)m0(1+v)n0and 3.30a
F(1)(w,z,0)=(1-χ)wm0zn0=(1-χ)(1+u)m0(1+v)n0, 3.30b

taking into account the change of variables in (3.8).

Next, we can infer e.g. from (3.10) and well-known properties of the incomplete Gamma function—see §8.2 of NIST Digital Library of Mathematical Functions —that solutions to (3.7) exist globally, i.e. for all s. In particular, given an arbitrary triple (s,u,v), we can apply the inverse flow of (3.7) to flow backward (u,v) for time s. We conclude that the characteristics of the operator t+(z-1)z+γ(w-1)w-γμ(z-1)ww=t+vv+γuu-γμ(u+1)vu, which are given by the orbits of system (3.7), foliate the (tuv)-coordinate space over the t=0-plane when the parameter along the characteristic (s) is identified with the time variable (t). We can therefore uniquely identify such a characteristic—interpreted as a fibre over the t=0-plane—by its base point. Because orbits of (3.7) provide a parametrisation of the underlying characteristics, the coordinates of that base point are given by (0,u0,v0), where (u0,v0) are the initial values of the corresponding orbit of (3.7). Hence, the conditions in (3.30) yield, on each characteristic,

F(0)s=0=χ(1+u0)m0(1+v0)n0and 3.31a
F(1)s=0=(1-χ)(1+u0)m0(1+v0)n0. 3.31b

As s=0 implies σ=0, we can apply the initial conditions in (3.31) to the solutions of the ODE system (3.19) in order to determine the free constants fi and gi in (3.27) and (3.28). Combining the power series expansion in (3.22) with (3.31), we obtain

F0(0)σ=0=χ(1+u0)m0(1+v0)n0, 3.32a
F0(1)σ=0=(1-χ)(1+u0)m0(1+v0)n0,and 3.32b
Fn(j)σ=0=0for alln1,withj=0,1, 3.32c

which implies

f0=χ(1+u0)m0(1+v0)n0, 3.33a
f1=0, 3.33b
f2=-(1-χ)(1+u0)m0(1+v0)n0λκ1(1-μv0)2u0-μv01-μv0, 3.33c
g1=(κ1(1-χ)-κ0χ)(1+u0)m0(1+v0)n0,and 3.33d
g2=-(1-χ)(1+u0)m0(1+v0)n0λκ11-μv0u0-μv01-μv0 3.33e

in (3.27) and (3.28).

Inverse transformation

Since the (tuv)-coordinate space is foliated by the characteristics of the operator t+vv+γuu-γμ(u+1)vu, any point (tuv) lies on a unique characteristic. Flowing backward along that characteristic to its intersection with the t=0-plane, we can determine the corresponding base point (0,u0,v0) by inverting the relations in (3.10). Since the dynamics of the v-coordinate do not depend on u, we may use (2.56) to express v0 in terms of t and v only. Taking the resulting expression as input for inverting (3.10a), we obtain the inverse characteristic transformation

(t,u,v)u0(t,u,v),v0(t,v), 3.34

with

u0(t,u,v)=e-γ(t+μve-t)ueγμv--γμvγ[Γ(1-γ,-γμve-t)-Γ(1-γ,-γμv)]and 3.35a
v0(t,v)=ve-t. 3.35b

Under Assumption 3.1, we can employ the power series expansion in (3.15), in combination with the recursive set of ODEs in (3.16) and initial conditions as in (3.17), as an alternative to (3.35a) to obtain u0 as a function of u, v (or v0), and σ. Rewriting the result as a power series in ε, we find

u0(σ,u,v)=μv1-μv+e(μv-1)σu-μv1-μv-εμve(μv-1)σ-1(1-μv)3+σ(1-μv)2+σ22e(μv-1)σu-μv1-μv+O(ε2); 3.36

naturally, (3.14) gives rise to

v0(σ,v)=v0n=0εn(-σ)nn!. 3.37

Since the independent variable in (3.16) is σ, and since s is naturally identified with t, we replace σ with tε in (3.36) to obtain a perturbative expansion for u0(t,u,v).

The solution to the PDE system (3.6) can now be found to satisfactory order in ε by applying the inverse transformation in (3.34) to the solutions given in (3.27) and (3.28), taking into account the values of fi and gi in (3.33). In other words,

F(j)(u,v,t)[as solution to (3.6)]=n=0εnFn(j)(u0,v0)=(u0(σ,u,v),ve-εσ)σ=tε, 3.38

where Fn(j) (j=0,1) on the right-hand side of the above expression denotes the solution to Eqs. (3.23a)–(3.26), with initial conditions as in (3.32).

Remark 3.5

The absence of any detailed analysis of the characteristic system in its slow formulation, Eq. (3.12), can be argued as follows, by considering the corresponding phase space, as depicted in Fig. 7.

  1. For arbitrary initial conditions (u0,v0), the dominant dynamics are fast, since the critical manifold C0 is normally repelling. In other words, solutions are generally repelled away from C0 under the fast dynamics.

  2. All orbits that have their initial conditions on the same Fenichel fibre are exponentially close (in ε) to each other near the slow manifold Cε that is associated to the critical manifold C0. Therefore, flowing backward from (uv) to (u0,v0)—as expressed through the inverse transformation in (3.34) that yields the corresponding PDE solution—may introduce exponentially large terms in the transformation, precluding any sensible series expansion.

Thus, although the construction of a composite ‘(initially) slow—(ultimately) fast’ expression of F(j) as a solution to systems (3.12) and (3.13) certainly makes sense from a dynamical systems perspective, the extreme lack of sensitivity of orbits on their initial conditions (u0,v0) may prevent such a composite expansion from being useful for obtaining solutions to the original PDE system (3.6).

Remark 3.6

In Sect. 3.3, explicit expressions are given for the expansion of F(0) up to O(ε2) only, cf. (3.27); similarly, F(1) is approximated to O(ε) in (3.28), for the sake of brevity. It is worthwhile to note that a lower bound on the order of the expansion is stipulated by the application; recall Sect. 1.1: the sampling time Δt can be considered as a minimum time interval over which the results of our analysis should be (reasonably) accurate. To that end, we have to compare Δt with ε=1γ, the parameter defining the fast time scale on which the above analytical results have been derived. We can then apply the classical theory of Poincaré expansions (Verhulst 2000) to infer that, if for example Δt=O(1)—which implies Δt=O(ε-1) in the fast time variable σ—the generating functions F(j) should at least be expanded up to O(ε2) for the resulting approximation to be accurate to O(ε).

Autoregulation

The inclusion of any type of autoregulation into system (3.6)—which is equivalent to the addition of model terms from Table 2 to the right-hand sides of the corresponding equations—precludes the direct application of the method of characteristics, as the resulting partial differential operators in Eqs. (3.6a) and (3.6b) do not coincide anymore. To resolve that complication, we follow the approach of Sect. 2, making the following assumption:

Assumption 3.7

We assume that the autoregulation rates aM, rM, aP, and rP, as defined in (3.2), are small in comparison with the protein decay rate d1; specifically, we write

aM=αMd1δ,rM=ρMd1δ,aP=αPd1δ,andrP=ρPd1δ, 3.39

where 0<δ<1 is sufficiently small.

Next, we expand the generating functions F(j) (j=0,1) as power series in δ; recall (2.11):

F(j)(z,w,t)=m=0δmFm(j)(z,w,t). 3.40

To demonstrate the procedure, we include mRNA autoactivation in (3.6), see again Table 2; the analysis of the remaining autoregulatory mechanisms can be performed in a similar fashion. Substitution of (3.40) now yields

t+(z-1)z+γ(w-1)w-γμ(z-1)wwFm(0)=-κ0Fm(0)+κ1Fm(1)-αMwwFm-1(0), 3.41a
t+(z-1)z+γ(w-1)w-γμ(z-1)wwFm(1)=κ0Fm(0)-κ1Fm(1)+λ(w-1)Fm(1)+αMwwFm-1(0); 3.41b

cf. (2.12). System (3.41) is then amenable to the method of characteristics. In fact, employing the same characteristics as in the unperturbed setting, recall (3.7), we find

t+(z-1)z+γ(w-1)w-γμ(z-1)ww=s, 3.42

while the partial differential operators in Table 2 transform into

ww=(u+1)uu0-1u0and 3.43a
zz=(v+1)vv0-1-uv0uu0-1u0+v0; 3.43b

here, u(s;u0,v0) and v(s;v0) are as given in (3.10). Thus, the mRNA autoactivation system (3.41) transforms to

sFm(0)=-κ0Fm(0)+κ1Fm(1)-αM(u+1)uu0-1u0Fm-1(0), 3.44a
sFm(1)=κ0Fm(0)-κ1Fm(1)+λuFm(1)+αM(u+1)uu0-1u0Fm-1(0). 3.44b

To obtain explicit solutions to system (3.44), we adopt Assumption 3.1 and revert to the fast time scale σ=sε to write the dynamical system (3.44) as a second-order ODE for Fm(0), which yields

σ2+ε(κ0+κ1-λu)σ-ε2κ0λuFm(0)=(ε2λu-εσ)αM(u+1)uu0-1u0Fm-1(0). 3.45

Using (3.44), we can express Fm(1) in terms of Fm(0) as

Fm(1)=1εκ1σFm(0)+εκ0Fm(0)+εαM(u+1)uu0-1u0Fm-1(0). 3.46

To solve (3.45) (recursively), we expand Fm(j) (j=0,1) in powers of ε:

Fm(j)(σ)=n=0εnFm,n(j)(σ); 3.47

recall Eq. (3.22). Together with the series expansion for u in (3.15), we thus obtain

σ2Fm,0(0)=0, 3.48
σ2Fm,1(0)=-(κ0+κ1-λu^0)σFm,0(0)-σGm-1,0, 3.49

and

σ2Fm,n(0)+(κ0+κ1-λu^0)σFm,n+1(0)-λk=0nu^n+1-kσ+κ0u^n-kFm,k(0)=-σGm-1,n+1(0)+λk=0nu^n-kGm-1,k(0) 3.50

for n2; compare with Eqs. (3.23a), (3.23b), and (3.24). The coefficients Gm,n(0) in the above expression are defined from an expansion of the autoregulation term as

αM(u+1)uu0-1u0Fm-1(0)=n=0εnGm-1,n(0)(u0,σ). 3.51

From (3.46), we obtain

σFm,0(0)=0 3.52

and

Fm,n(1)=1κ1σFm,n+1(0)+κ0κ1Fm,n(0)+1κ1Gm-1,n(0) 3.53

for n0; recall (3.25) and (3.26). To solve Eqs. (3.48) through (3.53) iteratively, we fix m—the order of the expansion in δ—and determine the solution to satisfactory order in n, the order of the expansion in ε. Then, we increase m to m+1 and take the result as input for the dynamics at order m+1. The resulting repeated iteration procedure yields an explicit expression for the generating functions F(j) (j=0,1) as double asymptotic series in both δ and ε.

The determination of appropriate initial conditions is largely analogous to the non-autoregulated case; see Sect. 3.4.1. However, with the inclusion of autoregulation into model B, we need to incorporate the possibility that χ(m0,n0) depends on the corresponding autoregulation rates. As in the case of model A, we expand χ(m0,n0) as a power series in δ:

χ(m0,n0)=m=0δmχm(m0,n0); 3.54

recall (2.51). In that case, the initial conditions for Fm,n(j) can be inferred from (3.32) to give

Fm,0(0)σ=0=χm(1+u0)m0(1+v0)n0for allm0, 3.55a
F0,0(1)σ=0=(1-χ0)(1+u0)m0(1+v0)n0, 3.55b
Fm,0(1)σ=0=-χm(1+u0)m0(1+v0)n0for allm1,and 3.55c
Fm,n(j)σ=0=0for allm0andn1,withj=0,1. 3.55d

Solutions to Eqs. (3.48)–(3.53) that incorporate the conditions in (3.55) for all types of autoregulation introduced in (3.2) can be found in Appendix B. Finally, our previous results on the inverse characteristic transformation in the non-autoregulated case from Sect. 3.4.2 can now be applied in a straightforward fashion to give solutions to the PDE system (3.6) with added autoregulation.

Summary of main result

To summarise Sect. 3, we combine the analysis of the previous subsections to state our main result.

Main result: The PDE system (3.6) can be solved for sufficiently large values of γ; see Assumption 3.1. Its solutions F(j)(w,z,t) (j=0,1) are expressed as power series in the small parameter ε=1γ; cf. (3.22). The coefficients Fn(j)(w,z,t) in these series, written in terms of the shifted variables u and v defined in (3.8), can be found by

  1. solving recursively the second-order ODEs (3.23a) through (3.24) and using the identities in (3.25) and (3.26), incorporating the initial conditions in (3.32);

  2. subsequently applying the inverse transformations in (3.36) and (3.37) to the resulting solutions;

  3. and, finally, substituting σ=tε.

To illustrate the procedure described above, we state the resulting explicit expressions for the leading-order solution to (3.6) in terms of the original variables w, z, and t here:

F0(0)(w,z,t)=χ[1+(z-1)e-t]n0{11+μ(1-z)+e-[1+μ(1-z)]tε×[w-11+μ(1-z)]}m0and 3.56a
F0(1)(w,z,t)=1-χχF0(0)(w,z,t). 3.56b

Note that the sum F0(0)+F0(1) corresponds precisely to the leading-order fast expansion found in Equation (21) of Bokes et al. (2012b).

If autoregulation as in (3.2) is incorporated into model B, the main result can be formulated as follows.

Main result (autoregulatory extension): The PDE system (3.6) incorporating any one type of autoregulation from Table 2 can be solved as long as γ is sufficiently large and δ is sufficiently small; see Assumptions 3.1 and 3.7, respectively. Its solutions F(j)(w,z,t) (j=0,1) are expressed as double power series in the small parameters δ and ε, viz.

F(j)=m=0n=0δmεnFm,n(j). 3.57

The coefficients Fm,n(j) in these series, written in terms of the shifted variables u and v defined in (3.8), can be found by

  1. solving recursively the second-order ODEs (3.48) through (3.50) for fixed m and using the identities in (3.52) and (3.53), incorporating the initial conditions in (3.55);

  2. increasing m to m+1, and repeating step (1a) until a sufficient accuracy in δ (and ε) is attained;

  3. subsequently applying the inverse transformations in (3.36) and (3.37) to the resulting solutions;

  4. and, finally, substituting σ=tε.

From generating function to propagator

The final step in our analytical method consists in reconstructing the probabilities Pn(j) (model A) and Pm,n(j) (model B), respectively, from the explicit expressions for the associated generating functions F(j) (j=0,1), which were the main analytical outcome of Sects. 2 and 3.

In principle, the relation between probabilities and probability-generating functions is clear from the definition of the latter, and is given in (2.6) and (3.5), respectively. Specifically, probabilities can be expressed in terms of derivatives of their generating functions as follows:

Pn(j)(t)=1n!nznF(j)(z,t)z=0(model A), 4.1a
Pm,n(j)(t)=1m!n!m+nwmznF(j)(w,z,t)(w,z)=(0,0)(model B). 4.1b

However, explicit expressions for the nth and (m+n)th order derivatives, respectively, of these generating functions become progressively unwieldy with increasing m and n. Indeed, from the expressions for the generating functions F(j) obtained previously, which combine (2.26) for specific initial conditions as in (2.54) with the inverse characteristic transformation in (2.56), it is clear that finding explicit expressions for derivatives of arbitrary order is very difficult indeed, if it is possible at all.1

To complete successfully the final step towards approximating propagators for parameter inference in the present setting, we abandon the requirement of deriving explicit expressions for the probabilities Pn(j) and Pm,n(j). Instead, we use the standard Cauchy integral formula for derivatives of holomorphic functions to write

Pn(j)(t)=12πiγAF(j)(z,t)zn+1dz(model A), 4.2a
Pm,n(j)(t)=1(2πi)2γBF(j)(w,z,t)wm+1zn+1dwdz(model B); 4.2b

here, γA is a suitably chosen contour around z=0, while γB is a (double) contour around (w,z)=(0,0).

The above expression of probabilities as integrals is well suited for an efficient numerical implementation, which is naturally incorporated into the realisation of the parameter inference scheme discussed in Sect. 1. From a numerical perspective, the integral formula in (4.2) has the additional advantage that the values of F(j) on (a discretisation of) the integral contours γA and γB, respectively, only have to be determined once to yield propagators for any values of m and n and fixed initial states m0 and n0. Moreover, we are free to choose the integration contours γA and γB, which allows us to accelerate the calculation of these integrals; see Bornemann (2011). Here, we note that the choice of circular integration contours with unit radius, and subsequent discretisation of those contours as M-sided and N-sided polygons, respectively, coincides with the ‘Fourier mode’ approach, as presented by Bokes et al. (2012a).

Remark 4.1

By introducing the Cauchy integral formula for derivatives of holomorphic functions in (4.2), we implicitly assume that the integration contours γA and γB are chosen such that they lie completely within the open neighbourhoods of the origin in C and C2, respectively, where the canonical complex extensions of the generating functions F(j)—which exist by the Cauchy-Kowalevski theorem—are holomorphic. In other words, γA and γB must be chosen such that any poles of F(j) lie outside of these integration contours. The expansion for u0 in (3.36) shows that this is not a moot point: the generating functions F(j) resulting for model B, as established in Sect. 3, will generically have a pole at v=1μ, i.e. at z=1+1μ. As μ is positive, by (3.4), choosing the z-contour of γB to be a circle with at most unit radius allows us to avoid that pole.

Incorporation of χ

In the course of the analysis presented in Sects. 2 and 3, the introduction of the parameter χ was necessary to obtain definite, explicit expressions for the generating functions as solutions to the PDE systems (2.7), (2.8), and (3.6); see Definitions 2.4 and 3.4. The successful implementation of these expressions in a parameter inference scheme requires us to decide how to incorporate that new parameter. We identify three options here.

  1. Before implementing parameter inference, we can marginalise over the new parameter χ to eliminate it altogether, using a predetermined measure dμ(χ), which adds an additional integration step to the requisite numerical scheme.

  2. We can make a choice for χ that is based on the specifics of the model under consideration. Thus, exploiting the Markov property of the stochastic models underlying (2.3), (2.5), and (3.3), we may use the switching rates and any autoregulation rates to express χ in model A as
    χ(n0)=cbcb+cf+an0(autoactivation), 4.3a
    χ(n0)=cb+rn0cb+rn0+cf(autorepression); 4.3b
    the corresponding expressions for model B read
    χ(m0,n0)=k1k0+k1(no autoregulation), 4.4a
    χ(m0,n0)=k1k0+k1+aMm0(mRNA autoactivation), 4.4b
    χ(m0,n0)=k1+rMm0k0+k1+rMm0(mRNA autorepression), 4.4c
    χ(m0,n0)=k1k0+k1+aPn0(protein autoactivation), 4.4d
    χ(m0,n0)=k1+rPn0k0+k1+rPn0(protein autorepression). 4.4e
  3. We can determine χ ‘experimentally’ by including the latter in the parameter set that is to be inferred in the (numerical) process of parameter inference.

Note that step 2 has been anticipated in the analysis of model A, by introducing the series expansion in (2.51). Indeed, by Assumption 2.2, we can expand χ(n0) as

χ=κbκb+κf+δαn0=κbκf+κbm=0δm-αn0κf+κbm(autoactivation), 4.5a
χ=κb+δρn0κb+κf+δρn0=κbκf+κb-κfκf+κbm=1δm-ρn0κf+κbm(autorepression). 4.5b

Likewise, when autoregulation is added to model B, Assumption 3.7 implies an expansion for χ(m0,n0) of the form

χ=κ1κ0+κ1+δαMm0=κ1κ0+κ1m=0δm-αMm0κ0+κ1m(mRNA autoactivation), 4.6a
χ=κ1+δρMm0κ0+κ1+δρMm0=κ1κ0+κ1-κ0κ0+κ1m=1δm-ρMm0κ0+κ1m(mRNA autorepression), 4.6b
χ=κ1κ0+κ1+δαPn0=κ1κ0+κ1m=0δm-αPn0κ0+κ1m(protein autoactivation), 4.6c
χ=κ1+δρPn0κ0+κ1+δρPn0=κ1κ0+κ1-κ0κ0+κ1m=1δm-ρPn0κ0+κ1m(protein autorepression). 4.6d

Discussion and outlook

In the present article, we have developed an analytical method for obtaining explicit, fully time-dependent expressions for the probability-generating functions that are associated to models for stochastic gene expression. Moreover, we have presented a computationally efficient approach which allows us to derive model predictions (in the form of propagators) from these generating functions, using the Cauchy integral formula. It is important to note that our method does not make any steady-state or long-evolution-time approximations. On the contrary, the perturbative nature of our approach naturally optimises its applicability over relatively short (or intermediate) time scales; see also Remark 3.6. As is argued in Sect. 1.1, such relatively short evolution times naturally occur in the calculation of quantities such as the log-likelihood, as defined in Eq. (1.1). Therefore, our analytical approach is naturally suited to an implementation in a Bayesian parameter inference scheme, such as is outlined in Sect. 1.1.

As mentioned in Sects. 2.2 and 3.3, the introduction of Assumptions 2.2 and 3.7 in our analysis of the systems of PDEs and ODEs that are obtained via the generating function approach is necessary for determining explicit expressions for the generating functions themselves. Therefore, we can only be certain of the validity of our approach if we assume that the autoregulation rates are small in comparison with other model parameters, as is done there. Moreover, in the analysis of model B, we have to assume that the protein decay rate is smaller than the decay rate of mRNA; recall Assumption 3.1. That assumption is valid for a large class of (microbial) organisms (Shahrezaei and Swain 2008a; Yu et al. 2006); however, it is by no means generic, as the two decay rates are often comparable in mammalian cells (Schwanhäusser et al. 2011; Vogel and Marcotte 2012). Since the accuracy of approximation of the explicit expressions for the generating functions derived here is quantified in terms of orders of the perturbation parameter(s), see e.g. Remark 3.6, violation of Assumption 2.2, 3.1, or 3.7 will decrease the predictive power of the results obtained by the application of the analytical method developed in the article.

The method which is described in Sect. 1.2, and outlined visually in Fig. 5, hence provides a generic framework for the analysis of stochastic gene expression models such as model A (Fig. 3 and Sect. 2.1) and model B (Fig. 4 and Sect. 3.1). Note that, for example, the steady-state and long-evolution-time approximations derived by Shahrezaei and Swain (2008a) could be extended to autoregulatory systems via the same approach. However, as is apparent from the (differences between the) analysis presented in Sects. 2 and 3, the ‘path to an explicit solution’ is highly model-dependent. The decision on which analytical techniques to apply, such as the perturbative expansion postulated in (3.22), has to be made on a case-by-case basis. The success of the method presented in the article fully depends on whether the resulting dynamical systems can be solved explicitly. To that end, it is highly beneficial that the systems (2.19) and (3.19) obtained here are linear, which is a direct consequence of the fact that all reactions described in the reaction schemes in (2.1) and (2.2), as well as in (3.1) and (3.2), are of first order. Inclusion of second-order reactions would introduce both nonlinear terms and second-order differential operators in the PDE systems for the corresponding generating functions, which would severely increase the complexity of these systems, thus preventing us from obtaining explicit solutions.

The method presented in this article, and the results thus obtained, can be seen, first and foremost, as the natural extension of previous work by Popović et al. (2016). Analytical results for the classes of models studied here can be found in several earlier articles. We mention the article by Shahrezaei and Swain (2008a), where a leading-order approximation was obtained in a long-evolution-time and steady-state limit. Bokes et al. (2012a) derived analytical expressions for stationary distributions in a model that is equivalent to that considered by Popović et al. (2016). Also for that model, a time-scale separation was exploited by Bokes et al. (2012b), in a manner that is similar to the present article, to obtain leading-order analytical expressions on both time scales. The model that is referred to as Model A in Sect. 2 was analysed in a steady-state setting by Iyer-Biswas and Jayaprakash (2014) via the Poisson integral transform. A similar model was studied by Hornos et al. (2005), were a generating function approach was used; making a steady-state Ansatz, the authors were able to obtain an explicit solution for the generating function in terms of special (Kummer) functions; see also NIST Digital Library of Mathematical Functions . The same model was later solved in a fully time-dependent context by Ramos et al. (2011), after a cleverly chosen variable substitution, in terms of another class of special (Heun) functions; cf. again NIST Digital Library of Mathematical Functions .

Other authors have attempted to solve several classes of CMEs directly, i.e. without resorting to generating function techniques or integral transforms. A noteworthy example is the work of Jahnke and Huisinga (2007) on monomolecular systems. Another, more recent example can be found in the work by Iserles and MacNamara (2017), where exact solutions are determined for explicitly time-dependent isomerisation models.

It is important to emphasise that the ‘time dependence’ referred to in the title of the present article is solely due to the dynamic nature of the underlying stochastic process, and that it hence manifests exclusively through time derivatives in the associated CMEs, such as e.g. in (2.3). In particular, none of the model parameters are time-dependent, as opposed to, for example, the system studied by Iserles and MacNamara (2017). The inclusion of such explicitly time-dependent parameters would be a starting point for incorporating the influence of (extrinsic) noise in the context of the model categories considered in the article.

The availability of analytical expressions for generating functions does, in principle, allow one to try to obtain insight into the underlying processes by studying the explicit form of said expressions, as has been done e.g. by Bokes et al. (Bokes et al. 2012a, b). However, the complex nature of the processes we analyse here seems to preclude such insights. For example, the integrals over confluent hypergeometric functions, which appear in (2.29), cannot themselves be efficiently expressed in terms of (other) special functions. Still, that complication does not necessarily pose an obstacle to the application we ultimately have in mind, i.e. to Bayesian parameter inference. As the last step in our method—the extraction of propagators from generating functions, see Sect. 4—is numerical, the precise functional form of the generating function is not of importance. The mere fact that an explicit expression can be obtained is sufficient for the application of the Cauchy integral formula, where these generating functions enter into the calculation of the appropriate integrals; see again Sect. 4.

The analytical approach explored in the article does not, of course, represent the only feasible way of obtaining numerical values for propagator probabilities, which can, in turn, serve as input for a Bayesian parameter inference scheme. For an example of a direct numerical method in which the Cauchy integral plays a central role, the reader is referred to the work by MacNamara (2015). Our main motivation for pursuing an analytical alternative is reducing the need for potentially lengthy numerical simulations. An efficient implementation of the resulting expressions can result in (significantly) reduced computation times; see, for example, the work by Bornemann (2011). The optimisation of the underlying numerical procedures is, however, beyond the scope of the present article in particular, and of our research programme in general.

The analytical results obtained thus far, as presented in the article, are ready for implementation in a Bayesian parameter inference framework. An analysis of the performance of the resulting approximations to the associated generating functions in the spirit of the article by Feigelman et al. (2015), where parameter inference is tested on simulated data based on specific stochastic models, is ongoing work. Moreover, the successful application of our analytical method to specific model categories, such as are represented by model A and model B, suggests several feasible expansions of the ‘model library’ for which explicit expressions for the corresponding generating functions can be constructed. Thus, stochastic models comprised of multiple proteins represent a natural next stage, bringing the analysis of toggle switch-type models within reach. In addition, one could begin exploring the vast field of gene regulatory networks by considering a simple two-protein system with, for example, activator-inhibitor interaction. Under the assumption of small interaction rates, the resulting PDE system for the associated generating function would be directly amenable to the analytical method described in the article. The analysis of these and similar systems could be a topic for future research.

Acknowledgements

The authors thank Peter Swain and Adri Olde Daalhuis (both University of Edinburgh) for their valuable input during discussions at earlier stages of the article. Moreover, the authors are grateful to three anonymous reviewers whose comments and suggestions improved the original manuscript. This work has been supported by the Leverhulme Trust, through Research Project Grant RPG-2015-017 (‘A geometric analysis of multiple-scale models for stochastic gene expression’).

Model A: explicit expressions

An explicit expression for (2.27) can be stated as follows:

F0(1)(w)=1κbwddwF0(0)+κfF0(0)=c1κfκb1F1(κf,1+κf+κb,w)+w1+κf+κb1F1(1+κf,2+κf+κb,w)-c2w-κf-κb1F1(-κb,1-κf-κb,w)+w1-κf-κb1F1(1-κb,2-κf-κb,w), A.1

which can be shown to be equal to (2.35) using §13.3(i) of NIST Digital Library of Mathematical Functions when c^1=c1κfκb and c^2=c2. Explicit expressions involving the integration limits c3 and c4 in (2.29), respectively, are obtained as

c3λv01F1(-κb,1-κf-κb,w^)g(w^)dw^=e-λv0{χ1(1+v0)n0λv0κb1-κf-κb1F1(1-κb,2-κf-κb,λv0)-1F1(-κb,1-κf-κb,λv0)α(w+λ)w+v0wv0F0(0)w=λv0} A.2

and

c4λv01F1(κf,1+κf+κb,w^)g(w^)w^-κf-κbdw^=-e-λv0(λv0)κf+κb{χ1(1+v0)n0(κf+κb)1F1(κf,1+κf+κb,λv0)+χ1(1+v0)n0λv0κf1+κf+κb1F1(1+κf,2+κf+κb,λv0)+1F1(κf,1+κf+κb,λv0)α(w+λ)w+v0wv0F0(0)w=λv0}, A.3

respectively, with F0(0) as in (2.26). Similarly, explicit expression that involve the integration limits c^3 and c^4 in (2.38), respectively, read

c^3λv01F1(1-κb,1-κf-κb,w^)h(w^)dw^=-e-λv0{χ1(1+v0)n0λv01F1(1-κb,1-κf-κb,λv0)-χ1(1+v0)n0λv0(1-κb)1-κf-κb1F1(2-κb,2-κf-κb,λv0)+1F1(1-κb,1-κf-κb,λv0)ρ(w+λ)w+v0wv0F0(1)w=λv0} A.4

and

c^4λv01F1(κf,1+κf+κb,w^)h(w^)w^-κf-κbdw^=e-λv0(λv0)κf+κb{χ1(1+v0)n0(κf+κb-λv0)1F1(1+κf,1+κf+κb,λv0)+χ1(1+v0)n0λv0(1+κf)1+κf+κb1F1(2+κf,2+κf+κb,λv0)+1F1(1+κf,1+κf+κb,λv0)ρ(w+λ)w+v0wv0F0(1)w=λv0}, A.5

respectively, with F0(1) as in (2.35).

Model B with autoregulation: explicit expressions

To leading order in δ, i.e. for m=0, the right-hand side of Eq. (3.45) vanishes; hence, the expressions in Eqs. (3.20) and (3.21) apply. In other words, we can identify F0,n(j) with Fn(j) as given in (3.27) and (3.28), for j=0,1 and n=0,1,2. The free constants fi and gi are as stated in (3.33), with χ replaced with χ0. In the following sections, we present explicit expressions for the first-order correction in δ—i.e. for m=1—for all autoregulatory scenarios listed in (3.2).

mRNA autoactivation

We can solve (3.45) to second order in ε, i.e. for n=2, applying the initial conditions in (3.55) to obtain

F1,0(0)=h0,M, B.1
F1,1(0)=h1,M+j1,Mσ+αMχ0m0(1+u0)m0-1(1+v0)n0e-(1-μv0)σ(1-μv0)2,and B.2
F1,2(0)=h2,M+j2,Mσ B.3
+αMm0(1+u0)m0-1(1+v0)n0κ1e-(1-μv0)σ1-μv011-μv0+σ-σ22(1-μv0)u0-μv01-μv0+αMm0(1+u0)m0-1(1+v0)n0χ0{e-(1-μv0)σ1-μv03μv0(1-μv0)2-σκ0+κ1-2μv01-μv0+σ22μv0+σ2(κ0+κ1)(1-μv0)u0-μv01-μv0}-(1+u0)m0(1+v0)n0χ1λκ1e(1-μv0)σ(1-μv0)2u0-μv01-μv0+σ22λκ1μv01-μv0-(κ0+κ1)2, B.4

which yields

F1,0(1)=1κ1j1,M+κ0h0,M+αMm0χ0(1+u0)m0-1(1+v0)n0u0-μv01-μv0and B.5
F1,1(1)=j2,M+κ0(h1,M+j1,Mσ)κ1-χ1κ1(1+u0)m0(1+v0)n0λκ1e(1-μv0)σ1-μv0u0-μv01-μv0+σλκ1μv01-μv0-(κ0+κ1)2-αMχ0κ1m0(1+u0)m0-1(1+v0)n0μv0(1-μv0)3+κ1e-(1-μv0)σ(1-μv0)2-σ(κ0+κ1)u0-μv01-μv0, B.6

where

h0,M=χ1(1+u0)m0(1+v0)n0, B.7
h1,M=-αMχ0m0(1+u0)m0-1(1+v0)n01(1-μv0)2, B.8
j1,M=-χ1(κ0+κ1)(1+u0)m0(1+v0)n0+αMχ0m0(1+u0)m0-1(1+v0)n0u0-μv01-μv0, B.9
h2,M=χ1(1+u0)m0(1+v0)n0λκ1(1-μv0)2u0-μv01-μv0-αMm0(1+u0)m0-1(1+v0)n01(1-μv0)3κ1+3χ0μv01-μv0,and B.10
j2,M=χ1(1+u0)m0(1+v0)n0λκ11-μv0u0-μv01-μv0+αMχ0m0(1+u0)m0-1(1+v0)n01(1-μv0)2κ0+κ1+μv01-μv0. B.11

mRNA autorepression

For the case of mRNA autorepression, we replace αM(u+1)uu0-1u0Fm-1(0) in Eqs. (3.45) and (3.46) with

-ρM(u+1)uu0-1u0Fm-1(1), B.12

which yields

F1,0(0)=k0,M, B.13
F1,1(0)=k1,M+l1,Mσ-ρM(1-χ0)m0(1+u0)m0-1(1+v0)n0e-(1-μv0)σ(1-μv0)2,and B.14
F1,2(0)=k2,M+l2,Mσ-ρMm0(1+u0)m0-1(1+v0)n0κ0e-(1-μv0)σ(1-μv0)211-μv0+σ-σ22u0-μv01-μv0+ρMm0(1+u0)m0-1(1+v0)n0(1-χ0){λe(1-μv0)σ(1-μv0)2u0-μv01-μv02+λe-(1-μv0)σ(1-μv0)3u0-μv01-μv0-(λ+3)μv0e-(1-μv0)σ(1-μv0)4-σe-(1-μv0)σ(1-μv0)2(λ+2)μv01-μv0-(κ0+κ1)-σ22μv0e-(1-μv0)σ(1-μv0)2+u0-μv01-μv02(κ0+κ1)-λμv01-μv0}-(1+u0)m0(1+v0)n0χ1λκ1e(1-μv0)σ(1-μv0)2u0-μv01-μv0+σ22λκ1μv01-μv0-(κ0+κ1)2+ρM(1+u0)m0(1+v0)n0(1-χ0)λ(1-μv0)2×e(1-μv0)σu0-μv01-μv0+e-(1-μv0)σ1-μv0. B.15

Hence, it follows that

F1,0(1)=1κ1l1M+κ0k0,M-ρMm0(1-χ0)(1+u0)m0-1(1+v0)n0u0-μv01-μv0and B.16
F1,1(1)=l2,M+κ0(k1,M+l1,Mσ)κ1-χ1κ1(1+u0)m0(1+v0)n0λκ1e(1-μv0)σ1-μv0u0-μv01-μv0+σλκ1μv01-μv0-(κ0+κ1)2+ρM(1+u0)m0(1+v0)n0(1-χ0)λκ1(1-μv0)×u0-μv01-μv0-11-μv0+ρM(1-χ0)m0(1+u0)m0-1(1+v0)n0[e-(1-μv0)σ(1-μv0)2-λκ1(1-μv0)2u0-μv01-μv0+μv0κ1(1-μv0)3+λκ1(1-μv0)u0-μv01-μv02-σκ1(κ0+κ1)u0-μv01-μv0], B.17

where

k0,M=χ1(1+u0)m0(1+v0)n0, B.18
k1,M=ρM(1-χ0)m0(1+u0)m0-1(1+v0)n01(1-μv0)2, B.19
l1,M=-χ1(κ0+κ1)(1+u0)m0(1+v0)n0-ρM(1-χ0)m0(1+u0)m0-1(1+v0)n0u0-μv01-μv0, B.20
k2,M=χ1(1+u0)m0(1+v0)n0λκ1(1-μv0)2u0-μv01-μv0-ρM(1+u0)m0+1(1+v0)n0(1-χ0)λ(1-μv0)2+ρM(1+u0)m0-1(1+v0)n0m0κ0(1-μv0)3+ρMm0(1+u0)m0-1(1+v0)n0(1-χ0)1(1-μv0)2×[(λ+3)μv0(1-μv0)2+λ1-μv0u0-μv01-μv0-λu0-μv01-μv02],and B.21
l2,M=χ1(1+u0)m0(1+v0)n0λκ11-μv0u0-μv01-μv0-ρM(1+u0)m0(1+v0)n0(1-χ0)λ1-μv0u0-μv01-μv0-11-μv0-ρM(1+u0)m0-1(1+v0)n0m0(1-χ0)11-μv0×[κ0+κ1-11-μv0+1(1-μv0)2-λ1-μv0u0-μv01-μv0+λu0-μv01-μv02]. B.22

Protein autoactivation

For the case of protein autoactivation, we replace αM(u+1)uu0-1u0Fm-1(0) in Eqs. (3.45) and (3.46) with

αP(v+1)vv0-1-uv0uu0-1u0+v0Fm-1(0), B.23

which yields

F1,0(0)=h0,P, B.24
F1,1(0)=h1,P+j1,Pσ-αPχ0m0μ(1+u0)m0-1(1+v0)n0+1×[e-(1-μv0)σ(1-μv0)3+σ22u0-μv01-μv0],and B.25
F1,2(0)=h2,P+j2,Pσ+αPn0(1+u0)m0(1+v0)n0-1σ22[2(κ0+κ1)χ0-κ1](1+v0)+χ0-αPm0(1+u0)m0-1(1+v0)n0μe-(1-μv0)σ(1-μv0)4×κ1(1+v0)+χ01+2v0+μv0(5+4v0)1-μv0+αPm0(1+u0)m0-1(1+v0)n0σμe-(1-μv0)σ(1-μv0)3×-χ0[1+(3+2v0)μ]v01-μv0+[(1-χ0)κ1-χ0κ0](1+v0)+αPm0(1+u0)m0-1(1+v0)n0μσ221(1-μv0)2×-χ0(1+v0)μv0e-(1-μv0)σ1-μv0+(1+v0)[2(κ0+κ1)χ0-κ1]+χ01+μv021-μv0+αPm0(1+u0)m0-1(1+v0)n0μσ36u0-μv01-μv0(1+v0)×[3χ0(κ0+κ1)-2κ1]+χ0(1-v0)-(1+u0)m0(1+v0)n0χ1λκ1e(1-μv0)σ(1-μv0)2u0-μv01-μv0+σ22λκ1μv01-μv0-(κ0+κ1)2. B.26

Hence, it follows that

F1,0(1)=1κ1j1,P+κ0h0,P+αPχ0κ1(1+u0)m0-1(1+v0)n0m0μ(1+v0)(1-μv0)2+n0u0-μv01-μv0+11-μv0and B.27
F1,1(1)=j2,P+κ0(h1,P+j1,Pσ)κ1 B.28
-χ1κ1(1+u0)m0(1+v0)n0λκ1e(1-μv0)σ1-μv0u0-μv01-μv0+σλκ1μv01-μv0-(κ0+κ1)2+αPn0(1+u0)m0(1+v0)n0σχ0κ0+κ1κ1+αPm0(1+u0)m0-1(1+v0)n0+1μχ0κ1κ1e-(1-μv0)σ(1-μv0)3+1+2μv0(1-μv0)4+σκ0+κ1(1-μv0)2+κ1σ22u0-μv01-μv0, B.29

where

h0,P=χ1(1+u0)m0(1+v0)n0, B.30
h1,P=αPχ0m0μ(1+u0)m0-1(1+v0)n0+11(1-μv0)3, B.31
j1,P=-χ1(κ0+κ1)(1+u0)m0(1+v0)n0-αPχ0(1+u0)m0(1+v0)n0n0+μm01+v01+u01(1-μv0)2, B.32
h2,P=χ1(1+u0)m0(1+v0)n0λκ1(1-μv0)2u0-μv01-μv0+αPm0(1+u0)m0-1(1+v0)n0μ(1-μv0)4×(1+v0)κ1+χ01+2v0+μv0(5+4v0)1-μv0,and B.33
j2,P=χ1(1+u0)m0(1+v0)n0λκ11-μv0u0-μv01-μv0+αPχ0μm0(1+u0)m0-1(1+v0)n0+11(1-μv0)3×2-(κ0+κ1)-31-μv0. B.34

Protein autorepression

For the case of protein autorepression, we replace αM(u+1)uu0-1u0Fm-1(0) in Eqs. (3.45) and (3.46) with

-ρP(v+1)vv0-1-uv0uu0-1u0+v0Fm-1(1), B.35

which yields

F1,0(0)=k0,P, B.36
F1,1(0)=k1,P+l1,Pσ+ρP(1-χ0)m0μ(1+u0)m0-1(1+v0)n0+1×e-(1-μv0)σ(1-μv0)3+σ22u0-μv01-μv0,and B.37
F1,2(0)=k2,P+l2,Pσ+ρP(1-χ0)λμ(1+u0)m0(1+v0)n0+11(1-μv0)3×e(1-μv0)σu0-μv01-μv0-e-(1-μv0)σ1-μv0+ρP(1-χ0)λn0(1+u0)m0(1+v0)n0e(1-μv0)σ(1-μv0)u0-μv01-μv0+ρPm0(1+u0)m0-1(1+v0)n0+11(1-μv0)3κ0μe-(1-μv0)σ1-μv0+ρP(1-χ0)m0(1+u0)m0-1(1+v0)n0+1λμe(1-μv0)σ(1-μv0)3u0-μv01-μv0×11-μv0-u0-μv01-μv0+ρPm0(1+u0)m0-1(1+v0)n0+1μe-(1-μv0)σ(1-μv0)4×-λu0-μv01-μv0+61-μv0+λμv01-μv0-5+4v01+v0+ρPσλμm0(1-χ0)(1+u0)m0-1(1+v0)n0+1e(1-μv0)σ(1-μv0)2×u0-μv01-μv02+ρPσμm0(1+u0)m0-1(1+v0)n0e-(1-μv0)σ(1-μv0)3[χ0(1+v0)(κ0+κ1)-(1+v0)κ1+(1-χ0)v0+μv0(3+2v0)+μv0λ(1+v0)1-μv0]+ρPσ22(1+u0)m0(1+v0)n0+1(1-χ0)λμ1-μv0×11-μv0-u0-μv01-μv0+ρPσ22n0(1+u0)m0(1+v0)n0-(1-χ0)11+v0+λμv01-μv0-(1-2χ0)(κ0+κ1)-κ1+ρPσ22μm0(1+u0)m0-1(1+v0)n0+1×(1-χ0)μv0e-(1-μv0)σ(1-μv0)3+ρPσ22μm0(1+u0)m0-1(1+v0)n0{[-(1-2χ0)(κ0+κ1)-κ1]
1+v0(1-μv0)2-(1-χ0)1+μv02(1-μv0)3+(1-χ0)(1+v0)×μv0(1-μv0)3+11-μv0u0-μv01-μv02}+ρPσ36μm0(1+u0)m0-1(1+v0)n0u0-μv01-μv0×[(-(1-3χ0)(κ0+κ1)-2κ1)(1+v0)+(1-χ0)v0-1+2λμv0(1+v0)1-μv0]-(1+u0)m0(1+v0)n0χ1λκ1e(1-μv0)σ(1-μv0)2u0-μv01-μv0×+σ22λκ1μv01-μv0-(κ0+κ1)2. B.38

It follows that

F1,0(1)=1κ1l1,P+κ0k0,P-ρPm0(1-χ0)(1+u0)m0-1(1+v0)n01κ1×m0μ(1+v0)(1-μv0)2+n011-μv0+u0-μv01-μv0and B.39
F1,1(1)=l2,M+κ0(k1,M+l1,Mσ)κ1+ρP(1-χ0)(1+u0)m0+1(1+v0)n0+1λμκ11(1-μv0)2+ρPn0(1-χ0)(1+u0)m0(1+v0)n01κ1×λ1-μv0u0-μv01-μv0-(κ0+κ1)σ+ρPμm0(1-χ0)(1+u0)m0-1(1+v0)n0+11κ1×{-3(1-μv0)4+1(1-μv0)3-κ1e-(1-μv0)σ+2+2λu0-μv01-μv0-σκ0+κ1(1-μv0)2-σ22κ1u0-μv01-μv0}-χ1κ1(1+u0)m0(1+v0)n0×λκ1e(1-μv0)σ1-μv0u0-μv01-μv0+σλκ1μv01-μv0-(κ0+κ1)2, B.40

where

k0,P=χ1(1+u0)m0(1+v0)n0, B.41
k1,P=-ρP(1-χ0)m0(1+u0)m0-1(1+v0)n0+1μ(1-μv0)3, B.42
l1,P=-χ1(κ0+κ1)(1+u0)m0(1+v0)n0-ρP(1-χ0)(1+u0)m0(1+v0)n0n0+μm01+v01+u01(1-μv0)2, B.43
k2,P=χ1(1+u0)m0(1+v0)n0λκ1(1-μv0)2u0-μv01-μv0+ρP(1-χ0)(1+u0)m0(1+v0)n0+1λμ(1-μv0)3×11-μv0-u0-μv01-μv0-ρP(1-χ0)n0(1+u0)m0(1+v0)n0λ(1-μv0)2u0-μv01-μv0+ρP(1-χ0)m0μ(1+u0)m0-1(1+v0)n0+1λ(1-μv0)3×u0-μv01-μv02-μv0(1-μv0)2-ρPm0(1+u0)m0-1(1+v0)n0μ(1-μv0)4×1-χ01-μv0[1+2v0+μv0(5+4v0)]+(1+v0)κ0,and B.44
l2,P=χ1(1+u0)m0(1+v0)n0λκ11-μv0u0-μv01-μv0-ρP(1+u0)m0+1(1+v0)n0+1(1-χ0)λμ(1-μv0)2-ρPn0(1+u0)m0(1+v0)n0(1-χ0)λ1-μv0u0-μv01-μv0+ρPμm0(1+u0)m0-1(1+v0)n0+1(1-χ0)1(1-μv0)3×-2-2λu0-μv01-μv0+31-μv0+κ0+κ1. B.45

Footnotes

1

In some specific cases, however, such explicit expressions can be found; see e.g. Popović et al. (2016).

References

  1. Bokes P, et al. Exact and approximate distributions of protein and mRNA levels in the low-copy regime of gene expression. J Math Biol. 2012;64(5):829–854. doi: 10.1007/s00285-011-0433-5. [DOI] [PubMed] [Google Scholar]
  2. Bokes P, et al. Multiscale stochastic modelling of gene expression. J Math Biol. 2012;65(3):493–520. doi: 10.1007/s00285-011-0468-7. [DOI] [PubMed] [Google Scholar]
  3. Bornemann F. Accuracy and stability of computing high-order derivatives of analytic functions by Cauchy integrals. Found Comput Math. 2011;11(1):1–63. doi: 10.1007/s10208-010-9075-z. [DOI] [Google Scholar]
  4. Bressloff PC (2014) Stochastic processes in cell biology, vol 41. In: Interdisciplinary applied mathematics, Chap 6. Stochastic gene expression and regulatory networks. Springer, pp 269–340. ISBN:978-3-319-08488-6. 10.1007/978-3-319-08488-6_6 [DOI]
  5. Courant R, Hilbert D (1962) Methods of mathematical physics, vol II: partial differential equations. Wiley. ISBN:978-0-471-50439-9. 10.1002/9783527617234 [DOI]
  6. Coutu DL, Schroeder T. Probing cellular processes by long-term live imaging—historic problems and current solutions. J Cell Sci. 2013;126:3805–3815. doi: 10.1242/jcs.118349. [DOI] [PubMed] [Google Scholar]
  7. Elowitz MB, et al. Stochastic gene expression in a single cell. Science. 2002;297(5584):1183–1186. doi: 10.1126/science.1070919. [DOI] [PubMed] [Google Scholar]
  8. Feigelman J, Popović N, Marr C. A case study on the use of scale separation-based analytical propagators for parameter inference in models of stochastic gene regulation. J Coupled Syst Multiscale Dyn. 2015;3(2):164–173. doi: 10.1166/jcsmd.2015.1074. [DOI] [Google Scholar]
  9. Feigelman J, et al. Analysis of cell lineage trees by exact Bayesian inference identifies negative autoregulation of Nanog in mouse embryonic stem cells. Cell Syst. 2016;3(5):480–490. doi: 10.1016/j.cels.2016.11.001. [DOI] [PubMed] [Google Scholar]
  10. Gardiner C (2009) Stochastic methods: a handbook for the natural and social sciences, 4th edn, vol 13. Springer series in synergetics. Springer. ISBN:978-3-540-70712-7. http://www.springer.com/978-3-540-70712-7
  11. Gillespie DT. Exact stochastic simulation of coupled chemical reactions. J Phys Chem. 1977;81(25):2340–2361. doi: 10.1021/j100540a008. [DOI] [Google Scholar]
  12. Hornos JEM, et al. Self-regulating gene: an exact solution. Phys Rev E Stat Nonlinear Biol Soft Matter Phys. 2005 doi: 10.1103/PhysRevE.72.051907. [DOI] [PubMed] [Google Scholar]
  13. Iserles A, MacNamara S (2017) Magnus expansions and pseudospectra of master equations. Preprint. arxiv:1701.02522
  14. Iyer-Biswas S, Jayaprakash C. Mixed Poisson distributions in exact solutions of stochastic autoregulation models. Phys Rev E Stat Nonlinear Biol Soft Matter Phys. 2014 doi: 10.1103/PhysRevE.90.052712. [DOI] [PubMed] [Google Scholar]
  15. Iyer-Biswas S, Hayot F, Jayaprakash C. Stochasticity of gene products from transcriptional pulsing. Phys Rev E Stat Nonlinear Biol Soft Matter Phys. 2009 doi: 10.1103/PhysRevE.79.031911. [DOI] [PubMed] [Google Scholar]
  16. Jahnke T, Huisinga W. Solving the chemical master equation for monomolecular reaction systems analytically. J Math Biol. 2007;54(1):1–26. doi: 10.1007/s00285-006-0034-x. [DOI] [PubMed] [Google Scholar]
  17. Karlebach G, Shamir R. Modelling and analysis of gene regulatory networks. Nat Rev Mol Cell Biol. 2008;9:770–780. doi: 10.1038/nrm2503. [DOI] [PubMed] [Google Scholar]
  18. Kuehn C (2015) Multiple time scale dynamics, vol 191. Applied mathematical sciences. Springer. ISBN:978-3-319-12315-8. 10.1007/978-3-319-12316-5 [DOI]
  19. Larson DR, et al. Real-time observation of transcription initiation and elongation on an endogenous yeast gene. Science. 2011;332(6028):475–478. doi: 10.1126/science.1202142. [DOI] [PMC free article] [PubMed] [Google Scholar]
  20. MacNamara S (2015) Cauchy integrals for computational solutions of master equations. In: Sharples J, Bunder J (eds) Proceedings of the 17th biennial computational techniques and applications conference, CTAC-2014, vol 56. ANZIAM Journal, pp C32–C51. http://journal.austms.org.au/ojs/index.php/ANZIAMJ/article/view/9345
  21. Muzzey D, van Oudenaarden A. Quantitative time-lapse fluorescence microscopy in single cells. Annu Rev Cell Dev Biol. 2009;25:301–327. doi: 10.1146/annurev.cellbio.042308.113408. [DOI] [PMC free article] [PubMed] [Google Scholar]
  22. NIST Digital Library of Mathematical Functions. http://dlmf.nist.gov/, Release 1.0.14 of 2016-12-21. Olver FWJ, Olde Daalhuis AB, Lozier DW, Schneider BI, Boisvert RF, Clark CW, Miller BR, Saunders BV (eds). http://dlmf.nist.gov/
  23. Popović N, Marr C, Swain PS. A geometric analysis of fast-slow models for stochastic gene expression. J Math Biol. 2016;72(1):87–122. doi: 10.1007/s00285-015-0876-1. [DOI] [PubMed] [Google Scholar]
  24. Raj A, van Oudenaarden A. Nature, nurture or chance: stochastic gene expression and its consequences. Cell. 2008;135(2):216–226. doi: 10.1016/j.cell.2008.09.050. [DOI] [PMC free article] [PubMed] [Google Scholar]
  25. Raj A, et al. Stochastic mRNA synthesis in mammalian cells. PLoS Biol. 2006;4(10):1707–1719. doi: 10.1371/journal.pbio.0040309. [DOI] [PMC free article] [PubMed] [Google Scholar]
  26. Ramos AF, Innocentini GCP, Hornos JEM. Exact time-dependent solutions for a self-regulating gene. Phys Rev E Stat Nonlinear Biol Soft Matter Phys. 2011 doi: 10.1103/PhysRevE.83.062902. [DOI] [PubMed] [Google Scholar]
  27. Schwanhäusser B, et al. Global quantification of mammalian gene expression control. Nature. 2011;473:337–342. doi: 10.1038/nature10098. [DOI] [PubMed] [Google Scholar]
  28. Shahrezaei V, Swain PS. Analytical distributions for stochastic gene expression. Proc Nat Acad Sci USA. 2008;105(45):17256–17261. doi: 10.1073/pnas.0803850105. [DOI] [PMC free article] [PubMed] [Google Scholar]
  29. Shahrezaei V, Swain PS. The stochastic nature of biochemical networks. Curr Opin Biotechnol. 2008;19(4):369–374. doi: 10.1016/j.copbio.2008.06.011. [DOI] [PubMed] [Google Scholar]
  30. Taylor M (2011) Partial differential equations I: basic theory. 2nd edn, vol 115. Applied mathematical sciences. Springer. ISBN:978-1-4419-7054-1. 10.1007/978-1-4419-7055-8 [DOI]
  31. Thattai M, van Oudenaarden A. Intrinsic noise in gene regulatory networks. Proc Nat Acad Sci USA. 2001;98(15):8614–8619. doi: 10.1073/pnas.151588598. [DOI] [PMC free article] [PubMed] [Google Scholar]
  32. Verhulst F (2000) Nonlinear differential equations and dynamical systems. Universitext, 2nd edn. Springer, Berlin. ISBN:978-3-540-60934-6. 10.1007/978-3-642-61453-8
  33. Vogel C, Marcotte EM. Insights into the regulation of protein abundance from proteomic and transcriptomic analyses. Nat Rev Genet. 2012;13:227–232. doi: 10.1038/nrg3185. [DOI] [PMC free article] [PubMed] [Google Scholar]
  34. Wilkinson DJ. Stochastical modelling for quantitative description of heterogenous biological systems. Nat Rev Genet. 2009;10:122–133. doi: 10.1038/nrg2509. [DOI] [PubMed] [Google Scholar]
  35. Young JW, et al. Measuring single-cell gene expression dynamics in bacteria using fluorescence time-lapse microscopy. Nat Protocols. 2011;7:80–88. doi: 10.1038/nprot.2011.432. [DOI] [PMC free article] [PubMed] [Google Scholar]
  36. Yu J, et al. Probing gene expression in live cells, one protein molecule at a time. Science. 2006;311(5767):1600–1603. doi: 10.1126/science.1119623. [DOI] [PubMed] [Google Scholar]

Articles from Journal of Mathematical Biology are provided here courtesy of Springer

RESOURCES