Skip to main content
PLOS Computational Biology logoLink to PLOS Computational Biology
. 2022 Nov 28;18(11):e1010734. doi: 10.1371/journal.pcbi.1010734

Efficient inference and identifiability analysis for differential equation models with random parameters

Alexander P Browning 1,2,3, Christopher Drovandi 1,2, Ian W Turner 1, Adrianne L Jenner 1,2, Matthew J Simpson 1,2,*
Editor: Ricardo Martinez-Garcia4
PMCID: PMC9731444  PMID: 36441811

Abstract

Heterogeneity is a dominant factor in the behaviour of many biological processes. Despite this, it is common for mathematical and statistical analyses to ignore biological heterogeneity as a source of variability in experimental data. Therefore, methods for exploring the identifiability of models that explicitly incorporate heterogeneity through variability in model parameters are relatively underdeveloped. We develop a new likelihood-based framework, based on moment matching, for inference and identifiability analysis of differential equation models that capture biological heterogeneity through parameters that vary according to probability distributions. As our novel method is based on an approximate likelihood function, it is highly flexible; we demonstrate identifiability analysis using both a frequentist approach based on profile likelihood, and a Bayesian approach based on Markov-chain Monte Carlo. Through three case studies, we demonstrate our method by providing a didactic guide to inference and identifiability analysis of hyperparameters that relate to the statistical moments of model parameters from independent observed data. Our approach has a computational cost comparable to analysis of models that neglect heterogeneity, a significant improvement over many existing alternatives. We demonstrate how analysis of random parameter models can aid better understanding of the sources of heterogeneity from biological data.

Author summary

Heterogeneity is a dominant factor in the behaviour of many biological and biophysical processes, and is often a primary source of the variability evident in experimental data. Despite this, it is relatively rare for mathematical models of biological systems to incorporate variability in model parameters as a source of noise. Therefore, methods for analysing whether model parameters and sources of variability are identifiable from commonly reported experimental data are relatively underdeveloped. As we demonstrate, such identifiability analysis is vital for model selection, experimental design, and gaining biological insights. In this work, we develop a fast, approximate framework for model calibration and identifiability analysis of mathematical models that incorporate biological heterogeneity through random parameters. Our method is highly flexible, and can be employed in both frequentist and Bayesian inference paradigms. Compared to alternative approaches, our approach is computationally efficient, with a computational cost comparable to analysis of standard models that neglect parameter variability.


This is a PLOS Computational Biology Methods paper.

Introduction

Heterogeneity is understood as a dominant factor in the behaviour of many biological and biophysical systems [13]. Mathematical analysis of these systems is often constrained to parameter-fitting of differential equation based models. In many cases, heterogeneity is neglected, with variability in the data assumed to be noise and incorporated through independent, probabilistic observation processes [48].

Mathematical models have long been an essential tool for understanding the behaviour of systems from quantitative and experimental data. Parameter estimation allows practitioners to quantify observed behaviour in terms of parameters that carry physical interpretations. Establishing whether model parameters can be identified from the quantity and quality of experimental data available is critical for tailoring model and data complexity to the scientific questions of interest [7, 912]. Furthermore, predictions from non-identifiable models can be unreliable [9]. Such identifiability analysis is well established for ordinary differential equation models [9, 10, 13], stochastic models [12, 1416], and, recently, partial differential equation models [8, 17]. There is, however, comparatively little guidance for identifiability analysis for models that explicitly incorporate heterogeneity in model parameters, limiting the ability of practitioners to identify and predict sources of biological variability.

In biological systems, heterogeneity might arise due to inter-experiment variability, gene expression [18], or patient-to-patient variability [19]. Even from tightly controlled experiments is data variability evident, potentially due to differences in cell behaviour between experiments. We demonstrate this in Fig 1A by showing results of an in vitro multicellular tumour spheroid experiment, using melanoma tumour spheroids generated from a single cell line and imaged using microscopy after seven days of growth [20]. Despite similarities in both size and morphology, spheroids are not identical. In Fig 1B, we summarise the experimental images with the most obvious measurement corresponding to the radius of a circle with the same cross-sectional area, and repeat this for ten spheroids collected from eight observation days (yielding 80 independent measurements). We also show predictions from a calibrated logistic model [21, 22], with a prediction interval capturing variability in the data through additive normal noise, a common assumption [9]. Key questions posed by the data in Fig 1 might relate to whether variability observed in the data is due to heterogeneity in the initial condition (spheroids are seeded with approximately 5000 cells), due to heterogeneity in the dynamic behaviour (differences in experimental conditions such as the concentration of nutrient might yield differences in the growth rate between spheroids), or due to extrinsic factors, including measurement noise.

Fig 1. Heterogeneity in experimental measurements of spheroid growth.

Fig 1

(A) Microscopy images of tumour spheroids grown from melanoma cell line WM983b from which a subset of measurements in (B) were taken. Spheroids were grown from 5000 cells, harvested, fixed, and imaged after nine days. Cells are transduced with fluorescent cell cycle indicators: colouring indicates cells in gap 1 (purple) and cells in gap 2 (green). In all but one spheroid, a lack of definition in the spheroid centre indicates the presence of a necrotic core. (B) Images are summarised by the radius of a circle with equivalent cross-sectional area as microscopy images. Shown are experimental measurements (blue discs and box plots) and a calibrated logistic model with prediction interval based on additive normal noise. Data from [20].

Differential equation models are widely used throughout biology, and have the potential to capture heterogeneity by relaxing the requirement that model parameters be fixed between measurements [23]. In the mathematical literature, such models are termed heterogeneous ordinary differential equations (ODEs) [24], random ODEs [25], or populations of models [26]. In the statistical literature, ODE-constrained hierarchical models, random-effects models, or non-linear mixed-effects models describe heterogeneity through a parameter hierarchy where model parameters have specified distributions that are themselves described by hyperparameters [23, 2730]. The literature is further split by a distinction between inference procedures that assume distributional forms [23, 2833] or those that do not [24, 34], when inferring the underlying parameter distributions from quantitative observations. For the non-linear models common to biology, inference for random parameter variants is often computationally costly—with cost that can increase significantly with the data sample size—and requires non-trivial selection of tuning parameters. Furthermore, there is very little guidance in the literature for assessing the identifiability, and hence ability of practitioners to determine the source of variability or even the benefit of considering parameter variability, using these classes of random parameter models.

Motivated by these observations, we develop a novel, approximate, and computationally efficient likelihood-based approach to inference and identifiability analysis of differential equation based models with random parameters based on an approximate moment-matched solution to the random parameter model [35]. To do this, we express the solution to the model about the parameter mean using a second order Taylor series expansion, which can be manipulated to obtain approximate expressions for the first three statistical moments of the model output distributions in terms of the statistical moments of the input (i.e., parameter) distributions. Our method may therefore be classified as a moment-matching method similar to methods routinely employed for stochastic fixed-parameters models such as the linear-noise approximation [3639]. We highlight that, although we assume a distributional form for the input parameter distributions, it is the statistical moments of the parameters that are inferred. Given the challenges of formulating high-dimensional distributions with possibly highly non-linear dependence structures, we focus our approach to low-dimensional data; for example, univariate or bivariate measurements collected independently at observation times, common in biology due to the challenges in collecting data and where samples are often destroyed for data to be collected [4042]. While not our focus, our method is applicable to data of any dimension (including, for example, time-series data) provided that dependent measurements are approximately multivariate normally distributed, or can be transformed to meet this requirement. The restriction is less strict for univariate and bivariate data, where we are also able to capture the skewness in the data. By leveraging techniques such as automatic differentiation to construct the Taylor series expansion, our approach provides a deterministic approximation to the data likelihood with comparable computational cost to the corresponding fixed-parameter model.

Our approach differs from many as we construct a surrogate likelihood directly from the approximate distributional solution to the random parameter model, alleviating the need to either infer individual-level parameters or marginalise the posterior in non-linear mixed-effects modelling [30] and Bayesian hierarchical modelling [29], respectively. The availability of a surrogate likelihood allows us to perform inference and identifiability analysis of random parameter models using the standard suite of tools, including profile likelihood [9], Fisher information [43], and Markov-chain Monte-Carlo [7]. Aside from assessing the identifiability of hyperparameters—parameters that relate to the distributional form of the random parameters in the dynamical model—we demonstrate our method by answering a number of questions specific to identifiability analysis of random parameter models. Namely, whether variation in model parameters can be distinguished from measurement noise; whether identifiability of a random-parameter model differs from that of a fixed-parameter model; and, finally, how the order of the moment-matching approximation affects parameter identifiability. To aid in the uptake of both random parameter models and their application to better interpret the variability ubiquitous to biological data, we provide a Julia module implementing our approach on Github.

Methods

We consider ODE state-space models of the form

dx(t,θ)dt=g(t,x(t,θ),θ),x(0,θ)=x0(θ), (1)

subject to the observation process

y(t,θ)=h(x(t,θ),θ), (2)

at times t{t1,t2,,tn}=T. Here, x(t,θ):R×RdRp is the state, g(t,x,θ):R×Rp×RdRp is the time-derivative of the state and θRd is a vector of parameters, traditionally assumed to be fixed between measurements [5, 21]. The observation process, y(t,θ):R×RdRq, represents potentially incomplete observations of the underlying state, characterised by h(x,θ):Rp×RdRq.

The focus of this work is random parameter dynamical models, that is, dynamical models where model parameters vary between observations such that θ is a random variable. In distinction to other classes of random or stochastic differential equations, we emphasise that θ does not depend on t. In this formulation, it is possible to incorporate a probabilistic observation process directly into Eq (2) (for example, normally distributed measurement error) through a sequence of random parameters εi contained in θ associated with each observation time ti. For instance, to model additive normal noise, we would set the ith component of the observation process to hi(x(ti,θ),θ)=h¯i(x(ti,θ),θ)+εi where εiN(0,σ2) captures noise and h¯(·) represents a noiseless observation from the model. We demonstrate both additive and multiplicative normal noise in this work, and highlight the flexibility gained by incorporating measurement error directly into the observation process through an additional random parameter.

Therefore, the model can be considered as a transformation of the random variable θ to randomly distributed observations, y. We denote a vector of dependent measurements

f(θ)=[f1(θ),,fn(θ)]. (3)

For time-series data, f may represent dependent observations taken from an entire dependent trace; for example, in the case of univariate observations at times t1, …, tn, we have y(t, θ) = y(t, θ) and f(θ) = [y(t1, θ), …, y(tn, θ)]. For multivariate observations, we concatenate these observations such that

f(θ)=[y(1)(t1,θ),,y(1)(tn,θ),,y(m)(t1,θ),,y(m)(tn,θ)], (4)

where y(k)(ti, θ) denotes a measurement from the kth component of y(ti, θ). For a tumour spheroid experiment, this might represent time-series radius measurements (for univariate observations) or radius and inner structure measurements (for multivariate observations) from a single spheroid throughout the course of the experiment. Alternatively, for so-called snapshot data, where observations are taken at each observation time independently, we consider a series of transformations that can be handled independently, f(1)(θ), f(2)(θ), … where f(i)(θ) = y(ti; θ). For a tumour spheroid experiment, f(i)(θ) might represent a radius measurement collected by terminating a single tumour spheroid experiment at time ti. The key difference between time-series and snapshot data is that in the case of the former, data from all time points are considered a single, dependent, multivariate measurement for which the covariance structure must be considered; whereas for the latter, data from each time point can be considered entirely independently, significantly reducing the dimensionality of the problem.

Approximate solution of the random parameter model

Only in very limited cases (specifically, where the inverse f−1 is tractable) can the density of f(θ) be calculated directly. Indeed, of primary interest in our work is the case where independent observations at a series of observation times are collected where it is highly likely that there are fewer observations than random model parameters, so it is likely that f−1 will not exist. Therefore, we build an approximate surrogate likelihood based on a Taylor series approximation to the moments of f(θ) given the moments of θ under the assumption that f is sufficiently smooth.

First, consider a univariate transformation, f(θ):RR of the random variable θR. To formulate expressions describing the moments of f(θ) (and hence, an approximate expression for the density of f(θ)), consider the Taylor expansion of f(θ) about θ=θ^,

f(θ)=f(θ^)+df(θ^)dθ(θ-θ^)+12d2f(θ^)dθ2(θ-θ^)2+. (5)

If we choose θ^=E(θ), then the expectation of Eq (5) yields an equation for E(f(θ)) to expressions relating to E(θ-θ^)=0 and E((θ-θ^)2)=V(θ) (the variance of θ). Similarly, each side of Eq (5) can be squared or raised to higher powers to obtain expressions relating to the variance and skewness of f(θ).

Eq (5) readily extends to transformations of multivariate random variables. For instance, consider now that θ=[θ1,θ2,,θd]Rd and that f(θ) = [f1(θ), …, fn(θ)]. An expression for the ith component, fi(θ), can be expressed in the following form using a Taylor expansion around θ=θ^,

fi(θ)=fi(θ^)+a=1dfi(θ^)θa(θa-θ^a)+12a=1db=1d2fi(θ^)θaθb(θa-θ^a)(θb-θ^b)+. (6)

While expectations of Eq (6) can still be taken, it is now more difficult to relate terms to the central moments of θ, particularly when Eq (6) is raised to higher powers. However, this task becomes clearer when Eq (6) is expressed in matrix or tensor notation: the terms relating to the second derivatives in Eq (6), for example, are related to the Frobenius inner product (i.e., sum of the component-wise product of two matrices or tensors) of the Hessian matrix and a matrix that becomes the covariance matrix when expectations are taken. This notation yields

fi(θ)=fi(θ^)+fi(θ^)·(θ-θ^)+12Hfi(θ^)M2(θ-θ^)+. (7)

Here, ∘ denotes the Frobenius inner product, Hfi(θ) is the Hessian matrix of fi(θ), i.e., a matrix with elements

[Hfi(θ^)]ab=2fi(θ^)θaθb, (8)

and M2 is an operator that returns the matrix formed by taking an outer product of a vector with itself, with elements given by

[M2(θ-θ^)]ab=(θa-θ^a)(θb-θ^b). (9)

Similarly, higher order operators are defined by M3, which returns a three-dimensional tensor, and M4 which returns a four-dimensional tensor. We form M2 using a generalisation of the Kronecker product, M2(θ-θ^)=(θ-θ^)(θ-θ^) and Mk(θ-θ^)=(θ-θ^)Mk-1(θ-θ^), where ⊗ is defined such that, for two vectors a and b,

ab=[a1ba2banb]. (10)

The operation is similarly defined for arguments in higher-dimensions, returning a tensor of dimensionality equal to the sum of the dimensions of both arguments. For brevity, we define a Kronecker power operator such that an refers to the operation performed on a by itself n times.

Defining ·=E(·) and noting that, in our notation, V(θ)=M2(θ-θ^) (and similar for higher order moments relating to the skewness tensor, S and kurtosis tensor, K), we can show that

fi(θ)fi(θ^)+V(θ)12Hfi(θ^), (11)
fi2(θ)fi2(θ^)+V(θ)(fi(θ^)2+fi(θ^)Hfi(θ^))+S(θ)(Hfi(θ^)fi(θ^))+K(θ)14Hfi(θ^)2, (12)
fi3(θ)fi3(θ^)+V(θ)32fi(θ^)(2fi(θ^)2+fi(θ^)Hfi(θ^))+S(θ)(fi(θ^)3+3fi(θ^)Hfi(θ^)fi(θ^))+K(θ)3(14fi(θ^)Hfi(θ^)2+12fi(θ^)2Hfi(θ^)). (13)

and

fi(θ)fj(θ)fi(θ^)fj(θ^)+V(θ)12fi(θ^)Hfj(θ^)+fj(θ^)Hfi(θ^)+2fi(θ^)fj(θ^)+S(θ)fi(θ^)Hfj(θ^)+fj(θ^)Hfi(θ^)+K(θ)14Hfi(θ^)Hfj(θ^). (14)

Note that we have applied the closure 〈Mk(θ)〉 = 0 for k ≥ 5. Formal derivations of Eq (7) and Eqs (11) to (14) are provided as supporting material (S1 File).

Eqs (11) to (14) provide approximate expressions for the mean vector with entries μi = 〈fi(θ)〉, covariance matrix with entries Σij = 〈fi(θ)fj(θ)〉 − 〈fi(θ)〉〈fj(θ)〉 and univariate skewnesses vector with entries ωi=(fi(θ)-μi)3/Σij3/2, of f(θ). From this, we construct an approximate density function for f(θ) using two approaches: one based on a multivariate normal distribution that matches the first two moments, and another based on a gamma distribution that matches the first three.

The normal or two moment approach approximates

f(θ)MvNormal(μ,Σ). (15)

The primary advantage of this approach is that we can form approximations without regard to the dimensionality of f(θ). Furthermore, it is overwhelmingly the case in the mathematical biology literature, for instance, that normality is assumed when calibrating dynamical models to experimental data [5, 8].

The gamma or three moment approach approximates marginal distributions with a shifted gamma distribution parameterised in terms of its mean, variance and skewness [32],

fi(θ)ShiftedGamma(μi,Σii,ωi). (16)

This approach is advantageous as it recaptures the normal approach in the limit ωi → 0, but allows more flexibility in terms of the shape of the distribution. The primary difficulty of the gamma approach is to construct an approximation to multivariate f(θ). We do this in the case that f(θ) is two-dimensional by correlating f1(θ) and f2(θ) using a Gaussian copula, a statistical object that correlates two random variables with arbitrary univariate distributions [44]. The correlation parameter in the copula, ρ˜, is chosen to match the approximate correlation calculated from the approximate moments. Denoting the skewnesses of the marginal distributions as ω1 and ω2, we compute the map (ω1,ω2,ρ)ρ˜ ahead of time using the rectangle rule for a wide range of skewness parameters |ωi| ∈ [0, 2]. The upper limit of 2 is chosen as the gamma distribution changes shape from non-monotonic (normal-like) to monotonic (exponential-like) at ω = 2.

We demonstrate these approximations in Fig 2 using the logistic model (Eq (24)). Additional comparisons, including a multivariate comparison, are provided in the supporting material (Figs A to C in S1 File). In all cases, the gamma approximation is clearly superior to the normal approximation, and provides a fast, accurate approximate approximation to the solution to the random parameter model.

Fig 2. Accuracy of approximate transformation.

Fig 2

We compare the accuracy of two approximate solutions to the non-linear transformation f(θ) (here, f(θ) is given by the solution to the logistic model and θ=[λ,K,r0], see Eq (24)). In (A) θ has a multivariate normal distribution with independent components; in (B) θ has a multivariate normal distribution with correlation between λ and K; and, in (C) θ has independent, translated-gamma components such that the marginals have relatively strong skewnesses of (ωλ,ωK,ωr0)=(1,-1,0.2). In all cases, we show a kernel density estimate produced from 105 samples, the normal approximation (blue dashed), and the translated gamma approximation (red dashed). In the supporting material (Table A in S1 File) we provide a statistical comparison between the simulated and approximate distributions using the Kolmogorov-Smirnov test.

Surrogate likelihood

We make the assumption that the parameters in the dynamical model, θ, are random variables with a distribution parameterised by ξ,

θD(ξ). (17)

For example, D(ξ) may represent a multivariate normal distribution with means, variances and covariances determined by ξ [45]. The only constraint on D(ξ) that we require is that analytical expressions for the moments of θ be available. For example, we can capture skewness in parameter inputs by describing θ as having independent components with translated gamma marginals, but cannot, in general, describe the dependence in θ using a copula. We can capture systems with distinct subpopulations by specifying D(ξ) as a finite mixture, in which case the transformation defined by the mathematical model is applied to each component of the mixture before the mixture is reapplied (in this case, ξ may include parameters relating to both the parameterisation of the mixture components and the mixture weights). Finally, parameters assumed to be constant can be modelled by assuming they follow a degenerate (i.e., point mass) distribution.

We can form an approximate expression for the likelihood of the data, {yn(ti)}n=1N using the approximate solution to the random parameter problem given in Eqs (15) and (16). For a given distribution θD(ξ), the moments of θ will depend on ξ; i.e., θ^=θ^(ξ), V(θ)=m2(ξ), etc, where mi(ξ) are tensor-valued functions of ξ defined by the specification of D(ξ). Therefore, the moments of f(θ) can also be expressed as functions of the hyperparameters, ξ, such that μ = μ(ξ), Σ = Σ(ξ), and ω = ω(ξ). We denote by pf(i)(yn(ti);ξ) the probability density function for yn(ti)f(i)(θ) given θD(ξ). For the normal approach, we have that

pf(i)(yn(ti);ξ)=ϕ(yn(ti);μ(ξ),Σ(ξ)), (18)

where ϕ(y; μ, Σ) is the density function for the multivariate normal distribution with mean μ and covariance matrix Σ. Therefore, the log-likelihood function for snapshot data is given by

(ξ)=tiTn=1Nlogpf(i)(yn(ti);ξ). (19)

While not a focus of the present work, the log-likelihood function for time-series data would, therefore, be given by

(ξ)=n=1Nlogpf(yn;ξ),

where yn includes a set of dependent measurements from all time-points simultaneously, as per Eq (4).

Inference

As our method provides an approximate likelihood function, we permit application of the full gamut of likelihood-based inference and identifiability techniques. We demonstrate our method using both a frequentist method based on profile likelihood [9], and a Bayesian method based on Markov-chain Monte Carlo (MCMC) [7, 11, 46].

Profile likelihood

We explore identifiability of model hyperparameters using the profile likelihood method [9]. First, we establish the maximum likelihood estimate (MLE) as the hyperparameter vector that maximises the log-likelihood function,

ξ^=argmaxξ(ξ). (20)

Secondly, the hyperparameter space is partitioned, ξ=[ϕ,λ], where ϕ is the variable to be profiled and λ contains the remaining parameters. Profile log-likelihoods, ^p are then computed for each value of ϕ by determining the supremum of the log-likelihood over λ relative to the MLE

^p(ϕ)=supλ(ϕ,λ)-(ξ^). (21)

We take the supremum of the log-likelihood function using the Nelder-Mead algorithm over a sufficiently large region to cover the true parameters over several orders of magnitude [47].

One interpretation of ^p(ϕ0) is that of the test statistic in a likelihood-ratio test for ϕ = ϕ0 [48]. Therefore, approximate 95% confidence intervals for each variable ϕ can be constructed by considering the region where the likelihood-ratio test yields p-values greater than α = 0.05, corresponding to the region where

^p(ϕ)>-Δ1,0.952-1.92, (22)

where Δν,1−α is the 1 − α quantile of the χ2(ν) distribution. Given this interpretation, we can quantify statistical evidence for the presence of variability in a model parameter by examining the profile likelihood in the limit that the variance goes to zero.

Markov-chain Monte Carlo

To obtain samples from the posterior distribution of model hyperparameters and hence a distribution that quantifies uncertainty in model predictions, we also perform analysis using a Bayesian MCMC approach [7, 46].

Before consideration of data, X, information about model hyperparameters is encoded in a prior distribution, p(ξ). We then update our knowledge of the parameters using the likelihood to obtain the posterior distribution,

p(ξ|X)exp((ξ;X))p(ξ). (23)

To keep our results consistent with those obtained using the profile likelihood approach, we take the prior distribution to be uniform over the region that covers the true parameters by several orders of magnitude. Therefore, the posterior distribution, Eq (23), is directly proportional to the likelihood function and the MLE corresponds precisely to the maximum a posteriori estimate (MAP); we find the MAP by maximising likelihood function using the Nelder-Mead algorithm [47].

We implement an adaptive MCMC algorithm based on the adaptive Metropolis algorithm from the AdaptiveMCMC package in Julia [49]. To obtain a posterior predictive distribution of model outputs (in our case, including the probability density function of random parameter distributions) by repeated simulation of the model at posterior samples obtained using MCMC.

Results

Using the surrogate likelihood based on the moment-matching approximation, we provide a didactic guide to assessing the identifiability of dynamical models with random parameters using three case studies. As our focus is on identifiability, and not model selection, we work using purely synthetically generated data and apply our statistical methodology to recover the true parameter values.

Logistic model

We first assess identifiability of the canonical logistic model [21]. The logistic model is ubiquitous in biology, ecology, and population dynamics. Our motivation is to describe the time-evolution of the radius of multicellular tumour spheroids (Fig 1) and determine if variability in the initial spheroid size, growth rate, and carrying capacity are identifiable and distinguishable from measurement noise.

Denoting the spheroid radius r(t), the logistic model is

dr(t)dt=λ3r(t)(1-r(t)R)subjecttor(0)=r0, (24)

with exact solution

r(t;θ)=R1+(Rr0-1)exp(-λ3t). (25)

Here, λ is the per-volume growth rate of the spheroid for r(t)/R ≪ 1 (the term λ/3 represents the corresponding per-radius growth rate), R is the carrying capacity radius, and r0 is the initial radius.

We consider data comprising independent measurements (for example, originating from experiments that must be destroyed to collect measurements) subject to additive normal noise such that

f(i)(θ)=r(ti;θ)+ε. (26)

Here, N = 10 measurements are taken at each ti = 2(i − 1), i = 1, 2, …, 8 (Fig 3A and 3B) and εN(0,σε2) represents homoscedastic additive normal measurement noise. The logistic model is parameterised by the random parameter vector θ = [r0, λ, R, ε]. We assess the identifiability for several different parametric forms of the distribution of θ.

Fig 3. Profile likelihood analysis for logistic model with random parameters.

Fig 3

We perform inference on a synthetic data set comprising N = 10 measurements at t = 0, 2, 4, …, 14 of the random parameter logistic model (i.e., 80 independent measurements). In (C) we treat the standard deviation of the measurement noise, σε as unknown, in (D) we assume σε is known (for example, pre-estimated), and in (E) we work with a misspecified model where we assume σε = 0, corresponding to a scenario where we assume all variability in the data is due to variability in mechanistic parameters. In (C–E) we compare profiles produced using a normal surrogate model (blue dashed) and gamma surrogate model (red dotted). In (C) we also take a standard inference approach, assuming that observations are normally distributed about model predictions and where parameter variability is neglected (the fixed parameter model). In (A,B) we show the data (blue), model mean (black) and 95% prediction interval at the MLE using (A) the fixed parameter model, and (B) the random parameter approach with a gamma surrogate. Also shown are the true parameter values (black vertical dotted) and a 95% confidence interval threshold (black horizontal dashed).

Independent normal random parameters

First, we explore identifiability of a model where θD(ξ) is multivariate normal with independent components, such that

r0~Nμr0,σr02,λ~Nμλ,σλ2,R~NμR,σR2,ε~N0,σε2. (27)

Therefore, ξ=[μr0,μλ,μR,lnσr0,lnσλ,lnσR,lnσε], where we infer the natural logarithm of the standard deviations to ensure positivity. We show synthetic data generated from this parameterisation of the random parameter logistic model in Fig 3A and 3B, with identifiability results shown in Fig 3C–3E for μλ = 1, μR = 300, μr0=50, σλ = 0.05, σR = 20, σr0=3, σε = 4.

We present profile likelihoods for each parameter in Fig 3C using both the normal and gamma approximations. To aid interpretation, we show the normalised profile likelihood along with the threshold for an approximate 95% confidence interval. Model predictions (mean and 95% prediction interval) produced using the MLE are shown in Fig 3B. As expected from existing analysis of the fixed-parameter logistic model [22], all three location parameters (i.e., the means of λ, R and r0) are identifiable; this can be seen by profile likelihoods with compact support above the threshold for a 95% confidence interval.

As with the location parameters, we find that the standard deviation of carrying capacity, σR, is also identifiable. We expect this since, for sufficiently large observation times, the solution to the logistic model is simply r(t) = R, meaning that experimental observations at these later times are simply observations from the distribution RN(μR,σR2+σε2). The most interesting result is that for σλ, which is only just identified (within 95% confidence) to a relatively compact region; repeating the exercise with a second set of synthetic data yields a profile likelihood for σλ similar to that for σε, suggesting one-sided identifiability, meaning that σλ is indistinguishable from zero (i.e., variability in λ cannot be detected). Results for σr0 also suggest at one-sided identifiability. Therefore, only variability in R is distinguishable from measurement noise, although given that results for σλ were borderline identifiable, we expect variability in λ to be detectable should it be either larger, or as more data become available.

In Fig 3B we also show results where the fixed-parameter model (i.e., parameters λ, R and r0 are assumed constant) is calibrated to the data from the random parameter model. This represents the standard approach to parameter inference, where variability in the data is typically assumed to comprise entirely of measurement error. Given that the variance of r0 and λ were indistinguishable from zero, this may also seem like a reasonable simplification. First, we see that this has relatively little impact on the point estimates for the location parameters, however does give less precise estimates (i.e., wider confidence intervals). As expected, the estimate for σε is larger than the true value, with the true value not contained with the 95% confidence interval; this is expected, since the ε must now capture both measurement error and parameter variability. Examining predictions from the fixed-parameter model in Fig 3A show that accounting for the variation in (at least) carrying capacity produces a much better representation of the variability in the data.

We explore two further scenarios in Fig 3D, where we assumed that measurement error is pre-estimated or known prior to inference, and Fig 3E, where measurement error is neglected (i.e., variability in the data only comes from variability in the parameters). Both cases yield similar results (in terms of point estimates and precision) for the location parameters. Intriguingly, perhaps because σε is relatively small, pre-estimating the measurement error has very little impact on the results for the variance parameters. Finally, neglecting measurement error produces a bias in the estimates for σr0 (which we expect, since r(0)N(μr0,σr02+σε2)).

In all cases examined for the logistic model with independent multivariate normal parameters, only very minor differences are observed between results that use the normal and gamma approximations, which we interpret to suggest that the third moment (captured by the gamma approximation, but not the normal approximation) contains very little information about parameter variability.

In the supporting material (Fig F in S1 File), we explore the ability of our method to infer parameter distributions that are not from the exponential family; namely, where the input distributions of the logistic model are independent and uniformly distributed. Despite a discrepancy in higher order moments between the approximate solutions and simulations, we are still able to accurately recover the moments of the input distributions.

Correlated and skewed random parameters

Next, we consider two scenarios relating to the complexity of D(ξ), the first where λ and R are correlated such that

(λ,R)MvNormal([μλμR],[σλ2ρλRσλσRρλRσλσRσR2]), (28)

and where r0 and ε are as described by Eq (27). In the second scenario, the growth rate λ is skewed such that

λShiftedGamma(μλ,σλ2,ωλ), (29)

where r0, R, and ε are described by Eq (27) and we set ρλR = 0.6 and ωλ = −1.5.

To assess identifiability of the additional parameter in each of the correlated and skewed models, we show profile likelihoods in Fig 4 for both the normal and gamma approximations, for various sample sizes (observations are taken at the same observation times as in Fig 3). We suppose that prior knowledge has constrained |ρλR| < 0.9 and −2 < ωλ < 1.

Fig 4. Profile likelihoods for an unknown correlation coefficient and growth rate distribution skewness.

Fig 4

(A) We infer hyperparameters from synthetic data where the model parameters (λ, R, r0, ε) are multivariate normal as in Fig 3, but with single unknown correlation Cor(λ, R) = ρλR = 0.6. (B) We infer hyperparameters from synthetic data where the model parameters are independent as in Fig 3, but where λ has a skewed distribution with unknown skewness ωλ = −1.5. Only (A) ρλR and (B) ωλ are profiled. In both cases, we produce results using synthetic data sets of size N = 10 (solid), N = 100 (dashed), and N = 1000 (dotted), where normal (blue) and gamma (red) surrogates are used (black horizontal dashed).

Given the lack of identifiability of many parameters in Fig 3, it is anticipated that both additional parameters will be unidentifiable for small sample sizes, as is the case for N = 10 observations per time point. Even for a relatively large sample size of N = 100 (corresponding to a total of 800 independent samples across all time points), it is only the sign of the skewness parameter that can be identified in the skewed model, whereas the direction of the correlation between λ and R cannot be identified to within a 95% confidence interval until a sample of size N = 1000 is reached.

Results for the correlated model are similar between the normal and gamma approximations, which we interpret to suggest that higher-order moments in the data do not provide significant additional information about the correlation parameter in the model. In contrast, the results between the normal and gamma approximations are striking; in Fig 4B the normal approximation gives misleading results that suggest that the skewness parameter is non-identifiable even for very large sample sizes.

Inference for a misspecified random parameter distribution

Finally, we explore how misspecification of a parameter distribution affects identifiability and model predictions. We consider two cases for the true parameter distribution, the first where λ has a strong negative skew given by Eq (29), and secondly where λ has a bimodal distribution, given by a normal mixture λ ∼ wλ1 + (1 − w2 where

λ1~Nμλ(1),σλ(1),λ2~Nμλ(2),σλ(2). (30)

A similar problem was previously explored by Banks et al. [50]. We set μλ(1)=0.9, μλ(1)=1.1, σλ(1)=σλ(2)=0.05 and w = 0.4 (Fig 5E). The bimodal growth rate might represent a situation where, i.e., multiple subpopulations or cell lines are present in the experimental data.

Fig 5. Inference and prediction where parameter distribution is misspecified.

Fig 5

We explore a case where the underlying growth rate distribution has (A–C) a skewed distribution with (μλ, σλ, ωλ) = (1, 0.05, −1.5), and (e–g) a bimodal distribution, modelled as the mixture wλ1 + (1 − w2 with λ1=N(0.9,0.052), λ2=N(1.1,0.052) and w = 0.4. To ensure identifiability, we use a large sample size of N = 1000 per time point. In (A,E) the true form of the growth rate distribution is used, whereas in (B,F) the growth rate distribution is misspecified and assumed to be normal. Shaded regions in (A,B,E,F) indicate 95% credible intervals for the density. In (C,G), predictions at the MAP estimates (equivalent to MLE) are compared to the data. A 95% prediction interval is shown for the true model (shaded) and the misspecified model (blue dashed), solid curves to the mean, and violin plots show the data. In (D,H), we compare predictions for the density from the true and misspecified models at t = 4.

Given that the results in Fig 4 suggest that large sample sizes are required to infer higher-order parameters, such as the skewness of the growth rate, we consider synthetic data generated with N = 1000 observations per observation time. We show violin plots of the synthetic data in Fig 5C and 5G for the skewed and bimodal scenarios, respectively. To explore uncertainty in the inferred parameter distributions (in contrast to hyperparameter uncertainty), we take a Bayesian approach to inference, and perform inference using MCMC. We perform the analysis using both the true distribution for λ (with additional hyperparameters as appropriate), and a misspecified model where λ is assumed to be normally distributed.

In Fig 5A and 5E we compare the true distribution with the MAP point estimate and credible intervals for the probability density function for λ using posterior samples obtained using MCMC for each model. Given the large sample size, we find that the distribution is identifiable in both cases, confirming our hypothesis from Fig 3 where we found the variance of λ to be only one-sided identifiable from a small sample size. In Fig 5B and 5F we show similar results for a misspecified model where λ is incorrectly assumed to be normally distributed. These results show that misspecification can sometimes yield over-confidence in the identifiability of parameter density functions: results in Fig 5B and 5F show a narrow 95% credible interval for the probability density function for λ that do not contain the true distribution. However, it is still possible to accurately infer the statistical moments of the parameter distributions despite misspecification. For example, the true bimodal distribution and inferred MAP normal distribution (Fig 5F) have similar means (1.020 and 1.016, respectively) and variances (1.22 × 10−2 and 1.21 × 10−2, respectively).

Results in Fig 5C and 5G, showing a 95% prediction interval for the data at the MAP for both the true and misspecified models, demonstrate that coarse-scale predictions from the misspecified model can be useful. We note, however, that this is not always the case; Banks et al. [50] show that misleading predictions can result when λ has a bimodal distribution where λ1 and λ2 are sufficiently different (in our case, they are relatively similar). In Fig 5D and 5H we show a finer-scale comparison of the predictions from each model by considering a comparison between the predicted probability density function for the tumour spheroid radius at t = 4 days, r(4) (MAP with credible intervals). For the skewed model, Fig 5D, predictions are similar between the data (kernel density estimate), true model and misspecified model. However, the misspecified model cannot capture the multimodality of the data at t = 4 days, which is captured by the true model (Fig 5). Both the true and misspecified models have similar non-negligible support and (from results in Fig 5G) comparable 2.5% and 97.5% quantiles. In the supporting material (Fig D in S1 File), we demonstrate how the quality of fit obtained from the misspecified model is poor in the case where subpopulations are more distinct (μλ(1)=0.7,μλ(2)=1.3).

Linear two-pool model

The transfer of chemical species between and from two-pools is used widely as a model of cholesterol transfer or urea decay [51, 52]. We consider that material transfers from species one, denoted X1, to species two, denoted X2, at rate k21 and decays from each pool at rates k1 and k2, respectively. That is, we consider the chemical model

X1k21X2k2,X1k1, (31)

which we describe using a coupled system of differential equations describing the time-rate of change of the concentration of material in each pool, x1(t) and x2(t), respectively, so that

dx1(t)dt=(k21+k1)x1(t),dx2(t)dt=k21x1(t)k2x2(t). (32)

We model a closed system subject to a known input at t = 0 such that x1(0) = x0 and x2(0) = 0.

We consider an inference problem where observations are taken from only the second pool and that the measurement error scales with the concentration. Therefore, we assume multiplicative normal noise, such that

f(ti)(θ)=x2(ti)ε. (33)

We further assume that the decay from each pool is due to a strictly chemical process such that k1 and k2 are constant, and that k21 is a normally distributed random variable. Variation in k21 between data might arise clinically from variability between patients. We incorporate this parameterisation into our framework by assuming that θ=[k1,k21,k2,ε] is a random parameter vector with independent components, where

k1~δ(μ1),k2~δ(μ2),k21~Nμ21,σ212,ε~N1,σ2. (34)

Here, δ(x) denotes a Dirac or degenerate distribution (we take δ(x) to be normally distributed with σ → 0 such that all central moments above the third are zero). For the linear two-pool model, we have that θD(ξ) with hyperparameters ξ=[μ1,μ21,μ2,lnσ21,lnσ]. We set μ1 = 0.7, μ21 = 0.6, μ2 = 0.4, σ21 = 0.1 and σ = 0.01, and generate synthetic data using N = 20 independent observations at t ∈ {0.5, 1.5, 2.5, 3.5, 5, 7} (Fig 6G). The solution to Eq (31) at the MLE is shown in Fig 6G, and the approximate solution based on both a normal and gamma approximation is given as supporting material (Fig B in S1 File) in addition to a statistical comparison (Table B in S1 File). Given that the distribution of material in the second pool is skewed at later times, we work only with the gamma approximation for analysis of the two-pool model.

Fig 6. Identifiability analysis for two-pool model with random parameters.

Fig 6

(A–E) Profile likelihoods for each hyperparameter. Also shown are the true values (vertical dotted) and the threshold for a 95% confidence interval (horizontal dashed). (F) Inferred distribution of k21 showing the distribution at the MAP (black) and a 95% credible interval for the density function. (G) Synthetic data (red discs) and model prediction based on the MLE showing the mean (black) and 95% prediction interval (grey).

Profile likelihood results in Fig 6A–6E show that all physical parameters are identifiable, including the variance in the transfer rate k21. This result is particularly interesting as we are able to identify the source of variance due to heterogeneity despite the variance of the measurement noise being only one-sided identifiable; we cannot distinguish a model with measurement noise from a model without measurement noise, but this has no impact on the identifiability of other model parameters. Results in Fig 6G show model predictions (mean and 95% prediction interval) computed using the MLE. Evident in Fig 6G is a key advantage of the random parameter approach—in contrast to the standard approach where variability is often assumed to originate from independent measurement noise—where our model produces not only average behaviour consistent with the data, but excellent predictions relating to the data variance.

To better visualise how well the unknown distribution of k21 is identified from the available data we repeat the identifiability analysis taking a Bayesian approach to obtain posterior samples using MCMC. In Fig 6F we show a predictive distribution of the density function for k21 (mean and 95% credible interval of the density function), showing that the distribution is identifiable and that relatively precise estimates are recovered.

Non-linear two-pool model

Finally, we consider a non-linear extension of the two-pool model where the transfer rate is not constant, but described by a non-linear Michaelis-Menten form, k(x1) = k21x1/(V21 + x1). That is, we consider the chemical model

X1k(x1)X2k2,X1k1, (35)

described by the system of differential equations

dx1(t)dt=(k21V21+x1(t)+k1)x1(t),dx2(t)dt=k21x1(t)V21+x1(t)k2x2(t). (36)

In contrast to the previous two case studies, an exact solution is not available for the non-linear two-pool model. Therefore, this case study provides an example of the flexibility of our approach: we can solve the non-linear two-pool model using an explicit numerical scheme [53] and use automatic differentiation [54] to calculate the necessary derivatives with minimal additional computational overhead.

We consider identifiability under two scenarios. In both cases, we collect bivariate (i.e., dependent) outputs from both pools, with measurements of pool one subject to multiplicative normal noise, and that of pool two subject to additive normal noise, such that

f(ti)(θ)=[x1(ti)ε1x2(ti)+ε2]. (37)

In the first scenario, we take N = 20 bivariate observations at several observation times; ti = 2i for i = 1, 2, …, 5 (observations at different observation times are independent). Synthetic data for the first scenario are shown in Fig 7A (bivariate data with the gamma approximation are shown in supporting material (Fig C in S1 File)). In the second, clinically and experimentally motivated scenario [55, 56], we consider that N = 100 observations are available from the single observation time t = 10. This second scenario represents a situation where, for example, the data collection method is invasive or possibly where patients must return to a clinic for data collection. Synthetic data for the second scenario are shown in Fig 7B. Given that univariate observations are skewed, we consider only the bivariate gamma approximation for analysis of the non-linear two-pool model. Results in Fig 7B show excellent agreement between the synthetic data and gamma approximation.

Fig 7. Synthetic data from the non-linear two-pool model.

Fig 7

(A) Data comprise N = 20 noisy observations of the concentration in each pool from five observation times. Also shown is the mean and a 95% prediction interval based on the approximate solution to the random parameter non-linear two-pool model. Bivariate data and solution to the random parameter problem are provided as supporting material (Fig C in S1 File) in addition to a statistical comparison for each marginal distribution (Table C in S1 File). (B) Data comprise N = 100 noisy observations from the single observation time t = 10 (blue discs). Also shown is the approximate solution to the random parameter problem using correlated gamma marginals (red solid), and the exact density based on 105 randomly sampled parameter values (grey filled).

The random parameter vector θ=[k1,k21,V21,k2,ε1,ε2] has independent components, where

k1~δ(μ1),k2~δ(μ2),K21~Nμ21,σ212,V21~NμV21,σV212,ε1~N1,σ12,ε2~N0,σ22. (38)

We set μ21 = 0.6, μV21=5, μ1 = 0.1, μ2 = 0.4, σ21 = 0.1, σV21=1, σ1 = σ2 = 0.01.

Profile likelihood results in Fig 8H show that all parameters are identifiable from data with multiple observation times, with the exception of the standard deviation of the additive normal noise process for pool 2; σ2 is one-sided identifiable (indistinguishable from zero). Despite the sample sizes being equivalent, parameter estimates are more precise from data with multiple observation times than from a single observation time.

Fig 8. Identifiability analysis for the non-linear two-pool model.

Fig 8

Profile likelihoods for each parameter where the data comprise N = 20 observations each from five observation times (blue) and N = 100 observations at a single observation time.

Results from data with a single observation time are more interesting. At first, it appears that all hyperparameters relating to the physical parameters are identifiable (the variance of V21 is border-line identifiable, and the variances of the measurement noise variables are one-sided identifiable). However, the profile likelihood is relatively flat around the MLE. Given that we have taken a moment-matching approach to inference, we explore this further by exploring the sensitivity matrix, or Fisher information matrix (FIM) of the function M(ξ):R8R7, which maps the hyperparameters to the moments of the output. The FIM is given by

S(ξ)=JM(ξ)JM(ξ) (39)

where JM(ξ) is the Jacobian of M(ξ). The FIM relates directly to the Hessian (i.e., curvature) of the log-likelihood under the assumption that observations of the moments are normally distributed. Furthermore, the rank of the FIM at the MLE gives insight into the local-identifiability of the model: for identifiability, we require that FIM be of full-rank (or equivalently, non-singular) [43]. Using automatic differentiation to find JM(ξ^) we find that S(ξ^) has one zero eigenvalue so that rank(S(ξ^))=7<8. Therefore, parameters are locally non-identifiable; we also see this from profile likelihood analysis in Fig 8. We have not in this work explored the connection between the identifiability of the fixed parameter model and that of the random parameter model. This question is particularly relevant for the single observation time example as we would not, in general, expect that the four biophysical parameters [k1,k21,V21,k2] be identifiable from a single two-dimensional output. The provision of deterministic expressions connecting the input and output moments (Eqs (11) to (14)) may allow more rigorous exploration of this question in future work.

From profile likelihood results in Fig 8 we also notice non-monotonic behaviour in the likelihood, particularly in Fig 8B and 8E. To explore this further, we take a Bayesian approach to identifiability analysis [7, 11] and explore the convergence of 12 independent MCMC chains of length 100,000, 11 initiated at randomly sampled regions of the prior, and one chain initiated at the true values. In Fig 9A we see that several chains converge to a region of the parameter space with relatively low likelihood, whereas several converge to a region with comparable log posterior density to the MAP. In Fig 9C–9J we explore the marginal density of chains that converge to a region where the mean log-posterior density from the final 60,000 iterations is within a 95% confidence level of the MAP. First, it is clear that results from the single chain initiated at the true value are different from the other chains: the likelihood is clearly multimodal, where regions of the parameter space where the mean or the variance of V21 is zero. We demonstrate this in Fig 9B by finding a second MAP for a model where σV21=0, showing that both models are indistinguishable.

Fig 9. MCMC results for non-linear two-pool model.

Fig 9

(A) MCMC was run for 105 iterations at twelve initial locations: eleven sampled from the prior and one at the true value. Coloured curves show convergence in the log posterior density. Also shown is the posterior density at the MAP (dashed), posterior density at the true value (dotted), a 95% threshold based on an asymptotic chi-squared distribution and the MAP (solid). The first 6 × 104 iterations were discarded as burn-in. Each colour corresponds to an independent MCMC chain. (B) Synthetic data (discs), approximate model solution at the MAP (orange solid), approximate model solution at a model where V21 has zero variance (red dotted), and approximate model solution at the true values (grey). (C–F) Marginal posterior densities for each parameter. Each colour corresponds to an independent MCMC chain.

Discussion

Deterministic differential equation models are routinely applied to analyse data in terms of parameters that carry physically meaningful interpretations. Traditionally, these models have fixed parameters that describe only the mean of experimental observations: variability in data is neglected, often assumed to originate from a noise process unrelated to the underlying dynamics (i.e., measurement error) [7, 9, 13]. Allowing model parameters to vary randomly according to probability distributions provides flexibility to account for the heterogeneity that plays an essential role in the emergent behaviour of many biological and biophysical systems. Methods for performing inference of these models, and consequentially assessing parameter identifiability, are traditionally limited by a computational cost that far exceeds that of the corresponding fixed-parameter model. In this work, we present a novel framework for identifiability analysis of differential equation models with random parameters with a computational cost comparable to that of the fixed parameter problem. Our approach is applicable to many existing workflows, since we use a standard class of deterministic, differential equation models, and provide an approximate expression for the likelihood function providing flexibility in terms of the statistical methods used for inference and identifiability analysis.

We approach the random parameter inference problem by specifying a distribution for model parameters, and infer hyperparameters that relate to each model parameter distribution. Notably, this approach increases the number of unknown parameters that have to be estimated from data, however also allows interpretation of additional information that may be available in higher-order statistical moments of the data. Identifiability analysis of the logistic model, for example, shows that the additional unknown parameters in the random parameter model (i.e., hyperparameters relating to the variance of each model parameter) do not yield greater uncertainty in the mean of the parameters as is often assumed when the number of unknown parameters in a model increases. In fact, we see in Fig 3C that applying the random parameter model yields more precise estimates of the average proliferation rate and initial spheroid radius. We attribute this, in part, to a more accurate specification of the observation variance: for the fixed parameter model, we assume that variability arises due to homoscedastic normal measurement error, which leads to both under- and over- dispersion at early and late times, respectively (Fig 3C). This can be avoided by allowing the variance to vary with time (for example, by specifying a functional form for the variance), or accounted for naturally using the random parameter model. We find that our model yields accurate predictions of the data variance despite non-identifiability of several hyperparameters which relate to the model parameter variances.

The computational cost and ease of implementation of our approach is comparable to the fixed parameter model, in contrast to approximate Bayesian computational methods [32, 33], which are computationally costly, and Bayesian hierarchical approaches [27, 28], which suffer from a parameter dimensionality that scales with sample size. We benchmark our approach using the non-linear two-pool model with a single observation time, finding that likelihood evaluations for the random parameter problem (850 μs) are comparable to timings for the fixed parameter problem (67 μs) once inefficiencies in our implementation are considered (for example, forming the four-dimensional kurtosis tensor K(θ) without exploiting significant sparsity accounts for 65% (550 μs) of the computation time). Computations were performed on an Apple M1 Pro chip. Overall, the second order Taylor series provides an adequate approximation to the models we consider, requiring evaluation of only the model mean, gradient, and Hessian: all of which can be obtained efficiently and with relative ease using automatic differentiation. While the two-moment normal approximation can yield similar results in cases where the data are not significantly skewed, the three-moment approximation provides better results for a wider range of models with only minor additional computational cost. The use of automatic differentiation [54] means that the code we provide for analysis is applicable to a broad class of potentially black-box deterministic models, with any measurement noise model, provided that model outputs are vector valued.

The primary limitation of our approach is that data must be adequately approximated with a normal or gamma distribution, or be expressible as a mixture of normal or gamma distributions. While this may seem restrictive, we note that it is often the case in the mathematical biology literature that data are assumed normally distributed about model predictions, which describe the data mean (or equivalently, models are calibrated using least-squares estimation) [5, 7, 22, 57]. This assumption can be assessed by examining the fit of the approximate distribution to the data at the MLE. In the supporting material (Fig E in S1 File), we demonstrate a pathological example where our model performs poorly, by approximating the solution to the logistic model with a strong Allee effect [58]. The distribution of the initial condition is chosen so that approximately 16% of model realisations lead to population extinction, whereas 84% lead to logistic growth to carrying capacity. The resultant distribution is bimodal and constrained to a finite interval, whereas the approximation is unimodal, has infinite support, and clearly cannot capture the data. As our approximations are constructed from a finite set of moments, our approach may also fail for high-dimensional data where the dependence structure may be highly non-linear and not adequately captured by a multivariate normal distribution; this is potentially the case with time-series data.

Two sources of variability that we do not consider include intrinsic variability arising, for example, from the chemical master equation, and uncertainty in the independent variable. The former can be captured in a differential equation framework through stochastic differential equations [2, 59], potentially allowing for our approximate approach to inference and identifiability analysis through a nested moment-matching approach [12, 60] that captures both intrinsic variability and variability in model parameters. The latter source of variability is clinically relevant; immunological data arising from study of COVID-19 and other infectious diseases [56], relates to highly heterogeneous biological processes, and the exact time of infection is typically unknown. By making a distributional assumption for the infection time, t, we can already apply our framework to calculate the conditional distribution of measurements p(x|t, θ). This time-dependent distribution can be constructed efficiently by assuming continuity and constructing an interpolation of the moments x over a range of measurement times, t. The joint distribution of measurements and observation time can then be analytically expressed

p(x,t|θ)=p(x|t,θ)p(t|θ), (40)

and a likelihood constructed that accounts for uncertain observation times, that are possibly dependent on θ.

Heterogeneity is ubiquitous to biology, playing an essential role in the behaviour of biological systems, and contributing to the variability present in biological data. In this work, we present a novel, computationally efficient, framework for inference and identifiability analysis for differential equation based models that incorporate heterogeneity through random parameters. We demonstrate how our framework can be applied to identify sources of biological variability from data, and produce both more precise parameter estimates and more accurate predictions with minimal additional computational cost compared to a fixed-parameter approach. Our framework is easy to implement and applicable to a wide range of models commonly employed throughout biology. A better understanding of heterogeneity in biology, aided by quantitative methods to extract heterogeneity from data, has potential to yield a better understanding of disease, more accurate predictions and an overall more holistic insight into biological behaviour.

Supporting information

S1 File. Supporting material document.

(PDF)

Acknowledgments

We thank Nikolas Haass and Gency Gunasingh for training A.P.B. to perform the tumour spheroid experiments that motivated this work.

Data Availability

All relevant data are within the manuscript and its Supporting information files.

Funding Statement

This work is supported by the Australian Research Council (ARC) https://www.arc.gov.au/ through a Discovery Grant to MJS (Grant number DP200100177) and a Future Fellowship to CD (Grant number FT210100260). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1. Elsasser WM. Outline of a theory of cellular heterogeneity. Proceedings of the National Academy of Sciences. 1984;81(16):5126–5129. doi: 10.1073/pnas.81.16.5126 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2. Wilkinson DJ. Stochastic modelling for quantitative description of heterogeneous biological systems. Nature Reviews Genetics. 2009;10(2):122–133. doi: 10.1038/nrg2509 [DOI] [PubMed] [Google Scholar]
  • 3. Gough A, Stern AM, Maier J, Lezon T, Shun TY, Chennubhotla C, et al. Biologically relevant heterogeneity: metrics and practical insights. SLAS Discovery. 2016;22(3):213–237. doi: 10.1177/2472555216682725 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4. Johnson JB, Omland KS. Model selection in ecology and evolution. Trends in Ecology & Evolution. 2004;19(2):101–108. doi: 10.1016/j.tree.2003.10.013 [DOI] [PubMed] [Google Scholar]
  • 5. Liang H, Wu H. Parameter estimation for differential equation models using a framework of measurement error in regression models. Journal of the American Statistical Association. 2008;103(484):1570–1583. doi: 10.1198/016214508000000797 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6. Raue A, Kreutz C, Theis FJ, Timmer J. Joining forces of Bayesian and frequentist methodology: a study for inference in the presence of non-identifiability. Philosophical Transactions of the Royal Society A. 2013;371(1984):20110544. doi: 10.1098/rsta.2011.0544 [DOI] [PubMed] [Google Scholar]
  • 7. Hines KE, Middendorf TR, Aldrich RW. Determination of parameter identifiability in nonlinear biophysical models: A Bayesian approach. The Journal of General Physiology. 2014;143(3):401–16. doi: 10.1085/jgp.201311116 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8. Simpson MJ, Baker RE, Vittadello ST, Maclaren OJ. Practical parameter identifiability for spatio-temporal models of cell invasion. Journal of The Royal Society Interface. 2020;17(164):20200055. doi: 10.1098/rsif.2020.0055 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9. Raue A, Kreutz C, Maiwald T, Bachmann J, Schilling M, Klingmüller U, et al. Structural and practical identifiability analysis of partially observed dynamical models by exploiting the profile likelihood. Bioinformatics. 2009;25(15):1923–1929. doi: 10.1093/bioinformatics/btp358 [DOI] [PubMed] [Google Scholar]
  • 10. Chis OT, Banga JR, Balsa-Canto E. Structural identifiability of systems biology models: a critical comparison of methods. PLOS One. 2011;6(11):e27755. doi: 10.1371/journal.pone.0027755 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11. Siekmann I, Sneyd J, Crampin E. MCMC can detect nonidentifiable models. Biophysical Journal. 2012;103(11):2275–2286. doi: 10.1016/j.bpj.2012.10.024 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12. Browning AP, Warne DJ, Burrage K, Baker RE, Simpson MJ. Identifiability analysis for stochastic differential equation models in systems biology. Journal of The Royal Society Interface. 2020;17:20200652. doi: 10.1098/rsif.2020.0652 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13. Eisenberg MC, Hayashi MAL. Determining identifiable parameter combinations using subset profiling. Mathematical Biosciences. 2014;256:116–126. doi: 10.1016/j.mbs.2014.08.008 [DOI] [PubMed] [Google Scholar]
  • 14. Munsky B, Trinh B, Khammash M. Listening to the noise: random fluctuations reveal gene network parameters. Molecular Systems Biology. 2009;5(1):318. doi: 10.1038/msb.2009.75 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15. Komorowski M, Costa MJ, Rand DA, Stumpf MPH. Sensitivity, robustness, and identifiability in stochastic chemical kinetics models. Proceedings of the National Academy of Sciences. 2011;108(21):8645–8650. doi: 10.1073/pnas.1015814108 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16. Fox ZR, Munsky B. The finite state projection based Fisher information matrix approach to estimate information and optimize single-cell experiments. PLOS Computational Biology. 2019;15(1):e1006365. doi: 10.1371/journal.pcbi.1006365 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17. Renardy M, Kirschner D, Eisenberg M. Structural identifiability analysis of age-structured PDE epidemic models. Journal of Mathematical Biology. 2022;84(1-2):9. doi: 10.1007/s00285-021-01711-1 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18. Elowitz MB, Levine AJ, Siggia ED, Swain PS. Stochastic gene expression in a single cell. Science. 2002;297(5584):1183–1186. doi: 10.1126/science.1070919 [DOI] [PubMed] [Google Scholar]
  • 19. Mathew D, Giles JR, Baxter AE, Oldridge DA, Greenplate AR, Wu JE, et al. Deep immune profiling of COVID-19 patients reveals distinct immunotypes with therapeutic implications. Science. 2020;369(6508):eabc8511. doi: 10.1126/science.abc8511 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20. Browning AP, Sharp JA, Murphy RJ, Gunasingh G, Lawson B, Burrage K, et al. Quantitative analysis of tumour spheroid structure. eLife. 2021;10:e73020. doi: 10.7554/eLife.73020 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21. Murray JD. Mathematical Biology. 3rd ed. Berlin: Springer-Verlag; 2002. [Google Scholar]
  • 22. Simpson MJ, Browning AP, Warne DJ, Maclaren OJ, Baker RE. Parameter identifiability and model selection for sigmoid population growth models. Journal of Theoretical Biology. 2022;535:110998. doi: 10.1016/j.jtbi.2021.110998 [DOI] [PubMed] [Google Scholar]
  • 23. Loos C, Moeller K, Fröhlich F, Hucho T, Hasenauer J. A hierarchical, data-driven approach to modeling single-cell populations predicts latent causes of cell-to-cell variability. Cell Systems. 2018;6(5):593–603.e13. doi: 10.1016/j.cels.2018.04.008 [DOI] [PubMed] [Google Scholar]
  • 24. Lambert B, Gavaghan DJ, Tavener SJ. A Monte Carlo method to estimate cell population heterogeneity from cell snapshot data. Journal of Theoretical Biology. 2021;511:110541. doi: 10.1016/j.jtbi.2020.110541 [DOI] [PubMed] [Google Scholar]
  • 25. Soong T. Random Differential Equations in Science and Engineering. vol. 103 of Mathematics in Science and Engineering; 1973. [Google Scholar]
  • 26. Lawson BAJ, Drovandi CC, Cusimano N, Burrage P, Rodriguez B, Burrage K. Unlocking data sets by calibrating populations of models to data density: A study in atrial electrophysiology. Science Advances. 2018;4(1):e1701676. doi: 10.1126/sciadv.1701676 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27. Huang Y, Liu D, Wu H. Hierarchical Bayesian methods for estimation of parameters in a longitudinal HIV dynamic system. Biometrics. 2006;62(2):413–423. doi: 10.1111/j.1541-0420.2005.00447.x [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28. Hasenauer J, Hasenauer C, Hucho T, Theis FJ. ODE constrained mixture modelling: a method for unraveling subpopulation structures and dynamics. PLOS Computational Biology. 2014;10(7):e1003686. doi: 10.1371/journal.pcbi.1003686 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29. Zechner C, Unger M, Pelet S, Peter M, Koeppl H. Scalable inference of heterogeneous reaction kinetics from pooled single-cell recordings. Nature Methods. 2014;11(2):197–202. doi: 10.1038/nmeth.2794 [DOI] [PubMed] [Google Scholar]
  • 30. Dharmarajan L, Kaltenbach HM, Rudolf F, Stelling J. A simple and flexible computational framework for inferring sources of heterogeneity from single-cell dynamics. Cell Systems. 2019;8(1):15–26.e11. doi: 10.1016/j.cels.2018.12.007 [DOI] [PubMed] [Google Scholar]
  • 31. Wang L, Cao J, Ramsay JO, Burger DM, Laporte CJL, Rockstroh JK. Estimating mixed-effects differential equation models. Statistics and Computing. 2014;24(1):111–121. doi: 10.1007/s11222-012-9357-1 [DOI] [Google Scholar]
  • 32. Browning AP, Ansari N, Drovandi C, Johnston APR, Simpson MJ, Jenner AL. Identifying cell-to-cell variability in internalization using flow cytometry. Journal of the Royal Society Interface. 2022;19(190):20220019. doi: 10.1098/rsif.2022.0019 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Drovandi C, Lawson B, Jenner AL, Browning AP. Population calibration using likelihood-free Bayesian inference. arXiv. 2022;
  • 34. Hasenauer J, Waldherr S, Doszczak M, Radde N, Scheurich P, Allgöwer F. Identification of models of heterogeneous cell populations from population snapshot data. BMC Bioinformatics. 2011;12(1):125. doi: 10.1186/1471-2105-12-125 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35. Wolfinger RD, Lin X. Two Taylor-series approximation methods for nonlinear mixed models. Computational Statistics & Data Analysis. 1997;25(4):465–490. doi: 10.1016/S0167-9473(97)00012-1 [DOI] [Google Scholar]
  • 36. Elf J, Ehrenberg M. Fast evaluation of fluctuations in biochemical networks with the linear noise approximation. Genome Research. 2003;13(11):2475–2484. doi: 10.1101/gr.1196503 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37. van Kampen NG. Stochastic processes in physics and chemistry. 3rd ed. Amsterdam: Elsevier; 2007. [Google Scholar]
  • 38. Grima R. An effective rate equation approach to reaction kinetics in small volumes: Theory and application to biochemical reactions in nonequilibrium steady-state conditions. The Journal of Chemical Physics. 2010;133(3):035101. doi: 10.1063/1.3454685 [DOI] [PubMed] [Google Scholar]
  • 39. Fröhlich F, Thomas P, Kazeroonian A, Theis FJ, Grima R, Hasenauer J. Inference for stochastic chemical kinetics using moment equations and system size expansion. PLOS Computational Biology. 2016;12(7):e1005030. doi: 10.1371/journal.pcbi.1005030 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40. Pandey A, Kulkarni A, Roy B, Goldman A, Sarangi S, Sengupta P, et al. Sequential application of a cytotoxic nanoparticle and a PI3K inhibitor enhances antitumor efficacy. Cancer Research. 2014;74(3):675–685. doi: 10.1158/0008-5472.CAN-12-3783 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41. Jenner A, Yun CO, Yoon A, Kim PS, Coster ACF. Modelling heterogeneity in viral-tumour dynamics: The effects of gene-attenuation on viral characteristics. Journal of Theoretical Biology. 2018;454:41–52. doi: 10.1016/j.jtbi.2018.05.030 [DOI] [PubMed] [Google Scholar]
  • 42. Crosley P, Farkkila A, Jenner AL, Burlot C, Cardinal O, Potts KG, et al. Procaspase-activating compound-1 synergizes with TRAIL to induce apoptosis in established granulosa cell tumor cell line (KGN) and explanted patient granulosa cell tumor cells in vitro. International Journal of Molecular Sciences. 2021;22(9):4699. doi: 10.3390/ijms22094699 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43. Rothenberg TJ. Identification in parametric models. Econometrica. 1971;39(3):577. doi: 10.2307/1913267 [DOI] [Google Scholar]
  • 44. Nelsen RB. An introduction to copulas. 2nd ed. Springer series in statistics. New York: Springer; 2006. [Google Scholar]
  • 45. Adlung L, Stapor P, Tönsing C, Schmiester L, Schwarzmüller LE, Postawa L, et al. Cell-to-cell variability in JAK2/STAT5 pathway components and cytoplasmic volumes defines survival threshold in erythroid progenitor cells. Cell Reports. 2021;36(6):109507. doi: 10.1016/j.celrep.2021.109507 [DOI] [PubMed] [Google Scholar]
  • 46. Gelman A, Carlin JB, Stern HS, Dunson DB, Vehtari A, Rubin DB. Bayesian Data Analysis. 3rd ed. CRC Press; 2014. [Google Scholar]
  • 47.Johnson SG. The NLopt module for Julia; 2021. Available from: https://github.com/JuliaOpt/NLopt.jl.
  • 48. Pawitan Y. In all likelihood: statistical modelling and inference using likelihood. Oxford: Oxford University Press; 2013. [Google Scholar]
  • 49. Vihola M. Ergonomic and reliable Bayesian inference with adaptive Markov chain Monte Carlo. Wiley StatsRef: Statistics Reference Online. 2020; p. 1–12. doi: 10.1002/9781118445112.stat08286 [DOI] [Google Scholar]
  • 50. Banks HT, Meade AE, Schacht C, Catenacci J, Thompson WC, Abate-Daga D, et al. Parameter estimation using aggregate data. Applied Mathematics Letters. 2020;100:105999. doi: 10.1016/j.aml.2019.105999 [DOI] [Google Scholar]
  • 51. Nestel PJ, Whyte HM, Goodman DS. Distribution and turnover of cholesterol in humans. Journal of Clinical Investigation. 1969;48(6):982–991. doi: 10.1172/JCI106079 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 52. Cobelli C, DiStefano JJ. Parameter and structural identifiability concepts and ambiguities: a critical review and analysis. American Journal of Physiology-Regulatory, Integrative and Comparative Physiology. 1980;239(1):7–24. doi: 10.1152/ajpregu.1980.239.1.R7 [DOI] [PubMed] [Google Scholar]
  • 53. Rackauckas C, Nie Q. DifferentialEquations.jl—A performant and feature-rich ecosystem for solving differential equations in Julia. Journal of Open Research Software. 2016;5(1). doi: 10.5334/jors.151 [DOI] [Google Scholar]
  • 54.Revels J, Lubin M, Papamarkou T. Forward-mode automatic differentiation in Julia. arXiv. 2016;.
  • 55. Kim PH, Sohn JH, Choi JW, Jung Y, Kim SW, Haam S, et al. Active targeting and safety profile of PEG-modified adenovirus conjugated with herceptin. Biomaterials. 2011;32(9):2314–2326. doi: 10.1016/j.biomaterials.2010.10.031 [DOI] [PubMed] [Google Scholar]
  • 56. Lucas C, Wong P, Klein J, Castro TBR, Silva J, Sundaram M, et al. Longitudinal analyses reveal immunological misfiring in severe COVID-19. Nature. 2020;584(7821):463–469. doi: 10.1038/s41586-020-2588-y [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 57. Girolami M. Bayesian inference for differential equations. Theoretical Computer Science. 2008;408(1):4–16. doi: 10.1016/j.tcs.2008.07.005 [DOI] [Google Scholar]
  • 58. Li Y, Buenzli PR, Simpson MJ. Interpreting how nonlinear diffusion affects the fate of bistable populations using a discrete modelling framework. Proceedings of the Royal Society A. 2022;478(2262):20220013. doi: 10.1098/rspa.2022.0013 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 59. Turelli M. Random environments and stochastic calculus. Theoretical Population Biology. 1977;12(2):140–178. doi: 10.1016/0040-5809(77)90040-5 [DOI] [PubMed] [Google Scholar]
  • 60. Warne DJ, Baker RE, Simpson MJ. Rapid Bayesian inference for expensive stochastic models. Journal of Computational and Graphical Statistics. 2021;31(2):512–528. doi: 10.1080/10618600.2021.2000419 [DOI] [Google Scholar]
PLoS Comput Biol. doi: 10.1371/journal.pcbi.1010734.r001

Decision Letter 0

Jason A Papin, Ricardo Martinez-Garcia

10 Oct 2022

Dear Professor Simpson,

Thank you very much for submitting your manuscript "Efficient inference and identifiability analysis for differential equation models with random parameters" for consideration at PLOS Computational Biology. As with all papers reviewed by the journal, your manuscript was reviewed by members of the editorial board and by several independent reviewers. The reviewers appreciated the attention to an important topic. Based on the reviews, we are likely to accept this manuscript for publication, providing that you modify the manuscript according to the review recommendations. Notice, however, that the reviewers made important comments on some choices in your analysis and how sensitive your results are to them. Please make sure that your revised submission addresses all these concerns.

Please prepare and submit your revised manuscript within 30 days. If you anticipate any delay, please let us know the expected resubmission date by replying to this email.

When you are ready to resubmit, please upload the following:

[1] A letter containing a detailed list of your responses to all review comments, and a description of the changes you have made in the manuscript. Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out

[2] Two versions of the revised manuscript: one with either highlights or tracked changes denoting where the text has been changed; the other a clean version (uploaded as the manuscript file).

Important additional instructions are given below your reviewer comments.

Thank you again for your submission to our journal. We hope that our editorial process has been constructive so far, and we welcome your feedback at any time. Please don't hesitate to contact us if you have any questions or comments.

Sincerely,

Ricardo Martinez-Garcia

Academic Editor

PLOS Computational Biology

Jason Papin

Editor-in-Chief

PLOS Computational Biology

***********************

A link appears below if there are any accompanying review attachments. If you believe any reviews to be missing, please contact ploscompbiol@plos.org immediately:

Reviewer's Responses to Questions

Comments to the Authors:

Please note here if the review is uploaded as an attachment.

Reviewer #1: Browning et al. present a new approach that enables parameter estimation and identifiability analysis for differential equation models incorporating heterogeneity. This approach allows for efficient inference and identifiability analysis by leveraging an approximate solution to the differential equation. Overall this work is widely applicable to differential equation models that explicitly model heterogeneity observed in the data via random parameters with parametric distributions. The efficient approach presented in this work enables critical analyses that would otherwise be computationally intractable for many applications.

Overall, I found that the authors clearly explain their method and provide adequate background and detail to enable a reader’s understanding. Additionally, the authors use three test cases to provide an excellent introduction to the identifiability analysis of this class of models. I believe this work will allow future users to conduct a similar analysis of their models.

I have no major changes to suggest, but I have a few minor points I would like the authors to address:

The proposed approach requires the user to specify parametric distributions for the model parameters. They show that misspecification of these distributions can lead to errors in the identifiability analysis. I am concerned about the limitations of this approach when prior knowledge of these distributions is limited. Can the authors address the possible limitations of assuming specific parametric distributions for model parameters and provide suggestions for when the distribution may be unknown?

How does the identifiability of a model parameter affect the identifiability of its distributions hyper-parameters? For instance, if one found that a parameter is non-identifiable via an a priori analysis, what should they expect to see in the identifiability analysis of that parameter’s hyper-parameters?

Does the identifiability of a parameter’s variance terms provide insight into the model's sensitivity to those parameters? For instance, in section 3.1.1, the authors state that the slight identifiability of the variance parameter for lambda indicates that variability in this lambda cannot be distinguished from measurement noise. Does this imply that the model is not particularly sensitive to lambda’s value?

Can the authors clarify how they determine if a parameter is identifiable from its 95% credible interval in the cases where they perform a Bayesian analysis? For example, in section 3.1.3, lines 393-395, the authors conclude that the distribution of parameter lambda is identifiable. However, it is unclear how they arrived at this conclusion from the results presented in figure 5.

The authors present their method with examples that have relatively few state variables and parameters. Can the authors comment on how this approach scales to models with greater numbers of parameters?

In addition to these points, I found several typos while reading the manuscript. Can the authors please correct the following typos and double-check the manuscript and supplemental materials:

The caption for figure 4 states “omega_lambda = -1.5” however, the corresponding text on line 357 states “omega_lambda = 1.5.” I believe the value on line 357 needs to be negative.

Line 437 says, “second tool.” I believe this should read “second pool.”

Reviewer #2: In their manuscript titled "Efficient inference and identifiability analysis for differential equation models with random parameters", Browning and colleagues introduce a new method for the calibtration of ODE models with random parameters. The model can be used for the description and inference of inter-individual heterogeneity, which is a very relevant problem in the current literature. The proposed method is novel in so far as that noise model is incorporated into the model-transformed random variable, and that the taylor approximation is applied differently than in similar methods such as the method of moments or van kampens system size expansion, which brings some advantages in terms of scalability (potentially at the cost of some accuracy). The paper is very well written and easy to follow, but there are some technical aspects that remain opaque (which I will go into more detail below). I generally like the approach, and definitely think that this paper should be published. However, I am some concerns that I describe below, but I have no doubt the authors will be able to adress them.

Major Points:

1) Embedding in the existing literature:

The paper is a bit heavy on statistical jargon which might make it difficult for readers with a stronger biological background to follow the paper. Specifically it would be great to give a bit more explanation about what the authors mean by random/fixed effects models. I am familiar with these terms in the context of Non-linear mixed effects models, which are likely to be what the authors call "hierarchical" models. Such models have been used in the context of biological models (see e.g., https://doi.org/10.1186/s12918-015-0203-x, https://doi.org/10.1371/journal.pcbi.1004706, https://doi.org/10.1371/journal.pone.0124050, https://doi.org/10.1038/s41540-018-0079-7 and references therein). Similarly the authors should contrast their approach to similar approaches such as https://doi.org/10.1016/j.cels.2018.12.007 or https://doi.org/10.1038/nmeth.2794.

Another range of approaches that seems to be relevant, but isn't really discussed are the moment closure approximations (https://doi.org/10.1063/1.3454685) or van Kampens system size expansion (https://doi.org/10.1063/1.3454685). Both use taylor expansion to approximate moments, but with respect to different variables, which would be helpful if mentioned in the paper. Usually these methods are employed for the description of stochastic models, but, using the approach described in https://doi.org/10.1016/j.cels.2018.04.008 which the authors also use, can also be applied to describe heterogeneity. Accordingly, the authors also should contrast their approach to (https://doi.org/10.1371/journal.pcbi.1005030) where moment closure and van Kampens approximation are used for paramete inference (in a stochastic modeling context, not a heterogeneity context, but the transfer is trivial.).

2) Snapshot vs Timecourse data:

The authors just briefly mention the issue of considering snapshot vs timecourse data (it would be good to mention these terms such that a more biological audience can also follo) in l142-145. However, this isn't really picked up in the remainder of the manuscript, but is quite relevant in practical terms. The key difference for timecourse data is that there is temporal correlation between simulations across timepoints. It is unclear to me how this is accounted for in the method that the authors present, as what the authors describe only looks like an rearrangement of indices.

3) Likelihood function:

I am still unsure whether I am fully grasping what the authors are actually doing. For me it would be quite helpful to have some visual depcition of how the approximation method that the authors are proposing actually works. Similarly, it would probably to explicitely write down the equation for the likelihood pf(y,xi). My understanding is that this would simply, in the case of the normal approach, be a multivariate normal probability density function with mean and standard deviation according to equations 10 and 11. I understand that the equations would be quite bulky, but (10) - (13) already pose a solid chance of scare the reader away ;).

4) Approximation Error:

The authors really only discuss the approximation error in figure 2 and the discussion and seem to forget about in the remainder of the results. It would be good to see some more investigation of the approximation error of the method in the more complex settings, to make sure that the findings are not the result of approximation errors.

Minor Points:

a) I am not sure what the argument about the inverse of f in line 149 is about. In the vast majority of cases f(theta) wont be available analytically because f itself is not available analytically in the first place.

b) l234 should probably read "log-likelihood function" instead of "likelihood function"

c) panel labels in the legend to figure 5 seem misarranged

d) for the parameterization of the covariance matrix D(\\xi) the authors may want to consult https://doi.org/10.1016/j.celrep.2021.109507 for some pointers on how to avoid overparameterization

e) non-identifiability and slopiness are not the same, see https://doi.org/10.1016/j.mbs.2016.10.009

f) it's a bit uncommen to introduce new data in the discussion, the authors may want to create a separate section discussing benchmarking and the model with a strong Allee effect

Reviewer #3: Browning et al present methods for the analysis of models with intrinsic variability in their parameters. Sources of noise in biology are ubiquitous, complex, and often neglected or at least under-appreciated in systems biology analyses of model identifiability and parameter inference. This is thus an important and timely contribution to the literature that will be widely useful. The inclusion of open-source code in Julia is an additional strength of the manuscript. I have only minor requests/suggestions for possible improvement to the paper. These are given below.

1. While the analysis of how skewness or bimodality affect identifiability is interesting, at least in my experience, a far more common occurrence in biological modeling is the choice of a prior that is particularly uninformative (e.g. uniform over large interval). Would such a prior choice affect the inference/identifiability results, and how? This would be a useful example.

2. I appreciate the advantages of & the need to strive for model simplicity, however, practically, two species is really a lower bound for realistic model sizes in sys bio. I think it would be really beneficial if an example (or even just some discussion without an example) of a larger (e.g. 3 species) model, in light of the analyses contained in this work: studying identifiability with random parameters & the impact of prior choice, etc.

3. In Fig 9, please could you improve the caption/description: it is a figure containing quite a lot of details and it is currently hard to figure out several details. Are colors indep MCMC chains? The different dashed/dotted lines in A are hard to distinguish. I cannot see orange solid lines in B. Also what does grey represent in B?

**********

Have the authors made all data and (if applicable) computational code underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data and code underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data and code should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data or code —e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

**********

PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

Reviewer #3: No

Figure Files:

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email us at figures@plos.org.

Data Requirements:

Please note that, as a condition of publication, PLOS' data policy requires that you make available all data used to draw the conclusions outlined in your manuscript. Data must be deposited in an appropriate repository, included within the body of the manuscript, or uploaded as supporting information. This includes all numerical values that were used to generate graphs, histograms etc.. For an example in PLOS Biology see here: http://www.plosbiology.org/article/info%3Adoi%2F10.1371%2Fjournal.pbio.1001908#s5.

Reproducibility:

 

To enhance the reproducibility of your results, we recommend that you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. Additionally, PLOS ONE offers an option to publish peer-reviewed clinical study protocols. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols

References:

Review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript.

If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice.

PLoS Comput Biol. doi: 10.1371/journal.pcbi.1010734.r003

Decision Letter 1

Jason A Papin, Ricardo Martinez-Garcia

14 Nov 2022

Dear Professor Simpson,

We are pleased to inform you that your manuscript 'Efficient inference and identifiability analysis for differential equation models with random parameters' has been provisionally accepted for publication in PLOS Computational Biology.

Before your manuscript can be formally accepted you will need to complete some formatting changes, which you will receive in a follow up email. A member of our team will be in touch with a set of requests.

Please note that your manuscript will not be scheduled for publication until you have made the required changes, so a swift response is appreciated.

IMPORTANT: The editorial review process is now complete. PLOS will only permit corrections to spelling, formatting or significant scientific errors from this point onwards. Requests for major changes, or any which affect the scientific understanding of your work, will cause delays to the publication date of your manuscript.

Should you, your institution's press office or the journal office choose to press release your paper, you will automatically be opted out of early publication. We ask that you notify us now if you or your institution is planning to press release the article. All press must be co-ordinated with PLOS.

Thank you again for supporting Open Access publishing; we are looking forward to publishing your work in PLOS Computational Biology. 

Best regards,

Ricardo Martinez-Garcia

Academic Editor

PLOS Computational Biology

Jason Papin

Editor-in-Chief

PLOS Computational Biology

***********************************************************

Reviewer's Responses to Questions

Comments to the Authors:

Please note here if the review is uploaded as an attachment.

Reviewer #1: The authors addressed all of my comments and concerns. Thank you!

Reviewer #2: The authors have satisfyingly addressed all of my concerns, so I would recommend the manuscript for publication.

Reviewer #3: In this revision the authors have addressed all of my previous concerns.

**********

Have the authors made all data and (if applicable) computational code underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data and code underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data and code should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data or code —e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

**********

PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

Reviewer #3: No

PLoS Comput Biol. doi: 10.1371/journal.pcbi.1010734.r004

Acceptance letter

Jason A Papin, Ricardo Martinez-Garcia

21 Nov 2022

PCOMPBIOL-D-22-01128R1

Efficient inference and identifiability analysis for differential equation models with random parameters

Dear Dr Simpson,

I am pleased to inform you that your manuscript has been formally accepted for publication in PLOS Computational Biology. Your manuscript is now with our production department and you will be notified of the publication date in due course.

The corresponding author will soon be receiving a typeset proof for review, to ensure errors have not been introduced during production. Please review the PDF proof of your manuscript carefully, as this is the last chance to correct any errors. Please note that major changes, or those which affect the scientific understanding of the work, will likely cause delays to the publication date of your manuscript.

Soon after your final files are uploaded, unless you have opted out, the early version of your manuscript will be published online. The date of the early version will be your article's publication date. The final article will be published to the same URL, and all versions of the paper will be accessible to readers.

Thank you again for supporting PLOS Computational Biology and open-access publishing. We are looking forward to publishing your work!

With kind regards,

Anita Estes

PLOS Computational Biology | Carlyle House, Carlyle Road, Cambridge CB4 3DN | United Kingdom ploscompbiol@plos.org | Phone +44 (0) 1223-442824 | ploscompbiol.org | @PLOSCompBiol


Articles from PLOS Computational Biology are provided here courtesy of PLOS

RESOURCES