Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2015 Jan 1.
Published in final edited form as: Inverse Probl Sci Eng. 2013 May 17;22(4):557–590. doi: 10.1080/17415977.2013.797973

Experimental Design for Vector Output Systems

HT Banks 1, KL Rehm 1
PMCID: PMC3929304  NIHMSID: NIHMS477740  PMID: 24563655

Abstract

We formulate an optimal design problem for the selection of best states to observe and optimal sampling times for parameter estimation or inverse problems involving complex nonlinear dynamical systems. An iterative algorithm for implementation of the resulting methodology is proposed. Its use and efficacy is illustrated on two applied problems of practical interest: (i) dynamic models of HIV progression and (ii) modeling of the Calvin cycle in plant metabolism and growth.

Keywords: Optimal design, inverse problems, optimal selection of observables and sampling times, HIV models, plant metabolism and growth

1 Introduction

In many scientific fields where mathematical modeling is utilized, mathematical models grow increasingly complex, containing possibly more state variables and parameters, over time as the underlying governing processes of a system are better understood and refinements in mechanisms are considered. Additionally, as technology invents and improves devices to measure physical and biological phenomena, new data become available to inform mathematical modeling efforts. The world is approaching an era in which the vast amounts of information available to researchers may be overwhelming or even counterproductive to efforts. We explore a framework based on the Fisher Information Matrix (FIM) for a system of ordinary differential equations (ODEs) to determine when an experimenter should take samples and what variables to measure when collecting information on a physical or biological process that is modeled by a dynamical system.

Inverse problem methodologies are discussed in [10] in the context of dynamical system or mathematical model parameter estimation when a sufficient number of observations of one or more states (variables) are available. The choice of method depends on assumptions the modeler makes on the form of the error between the model and the observations (the statistical model). The most prevalent source of error is observation error, which is made when collecting data. (One can also consider model error, which originates from the differences between the model and the underlying process that the model describes. But this is often quite difficult to quantify.) Measurement error is most readily discussed in the context of statistical models. The three techniques commonly addressed are maximum likelihood estimators (MLE), used when the properties of the error distribution are known; ordinary least squares (OLS), for error with constant variance across observations; and generalized least squares (GLS), used when the variance of the data can be expressed as a nonconstant function. Uncertainty quantification is also described for optimization problems of this type, namely in the form of observation error covariances, standard errors, residual plots, and sensitivity matrices. Techniques to approximate the variance of the error are also included in these discussions.

Experimental design using the Fisher Information Matrix (FIM), which is based on sensitivity matrices, is described in [11] for the case of scalar data. Sensitivity matrices are composed of functions that relate the change in a variable to the change in the parameter. The first order quantifications of these relations are called traditional sensitivity functions and are useful in suggesting when a variable should be sampled to get the most information for estimating a particular parameter, especially when the first order sensitivity functions are used in conjunction with the so-called second order sensitivity functions. This work also examines the usefulness of generalized sensitivity functions [18], which are calculated using the FIM, that are known to describe how information about the parameters is distributed across time for each variable. Both types of sensitivity functions are then used in numerical simulations to determine the optimal final time for an experiment of a process described by a logistic curve.

In [12], the authors develop an experimental design theory using the FIM to identify optimal sampling times for experiments on physical processes (modeled by an ODE system) in which scalar or vector data will be taken. The experimental design technique developed is applied in numerical simulations to the logistic curve, a simple ODE model describing glucose regulation and a harmonic oscillator example.

In addition to when to take samples, the question of what variables to measure is also very important in designing effective experiments, especially when the number of state variables is large. Use of such a methodology to optimize what to measure would further reduce testing costs by eliminating extra experiments to measure variables neglected in previous trials. In [5], the best set of variables for an ODE system modeling the Calvin cycle [19] is identified using two methods. The first, an ad-hoc statistical method, determines which variables directly influence an output of interest at any one particular time. A model using a subset of variables is determined via multivariate linear regression, and the efficacy of the model is then measured using the Akaike Information Criterion. The variables that appear in the best models at the most time points are then identified as the most important to measure. Such a method does not utilize the information on the underlying time-varying processes given by the dynamical system model. Extension of the second method first suggested in [5], based optimal design ideas, is the subject of our presentation here.

Building on the theory in [12] and [5], we formulate a previously unexplored optimal design problem to determine not only the optimal sampling variables out of a finite set of possible sampling variables but also the optimal sampling time distribution given a fixed final time. We compare the SE-optimal design introduced in [11] and [12] with the well-known methods of D-optimal and E-optimal design on a six-compartment HIV model [2] and a thirty-one dimensional model of the Calvin Cycle [19]. Such models where there may be a wide range of variables to possibly observe are not only ideal on which to test our proposed methodology, but also are widely encountered in applications.

2 Mathematical Background

2.1 Mathematical and statistical models

We explore our experimental design questions using a mathematical model

dxdt(t)=g(t,x(t;θ),q),t[t0,tf]x(t0;θ)=x0 (1)

where x⃗(t; θ⃗) is the vector of state variables of the system generated using a parameter vector θ⃗ = (x⃗0; q⃗) ∈ ℝp, p = m + r, that contains m initial values and r system parameters listed in q⃗, g⃗ is a mapping ℝ1+m+r → ℝm, t0 0 is the initial time, and tf < ∞ is the final time. We define an observation process

f(t;θ)=Cx(t;θ), (2)

where C is an observation operator that maps ℝmN, where N is the number of variables observed at a single sampling time. If we were able to observe all states, each measured by a different sampling technique, then N = m and C = Im×m; however, this is most often not the case because of the impossibility of or the expense in measuring all state variables. In other cases (such as the HIV example below) we may be able to directly observe only combinations of the states.

In order to discuss the amount of uncertainty in parameter estimates, we formulate a statistical model [10] of the form

Y(t)=f(t;θ0)+E(t),t[t0,tf], (3)

where θ⃗0 is the hypothesized true values of the unknown parameters and Inline graphic is a vector random process that represents observation error for the measured variables. We make the standard assumptions:

E(E(t))=0,t[t0,tf],Var(E(t))=V0(t)=diag(σ0,1(t)2,σ0,2(t)2,,σ0,N(t)2),t[t0,tf],Cov(Ei(t)Ei(s))=σ0,i(t)2δ(t-s),s,t[t0,tf],Cov(Ei(t)Ej(s))=0,ij,s,t[t0,tf],

where δ(0) = 1 and δ(t) = 0 for t ≠ 0. Realizations of the statistical model (3) are written

y(t)=f(t;θ0)+ε(t),t[t0,tf].

When collecting experimental data, it is often difficult to take continuous measurements of the observed variables. Instead, we assume that we have n observations at times tj, j = 1, …, n, t0t1 < t2 < …< tntf. We then write the observation process (2) as

f(tj;θ)=Cx(tj;θ),j=1,2,,n, (4)

the discrete statistical model as

Yj=f(tj;θ0)+E(tj),j=1,2,,n, (5)

and a realization of the discrete statistical model as

yj=f(tj;θ0)+ε(tj),j=1,2,,n.

If we were given θ⃗0, we could solve (1) for x⃗ (t; θ⃗0), a process known as solving the forward problem. Alternatively, if we had a set of data y⃗j, j = 1, 2, …, n, we could estimate θ⃗0 in a process known as solving the inverse problem. We will use this mathematical and statistical framework to develop a methodology to identify sampling variables that provide the most information pertinent to estimating a given set of parameters and the most informative times at which the samples should be taken.

2.2 Formulation of the Optimal Design Problem

Several methods exist to solve the inverse problem. A major factor [10] in determining which method to use is additional assumptions made about Inline graphic(t). It is common practice to make the assumption that realizations of Inline graphic(t) at particular time points are independent and identically distributed (i.i.d.). If, additionally, the distributions describing the behavior of the components of Inline graphic(t) are known, then maximum likelihood methods may be used to find an estimate of θ⃗0. On the other hand, if the distributions for Inline graphic(t) are not known but the variance V0(t) (also unknown) is assumed to vary over time, weighted least squares methods are often used. We propose an optimal design problem formulation using a generalized weighted least squares criterion.

Let Inline graphic([t0, tf]) denote the set of all bounded distributions on the interval [t0, tf]. We consider the generalized weighted least squares cost functional for systems with vector output

JWLS(y,θ)=t0tf[y(t)-f(t;θ)]TV0-1(t)[y(t)-f(t;θ)]dP1(t), (6)

where P1(t) ∈ Inline graphic([t0, tf]) is a general measure on the interval [t0, tf]. For a given continuous data set y⃗(t), we search for a parameter θ̂ that minimizes JWLS(y⃗, θ⃗).

We next consider the case of observations collected at discrete times. If we choose a set of n time points τ = {tj}, j = 1, 2, …, n, where t0t1 < t2 < …< tntf and take

P(t)=Pτ=j=1nδtj, (7)

where δa represents the Dirac delta distribution with atom at a, then the weighted least squares criterion (6) for a finite number of observations becomes

JWLSn(y,θ)=j=1n[y(tj)-f(tj;θ)]TV0-1(tj)[y(tj)-f(tj;θ)].

To select a useful distribution of time points and set of observation variables, we introduce the N by p sensitivity matrices [f(t;θ)θ] and the m by p sensitivity matrices [x(t;θ)θ] that are determined using the differential operator in row vector form (θ1, θ2, …, θp) represented by ∇θ⃗ and the observation operator defined in (2),

θf(t,θ)=f(t;θ)θ=Cx(t;θ)θ=C(x1(t;θ)θ1x1(t;θ)θ2x1(t;θ)θpx2(t;θ)θ1x2(t;θ)θ2x2(t;θ)θpxm(t;θ)θ1xm(t;θ)θ2xm(t;θ)θp)=Cθx(t;θ). (8)

Using the sensitivity matrix ∇θ⃗f (t, θ⃗0), we may formulate the Generalized Fisher Information Matrix (GFIM). Consider the set Inline graphic ⊂ ℝm of admissible observation maps and let Inline graphic( Inline graphic) represent the set of all bounded distributions P2(c) on Inline graphic. Then the GFIM may be written

F(P1,P2,θ0)t0tfC1σ2(t,c)θf(t;θ0)θf(t;θ0)dP2(c)dP1(t) (9)
=t0tfC1σ2(t,c)θ(cx(t;θ0))θ(cx(t;θ0))dP2(c)dP1(t). (10)

Taking N different sampling maps in Inline graphic represented by the m-dimensional row vectors ck, k = 1, 2, …, N, we construct the discrete distribution on Inline graphic

PC=k=1Nδck, (11)

where δa represents the Dirac delta distribution with atom at a. Using PC in (10), we obtain the GFIM for multiple discrete observation methods taken continuously over [t0, tf],

F(P1,PC,θ0)=t0tfk=1N1σ2(t,ck)θ(ckx(t;θ0))θ(ckx(t;θ0))dP1(t)=t0tfk=1N1σ2(t,ck)θx(t;θ0)ckckθx(t;θ0)dP1(t)=t0tfk=1Nθx(t;θ0)ck1σ2(t,ck)ckθx(t;θ0)dP1(t)=t0tfθx(t;θ0)k=1N(ck1σ2(t,ck)ck)θx(t;θ0)dP1(t)=t0tfθx(t;θ0)(CV0-1(t)C)θx(t;θ0)dP1(t), (12)

where C = (c1, c2, …, cN) ∈ ℝN×m is the observation operator in (2) and (4) and V0(t) ∈ ℝN×N is the covariance matrix as described in (3). Applying the distribution Pt as described in (7) to the GFIM (12) for discrete observation operators measured continuously yields the discrete p × p Fisher Information Matrix (FIM) for discrete observation operators measured at discrete times

F(τ,C,θ0)=F(Pτ,PC,θ0)=j=1nθx(tj;θ0)CV0-1(tj)Cθx(tj;θ0). (13)

This describes the amount of information about the p parameters of interest that is captured by the observed quantities described by the sampling maps ck, k = 1, 2, …, N, listed in C, when they are measured at the time points in τ.

The questions of determining the best (in some sense) C and τ are important questions in the optimal design of an experiment. Recall that the set of time points τ has an associated distribution P1(τ) = PτInline graphic ([t0, tf]), where Inline graphic([t0, tf]) is the set of all bounded distributions on [t0, tf]. Similarly, the set of sampling maps {ck} has an associated bounded distribution PC Inline graphic ( Inline graphic). Define the space of bounded distributions Inline graphic([t0, tf] × Inline graphic) = Inline graphic([t0, tf]) × Inline graphic( Inline graphic) with elements P = (Pτ, PC) Inline graphic. Without loss of generality, assume that Inline graphic ⊂ ℝm is closed and bounded, and assume that there exists a functional Inline graphic: ℝp×p+ of the GFIM (10). Then the optimal design problem associated with Inline graphic is selecting a distribution Inline graphic such that

J(F(P^,θ0))=minPPJ(F(P,θ0)), (14)

where Inline graphic depends continuously on the elements of Inline graphic(P, θ⃗0).

The Prohorov metric [16] provides a general theoretical framework for the existence of and approximation in Inline graphic([t0, tf] × Inline graphic) (a general theoretical framework is developed in [7, 11]). The application of the Prohorov metric to optimal design problems formulated as (14) is explained more fully in [11]: briefly, define the Prohorov metric ρ on the space Inline graphic([t0, tf] × Inline graphic), and consider the metric space ( Inline graphic([t0, tf] × Inline graphic), ρ). Since [t0, tf] × Inline graphic is compact, ( Inline graphic([t0, tf] × Inline graphic), ρ) is also compact. Additionally, by the properties of the Prohorov metric, (P([t0, tf]×C), ρ) is complete and separable. Therefore an optimal distribution exists and may be approximated by a discrete distribution.

The formulation of the cost functional (14) may take many forms. We focus on the use of traditional optimal design methods, D-optimal, E-optimal, or SE-optimal design criteria, to determine the form of Inline graphic. Each of these design criteria are functions of the inverse of the FIM (assumed hereafter to be invertible) defined in (13).

In D-optimal design, the cost functional is written

JD(F)=det(F(τ,C,θ0)-1)-1det(F(τ,C,θ0)).

By minimizing Inline graphic, we minimize the volume of the confidence interval ellipsoid describing the uncertainty in our parameter estimates. Since F is symmetric and positive semi-definite, Inline graphic(F) ≥ 0. Additionally, since F is assumed invertible, Inline graphic(F) ≠ 0, therefore, Inline graphic: ℝp×p → ℝ+.

In E-optimal design, the cost functional is Inline graphic is the largest eigenvalue of (F(τ, C, θ⃗0))−1, or equivalently

JE(F)=max1eig(F(τ,C,θ0)).

To obtain a smaller standard error, we must reduce the length of the principal axis of the confidence interval ellipsoid. Since an eigenvalue λ solves det(FλI) = 0, an eigenvalue of λ = 0 would mean det(F) = 0, or that F is not invertible. Since F is positive definite, all eigenvalues are therefore positive. Thus Inline graphic: ℝp×p ∈ ℝ+.

In SE-optimal design, Inline graphic is a sum of the elements on the diagonal of (F(τ, C, θ⃗0))−1 weighted by the respective parameter values [11, 12], written

JSE(F)=i=1p(F(τ,C,θ0)))i,i-1θ0,i2.

Thus in SE-optimal design, the goal is to minimize the sum of squared errors of the parameters normalized by the true parameter values. As the diagonal elements of F−1 are all positive and all parameters are assumed non-zero in θ⃗ ∈ ℝp, Inline graphic: ℝp×p ∈ ℝ+.

In [12], it is shown that the D-, E-, and SE-optimal design criteria select different time grids and yield different standard errors. We expect that these design cost functionals will also choose different observation variables (maps) in order to minimize different dimensions of the confidence interval ellipsoid.

3 Algorithm and optimization constraints

In most optimal design problems, there is not a continuum of measurement possibilities that may be used; rather, there are N* < ∞ possible observation maps c. Denote this set as Inline graphic ⊂ ℝm. While we may still use the Prohorov metric-based framework to guarantee existence and convergence of (14), we have a stronger result first proposed in [5] that is useful in numerical implementation. Because Inline graphic is finite, all probability distributions made from the elements of Inline graphic have the form P2({ck})=PC=k=1Nδck for a fixed N. Moreover, the set of all distributions that use N sampling methods P2N(CN) is also finite. For a fixed distribution of time points Pτ, we may compute using (13) the set of all possible FIM F (Pτ, PC, θ⃗) that could be formulated from cInline graphic. By the properties of matrix multiplication and addition, this set is also finite. Then the functional (14) applied to all F in the set produces a finite set contained in ℝ+. Because this set is finite, it is well-ordered by the relation ≤ and therefore has a minimal element. Thus for any distribution of time points Pτ, we may find at least one solution P^CP2N(CN). Moreover, C may be determined by a search over all matrices C = (c1, c2, …, cN ) formed by N elements from Inline graphic. Therefore, for a fixed Pτ and NN* < ∞ a global optimal set of N sampling methods may be determined.

Due to the computational demands of performing nonlinear optimization for n time points and N observation maps (for a total of n + N dimensions), and to the difference in techniques between searching for an optimal PC in the finite set Inline graphic and searching for an optimal distribution of n sampling times, we instead solve the coupled set of equations

C^=argmin{CPCP2N(CN)}J(F(τ^,C,θ0)) (15)
τ^=argmin{τPτP1([t0,tf])}J(F(τ,C^,θ0)), (16)

where C ∈ ℝN×m represents a set of N sampling maps and τ={tj}j=1n, t0t1 < t2 < … < tntf, is an ordered set of n sampling times. These equations are solved iteratively as

C^i=argmin{CPCP2N(CN)}J(F(τ^i-1,C,θ0)) (17)
τ^i=argmin{τPτP1([t0,tf])}J(F(τ,C^i,θ0)), (18)

where Inline graphic is the D-, E-, or SE-optimal design criterion. We begin by solving for Ĉ1 where τ̂0 is specified by the user. The system (17)–(18) is solved until | Inline graphic(F(τ̂i, Ĉi, θ⃗0) − Inline graphicF(τi−1, Ĉi−1, θ̂0) | < ε or until Ĉi = Ĉi−1. For each iteration, (17) is solved using a global search over all possible C. Since the sensitivity equations cannot be easily solved for in the models chosen to illustrate our method, we use a modified version of tssolve.m by A. Attarian [4], which implements the myAD package developed by M. Fink [14].

Solving (18) requires using a nonlinear constrained iterative optimization algorithm. While MATLAB’s fmincon is a natural choice for such problems, as reported in [12], it does not perform well in this situation. Instead, we use SolvOpt developed by A. Kuntsevich and F. Kappel [15] (which utilizes a modified version of Shor’s r-algorithm) to search for an optimal distribution local to an initial uniformly spaced time point distribution. There exist many types of constraints that may be placed on this optimization. The four constraints used in [12] are

  • (C1)

    Optimize all time points such that t0t1t2 ≤ … ≤ tntf. This method optimizes n time points.

  • (C2)

    The initial and final time points are fixed as t1 = t0 and tn = tf. The routine then optimizes over the remaining n − 2 time points such that tjtj+1.

  • (C3)

    Optimize the time steps νj ≥ 0. Fix t1 = t0 and tn = tf, and the remaining time points may be found by tj+1 = tj + νj, j = 1, 2, …, n − 2. Additionally, tf-t0j=1n-2νj. This routine also optimizes over n − 2 variables.

  • (C4)

    Optimize the time steps νj ≥ 0. Fix t1 = t0. The remaining time points may be found by tj+1 = tj + νj, j = 1, 2, …, n − 1 such that tf-t0=j=1n-1νj. This method optimizes over n − 1 variables.

We select for our use two constraints, (C2) and (C3), to reduce the complexity of the problem of time point distribution selection. Once either of the convergence requirements are met and Ĉ and are determined, we compute standard errors using the asymptotic theory described in the next section.

4 Standard Errors

In order to compare the ability of different optimal design criteria to minimize uncertainty in parameter estimation, we compute the standard errors associated with these parameters. We begin by selecting an ODE system, a nominal set of parameters θ⃗ that we would estimate, the start and end times of the experiment t0 and tf, the number N of sampling times n, and the number of observation maps we wish to use. After an optimal t and C are determined according to one of the three previously described optimal design methods, we compute the standard errors for the parameters in θ⃗. There are multiple techniques [12] available to compute standard errors; here we choose to use asymptotic theory due to its ease of implementation.

4.1 Asymptotic Theory for Standard Errors

If we assume that the covariance matrix V0(t) is constant over time (V0(t)V0=Var(E(tj))=diag(σ0,12,σ0,22,,σ0,N2)), then we may use an ordinary least squares (OLS) framework to estimate standard errors. Once an optimal τ={tj}j=1N, t0t1 < t2 < … < tntf, and C are determined, we obtain data from an experiment or simulate data {yj} as a realization of the random process {Yj} described in (5), and then we estimate the parameters in θ⃗ by solving the inverse problem using the OLS criterion [10]. The discrete OLS estimator is defined as

θOLS=argminθΘj=1n[Yj-f(tj;θ)]TV0-1[Yj-f(tj;θ)], (19)

where Θ is the set of all possible values of θ⃗, such that each element of the difference vector Y⃗jf⃗(tj; θ⃗) is weighted using the variance of its corresponding sampling maps. One realization of this problem using data y⃗j, j = 1, 2, …, n, is written

θ^OLS=argminθΘj=1n[yj-f(tj;θ)]TV0-1[yj-f(tj;θ)]. (20)

However, calculating θ̂OLS still requires the unknown V0. If the number of parameters p of a system is sufficiently small and number of observations n large so that p < n, then we may calculate the bias adjusted estimate of the variances

V0V^=diag(1n-pj=1n[yj-f(tj;θ)][yj-f(tj;θ)]T), (21)

and find the estimate of θ⃗0 using

θ0θ^OLS=argminθΘj=1n[yj-f(tj;θ)]TV^-1[yj-f(tj;θ)]. (22)

Therefore, finding θ⃗OLS when V0 is unknown requires solving the coupled system of equations (21) and (22).

We may utilize the asymptotic properties of the OLS minimizer (19) to learn about the behavior of the model (1) and (3). As the number of samples n → ∞, θ⃗OLS has the following properties [10, 13, 17]

θOLS~N(θ0,0n)N(θ^OLS,^n),

where

0n(j=1nχjT(θ0)V0-1χj(θ0))-1 (23)

is the p × p covariance matrix, and χj(θ)=χjn(θ)=θf(tj;θ0) is the N × p matrix

χj(θ)=χjn(θ)=(f1(tj;θ)θ1f1(tj;θ)θ2f1(tj;θ)θpfN(tj;θ)θ1fN(tj;θ)θ2fN(tj;θ)θp). (24)

The approximation Σ̂n to the covariance matrix 0n is

0n^n=(j=1nχjT(θ^OLS)V^-1χj(θ^OLS))-1. (25)

We may use Σ̂n to approximate the standard errors of each parameter in θ̂OLS. For the kth element of θ̂OLS, written θ̂OLS,k, the asymptotic standard error is

ASE(θ0,k)=^0,kknASE(θ^OLS,k)=^kkn,

where ^0,kkn is the element in the kth row and kth column of 0n, and ^kkn is the corresponding element in Σ̂n. Additionally, since the FIM is defined to be the inverse of the covariance matrix, we may approximate the FIM using (25) by F(τ, C, θ⃗0) ≈ F(τ̂, Ĉ, θ̂OLS) = (Σ̂n)−1.

5 Example 1: HIV model

After validating code results for time point selection on simple models such as those in [12], we examined the performance of both the time point and observable operator selection algorithms on the log-scaled version of the HIV model developed in [1]. While the analytic solution of this system cannot easily be found, this model has been studied, improved, and successfully fit to and indeed validated with several sets of longitudinal data using parameter estimation techniques [2, 3]. Additionally, the sampling or observation operators used to collect data in a clinical setting as well as the relative usefulness of these sampling techniques are known [2]. The model includes uninfected and infected CD4+ T-cells, called type 1 target cells (T1 and T1, respectively), uninfected and infected macrophages (subsequently determined to more likely be resting or inactive CD4+ T-cells), called type 2 target cells (T2 and T2), infectious free virus VI, and immune response E produced by cytotoxic T-lymphocytes CD8+. The HIV model with treatment function u(t) is given by

T.1=λ1-d1T1-(1-ε1u(t))k1VIT1T.2=λ2-d2T2-(1-fε1u(t))k2VIT2T.1=(1-ε1u(t))k1VIT1-δT1-m1ET1T.2=(1-fε1u(t))k2VIT2-δT2-m2ET2V.=NTδ(T1+T2)-[c+(1-ε2u(t))103k1T1+(1-fε1u(t))103k2T2]VIE.=λE+bET1+T2T1+T2+KbE-dET1+T2T1+T2+KdE-δEE. (26)

The log-scaled HIV model, which is used to alleviate difficulties due to the large differences in scales of the variables and the parameters, is

x.1=10-x1ln(10)(λ1-d110x1-(1-ε1u(t))k110x1+x5)x.2=10-x2ln(10)(λ2-d210x2-(1-fε1u(t))k210x5+x2)x.3=10-x3ln(10)((1-ε1u(t))k110x5+x1-δ10x3-m110x6+x3)x.4=10-x4ln(10)((1-fε1u(t))k210x5+x2-δ10x4-m210x6+x4) (27)
x.5=10-x5ln(10)((1-ε2u(t))103NTδ(10x3+10x4)-c10x5-(1-ε1u(t))ρ1103k110x1+x5-(1-fε1u(t))ρ2103k210x2+x5)x.6=10-x6ln(10)(λE+bE(10x3+10x4)10x3+10x4+Kb10x6-dE(10x3+10x4)10x3+10x4+Kd10x6-δE10x6), (28)

where T1 = 10x1, T2 = 10x2, T1=10x3,T2=10x4, VI = 10x3, and E = 10x6.

In [2] this model’s parameters are estimated for each of 45 different patients who were in a treatment program for HIV at Massachusetts General Hospital (MGH). These individuals experienced interrupted treatment schedules, in which the patient did not take any medication for viral load control. Seven of these patients adhered to a structured treatment interruption schedule planned by the clinician. We use the parameters estimated to fit the data of one of these patients, Patient 4 in [2], as our “true” parameters θ⃗0 for this model. This patient was chosen because the patient continued treatment for an extended period of time (1919 days) and the corresponding data set contains 158 measurements of viral load and 107 measurements of CD4 cell count ( T1+T1) that exhibit a response to interruption in treatment, and the estimated parameters yield a model exhibiting trends similar to that in the data.

Sensitivity analysis performed in [1] and parameter subset selection carried out in [8] identified subsets of the 20 parameters and six initial conditions that would likely be reliably estimated when solving the corresponding inverse problems. Based on their results, we treat the subset of parameters θ⃗ = (λ1, d1, ε1, k1, ε2, NT, c, bE, x10,x20,x50) as unknown and fix all other parameters. The values for the fixed parameters and θ⃗0 are computed in [2] and given in Table 1. It is important to choose parameters to which the model is sensitive; a poor choice for the components of θ⃗ will negatively affect sensitivity matrices and may lead to a near-singular FIM. Based on currently available measurement methods, we allow the possible types of observations including (1) infectious virus x5, (2) immune response x6, (3) total activated CD4 cells x1 + x3, and (4) type 2 (resting or unactivated) target cells x2 + x4, each with an assumed error variance of 10% of the initial variable values given by x⃗0 = (3.0799, 1.2443, 1.7899, −0.2150, 5.9984, −0.7251).

Table 1.

Values of parameters in the HIV model (26).

parameter value unit parameter value unit
λ1 4.633
cellsμml-blood·day
λ2 0.1001
cellsμml-blood·day
d1 0.004533
1day
d2 0.02211
1day
k1 1.976e-6
mlvirions·day
k2 5.529e-4
mlvirions·day
ε1 100.6017 - ε2 100.5403 -
m1 0.02439
μmlcells·day
m2 0.013099
μmlcells·day
δ 0.1865
1day
c 19.36
1day
f 100.53915 - NT 19.04
virionscell
λE 0.009909
cellsμml-blood·day
δE 0.0703
1day
bE 0.09785
1day
dE 0.1021
1day
Kb 0.3909
cellsμml-blood
Kd .8379
cellsμml-blood

5.1 HIV observable selection results, times fixed

To simulate the experience of a clinician gathering data as a patient regularly returns for scheduled testing, we fix the sampling times and choose the optimal observables. We consider choices of N = 1, 2, 3 sampling operators out of the four possible observables, all N of which will be measured at n = 51, 101, 201, 401 evenly spaced times over 2000 days, corresponding to measurements every 40, 20, 10, and 5 days, respectively. The N optimal sampling maps are determined for each of the three optimal design criteria, and the asymptotic standard errors (ASE) are calculated after carrying out the corresponding inverse problem calcualtions.

Tables 2, 3, and 4 display the optimal observation operators determined by the D-, E-, and SE-optimal design cost functions, respectively, as well as the lowest and highest ASE. In all three optimal design criteria, there was a distinct best choice of observables (listed in Tables 2, 3, and 4) for each pair of n and N. When only N = 1 observable could be measured, each design criterion consistently picked the same observable for all n; similarly, at N = 2, both the D-optimal and SE-optimal design criteria were consistent in their selection over all n, and E-optimal only differed at n = 401. Even at N = 3, each optimal design method specified at least two of the same observables at all n.

Table 2.

Number of observables, number of time points, observables selected by D-optimal cost functional, and the minimum and maximum standard error and associated parameter for the parameter subset in the HIV model (27).

N n Observables min(ASE) max(ASE)
1 51 x1 + x3 ASE(λ1) = 0.18 ASE(λE) = 6.35
1 101 x1 + x3 ASE(λ1) = 0.13 ASE(λE) = 4.40
1 201 x1 + x3 ASE(λ1) = 0.091 ASE(λE) = 3.03
1 401 x1 + x3 ASE(λ1) = 0.065 ASE(λE) = 2.13
2 51 x1 + x3, x2 + x4 ASE(λ1) = 0.095
ASE(x50)=2.31
2 101 x1 + x3, x2 + x4 ASE(λ1) = 0.068
ASE(x50)=1.52
2 201 x1 + x3, x2 + x4 ASE(λ1) = 0.047
ASE(x50)=1.26
2 401 x1 + x3, x2 + x4 ASE(λ1) = 0.033
ASE(x50)=1.13
3 51 x6, x1 + x3, x2 + x4 ASE(λE) = 0.045
ASE(x50)=2.26
3 101 x6, x1 + x3, x2 + x4 ASE(λE) = 0.032
ASE(x50)=1.43
3 201 x6, x1 + x3, x2 + x4 ASE(λE) = 0.022
ASE(x50)=1.23
3 401 x6, x1 + x3, x2 + x4 ASE(λE) = 0.015
ASE(x50)=1.12

Table 3.

Number of observables, number of time points, observables selected by E-optimal cost functional, and the minimum and maximum standard error and associated parameter for the parameter subset in the HIV model (27).

N n Observables min(ASE) max(ASE)
1 51 x5 ASE(d1) = 0.27 ASE(λE) = 4.27
1 101 x5 ASE(d1) = 0.19 ASE(λE) = 3.11
1 201 x5 ASE(d1) = 0.13 ASE(λE) = 2.27
1 401 x5 ASE(d1) = 0.094 ASE(λE) = 1.63
2 51 x5, x2 + x4 ASE(d1) = 0.12 ASE(λE) = 2.18
2 101 x5, x2 + x4 ASE(d1) = 0.095 ASE(λE) = 1.52
2 201 x5, x2 + x4 ASE(d1) = 0.065 ASE(λE) = 1.10
2 401 x5, x1 + x3 ASE(λ1) = 0.042 ASE(λE) = 0.86
3 51 x5, x6, x1 + x3 ASE(λE) = 0.045
ASE(x50)=0.77
3 101 x5, x6, x1 + x3 ASE(λE) = 0.032
ASE(x50)=0.73
3 201 x5, x6, x1 + x3 ASE(λE) = 0.022
ASE(x50)=0.69
3 401 x5, x1 + x3, x2 + x4 ASE(λ1) = 0.032
ASE(x50)=0.65

Table 4.

Number of observables, number of time points, observables selected by SE-optimal cost functional, and the minimum and maximum standard error and associated parameter for the parameter subset in the HIV model (27).

N n Observables min(ASE) max(ASE)
1 51 x5 ASE(d1) = 0.27 ASE(λE) = 4.27
1 101 x5 ASE(d1) = 0.19 ASE(λE) = 3.11
1 201 x5 ASE(d1) = 0.13 ASE(λE) = 2.27
1 401 x5 ASE(d1) = 0.094 ASE(λE) = 1.63
2 51 x1 + x3, x2 + x4 ASE(λ1) = 0.095
ASE(x50)=2.31
2 101 x1 + x3, x2 + x4 ASE(λ1) = 0.068
ASE(x50)=1.52
2 201 x1 + x3, x2 + x4 ASE(λ1) = 0.047
ASE(x50)=1.26
2 401 x1 + x3, x2 + x4 ASE(λ1) = 0.033
ASE(x50)=1.13
3 51 x6, x1 + x3, x2 + x4 ASE(λE) = 0.045
ASE(x50)=2.26
3 101 x6, x1 + x3, x2 + x4 ASE(λE) = 0.032
ASE(x50)=1.49
3 201 x5, x6, x1 + x3 ASE(λE) = 0.022
ASE(x50)=0.69
3 401 x5, x6, x1 + x3 ASE(λE) = 0.015
ASE(x50)=0.67

The observables that were rated best changed between criteria, affirming the fact that each optimal design methods minimizes different aspects of the standard error ellipsoid. At N = 1 observable, D-optimal selects the CD4 cell count while E-optimal and SE-optimal choose the infectious virus count. As a result, the min(ASE) calculated for a parameter estimation problem using the D-optimal observables is approximately 1/3 lower than the min(ASE) of E- and SE-optimal for all tested time point distributions. Similarly, the max(ASE) calculated for E- and SE-optimal designed parameter estimation problems is approximately 1/3 lower than that of D-optimal. Thus at N = 1, based on minimum and maximum asymptotic standard errors, there is no clear best choice of an optimal design cost function.

At N = 2 allowed observables, both the D- and SE-optimal cost functions are minimized by an observation operator containing both activated CD4 cells (type 1 target cells) and type 2 target cells. The E-optimal cost function still favors infectious virus count in addition to type 2 target cells (at n = 51, 101, 201) or type 1 target cells (at n = 401). As a result, max(ASE) in the E-optimal design parameter estimation problem is 0% – 20% lower than that of D- or SE-optimal; however, the D- and SE-optimal designs reduce min(ASE) by 20% – 30% of the min(ASE) of E-optimal. Therefore at N = 2, the E-optimal cost functional would be preferable if the largest ASE’s were to be reduced, but for the best overall improvement (as measured by percent reduction from the E-optimal ASE), D- and SE-optimal are recommended.

When selecting N = 3 observables, each of the three design criteria select many of the same observables. This is to be expected as N* = 4 in this simulation. For n = 51, 101, 201, both total CD4 cell count and immune response E(t) are selected by all design criteria. The D-optimal criterion also chooses type 2 cell count, so the lack of information on virus count as measured by x5(t) leads to its high max(ASE), ASE(x50). E-optimal (and at larger n, SE-optimal) choose to measure infectious virus count, reducing ASE(x50) and thus reducing the max(ASE) by more than 50%. While at low n, E-optimal has the lowest min(ASE) and max(ASE), SE-optimal performs better at high n, so when selecting N = 3 observables, the number of time points n may affect which optimal design cost function performs best.

5.2 HIV observables and time point selection results

When taking samples on a uniform time grid, D-, E-, and SE-optimal design criteria all choose observation operators that yield favorable ASE’s, with some criteria performing best under certain circumstances. For example, the SE-optimal observables at n = 401, N = 3 yield the smallest standard errors; however, for all other values of n at N = 3, E-optimal performs best. At N = 2, E-optimal is a slightly weaker scheme. The examples in [12] also reveal that D-, E-, and SE-optimal designs are all competitive when only selecting time points for several different models. Now we wish to investigate the performance of these three criteria when selecting both an observation operator and a sampling time distribution using the algorithm described by equations (17) and (18).

To maintain consistency across trials while slightly simplifying the parameter estimation problem, we allow the set of six parameters and three initial conditions θ⃗ = (λ1, d1, k1, NT, c, bE, x10,x20,x50) to be treated as unknowns and fix all other parameters. We again allow the possible observations of (1) infectious virus x5, (2) immune response x6, (3) CD4 cells x1 + x3, and (4) type 2 target cells x2 + x4, each with an assumed error variance of 5% of the initial variable values given by x⃗0. To reflect the often limited resources of a clinical trial, we allow N = 2 observation maps to be included in the observation operators C and examine the distribution of time points if n = 35 or n = 105 samples (consisting of all observables in the observation operator) which may be taken from t0 = 0 through tf = 1460. We begin all simulations with uniformly spaced sample times, and use either time grid constraint C2 or C3. Both constraints assume that samples are taken at t0 and tf, so we in effect are optimizing the remaining 33 or 103 observation times.

When choosing N = 2 observables and distributing n = 35 time points using constraint C2, the D-optimal cost function yields the lowest ASE for the parameters (λ1, d1, NT, c, bE, x10), but the SE-optimals ASE are on the same order of magnitude (Table 5). Both the D- and SE-optimal cost functions selected the same observables, and they both were minimized by similar distributions of time points (Figure 1). The E-optimal design leads in the smallest ASE for (k1, x20,x50), with an ASE(x50) that is half that of D-optimal and SE-optimal. E-optimal, however, trails D- and SE-optimal in the accuracy of an estimate for bE, with an ASE(bE) that is larger by one order of magnitude. The large difference in ASE’s for these two parameters may be related to the observables selected by each design criterion: E-optimal design chooses x5 (the variable most closely related to x50) as an observable when D- and SE-optimal choose x6 (whose right hand side contains bE). All three also chose x2 + x4. The distribution of time points selected under each optimal design criterion focuses on regions very soon before or after a change in behavior of the selected observables (Figure 1). This may indicate that fewer sampling points may be needed for parameter estimates of similar accuracy. For the experimental design constraints of 35 time point allocated using constraint C2, the D-optimal cost functional is best for most parameters, but E-optimal is best for some parameters.

Table 5.

Approximate asymptotic standard errors calculated by asymptotic theory (25) using optimally spaced n = 35 time points under constraints C2 (left set of 3) and C3 (right set of 3) and optimally selected N = 2 observables for parameters of interest θ⃗ of the HIV model (27). Smallest ASE per parameter is highlighted using bold font.

Time C2 Optimal C3 Optimal

Method D E SE D E SE

Observe x6, x2 + x4 x5, x2 + x4 x6, x2 + x4 x6, x2 + x4 x5, x2 + x4 x6, x2 + x4
ASE(λ1) 0.1077 0.2501 0.1164 0.1048 0.1297 0.1215
ASE(d1) 0.0737 0.1523 0.0972 0.0683 0.0759 0.0817
ASE(k1) 0.0967 0.0851 0.1404 0.1085 0.0920 0.1174
ASE(NT ) 0.2200 0.2722 0.2263 0.2257 0.2784 0.2398
ASE(c) 0.2455 0.2916 0.2505 0.2517 0.3098 0.2650
ASE(bE) 0.0480 0.7788 0.0500 0.0476 0.8182 0.0494
ASE(x10)
0.1148 0.1275 0.1366 0.1201 0.1081 0.1292
ASE(x20)
0.2657 0.2456 0.2597 0.2635 0.2469 0.2654
ASE(x50)
0.9450 0.4798 0.9234 0.9303 0.4629 0.9473

Figure 1.

Figure 1

Solution of (27) plotted using the observables log(10x1+10x3) (upper left), log(10x2+10x4) (upper right), x5 (lower left), and x6 (lower right). Plotted on top of each curve are the D-optimal (circle), E-optimal (square), and SE-optimal (x) n = 35 times under constraint C2.

The ASE calculated for choosing N = 2 observables and distributing n = 35 time points using constraint C3 are very similar (Table 5), and the selected observables remain unchanged; however, the optimal distribution of time points under each design criterion (Figure 2) is more uniform than when using constraint C2 (Figure 1). For small n with N = 2 observables under constraint C3, D-optimal design is best for most parameters, but E-optimal is best for the remaining few. While the choice of a particular time point distribution constraint does not greatly impact the calculated ASEs for the paramters of interest, the constraint C3 yields a measurement schedule that may be more feasible for a patient to follow in that it does not call for as many days of multiple observations.

Figure 2.

Figure 2

Solution of (27) plotted using the observables log(10x1+10x3) (upper left), log(10x2+10x4) (upper right), x5 (lower left), and x6 (lower right). Plotted on top of each curve are the D-optimal (circle), E-optimal (square), and SE-optimal (x) n = 35 times under constraint C3.

Choosing N = 2 observables and distributing n = 105 time points using constraint C2 leads to a different pattern in which optimal design criterion is best for which parameter. For the more dense time point distribution, SE-optimal design is a much stronger candidate against D- and E-optimal (Table 6). It yields the lowest ASE’s for (λ1, NT, c, bE), while E-optimal is best for (k1, x10,x20,x50) and D-optimal is best for (d1, bE). The D- and SE-optimal cost functions again choose the same observables of x6 and x2 + x4, so their ASE’s for all parameters are similar. In this scenario, the E-optimal design has the largest percent reduction in ASE from those of D- and SE-optimal for the parameters x10 and x50. E-optimal also changed its selected observables to x1 + x3 and x2 + x4. The distribution of time points for the D-optimal design criteria appear very close to uniform (Figure 3), and the distributions for E- and SE-optimal are clustered near local maxima, local minima, and other large changes in behavior of the observables; however, even slight differences between the distributions of the E- and SE-optimal costs functions are visible. Consider the graph of infectious virus count in Figure 3. There is a cluster of SE-optimal times between t = 1100 and t = 1250, but no E-optimal times occur during that interval. Very soon after, between t = 1250 and t = 1400, there is a cluster of E-optimal times but no SE-optimal times. These clusters may indicate that either these periods in the patient’s treatment history are key to characterizing the parameters or that fewer time points may be adequate to obtain sufficiently accurate parameter estimates.

Table 6.

Approximate asymptotic standard errors calculated by asymptotic theory (25) using optimally spaced n = 105 time points under constraints C2 (left set of 3) and C3 (right set of 3) and optimally selected N = 2 observables for parameters of interest θ⃗ of the HIV model (27). Smallest ASE per parameter is highlighted using bold font.

Time C2 Optimal C3 Optimal

Method D E SE D E SE

Observe x6, x2 + x4 x1 + x3, x2 + x4 x6, x2 + x4 x6, x2 + x4 x1 + x3, x2 + x4 x6, x2 + x4
ASE(λ1) 0.0687 0.0618 0.0607 0.0674 0.0475 0.0729
ASE(d1) 0.0429 0.0541 0.0470 0.0441 0.0395 0.0475
ASE(k1) 0.0658 0.0360 0.0681 0.0629 0.0388 0.0739
ASE(NT ) 0.1599 0.1437 0.1394 0.1192 0.1351 0.1423
ASE(c) 0.1785 0.1591 0.1551 0.1325 0.1500 0.1578
ASE(bE) 0.0281 0.4560 0.0281 0.0266 0.4637 0.0299
ASE(x10)
0.0810 0.0451 0.0793 0.0572 0.0490 0.0818
ASE(x20)
0.2537 0.2004 0.2296 0.1823 0.2244 0.2330
ASE(x50)
0.7042 0.4384 0.6825 0.0528 0.4491 0.6571

Figure 3.

Figure 3

Solution of (27) plotted using the observables log(10x1+10x3) (upper left), log(10x2+10x4) (upper right), x5 (lower left), and x6 (lower right). Plotted on top of each curve are the D-optimal (circle), E-optimal (square), and SE-optimal (x) n = 105 times under constraint C2.

The strength of SE-optimal design for large n does not hold for time grid constraint C3 (Table 6). E-optimal design provides the lowest ASE for the parameters (λ1, d1, k1, x10), D-optimal is best for (NT, c, bE, x20,x50), and the ASE calculated using the SE-optimal designed experiment is often the largest. The observables selected in this case are the same as the n = 105, constraint C2 case. The optimal time point distributions determined under all three design criterion are near uniform (Figure 4). Small differences between the distributions may be observed when the functions have a slope of high magnitude, indicating that the distributions are not exactly uniform. As the time point optimization routine is started with a uniform time point distribution, this may indicate that when samples are taken at many times during an experiment, a uniform distribution is somewhat near optimal for model (27) or, more significantly, that a uniform distribution is not an appropriate initial distribution to use to obtain the optimal time point distribution.

Figure 4.

Figure 4

Solution of (27) plotted using the observables log(10x1+10x3) (upper left), log(10x2+10x4) (upper right), x5 (lower left), and x6 (lower right). Plotted on top of each curve are the D-optimal (circle), E-optimal (square), and SE-optimal (x) n = 105 times under constraint C3.

Between D- and E-optimal, there is no clear leader for optimal time point and observable selection in these examples for system (27). The SE-optimal criterion is useful in estimating some parameters under the constraint C2, and in most cases yields standard errors on the same order of magnitude of the leading optimal design criterion. The optimal design algorithm successfully identifies observables and a time distribution that optimizes an aspect of the Fisher Information Matrix as determined by the choice of optimal design cost functional; however, the time point distributions exhibit tendencies to cluster near times when the slopes of function solutions are changing. We proceed to examine the algorithm’s performance when presented with a system that allows many possible observables and does not require as many sampling times.

6 Example 2: Zhu’s Calvin Cycle model

The second model we use as an example is characteristic of the large differential equation systems that often appear in industrial problems. In [19], Zhu et al., present an ODE model for the Calvin Cycle in fully grown spinach. The Calvin Cycle, part of the light-independent reactions in photosynthesis, plays an important role in plant carbon fixation, which leads to the growth of the plant. This model contains 165 parameters and initial conditions, 31 ODEs, and 7 concentration balance laws and involves 38 state variables (metabolite concentrations in different parts of the cell) as well as a calculation for photosynthetic CO2 uptake rate. The metabolites used in the model are denoted by RuBP, PGA, DPGA, T3P, FBP, E4P, S7P, ATP, SBP, NADPH, HexP, PenP, NADHc, NADc, ADPc, ATPc, GLUc, KGc, ADP in photorespiration, ATP in photorespiration, GCEA, GCA, PGCA, GCAc, GOAc, SERc, GLYc, HPRc, GCEAc, T3Pc, FBPc, HexPc, F26BPc, UDPGc, UTPc, SUCP, SUCc, and PGAc, and the parameters are mainly initial conditions, maximum reaction velocities, and Michaelis-Menten constants for reaction substrates, products, activators, and inhibitors. The ‘c’ following some metabolite names indicates that the model compartment corresponds to the metabolite concentration in cytosol; compartment names lacking a ‘c’ are the metabolite in the chloroplast stroma. The full system of equations and parameter values may be found in the appendices of [19]. While the model has not been validated with data as completely as the family of HIV models discussed above, it is representative of the models used to describe plant enzyme kinetics and utilizes well-documented Michaelis-Menten enzyme kinetic model formulations [6].

In [5], sets of the optimal 3, 5, 10, and 15 metabolites are identified as the most useful to measure in an experiment in order to estimate a subset of 6 parameters, θ⃗a = [KM11, KM521, KI523, KC, KM1221, KM1241] with true values θ⃗a0 = [0.0115, 0.0025, 0.00007, 0.0115, 0.15, 0.15], and a subset of 18 parameters, θ⃗c =[RuBP0, SBP0, KM11, KM13, KI13, KE4, KM9, KM131, KI135, KE22, KM511, KM521, KI523, KC, KM1221, KM1241, V9, V58] with true values θ⃗c0 = [2, 0.3, 0.0115, 0.02, 0.075, 0.05, 0.05, 0.05, 0.4, 0.058, 0.02, 0.0025, 0.00007, 0.0115, 0.15, 0.15, 0.3242, 0.0168]. The simulation was set in the framework of an experiment run for 3000 seconds over which 11 samples are taken at evenly spaced times for the optimization of estimates for θ⃗a and 21 samples for estimation of θ⃗c.

We use a similar set up for our simulations in order to judge the ability of the time and variable selections to minimize the asymptotic standard errors of both θ⃗a and θ⃗c using N = 5 and N = 10 observables and n = 11 time points over the time interval t ∈ [0, 3000]. Each metabolite is assigned a variance of 5% of its initial value, and to reduce computation time, all metabolites that do not change in concentration over time (dx/dt = 0) are excluded from the search algorithm. As it is often possible to measure a particular metabolite’s concentration in plant tissue, we allow Inline graphic to be composed of 28 vectors in ℝ1×40 that are composed of a one in the element corresponding to the position of the differential equation describing the dynamics of the metabolite in the vector of model ODEs and zero elsewhere.

We compare the ASE calculated for n = 11 evenly spaced measurements over the interval t ∈ [0, 3000] against the optimally spaced n = 11 time points according to constraints C2 and C3 and plot the D-, E- and SE-optimal time point distributions on the solutions of selected compartments in the Zhu model of [19]. The four compartments chosen for graphing are carbon uptake rate A(t) (units μmol m−2s−1), which is a rate calculated from the output of the model that describes the plant’s efficiency in using the available environmental resources to develop; adenosine triphosphate (ATP) in chloroplast stroma, which is a well-known metabolite active in photosynthesis; sucrose, a sugar, in cytosol (SUCc); and Ribulose-1,5-bisphosphate (RuBP), a metabolite connected to the enzyme RuBisCO, which is essential to carbon fixation.

The simplest scenario we test is the selection of the optimal N = 5 observables (metabolites) and n = 11 time points to use when estimating the six parameters in θ⃗a. For all three optimal design methods, the optimal observables determined under the uniform grid were also determined to be optimal after the time point distribution is optimized under constraints C2 and C3. These observables are listed in the upper portion of Table 7. All three optimal design methods identify the observables PGA, SERc, and F26BPc, indicating that these three metabolites may be central to accurate estimates of the parameters in θ⃗a; moreover, the E-optimal and SE-optimal cost functions selected the same set of five observables.

Table 7.

Top: Optimal 5 observables chosen by each optimal design criterion when estimating the parameters θ⃗a of the model in [19]. Bottom: Approximate asymptotic standard errors calculated using asymptotic theory (25) for each parameter in θ⃗a using the optimal 5 observables with time point constraints of uniform spacing, constraint C2, and constraint C3. Smallest ASE for each parameter per time point constraint is highlighted in bold font.

Method Observables

D-opt PGA, T3P, GOAc, SERc, F26BPc
E-opt PGA, SERc, T3Pc, FBPc, F26BPc
SE-opt PGA, SERc, T3Pc, FBPc, F26BPc
Time Uniform C2 Optimal C3 Optimal

Method D E SE D E SE D E SE
ASE(KM11) 0.0014 0.0356 0.0356 0.0015 0.0033 8.7e-4 5.8e-4 0.0018 0.0018
ASE(KM521) 4.1811 4.0086 4.0086 0.1324 0.1078 0.6805 0.0750 0.1349 0.0532
ASE(KI523) 0.1173 0.1125 0.1125 0.0037 0.0030 0.0191 0.0012 0.0038 0.0015
ASE(KC) 0.0012 0.0353 0.0353 0.0012 0.0033 0.0019 0.0011 0.0018 0.0018
ASE(KM1221) 0.2805 1.3065 1.3065 0.2742 0.4330 0.5798 0.1460 0.5127 0.5136
ASE(KM1241) 0.2508 1.1764 1.1764 0.2451 0.3874 0.5183 0.1305 0.4582 0.4592

The similarity in results, however, does not continue through the selected time point distributions. Under constraint C2 (Figure 5), the D-optimal distribution is loosely clustered about the center of the time interval, the E-optimal time points are clustered about t = 250 seconds, and SE-optimal chooses a small cluster of time points around the initial bump and allows a few samples after the function reaches a steady state. Using constraint C3 (Figure 6), all three optimal design criterion chose a majority of their sampling times before t = 600 and allow only a few sampling times after the system reaches a steady state. The optimization of time point distributions yields improved asymptotic standard errors from those of the uniform distribution - sometimes by an order of magnitude or more. For all three time point constraints, D-optimal yields the smallest ASE’s for the most number of parameters, and SE-optimal yields the smallest ASE’s for the second most number of parameters. Both optimal design criteria perform better with the C3 optimal times than the C2 optimal times. Therefore, in this simple case, using either the D- or SE-optimal design criterion with time point distribution constraint C3 would yield the best results.

Figure 5.

Figure 5

Solutions of selected state variables in the Zhu model [19]. Plotted on top of each curve are the D-optimal (circle), E-optimal (square), and SE-optimal (x) n = 11 sampling times under constraint C2 when sampling the optimal 5 observables to estimate θ⃗a (Table 7). Top Left: Carbon uptake rate A(t); Top Right: ATP; Bottom Left: SUCc; Bottom Right: RuBP.

Figure 6.

Figure 6

Solutions of selected state variables in the Zhu model [19]. Plotted on top of each curve are the D-optimal (circle), E-optimal (square), and SE-optimal (x) n = 11 sampling times under constraint C3 when sampling the optimal 5 observables to estimate θ⃗a (Table 7). Top Left: Carbon uptake rate A(t); Top Right: ATP; Bottom Left: SUCc; Bottom Right: RuBP.

The next scenario is the selection of the optimal N = 10 observables and n = 11 time points to use when estimating the six parameters in θ⃗a. For all three optimal design methods, the optimal observables determined under the uniform grid were also determined to be optimal after the time point distribution is optimized under constraints C2 and C3. These observables are listed in the upper portion of Table 8. Of the 28 possible observables, 20 were selected by at least one optimal design criterion, 8 of of which (DPGA, T3P, E4P, S7P, ATP, GCA, SERc, T3Pc) were selected by two criteria and one, GOAc, was selected by all three. The effect of adding five observables does not have a large effect on the estimated ASE for the six parameters of interest. The ASE’s listed in Table 8 for the N = 10 observable case are on the same order as those in Table 7 for the N = 5 observable case for each time point constraint.

Table 8.

Top: Optimal 10 observables chosen by each optimal design criterion when estimating the parameters θ⃗a of the model in [19]. Bottom: Approximate asymptotic standard errors calculated using asymptotic theory (25) for each parameter in θ⃗a using the optimal 10 observables with time point constraints of uniform spacing, constraint C2, and constraint C3. Smallest ASE for each parameter per time point constraint is highlighted in bold font.

Method Observables

D-opt RuBP, PGA, T3P, E4P, SBP, GOAc, SERc, GLYc, T3Pc, F26BPc
E-opt DPGA, T3P, S7P, ATP, NADPH, NADHc, GCEA, GCA, GOAc, HexPc
SE-opt DPGA, FBP, E4P, S7P, ATP, HexP, GCA, GOAc, SERc, T3Pc
Time Uniform C2 Optimal C3 Optimal

Method D E SE D E SE D E SE
ASE(KM11) 0.0011 0.0013 0.0013 5.9e-4 0.0013 0.0013 0.0011 0.0013 0.0013
ASE(KM521) 4.0426 3.9710 3.9710 0.1104 0.7815 0.0588 0.1473 0.2715 0.0588
ASE(KI523) 0.1134 0.1114 0.1114 0.0031 0.0221 0.0016 0.0041 0.0076 0.0016
ASE(KC) 0.0009 0.0010 0.0010 9.1e-4 0.0010 0.0010 9.5e-4 0.0010 0.0010
ASE(KM1221) 0.2801 0.6740 0.6740 0.2531 0.6205 0.5808 0.3045 0.6136 0.6961
ASE(KM1241) 0.2504 0.6027 0.6027 0.2264 0.5548 0.5193 0.2722 0.5481 0.6227

The optimal time point distributions also differ from those selected under the five observable case. Under constraint C2 (Figure 7), the D-optimal distribution is loosely clustered near the boundaries of the time interval while the E- and SE-optimal time points are scattered over the whole interval. Using constraint C3 (Figure 8), the D-optimal times are spread throughout the middle half of the time interval, E-optimal over the first half, and SE-optimal over the second half. While there are no easily discernible patterns between the distributions chosen under the C2 and C3 constraints, the optimization of time point distributions again yield improved asymptotic standard errors from those of a uniform time distribution. For all three time point constraints, D-optimal yields the smallest ASE’s for the most number of parameters, and SE-optimal yields the smallest ASE’s for the second most number of parameters. Other than for the parameters it estimates best, SE-optimal is often the worst criterion. The time point distribution under constraint C2 allows smaller ASE’s for the N = 10 observable variables case. Therefore using the D-optimal design criterion with time point distribution constraint C2 would yield the best results.

Figure 7.

Figure 7

Solutions of selected state variables in the Zhu model [19]. Plotted on top of each curve are the D-optimal (circle), E-optimal (square), and SE-optimal (x) n = 11 sampling times under constraint C2 when sampling the optimal 10 observables to estimate θ⃗a (Table 8). Top Left: Carbon uptake rate A(t); Top Right: ATP; Bottom Left: SUCc; Bottom Right: RuBP.

Figure 8.

Figure 8

Solutions of selected state variables in the Zhu model [19]. Plotted on top of each curve are the D-optimal (circle), E-optimal (square), and SE-optimal (x) n = 11 sampling times under constraint C3 when sampling the optimal 10 observables to estimate θ⃗a (Table 8). Top Left: Carbon uptake rate A(t); Top Right: ATP; Bottom Left: SUCc; Bottom Right: RuBP.

The addition of five observables for θ⃗a does not greatly impact most of the calculated ASE’s. For the time point constraints C2 and C3, all but one of the minimum ASE’s remain on the same order of magnitude, and some of the ASE’s increase when N = 10 observables are allowed from when N = 5 observables are allowed. This may indicate that for a small parameter set such as θ⃗a, only a small amount of information is needed to obtain the best possible results; adding extra information (through additional observables) does not further improve the ASE’s.

Now consider the parameter vector θ⃗c. We identify the optimal N = 5 observables and the optimal distribution of n = 11 time points using the D-, E-, and SE-optimal cost functionals. Three metabolites, RuBP, SBP, and F26BPc, are selected by all three cost functionals, indicating that these three observables may be important to measure to accurately estimate a large number of the parameters in θ⃗c. The inclusion of RuBP0 and SBP0 in θ⃗c may also heavily influence the selection of observables. Many of the other metabolites selected by at least one of the cost functionals, namely, PGA, SERc, and T3Pc, were also selected for θ⃗a in the experimental setup of N = 5 observables and n = 11 time points, indicating that they are still relevant to some of the parameters in θ⃗c.

Using the uniform time point distribution yields standard errors that are up to six order of magnitude larger than the parameter (KI523 = 0.00007) and are typically two or three orders of magnitude larger than the parameter value (Table 9). Constraint C2 also leads to large standard errors when using the D- and E-optimal cost function. The E-optimal standard errors, in fact, do not change from those calculated using the uniform time point distribution, and as shown in Figure 9, the E-optimal time distribution under constraint C2 is uniform. The SE-optimal design criterion performs the best when using constraint C2 and produces much smaller standard errors, though the standard errors are still larger than the parameter values by orders of magnitude for most parameters. Both D- and SE-optimal perform favorably under constraint C3: the D-optimal, C3 design produces smaller ASE’s than the SE-optimal, C2 design, and the SE-optimal, C3 design yields the smallest ASE’s for almost all parameters. Under the constraint C3 and the SE-optimal cost functional, the calculated ASE’s are zero to one orders of magnitude larger than the parameter values.

Table 9.

Top: Optimal 5 observables chosen by each optimal design criterion when estimating the parameters θ⃗c of the model in [19]. Bottom: Approximate asymptotic standard errors calculated using asymptotic theory (25) for each parameter in θ⃗c using the optimal 5 observables with time point constraints of uniform spacing, constraint C2, and constraint C3. Smallest ASE for each parameter per time point constraint is highlighted in bold font.

Method Observables

D-opt RuBP, PGA, SBP, SERc, F26BPc
E-opt RuBP, SBP, PenP, SERc, F26BPc
SE-opt RuBP, SBP, GCEAc, T3Pc, F26BPc
Time Uniform C2 Optimal C3 Optimal

Method D E SE D E SE D E SE
ASE(RuBP0) 0.3162 0.3162 0.3162 0.3162 0.3162 0.2236 0.2235 0.3162 0.2222
ASE(SBP0) 0.1225 0.1225 0.1225 0.1225 0.1225 0.0866 0.0866 0.1225 0.0863
ASE(KM11) 10.81 17.08 15.66 12.93 17.08 0.0658 0.0373 20.63 0.0098
ASE(KM13) 69.74 116.02 88.10 68.45 116.02 1.1838 0.1084 147.67 0.0703
ASE(KI13) 39.68 45.02 71.89 53.46 45.03 4.8270 2.9386 56.67 0.9743
ASE(KE4) 23.33 23.97 15.12 12.89 23.97 0.9840 0.1781 48.76 0.0627

ASE(KM9) 11.35 19.67 8.7654 3.5664 19.67 1.5287 0.3212 25.59 0.1098
ASE(KM131) 109.58 7.8306 153.81 100.08 7.8306 1.1186 0.3684 7.9849 0.1399
ASE(KI135) 1.31e3 147.86 2.47e3 1.14e3 147.86 13.78 4.4908 115.39 1.7931
ASE(KE22) 26.15 47.41 27.36 21.92 47.41 0.1794 0.0941 40.85 0.0448
ASE(KM511) 4.3991 4.9372 0.1594 1.7103 4.9372 0.0050 0.1632 9.8977 0.0018
ASE(KM521) 330.93 402.73 18.64 396.04 402.73 8.4152 6.1770 377.72 0.1200

ASE(KI523) 9.2952 11.30 0.5225 11.12 11.30 0.2361 0.1736 10.60 0.0034
ASE(KC) 1.0948 1.5200 4.9411 0.9574 1.5200 0.5559 0.0789 2.4448 0.0292
ASE(KM1221) 40.12 61.53 621.18 49.80 61.53 14.24 0.7375 84.00 2.0692
ASE(KM1241) 36.04 55.30 557.93 44.71 55.30 12.91 0.6637 75.50 1.8290
ASE(V9) 28.26 47.99 38.76 7.9788 47.99 1.1204 0.1492 71.89 0.1373
ASE(V58) 0.5878 1.0247 0.1593 0.8311 1.0247 0.0866 0.0116 1.7824 7.8e-4

Figure 9.

Figure 9

Solutions of selected state variables in the Zhu model [19]. Plotted on top of each curve are the D-optimal (circle), E-optimal (square), and SE-optimal (x) n = 11 sampling times under constraint C2 when sampling the optimal 5 observables to estimate θ⃗c (Table 9). Top Left: Carbon uptake rate A(t); Top Right: ATP; Bottom Left: SUCc; Bottom Right: RuBP.

The favorable performance of the SE-optimal design criterion under constraint C2 may be attributed to its selected time point distribution (Figure 9). The SE-optimal distribution is the only one that places a large number of time points (5 out of 11) before the solution approaches a steady state at t = 250. The E-optimal design, as mentioned previously, selects the uniform distribution, and the D-optimal design suggests sampling after the solution has achieved a steady state. When using constraint C3, the E-optimal design selects a distribution that maintains the uniform spacing between points in the interior of the time interval (0,3000) (Figure 10), again leading to poor ASE’s. Both the D- and SE-optimal designs place the majority of the 11 time points in the first 1/3 of the time interval, thus capturing more of the dynamical system’s behavior before it achieves a steady state. These front-heavy distributions greatly reduce the estimated ASE’s from those calculated using a uniform time distribution.

Figure 10.

Figure 10

Solutions of selected state variables in the Zhu model [19]. Plotted on top of each curve are the D-optimal (circle), E-optimal (square), and SE-optimal (x) n = 11 times under constraint C3 when sampling the optimal 5 observables to estimate θ⃗c (Table 9). Top Left: Carbon uptake rate A(t); Top Right: ATP; Bottom Left: SUCc; Bottom Right: RuBP.

The last case tested is designing experiments using each of the three cost functionals to estimate θ⃗c when allowed N = 10 observables and n = 11 time points. As in the case of only allowing five observables for estimation of θ⃗a, the metabolites RuBP, SBP, and F26BPc are selected by all three optimal design criteria. Additionally, PenP and SERc are selected by all three. A total of 15 different metabolites are chosen across the three optimal design criteria, of which only five are chosen by a single criterion and five selected by two criteria (Table 10). The three optimal design criteria show more agreement on which 10 observables are important to θ⃗c than which 10 are important to θ⃗a (in that case, a total of 20 different metabolites are chosen). This indicates that the three design criterion agree upon observables that are central to understanding the behavior of the model but also select observables that help minimize their associated cost functionals.

Table 10.

Top: Optimal 10 observables chosen by each optimal design criterion when estimating the parameters θ⃗c of the model in [19]. Bottom: Approximate asymptotic standard errors calculated using asymptotic theory (25) for each parameter in θ⃗c using the optimal 10 observables with time point constraints of uniform spacing, constraint C2, and constraint C3. Smallest ASE for each parameter per time point constraint is highlighted in bold font.

Method Observables

D-opt RuBP, T3P, SBP, PenP, GCEA, GOAc, SERc, GLYc, T3Pc, F26BPc
E-opt RuBP, PGA, T3P, E4P, S7P, SBP, PenP, GCEA, SERc, F26BPc
SE-opt RuBP, PGA, SBP, PenP, GOAc, SERc, T3Pc, FBPc, HexPc, F26BPc
Time Uniform C2 Optimal C3 Optimal

Method D E SE D E SE D E SE
ASE(RuBP0) 0.3162 0.3162 0.3162 0.2236 0.3154 0.2236 0.2235 0.3132 0.3126
ASE(SBP0) 0.1225 0.1225 0.1225 0.0866 0.1224 0.0866 0.0866 0.1223 0.1221
ASE(KM11) 0.3753 0.8641 0.6406 0.5256 0.0245 0.6245 0.0909 0.0143 0.0116
ASE(KM13) 2.0637 4.6387 3.9718 2.9305 0.0576 3.9110 0.4995 0.0691 0.0620
ASE(KI13) 2.1607 2.6684 9.7853 3.2408 1.6959 9.4439 0.8757 1.2024 1.1163
ASE(KE4) 0.0111 0.0137 1.5091 0.1765 0.0115 1.4072 0.0079 0.0109 0.0359

ASE(KM9) 0.2489 0.2534 1.5764 0.5398 0.1759 1.5552 0.1727 0.1409 0.1264
ASE(KM131) 2.6863 2.2992 3.0282 4.2942 0.1250 2.9806 0.2176 0.0935 0.2603
ASE(KI135) 39.00 33.38 46.42 62.39 1.7950 45.67 3.1539 1.3321 3.6831
ASE(KE22) 0.0539 0.0925 2.2836 1.1505 0.0342 2.1316 0.0140 0.0125 0.0497
ASE(KM511) 0.0202 1.6256 0.0038 0.4271 0.0735 0.0038 0.0047 0.0176 0.0015
ASE(KM521) 8.6952 17.63 4.3969 133.86 1.3817 4.7393 2.6856 0.5071 0.1681

ASE(KI523) 0.2441 0.4942 0.1234 3.7574 0.0385 0.1330 0.0754 0.0142 0.0047
ASE(KC) 0.0514 0.0632 0.2632 0.0770 0.0402 0.2538 0.0216 0.0285 0.0303
ASE(KM1221) 0.3047 2.0413 0.3358 0.3230 0.7299 0.3486 0.2883 0.5776 0.2581
ASE(KM1241) 0.2726 1.8273 0.2953 0.2890 0.6535 0.3072 0.2579 0.5173 0.2313
ASE(V9) 0.2399 0.2467 2.4578 0.4956 0.1685 2.2104 0.1641 0.1393 0.1414
ASE(V58) 0.0236 0.1514 0.0030 0.5100 0.0078 0.0031 0.0052 0.0013 6.9e-4

Similar to the 5 observable case for θ⃗c, the ASE’s calculated when using a time point distribution optimized under the C2 or C3 constraints are smaller than those of the uniform time point distribution (though the difference is much less pronounced). While the ASE’s from the D- and SE-optimal criteria are typically smallest under the uniform time point distribution, the E-optimal designed experiment yields the smallest ASE’s for the most number of parameters under constraint C2 (Table 10). The ASE’s for the D-optimal experiment under constraint C2 are typically larger than those for the D-optimal experiment with a uniform time distribution, and the SE-optimal ASE’s see only marginal improvements. For parameters when the E-optimal ASE’s are not as small as those for the D- and SE-optimal designs, the ASE’s are often still on the same order. These ASE’s are still orders of magnitude larger than the values of some parameters. The minimum ASE’s under constraint C3 are smaller than those under constraint C2 (and are notable improvements over those from the uniform time point distribution), but most are improved by only one order of magnitude or less. The D- and SE-optimal cost functions yield the smallest standard errors for more parameters under C3 than C2, as well.

The performance improvements seen in the D- and SE-optimal experimental designs from the C2-optimal time point distributions to the distributions optimal under constraint C3 can be easily explained. Under constraint C2, neither criterion produced clusters of time points (Figure 11). The distributions generated by both criteria are more similar to the uniform distribution that was used as the initial seed for the time point distribution optimization than to the highly clustered distributions seen for other selections of n and N (such as the one shown in Figure 10). The D-optimal distribution only contains one time point before t = 500, as well, thus ignoring the early dynamics of the system. The E-optimal time point distribution contains clusters at the very beginning (t < 200) and very end (t > 2900) of the time interval as well as a loose cluster between t = 1500 and t = 2000.

Figure 11.

Figure 11

Solutions of selected state variables in the Zhu model [19]. Plotted on top of each curve are the D-optimal (circle), E-optimal (square), and SE-optimal (x) n = 11 sampling times under constraint C2 when sampling the optimal 10 observables to estimate θ⃗c (Table 10). Top Left: Carbon uptake rate A(t); Top Right: ATP; Bottom Left: SUCc; Bottom Right: RuBP.

All three optimal design criteria produce non-uniform distributions under constraint C3 (Figure 12). The time point distributions determined using the E- and SE-optimal cost functions place a majority of their time points before t = 1000, capturing the dynamics early in the system solution. The D-optimal distribution also concentrates its sampling times before t = 1000, but does not as tightly cluster the time points. All three cost functionals under time point constraint C3 produce smaller ASE’s than those calculated with a uniform time point distribution (Table 10). The distributions generated under both constraints C2 and C3 indicate that heavily weighting time periods when the system is changing its pattern of behavior (such as transitioning from oscillations to a steady state) produce smaller standard errors than uniform or near-uniform distributions.

Figure 12.

Figure 12

Solutions of selected state variables in the Zhu model [19]. Plotted on top of each curve are the D-optimal (circle), E-optimal (square), and SE-optimal (x) n = 11 times under constraint C3 when sampling the optimal 10 observables to estimate θ⃗c (Table 10). Top Left: Carbon uptake rate A(t); Top Right: ATP; Bottom Left: SUCc; Bottom Right: RuBP.

Our computations using the Zhu model of [19] for the Calvin cycle indicate that both the selection of observables and the distribution of sampling times across the experiment affect the estimated asymptotic standard errors – sometimes by several orders of magnitude. While the ASE’s calculated for all exercises were often larger than the parameter values, this may be related to the model’s tendency to quickly reach a steady state.

Overall, the optimal design cost functions performed best under time point constraint C3, followed by C2 and finally the uniform distribution. While in some examples a particular cost functional would do particularly well (such as D-optimal in the θ⃗a, 10 observable case and SE-optimal in the θ⃗c, 5 observable case), it is hard to determine apriori – or even using the results of a uniform time distribution – which cost functional will perform best for a particular set of parameters and observables. The reduction in ASE’s gained from adding additional observables is also difficult to predict. Adding five observables to better estimate θ⃗a did not result in noticeably smaller estimated asymptotic standard errors as calculated by asymptotic theory. The improvement may be better quantified using a different method to calculate asymptotic standard errors such as Monte Carlo simulations or bootstrapping, but that is beyond the scope of this initial investigation.

7 Conclusion

Expanding the efforts reported in [12] on time point distribution selection using the D-, E-, and SE-optimal design criteria, we introduce a new methodology and algorithm for selecting both optimal observables and an optimal time point distribution. While the D- and E-optimal cost functions are well established in the literature, the SE-optimal design method is relatively new but competitive. We compare the abilities of these three design criteria to reduce the estimated asymptotic standard errors of selected subsets of parameters for two ordinary differential equation models representative of ODE systems used in industrial applications: the log-scaled HIV model (27) and the model of [19] for the Calvin cycle. These examples suggest the strengths of each design method in selecting appropriate observable variables and sampling times.

Collecting patient data for the treatment of HIV is limited by the types and expenses of assays available. In efforts with experimental data [2], parameter estimation for the HIV model (27) has only been performed with one or two observed quantities. Our tests for determining the optimal sets of 1, 2, and 3 observables when data is collected at uniformly distributed times show that the SE-optimal design criterion performs very similarly to the D-optimal criterion; moreover, a small number of observables may be compensated for by allowing more sampling times. While only measuring one observable may not quite be adequate for accurate parameter estimates, taking either 201 or 401 measurements of two observables yielded similar standard errors when compared to an experiment in which three observables are sampled 51–101 times.

For the HIV model, we compared the performance of all three design criteria under time point distribution constraints C2 and C3. When determining the optimal distribution of 35 time points, D-optimal yields the smallest ASE’s for the most parameters, followed by E-optimal, for both time point constraints. When determining the optimal distribution of 105 time points, however, D-, E-, and SE-optimal each yield the smallest ASE’s for some of the parameters under C2, and both D-and E-optimal are strong performers under constraint C3. The SE-optimal standard errors, while not the lowest, are on the same order of magnitude as those of D- and E-optimal. Therefore for our selected parameter values, D- and E-optimal more reliably yield the smallest standard errors.

The Zhu model of [19] is representative of experimental environments where a great number of state variables may possibly be measured but the costs of measurements limit the number of observed variables that may be sampled. We compared the estimated asymptotic standard errors calculated using the observables and time point distributions determined with the D-,E-, and SE-optimal design methods. For the small parameter subset θ⃗a, a small number of observables is sufficient to obtain as small ASE’s as possible - adding more observables does not improve the ASE’s generated for any of the design methods. Optimizing the time point distribution under either constraint C2 or C3 improves the ASE’s by up to two orders of magnitude. At both 5 and 10 observables, the D-optimal design provides the smallest ASE’s for the most parameters, followed by SE-optimal. The E-optimal standard errors, however, are often on the same order of magnitude, indicating that it is still competitive.

When testing the treatment of the larger parameter subset θ⃗c of the Zhu model by each of the three design methods, we found that while an excellent selection of observables and time points may still yield small standard errors (as with the SE-optimal design method with 5 observables under constraint C3), adding more information through more observables reduces the calculated ASE’s in the uniform time distribution. At 5 observables, the SE-optimal cost function performs the best and E-optimal time distribution optimization fails to venture far from the uniform grid initially used; at 10 observables, E-optimal performs best under constraint C2 and all three design criteria perform the best for different parameters under constraint C3. Thus for the Zhu model, it is difficult to predict which optimal design criterion will perform best in a particular case.

While the performance of each optimal design criteria is highly dependent upon the ODE system, parameter subset, number of observables allowed, number of time points allowed, and even constraints imposed on the time point distribution selection, the examples performed using the HIV model (27) and the Zhu model [19] demonstrate that the D-, E-, and SE-optimal design methods are all competitive and useful. Moreover, in the Zhu model examples, selection of optimal time points can reduce the estimated asymptotic standard errors of parameters by several orders of magnitude, thus providing data that is more useful in a parameter estimation problem without taking more samples or measuring more observables. Thus the new methodology developed and illustrated here can be an important tool in designing experimental protocols for obtaining data to be used in estimating parameters in complex dynamic models.

Acknowledgments

This research was supported in part (HTB and KLR) by grant number R01AI07 1915-07 from the National Institute of Allergy and Infectious Diseases and in part (KLR) by Fellowships from the CRSC and CQSB. The authors are grateful to Dr. Laura Potter of Bioinformatics, Syngenta Biotechnology, Inc., Research Triangle Park, NC, for her numerous conversations on plant biology experimentation and for initially providing, among a number of other references, the reference [19].

References

  • 1.Adams BM, Banks HT, Davidian M, Kwon H, Tran HT, Winfree SN, Rosenberg ES. HIV dynamics: Modeling, data analysis, and optimal treatment protocols, CRSC Technical Report CRSC-TR04-05, NCSU, February 2004. J Comp Appl Math. 2005;184:10–49. [Google Scholar]
  • 2.Adams BM, Banks HT, Davidian M, Rosenberg ES. Model fitting and prediction with HIV treatment interruption data, CRSC Technical Report CRSC-TR05-40, NCSU, October 2005. Bull Math Biol. 2007;69:563–584. doi: 10.1007/s11538-006-9140-6. [DOI] [PubMed] [Google Scholar]
  • 3.Adams BM. PhD Thesis. NC State Univ; 2005. Non-parametric Parameter Estimation and Clinical Data Fitting with a Model of HIV Infection. [Google Scholar]
  • 4.Attarian A. tssolve.m. Retrieved August 2011, from http://www4.ncsu.edu/arattari/
  • 5.Avery M, Banks HT, Basu K, Cheng Y, Eager E, Khasawinah S, Potter L, Rehm KL. Experimental design and inverse problems in plant biological modeling, CRSC Technical Report CRSC-TR11-12, NCSU, October 2011. J Inverse and Ill-posed Problems. doi: 10.1515/jiip-2012-0208. [DOI] [Google Scholar]
  • 6.Banks HT. Modeling and Control in the Biomedical Sciences. In: Levin S, editor. Lecture Notes in Biomathematics. Vol. 6. Springer-Verlag; Berlin Heidelberg New York: 1975. [Google Scholar]
  • 7.Banks HT, Bihari KL. Modeling and estimating uncertainty in parameter estimation, CRSC Technical Report CRSC-TR99-40, NCSU, December 1999. Inverse Problems. 2001;17:95–111. [Google Scholar]
  • 8.Banks HT, Cintrón-Arias A, Kappel F. Mathematical Model Development and Validation in Physiology: Application to the Cardiovascular and Respiratory Systems, Lecture Notes in Mathematics, Mathematical Biosciences Subseries. Springer-Verlag; 2012. Parameter selection methods in inverse problem formulation, CRSC Technical Report CRSC-TR10-03, NCSU, revised November 2010. to appear. [Google Scholar]
  • 9.Banks HT, Davidian M, Hu S, Kepler G, Rosenberg E. Modeling HIV immune response and validation with clinical data, CRSC Technical Report CRSC-TR07-09, NCSU, March 2007. J Biological Dynamics. 2008;2:357–385. doi: 10.1080/17513750701813184. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Banks HT, Davidian M, Samuels JR, Sutton KL. An inverse problem statistical methodology summary, CRSC Technical Report CRSC-TR08-01, NCSU, January 2008; Chapter 11. In: Gerardo, Chowell, et al., editors. Statistical Estimation Approaches in Epidemiology. Springer; Berlin Heidelberg New York: 2009. pp. 249–302. [Google Scholar]
  • 11.Banks HT, Dediu S, Ernstberger SL, Kappel F. Generalized sensitivities and optimal experimental design, CRSC-TR08-12, September, 2008, (Revised), November, 2009. J Inverse and Ill-posed Problems. 2010;18:25–83. [Google Scholar]
  • 12.Banks HT, Holm K, Kappel F. Comparison of optimal design methods in inverse problems, CRSC Technical Report CRSC-TR10-11, NCSU, July 2010. Inverse Problems. 2011;27:075002. doi: 10.1088/0266-5611/27/7/075002. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Davidian M, Giltinan D. Nonlinear Models for Repeated Measurement Data. Chapman & Hall; London: 1998. [Google Scholar]
  • 14.Fink M. myAD. Retrieved August 2011, from http://www.mathworks.com/matlabcentral/fileexchange/15235-automatic-differentiation-for-matlab.
  • 15.Kuntsevich A, Kappel F. SolvOpt. Retrieved July 2011, from http://www.kfunigraz.ac.at/imawww/kuntsevich/solvopt/
  • 16.Prohorov YV. Convergence of random processes and limit theorems in probability theory. Theor Prob Appl. 1956;1:157–214. [Google Scholar]
  • 17.Seber GAF, Wild CJ. Nonlinear Regression. Wiley; New York: 1989. [Google Scholar]
  • 18.Thomaseth K, Cobelli C. Generalized sensitivity functions in physiological system identification. Ann Biomed Eng. 1999;27:607–616. doi: 10.1114/1.207. [DOI] [PubMed] [Google Scholar]
  • 19.Zhu XG, de Sturler E, Long SP. Optimizing the distribution of resources between enzymes of carbon metabolism can dramatically increase photosynthetic rate: a numerical simulation using an evolutionary algorithm. Plant Physiology. 2007;145:513–526. doi: 10.1104/pp.107.103713. [DOI] [PMC free article] [PubMed] [Google Scholar]

RESOURCES