Skip to main content
BMC Systems Biology logoLink to BMC Systems Biology
. 2011 Feb 21;5:30. doi: 10.1186/1752-0509-5-30

A simple work flow for biologically inspired model reduction - application to early JAK-STAT signaling

Tom Quaiser 1,2, Anna Dittrich 3,4, Fred Schaper 3,4, Martin Mönnigmann 1,
PMCID: PMC3050741  PMID: 21338487

Abstract

Background

Modeling of biological pathways is a key issue in systems biology. When constructing a model, it is tempting to incorporate all known interactions of pathway species, which results in models with a large number of unknown parameters. Fortunately, unknown parameters need not necessarily be measured directly, but some parameter values can be estimated indirectly by fitting the model to experimental data. However, parameter fitting, or, more precisely, maximum likelihood parameter estimation, only provides valid results, if the complexity of the model is in balance with the amount and quality of the experimental data. If this is the case the model is said to be identifiable for the given data. If a model turns out to be unidentifiable, two steps can be taken. Either additional experiments need to be conducted, or the model has to be simplified.

Results

We propose a systematic procedure for model simplification, which consists of the following steps: estimate the parameters of the model, create an identifiability ranking for the estimated parameters, and simplify the model based on the identifiability analysis results. These steps need to be applied iteratively until the resulting model is identifiable, or equivalently, until parameter variances are small. We choose parameter variances as stopping criterion, since they are concise and easy to interpret. For both, the parameter estimation and the calculation of parameter variances, multi-start parameter estimations are run on a parallel cluster. In contrast to related work in systems biology, we do not suggest simplifying a model by fixing some of its parameters, but change the structure of the model.

Conclusions

We apply the proposed approach to a model of early signaling events in the JAK-STAT pathway. The resulting model is not only identifiable with small parameter variances, but also shows the best trade-off between goodness of fit and model complexity.

Background

Mathematical models and simulations are central to systems biology. Any mathematical model is only as reliable as the numerical values assigned to its biological parameters, however. Since biological parameters, such as kinetic constants, can often not be measured directly, they must be determined indirectly with parameter estimation (PE) methods. Naturally, results obtained with PE are the more precise, the greater the information content of the experimental data.

It is well-known that there may exist parameters in a model that cannot be estimated by PE at all, i.e. not with a finite error (see e.g. [1-5]). We can asses whether the unknown model parameters of a model may be determined by PE by testing the identifiability of the model. A number of useful definitions of identifiability exist (see [6] for a review). Two concepts are important in the context of the present paper: structural identifiability and at-a-point identifiability. Unfortunately, methods for structural identifiability testing ([5,7-16]) are not widely used for large nonlinear models due to either the computational complexity or the lack of mature computer implementations [17].

Methods for at-a-point identifiability testing [1,3,4,18-22], in contrast, are easier to implement and more widely applicable. Due to their size and nonlinearity, the models treated in the present paper are analyzed with methods for at-a-point identifiability. In the remainder of this article the term identifiability therefore refers to at-a-point identifiability (see Methods section for a definition). We use the eigenvalue method for identifiability testing, because it proved to be computationally efficient and precise in a recent comparison to other methods [2].

If a mathematical model turns out not to be identifiable for the accessible experimental data, we may attempt to replace it by a simplified one. Many methods for model reduction of nonlinear models exist. These include but are not limited to techniques based on detecting and decomposing different time-scales [23-25], sensitivity analysis [21,26,27] and balanced truncation [28-30]. In this contribution our aim is to find an identifiable model. Therefore we propose an identifiability based approach for model reduction. Our iterative approach is not fully automated, but depends on a biologically skilled modeler. Even though the modeler is guided by the proposed approach, he remains in control of the simplification throughout the entire procedure. The work flow proceeds by the following steps: 1) parameter estimation, 2) ranking estimated parameters according to their identifiability, 3) calculating parameter variances, and 4) simplifying those parts of the model that contain the least identifiable parameters. These steps are iterated until an identifiable model results. In the presented case study typical simplifications in step 4) amount to lumping and neglecting reactions that do not affect any model output. We stress that we do not enforce identifiability by merely fixing unidentifiable parameters to values that result in good fits. While parameter fixing helps investigating which parameters could be estimated within the set of unknown parameters [3,31-34], the resulting parameter values must be considered with great care. After all, the values of the fixed unidentifiable parameters remain to be unknowns in these models. In fact, the unidentifiability of these parameters implies infinite error bars have to be assigned to them. Parameter estimation with random number based optimization algorithms can be particularly misleading in this context. These algorithms often suggest there exists an optimization result, even if the PE did not converge to a solution that fulfills optimality criteria. The convergence properties of gradient based optimization methods are, in contrast, closely linked to identifiability. Briefly speaking, convergence of gradient based methods is usually tested by checking first order optimality conditions, which are a necessary condition for local optimality of the least squares parameter estimation to Gaussian approximation and, in turn, a necessary condition for local identifiability of the model (see [2,35] and Methods).

The proposed model simplification work flow is applied to a model of Janus kinase (JAK) and signal transducer and activator of transcription (STAT) signaling, which is based on the model by Yamada et al. [36]. The JAK-STAT pathway is of biological importance because it is involved in several key cellular processes such as inflammation, cellular proliferation, differentiation and apoptosis.

We do not use the full JAK-STAT model proposed by Yamada et al. [36], since even when all species can be measured noise-free and at high frequency, the full model is by far too detailed [2]. This problem would be even more pronounced if realistic experimental conditions were assumed. Instead of the original JAK-STAT model we use a truncated model that focuses on the early signaling phase (first 15 minutes) of the JAK-STAT signaling pathway before transcriptional feedback occurs.

We apply the proposed iterative work flow to the JAK-STAT model until eventually an identifiable model results. After each iteration step the unknown parameters of the resulting new model are determined by maximum likelihood estimation using simulated data. Consequently, the work flow is applicable in silico, i.e. the proposed identifiability analysis may be carried out based on experimental data, but it may also be applied prior to any laboratory measurements. If no experimental data are available, it is important to still incorporate realistic assumptions on which biological quantities could in principle be measured, as well as on which measurement error must be expected. These assumptions on the availability and precision of data are crucial, since they heavily influence identifiability properties.

The proposed work flow extends our earlier work [2], where we assumed that all state variables of the model can be measured to arbitrary precision. For the purpose of the present paper it is important to incorporate stricter assumptions on measurability, sampling rates, and error bars. Moreover, the identifiability criterion used in [2] arguably is difficult to interpret. Here we use a variance based criterion that is both concise and easy to interpret. Finally, the work flow proposed here improves upon the identifiability method used in [2] in that it is iterative and uses multi-start parameter estimation to mitigate the problem of convergence to local minima in non convex least squares optimization.

Methods

The first part of this section focuses on the description of the JAK-STAT model. After introducing the general system class, the particular JAK-STAT model used throughout the paper is discussed. The second part of the section introduces the proposed model simplification workflow and its building blocks.

Modeling

System class

We consider mathematical models M of the form

x˙(t)=f(x(t),p,u(t)),x(0)=x0y(t)=h(x(t),p,u(t)), (1)

where xnx, pnp, unu and yny denote state variables, parameters, inputs and outputs, respectively. The functions f and h are assumed to be smooth. One component pi of the parameter vector p corresponds to one biological parameter such as a kinetic constant. In our example the initial conditions are known. Note that this might not always be the case. The proposed approach can easily be extended to systems with estimated initial conditions (see Additional file 1 supplementary text 1).

We stress that it is important to distinguish state variables x from output variables y. In a simulation, all state variables x can be determined to practically arbitrary precision and recorded at arbitrary sampling rates. In a laboratory experiment, in contrast, many state variables can usually not be measured at all, or not within any relevant precision. Moreover, some state variables can often not be measured separately from others, but only as a sum or, more generally, linear or nonlinear combination. The output variables y are introduced to distinguish those quantities that could in principle be measured in a laboratory from the state variables. In terms of the output values y, a laboratory or computer experiment results in values at successive points 0=t0<t1< <tnt in time

y(ti),y˜(ti)i=1,...,nt, (2)

where y(t) and y˜(t) denote simulated output data and experimental data, respectively. Up to measurement noise the outcomes of both a real and a simulated experiment are uniquely determined by the initial conditions x(0) = x0 and the values of the inputs u(t) from t = 0 to the final time t=tnt. Input functions u(t), t[0,tnt] also have to be chosen with care, since computer simulations can be run for a much larger class of functions u(t) than can be realized experimentally. Both the choice of u(t) and y are discussed in detail below.

We use a variable order, variable step size, backward differentiation formula based numerical integrator (DDASPK [37]) to obtain the solution of equation (1). The derivatives of the state variables with respect to the parameters are calculated by adding so called sensitivity equations to equation (1) and integrating the extended equations system (e.g. see [38]).

We focus on the events that occur in the JAK-STAT pathway during the first 15 minutes after receptor activation. This assumption results in a considerable simplification, since transcriptional feedback and protein synthesis need not be modeled. The model of the JAK-STAT pathway published by Yamada et al. [36] serves as a reference in the present work. Figure 1 sketches those parts of the model published by Yamada et al. [36] that are needed to describe the first 15 minutes after receptor activation. Its signaling steps can briefly be described as follows. In the first step, the receptor (R) associates with JAK to form the R_JAK complex. Binding of interferon-γ (IFN) to R_JAK creates the receptor complex IFN_R_JAK, which is able to dimerize. JAK may phosphorylate the dimerized receptor complex, and, as a consequence, the active receptor complex (IFN_R_JAKPhos_2) is formed. Either SH2 domain-containing tyrosine phosphatase 2 (SHP2) or cytoplasmic STAT1 (STAT1c) can bind to this complex. In the first case, the receptor complex is deactivated. In the latter case, the activated receptor phosphorylates STAT1c, a prerequisite to STAT1c dimerization. The remaining parts of the model describe how STAT1c can be dephosphorylated by the cytoplasmic phosphatase PPX, and how STAT1c monomers and dimers can be converted into each other.

Figure 1.

Figure 1

Truncated JAK-STAT model. Protein species are depicted as colored rectangles. Arrows between the protein species describe association or dissociation reactions, which follow mass action kinetics. The names of kinetic parameters are written next to the corresponding reaction arrow. A mathematical representation of the model is given in Additional file 1 table S2a and S2b.

Choice of input function u(t) and outputs yi

As pointed out above, input functions u(t), t[0,tnt] and output variables yi have to be chosen with care, whenever simulations are supposed to mimic conditions of realistic laboratory experiments. We choose an input function u(t) that represents a pulse stimulation with IFN,

u(t)=IFN(t)={1,ift7min0,else, (3)

where IFN is assumed to be removed completely from the medium in a washing step at time t = 7 min. Throughout the paper we consider the outputs y1,...,y4 illustrated in Figure 2. Since it is reasonable to assume that these outputs could be measured every minute in a laboratory experiment, we record their simulated values at times t = 0 min,..., 15 min and discard all other simulated values. The outputs y1,...,y4 correspond to the following quantities: the sum of the concentrations of all phosphorylated STAT1 molecules regardless of their binding or dimerization status (y1); the concentration of all activated JAK molecules regardless of their binding status (y2); the concentration of all STAT1 dimers regardless of their binding and phosphorylation status (y3); and the concentration of all STAT1 monomers regardless of their binding or phosphorylation status (y4).

Figure 2.

Figure 2

Model outputs.

By choosing the output y1 as defined in Figure 2 we implicitly assume that phosphorylated STAT1 molecules can experimentally be distinguished from unphosphorylated ones. Similarly, measuring y2 requires distinguishing phosphorylated JAK molecules from unphosphorylated ones. We assume the necessary measurements can be accomplished by western blotting with phospho-specific antibodies. Furthermore, output y3 requires distinguishing dimeric from monomeric STAT1 species. This can be accomplished, if we assume that native SDS-gels, which separate dimers from monomers, are used for Western blot measurements. Finally, we assume that relative phosphorylated JAK and phosphorylated STAT1 concentrations from Western blot measurements can be converted into absolute concentrations by immunoprecipitating the phosphorylated protein with phospho-specific antibodies and parallel Western blotting of a calibrator protein (recombinant JAK or STAT1) with known concentration.

Since Western blot experiments are known to have large standard deviations, we choose a standard deviation of 20% [31]. The choice of experimental noise is critical for the results of any analysis of the parameter variance, since larger measurement noise leads to larger parameter variance.

We stress again that all values yi(tj) are simulated values throughout the paper. As pointed out above, however, it is crucial to choose outputs, inputs, and the sampling time which mimic real experimental conditions.

Model simplification work flow

The model simplification work flow combines delocalized parameter estimation, identifiability testing, variance analysis, and goodness of fit testing. We describe these building blocks first. Note that the work flow can either be started with simulated data from a reference model or with real data.

Parameter estimation

Assuming normally distributed and independent observational noise, the solution of the optimization problem

p^=argminpχ2(p),where (4)
χ2(p)=i=1nyj=1nt(y˜i(tj)yi(tj,p,x0)σij)2, (5)

results in the maximum likelihood estimate (MLE) of p, where σij denotes the standard deviation of data point y˜i(tj). The MLE is calculated numerically with the gradient-based sequential quadratic programming software NPSOL5.0 [39]. Since gradient-based solvers for nonlinear problems are attracted by local optima, results strongly depend on starting values. We solve the optimization problem for many starting values on a parallel computing cluster to mitigate this well-known problem. Starting values are sampled with Latin Hypercube Sampling (LHS) [40] (see Additional file 1 supplementary text 2) for an efficient coverage of the sampled space. We will refer to this delocalized parameter estimation approach as multi-start estimation for short.

Identifiability

The parameter pk of the model (1) is called globally (locally) structurally identifiable, if for any admissible input u(t), p' and p" (p" from some neighborhood of p'), the equality y(p', u) = y(p", u) implies pk'=pk". Global (local) at-a-point identifiability is defined analogously for a fixed reference parameter p'. In the remainder of the article the term identifiability refers to at-a-point local identifiability.

It can be shown that identifiability of a model implies uniqueness of the minimizer p^ of (4) [4,35]. Conversely, a model is not identifiable if p^ minimizes (4) and there exists a p˜p^ such that χ2(p˜)=χ2(p^). We can locally test if such a p^ exists by analyzing the neighborhood of a minimum p^ of (4) using the quadratic approximation of χ2

χ2(p^)12ΔpTHΔp,

where Δp=p˜p^ and H is the Hessian matrix of (5). In Gaussian approximation the components of H read

Hkl=i=1nyj=1ntyi(tj,p,x0)pk1σij2yi(tj,p,x0)pl=(STWS)kl, (6)

where W is the inverse of the measurement variance matrix. For any fixed ϵχ2 the inequality ΔpTHΔpϵχ2 describes an ellipsoid that approximates the ϵ-uncertainty region around p^. The axes of this ellipsoid coincide with the eigenvectors ui of H. The length of the ith axis is proportional to 1/λi, where λi denotes the eigenvalue of eigenvector ui. Since H is positive semidefinite in Gaussian approximation, its eigenvalues are nonnegative. If any eigenvalue is close to zero, λi ≈ 0, the ellipsoid that represents the ϵ-uncertainty region is very elongated in the direction of the corresponding eigenvector ui. This implies that parameters far apart have almost the same χ2 value, or equivalently, at least one parameter has a large variance. Correspondingly, the existence of one or more λi = 0 implies at least one parameter has an infinite variance.

We use an eigenvalue-based approach [2,32] to rank model parameters pi according to their identifiability. This method identifies the smallest eigenvalue λmin of H, searches for the largest component in the corresponding eigenvector umin, and fixes the corresponding parameter to its estimated value p^i. The Hessian is then recomputed with respect to all parameters that have not been fixed, and the procedure is repeated until all parameters are fixed. The order in which parameters are fixed by this approach corresponds to the identifiability ranking of parameters, with the first parameter in the ranking being the least identifiable and the last parameter being the most identifiable one [2,32]. We stress that, in contrast to [2], we do not use the eigenvalue method to distinguish the identifiable parameters from the unidentifiable ones, but only to rank all parameters according to their identifiability. The resulting ranking is then used in the variance analysis as described in the next section. In particular we do not need a cutoff value for small eigenvalues as introduced in [2].

Variance analysis

Intuitively, a model is locally identifiable at p^ if the variances σ2(pi) of all parameters are small. A variance based identifiability criterion, which is stated more formally below, is preferred over the eigenvalue criterion [2,32] for its simplicity and intuitive interpretation. Variances σ2(pi) are determined by running parameter estimations (4) for a large number N of starting points in the neighborhood of a minimizer p^ of (4), accepting those estimation results with small P-value (defined below), and calculating the mean and variances for the resulting set of accepted parameter estimates. More precisely, we consider an estimation result to be acceptable if it has a significantly small P(χ2|v) value [41] (≤ 0.1%, see Results and Discussion), where P(χ2|v) is the probability for the value of the χ2-distribution with v = nynt - np degrees of freedom to be less than an observed χ2-value. The variance σ2(pi) and the mean p¯i are calculated from the accepted estimates according to the following equations.

σ2(pi)=1naccept1j=1naccept(Qijp¯i)2, (7)
p¯i=1nacceptj=1nacceptQij, (8)

where naccept is the number of accepted estimates, and Q is an np × naccept-matrix that contains these estimates in its columns. Since both variance and standard deviation depend on the mean, we use the scale invariant coefficients of variation v(pi)=σ(pi)/p¯i[42] to compare variances of different parameters. A v(pi) value of 1, for example, implies the estimates for parameter pi exhibit 100% standard deviation relative to the mean.

We consider a model to be identifiable, if the coefficients of variation v(pi), i = 1,...,np, of all parameters are smaller than a bound v¯ (smaller than v¯=0.01 deviation, see Results and Discussion). A value of v¯=0.01 implies the parameter estimation step resulted in a relative standard deviation of 1% or less for all parameters. Note that this interpretation of v¯ is only valid if the parameter estimation runs cover a sufficiently large neighborhood of the candidate value p^. If the sampled neighborhood is too small, the estimates converge to parameter values that cover the entire sampled neighborhood, and the neighborhood needs to be enlarged. In this case, v(pi) is a not an upper but a lower bound for the coefficient of variation of parameter pi. Clearly, this lower bound cannot be used to infer identifiability, but it can be used to infer that a model is not identifiable.

Both v¯ and the significance level for P(χ2|v) obviously are adjustable parameters. As discussed in the section entitled Results and Discussion, however, the chosen values are not critical. Furthermore, we stress that their values (0.01 and 0.1%, respectively) are conservative in any case.

Testing goodness of fit

The purpose of the identifiability tests and variance analyses is to ensure a balance between the number of unknown parameters that need to be estimated on the one hand, and the information content of the data used for their estimation on the other hand. Essentially, we consider model complexity and information content of data to be balanced, if the parameter estimation yields unique values for all estimated parameters, or equivalently, sufficiently small coefficients of variation for all estimated parameters. Besides guaranteeing the uniqueness of the estimated parameters, however, we also need a measure for the quality of the fit of different models to the data. We use a criterion based on Akaike's information criterion [43] for this purpose. The AIC reads

AIC=2ln(L(g(p^)|data)+2np, (9)

where L(g(p^)|data) is the likelihood of model g(p^) given the measurement data. For brevity L(g(p^)|data) will be denoted as L. In case of χ2-parameter estimation L is given by

L=i=1nyj=1nt(12πσij2)12exp(i=1nyj=1nt(y˜i(tj)yi(tj))22σij2),

and ln(L) can be written as

ln(L)=ln(i=1nyj=1nt(12πσij2)12)i=1nyj=1nt(y˜i(tj)yi(tj))22σij2=Cχ2/2. (10)

For a review see [6,44]. In equation (10) C denotes a model independent constant, which can be neglected when comparing AIC values of different models. Combining equations (9) and (10) the AIC criterion can be written as: AIC = χ2 + 2np. We use a variant of AIC, AICc [45], which is well suited for estimation with small data sets. The results obtained with AICc are compared to other established variants of the AICc criterion, the AICc difference Δk and the AICc weights wk. Δk describes the difference between the AICc value of model Mk and the model with the smallest AICc value. The value of wk can be interpreted as a probability that model Mk is the best model within a set of alternative models. A closer description of the criteria can be found in Additional file 1 supplementary text 3. The interested reader is referred to [44] for more details.

Model simplification work flow

Having presented its building blocks, the work flow for the model simplification can now be stated. We assume the original model M0 is of the form (1). This implies the output variables y have been chosen and initial conditions x(0) = x0 are given.

0. Choose the input function u(t), t[0,tnt] Choose an upper bound v¯ for the coefficients of variation as introduced in the section Variance analysis. Set the iteration counter to k = 0.

1. Calculate an estimate p^k for Mk as described in the section entitled Parameter estimation.

2. Calculate the Hessian matrix of Mk evaluated at the estimated values p^k and rank the parameters of Mk as described in the section entitled Identifiability analysis.

3. For i ∈ {1,..., np} in the order that results from step 2, calculate the coefficient of variation v(pi) as described in the section entitled Variance analysis. If v(pi)>v¯, go to step 4. If v(pi)v¯ for all i = 1,..., np, terminate.

4. Simplify those parts of the model that involve the least identifiable parameters. Refer to the resulting model as Mk+1, increment k, and go to step 1.

As pointed out before [2], the model simplification step (step 4) cannot be carried out automatically, but requires insight into the model. Typical simplifications consists of lumping of reactions or introducing simplified reactions.

Note that the work flow does not guarantee that the smallest eigenvalue of the Hessian matrix increases in every iteration. Since the model structure is changed in step 4, the size and structure of the Hessian matrix changes. Consequently, there is no one-to-one relationship between the eigenvalues of the Hessian before and after a simplification step. In practice, however, a decrease of the smallest eigenvalue from one simplification step to the next rarely occurs. In the example treated here, a decrease occurs in step 5 from λmin = 6.9 · 10-7 to λmin = 1.5 · 10-10. This decrease is outweighed by an overall increase from λmin = 1.9 · 10-15 before the first simplification step to λmin = 1.2 · 10-6 after the last simplification step. Our improved work flow differs from the one proposed in [2] in several aspects. Most importantly, the single step model simplification described in [2] is extended to a multi-step model simplification work flow in which new estimations are carried out after each model simplification step. Secondly, we propose a variance based stopping criterion, which can be interpreted more easily than the eigenvalue bound used in [2]. For both the parameter estimation and the variance analysis step we use multi-start estimations. Finally, the parameter estimation and identifiability analysis are carried out with respect to outputs y here. In contrast, trajectories of all state variables xi(t), i = 1,..., nx were assumed to be available in [2]. As pointed out above it is crucial to consider outputs instead of state variables if simulations are supposed to mimic realistic laboratory experiments.

Results and Discussion

In this section we apply the model simplification work flow to the truncated JAK-STAT model from Figure 1. We refer to this model and the parameter values published by Yamada et al. [36] as M0 and p^0, respectively. We carry out a sequence of 6 simplification steps. The resulting models and the corresponding estimated parameter values are denoted by M1,...,M6 and p^1,...,p^6, respectively. All models are described in detail in Additional file 1.

All parameter estimation steps (step 1 of the work flow) are carried out in logarithmic parameter space to efficiently cover a large search area (cf. [31]). When estimating parameters pk for model Mk, estimated parameter values p^k1 of Mk-1 serve as reference values. More specifically, starting points pk for multi-start parameter estimation are sampled within four magnitudes around these reference values, i.e., in the ranges p^ik1102p^ikp^ik1102, i = 1,...,np. Data points for the estimation are generated once by simulating the reference model M0 and recording the output values y(tj), t0 = 0 min,..., t15 = 15 min. As proposed in [31] we assume 20% standard deviation for experiments and set σij in Eq. (6) accordingly. For the variance analysis (step 3 of the work flow) we choose the boundaries p^i1012pip^i1012 and a significance level for P(χ2|v) of 0.1%. Both choices turned out not to be critical as discussed at the end of the section. Finally, we set v¯=0.01. This implies we consider two values for the same parameter to be equal, if these values deviate by a relative error of 1% or less.

In each parameter estimation step and in each variance analysis step 1000 starting points are sampled. The best estimates of each work flow iteration are listed in Additional file 1 table S1.

Simplification 1: Neglecting the STAT1Phos reassociation to the activated receptor

Values p^0 for the parameters of model M0 are available from the literature [36] and therefore need not be estimated in the first simplification step. The work flow is therefore started with the identifiability analysis (step 2). This analysis yields a smallest eigenvalue of 1.9 · 10-15, which is a strong indication that the model is not identifiable. This result is corroborated in step 3 (variance analysis), where a high coefficient of variation, v(kf7) ≥ 0.63, results for the least identifiable parameter, kf7. We can only infer a lower bound on v(kf7), since the parameter values estimated in the variance analysis step span the entire range of the starting values. As discussed in the section entitled Variance analysis, this indicates that starting values would have to be sampled from a larger parameter space region, if a value for v rather than a lower bound was necessary. A lower bound suffices, however, to infer that the variances are too large to consider the model to be identifiable. Consequently, we attempt to simplify those parts of the model that involve the least identifiable parameter kf7 (step 4).

The parameter kf7 describes the association of phosphorylated STAT1c to the activated receptor complex (see Figure 3). As discussed in [2] this re-association of previously phosphorylated STAT1c can be neglected, since the phosphorylation of cytoplasmic STAT1c by the activated receptor is the key event rather than the reassociation of the previously phosphorylated STAT1c. Moreover, phosphorylated STAT1cPhos monomers are driven to homodimerize. Therefore, only few such monomers are available to react with the activated receptor. By removing kf7, the backwards reaction with kinetic constant kd7 and the species IFN_R_JAKPhos_2_STAT1cPhos become obsolete and are removed, too. Since the initial value of IFN_R_JAKPhos_2_STAT1cPhos is zero in the reference model, this species can be removed from the model without the need for adjusting the initial values of the remaining state variables. We reduce our model from 17 states and 25 parameters to 16 states and 23 parameters. We refer to the resulting model as M1.

Figure 3.

Figure 3

Visualization of the first three model simplifications. Removed components are shown in gray. Red numbers indicate the iteration k of the work flow in which the reaction has been removed. Unidentifiable parameters are stated next to the respective reaction. In the first iteration the reactions containing kf7, kd7 and the species IFN_R_JAKPhos_2_STAT1cPhos have been removed. In the second iteration the dissociation reactions containing the parameters kd5, kd9, kd11, and kd24 have been removed. In the third iteration the model was simplified by deleting the species R and JAK and the corresponding dissociation and association reactions. A mathematical representation of the models created in each iteration is given in Additional file 1.

Simplification 2: Neglecting the dissociation of high affinity complexes before phosphorylation or dephosphorylation can occur

The unknown model parameters of the simplified model M1 are determined by carrying out a multi-start parameter estimation (step 1 of the work flow). The best estimate of the multi-start parameter estimation p^1 results in a χ2 value of 8.7 · 10-6 and an AICc value of 77 (see Table 1). Both values will be discussed below once a comparison with other models is possible.

Table 1.

Identifiability ranking and results of multi-start estimation and variance analysis

Model Top ten parameters of the identifiability ranking χ2 AICc v(pi)
M0 kf7, kf9, kd11, kd9, kf24, kf13, kd5, kd13, kd7, kf1 none none v(kf7) ≥ 0.63 >v¯
M1 kd5, kf9, kd11, kf13, kf1, kf25, kd13, kd9, kd1, kd24 8.7 · 10-6 77 v(kd5) ≥ 0.6 >v¯
M2 kd1, kf25, kf1, kf10, kd3, kf12, kf13, k24, kd13, k11 1.0 · 10-4 57 v(kd1) ≥ 0.59 >v¯
M3 kf25, kf10, kf12, k24, kd3, kf13, kd13, k11, k9, kf3 9.5 · 10-5 49 v(kf25) ≥ 0.75 >v¯
M4 k5, kf10, kd3, k9, kf3, kd8, kf4, k11, kf8, kd2 2.7 · 10-3 31 v(k5) ≥ 0.47 >v¯
M5 k11, kf10, kd2, kf3, kd3, kd8, k9, kf4, kf8, kf2 7.6 · 10-1 28 v(k11) ≥ 0.18 >v¯
M6 kf10, kd2, kf3, kd3, kd8, k9, kf4, kf8, kf2, k5new 7.6 · 10-1 25 v(pi) ≤0.0091 <v¯

The first parameter in each ranking is the least identifiable one. No χ2 and AICc values are given for M0, since the model parameters are not estimated but taken from the literature [36].

The identifiability analysis (step 2) and the variance analysis (step 3) result in a smallest eigenvalue of 6.7 · 10-11 and in v(kd5) ≥ 0.6 for the least identifiable parameter kd5, respectively. Note that a lower bound results for v for the same reason as in the first simplification step. Again this lower bound suffices to infer that the model is not identifiable. Since the identifiability analysis ranks parameter kd5 as the least identifiable parameter, we attempt to simplify the reactions that are affected by kd5.

The parameter kd5 describes the dissociation of unphosphorylated STAT1c from the activated receptor complex. We assume the affinity of STAT1c for the active receptor complex is high [46], or equivalently, phosphorylation of STAT1c on average takes place much faster than the dissociation of unphosphorylated STAT1c. Therefore, the dissociation of unphosphorylated STAT1c is a rare event and the dissociation reaction can be removed from the model.

The same line of argumentation holds for the parameters kd11, kd9, and kd24. kd11 (kd24) describes the dissociation of STAT1cPhos (STAT1cPhos2) from PPX before PPX-induced dephosphorylation occurs. kd9 belongs to the dissociation reaction of SHP2 from the activated receptor that takes place before SHP2 dephosphorylates the activated receptor. For consistency we renamed kf5, kf11, kf9, and kf24 to k5, k11, k9, and k24, respectively. Since no state variables are removed in this simplification the initial conditions are taken from the previous model.

While the resulting model M2 has the same number of state variables (16) as the previous model M1, the number of parameters dropped from 23 in M1 to 19 in the new model M2.

Simplification 3: Assuming JAK and the receptor are preassociated

The third simplification step starts with the multi-start estimation of the parameters p2 of M2 (step 1 of the work flow). These parameters can be estimated with a χ2 value of 1.0 · 10-4. The AICc value decreases from 77 to 57, which implies that the simplifications made so far improve the balance between model detailedness and quality of fit (Table 1).

The identifiability analysis (step 2) indicates, however, that M2 with a smallest eigenvalue of 1.2 · 10-9 is not identifiable. This result is confirmed by the variance analysis (step 3), which results in v(kd1) ≥ 0.59 for the least identifiable parameter kd1 (Table 1). This value of v(kd1) is a lower bound only; see the discussion in the first and second simplification steps. The results of the identifiability analysis and the variance analysis suggest to simplify the reactions that kd1 is involved in.

kd1 is the kinetic constant for the dissociation of JAK from a single receptor. If we remove this reaction, we can further simplify the model by assuming that JAK and the receptor are associated before the signal arrives [47,48] (see Figure 3). The changes amount to assuming that the association of JAK to R is much faster than the dissociation of R_JAK. Under this assumption JAK (JAK(0) = 12 nM) and R (R(0) = 12 nM) initially only exist in complex, therefore the initial condition of R_ JAK has to be changed from 0 to 12 nM.

The resulting model, which we refer to as M3, comprises 14 state variables and 17 parameters.

Simplification 4: Omitting PPX mediated STAT1cPhos dephosphorylation

The multi-start parameter estimation for the parameters p3 of M3 results in χ2 = 9.5 · 10-5 (step 1). The AICc value improves from 57 to 49 (cf. Table 1). Step 2 of the work flow ranks kf25 as the least identifiable parameter with a corresponding eigenvalue of 7.0 · 10-9, which indicates that kf25 is not identifiable. Step 3 yields a lower bound on the variation, v(kf25) ≥ 0.75 for kf25. (cf. Table 1). While providing only a lower bound, this result corroborates that the model needs to be simplified with respect to the reaction kf25 is involved in.

kf25 describes the rate of the dephosphorylation of STAT1cPhos_2 by PPX (cf. Figure 4). STAT1cPhos can be dephosphorylated by PPX along either of two routes. 1) PPX binds and dephosphorylates the phosphorylated STAT1c monomer or 2) PPX binds the homodimer of phosphorylated STAT1c and subsequently dephosphorylates one of the two STAT1c proteins. In the second case the heterodimer STAT1c_STAT1cPhos is created. Since neither STAT1c heterodimers, nor the phosphatase PPX have been experimentally validated we propose to considerably simplify this part of the model (see Figure 4). We remove PPX and all its complexes from the model and assume a simple first order kinetic for the conversion of the phosphorylated dimer into two unphosphorylated monomers. We stress that removing PPX is a drastic simplification. The fact that the PPX-related parameters kf12 and k24 are ranked third and fourth in the identifiability ranking (cf. Table 1) further suggests that this simplification is reasonable. Since not only PPX (PPX(0) = 50 nM) but also all PPX bound species are removed, the initial conditions of the remaining species do not change.

Figure 4.

Figure 4

Simplification of the PPX part leads to model M4. The same notation as in Figure 3 is used. Additionally, a red colored parameter name indicates, the introduction of a new parameter. In this simplification step PPX and all corresponding reactions have been removed from the model. In order to still account for STAT1cPhos_2 dissociation and phosphorylation, we added a new dissociation reaction containing the parameter k11new. The mathematical description of model M4 is given in Additional file 1 table S6a and S6b.

The resulting model M4 is visualized in Figure 4. The number of state variables and parameters drops from 14 to 10 and from 17 to 12, respectively.

Simplification 5: Combining STAT1c-receptor complex formation, STAT1c phosphorylation, and complex dissociation

The parameter values p^4 can be estimated with a χ2 value of 2.7 · 10-3 (step 1 of the work flow). The AICc value improved from 49 to 31 (cf. Table 1).

The identifiability analysis ranks k5 as the least identifiable parameter with a corresponding eigenvalue of 6.9 · 10-7 (step 2). The variance analysis results in v(k5) ≥ 0.47 for k5 (step 3), where v(k5) is a lower bound for the same reason as in the previous simplification steps. Both the smallest eigenvalue and the large lower bound on v(k5) indicate that model M4 is not identifiable.

The reactions around k5 are simplified as sketched in Figure 5. Essentially, this simplification combines the formation of the STAT1c-receptor complex, the phosphorylation of STAT1c, and the dissociation into one step. As a consequence, STAT1c phosphorylation is modeled as a second order kinetic in which the activated receptor is not consumed. This implies that STAT1c phosphorylation still depends on the concentration of both the activated receptor complex and STAT1c. In this simplification the species IFN_R_JAKPhos_2_STAT1c is removed. Since initially this species does not exist, its removal has no influence on the remaining initial conditions.

Figure 5.

Figure 5

Visualization of simplifications 5 and 6. See Figure 3 and 4 for comments on the notation. In iteration 5 the association and dissociation plus phosphorylation of the active receptor IFN_R_JAKPhos_2_STAT1c and STAT1c are replaced by a simplified reaction (highlighted in yellow). In the simplified reaction with the new parameter k5new, phosphorylation of STAT1c by the active receptor is modeled without considering complex formation of both species. In the last iteration the dissociation and dephosphorylation of STAT1cPhos_y2 is removed from the model. A mathematical description of the models created in steps 5 and 6 are given in Additional file 1.

The resulting Model M5 has 9 state variables and 11 parameters.

Simplification 6: Cytoplasmic STAT1cPhos dephosphorylation can be omitted when modeling only the first 15 minutes of signaling

The multi-start parameter estimation for the parameters p5 of model M5 yields a χ2 value of 7.6 · 10-1 (step 1). The AICc value improves from 31 to 28 (cf. Table 1).

The identifiability analysis (step 2) yields a smallest eigenvalue of 1.5 · 10-10 caused by k11, which indicates that model M5 is not identifiable. This result is corroborated by a large lower bound on the coefficient of variation of k11, v(k11) ≥ 0.18 (step 3). Step 3 does not yield a value but only a lower bound for the coefficient of variation for the same reasons as discussed in the previous simplification steps. k11 is the kinetic constant of the dephosphorylation of STAT1c, a reaction we cannot simplify further. In order to create an identifiable model, we must remove this reaction (see Figure 5), which amounts to ignoring cytoplasmic dephosphorylation in the first 15 minutes of signaling. We stress this does not imply that no cytoplasmic dephosphorylation exists. Note that a similar conclusion was drawn by Yamada et al. [36], who state that dephosphorylation of cytoplasmic STAT by PPX is of minor importance in the pathway. Since no states are removed in this step, the initial conditions are taken from the previous model. While the number of state variables remained unchanged, the number of parameters of M6 decreased from 11 to 10.

Properties of the final model M6

The multi-start parameter estimation of parameter values p6 of model M6 results in a χ2 value of 7.6 · 10-1. Note this is the same value as in the previous simplification step. The AICc value, in contrast, improved from 28 to 25 (cf. Table 1).

Model M6 turns out to be identifiable in steps 2 and 3 of the work flow. While this is not clearly apparent from the smallest eigenvalue (1.2 · 10-6), the coefficient of variation is now bounded above by v¯=0.01 for all parameters. This implies that the multi-start parameter estimation conducted in the variance analysis resulted in parameter values that are within 1% of the respective mean p¯i for all parameters pi of the model.

Assessing the quality-of-fit with Akaike's information criterion and related criteria

The AICc values improved in all simplification steps. In addition to the AICc values we report AIC values, AICc differences Δk, and AIC weights wk in Table 2. In summary, all criteria agree in the sense that the final model M6 is the best model with respect to the quality-of-fit. We note that the improvement of AIC and AICc values is quite low in the last step while the last model has a drastically higher probability of being the "best" model in the sense of its AIC weight w6 = 78%. We conclude that the resulting simplified model M6 is both identifiable, and it balances quality-of-fit and model detailedness.

Table 2.

Summary of model selection statistics

Model ny np AIC AICc Δk wk
M1 16 23 46 77 51 5.3 · 10-12
M2 16 19 38 57 32 9.9 · 10-8
M3 14 17 34 49 23 0.67 · 10-5
M4 10 12 24 31 5.4 0.52 · 10-1
M5 9 11 23 28 3.0 0.17
M6 9 10 21 25 0 0.78

Choice of parameter boundaries and the significance level in the variance analysis

The starting values for the variance analysis must be chosen from a sufficiently large parameter space neighborhood of the nominal values p^. If the estimates obtained in the variance analysis converge to values that cover the entire sampled neighborhood, the coefficient of variation may be limited by the choice of the neighborhood. Since enlarging the sampling neighborhood can only result in equal or larger coefficients of variation, the v values obtained with a too small neighborhood are lower bounds on the actual coefficients of variation.

In all but the last simplification step the estimates obtained in the variance analyses cover the entire sampling neighborhood (data not shown). The resulting lower bounds on v are sufficiently large (cf. Table 1), however, to infer that the respective models are not identifiable. In the final simplification step, in contrast, the sampling neighborhood is large enough, since the estimates obtained in the variance analysis are tightly packed within 1% error (v¯=0.01, cf. Table 1) around the nominal value. The choice of the sampling neighborhood is therefore not critical in any simplification step.

We claim the chosen value of 0.1% for the significance level is not critical, either, since the work flow terminates after the same number of simplification steps for a large range of values, which can be seen as follows. The values of v do not change if the significance level is increased in the sixth simplification step, since all converged parameter estimates for p^6 (970 out of 1000) are already accepted for the chosen value 0.1%. Smaller, i.e. stricter, significance levels well below 10-10, on the other hand, do not change the coefficients of variation, either (data not shown). In the first through fifth simplification step, the coefficients of variation increase for larger, i.e. less strict, significance levels than 0.1%. Since the coefficients of variation are large enough to necessitate another iteration through the work flow already, increasing the significance level does not have any effect. On the other hand, decreasing the significance level to value as low as 10-9 does not affect the v-values for M1 through M5 (data not shown).

Conclusions

We presented a work flow for model simplification and demonstrated its use by simplifying an overly detailed model for JAK-STAT signal transduction. Here, we used simulated data from a model taken from the literature. However, the method is intended to be used with real data. The work flow can be used to ensure model quality with respect to two important aspects: 1) Given an unidentifiable model, the work flow results in a model that is locally identifiable with small coefficients of variation for all parameters and 2) the balance of goodness of fit and the amount of model detail - as quantified by AIC based criteria - is better than for the original model and intermediate models that arise in the work flow. The first point is addressed by applying both local identifiability analysis and local multi-start parameter estimation with subsequent analysis of the variance of acceptable estimates. The second point is taken care of by calculating the information criteria, based on AIC, that provide a measure for the information loss that occurs when reducing model complexity.

The proposed work flow does not automate model simplifications, but the identifiability ranking of parameters only provides hints on which parts of the model are too detailed. Nevertheless, the simplifications found for the JAK-STAT example lend themselves to biologically sound interpretations. For example, dissociation reactions are neglected for high affinity complexes in the simplified model (second simplification), JAK and the receptor (R) are assumed to be preassociated (third simplification), and cytoplasmic dephosphorylation turns out not to be relevant for the considered outputs (fourth and sixth simplification).

Several other authors suggest to fix unidentifiable parameters and to estimate the remaining ones [3,31-34]. In contrast, the model structure is simplified in the present paper until all parameters are identifiable, i.e. all parameters can be estimated with small error bars. Note that changes in the model structure imply structural changes of the Hessian matrix. Due to these changes together with the local character of the identifiability method, λmin may decrease after one simplification step. In practice, however, this is a rare event, which is outweighed by a prominent increasing trend of λmin.

We stress that the presented case study involves generating output data points yi(tj) by simulations carried out with the unidentifiable reference model M0. In this sense, both parameter fixing, and the approach as used in the case study involve unidentifiable models. However, when real experimental data is used, our approach is free of this defect. While more involved than parameter fixing, the proposed approach results in models with a complexity that is consistent with the experimentally accessible data.

Authors' contributions

TQ and MM developed the presented work flow. TQ implemented the method and accomplished the model simplification of the JAK-STAT pathway. AD and FS designed the hypothetical experiments and made critical suggestions to model simplifications. All authors drafted, read and approved the final manuscript.

Supplementary Material

Additional file 1

The supplementary pdf file accompanying this article contains the supplementary tables S1-S8b and supplementary texts 1-3.

Click here for file (81.6KB, PDF)

Contributor Information

Tom Quaiser, Email: tom.quaiser@avt.rwth-aachen.de.

Anna Dittrich, Email: Anna.Dittrich@rwth-aachen.de.

Fred Schaper, Email: fred.schaper@ovgu.de.

Martin Mönnigmann, Email: martin.moennigmann@rub.de.

Acknowledgements

We gratefully acknowledge funding of TQ and AD by Deutsche Forschungsgemeinschaft (DFG) under grant MO 1086/5-1 and grant SCHA 785/4-1, respectively.

References

  1. Quaiser T, Marquardt W, Mönnigmann M. Local Identifiability Analysis of Large Signaling Pathway Models. 2nd Conference Foundation of Systems Biology in Engineering, Proceedings Plenary and Contributed Papers. 2007. pp. 465–470.
  2. Quaiser T, Mönnigmann M. Systematic identifiability testing for unambiguous mechanistic modeling-application to JAK-STAT, MAP kinase, and NF-kappaB signaling pathway models. BMC Syst Biol. 2009;3:50. doi: 10.1186/1752-0509-3-50. [DOI] [PMC free article] [PubMed] [Google Scholar]
  3. Hengl S, Kreutz C, Timmer J, Maiwald T. Data-Based Identifiability Analysis of Nonlinear Dynamical Models. Bioinformatics. 2007. [DOI] [PubMed]
  4. Vajda S, Rabitz H, Walter E, Lecourtier Y. Qualitative and quantitative identifiability analysis of nonlinear chemical kinetic-models. Chem Eng Commun. 1989;83:191–219. doi: 10.1080/00986448908940662. [DOI] [Google Scholar]
  5. Xia X, Moog CH. Identifiability of nonlinear systems with application to HIV/AIDS models. IEEE T Automat Contr. 2003;48(2):330–336. doi: 10.1109/TAC.2002.808494. [DOI] [Google Scholar]
  6. Walter E, Pronzato L. Identification of parametric models from experimental data. Springer, Berlin; 1997. [Google Scholar]
  7. Pohjanpalo H. System identifiability based on power-series expansion of solution. Math Biosci. 1978;41(1-2):21–33. doi: 10.1016/0025-5564(78)90063-9. [DOI] [Google Scholar]
  8. Godfrey K, DiStefano J. Identifiability of Parametric Models. Pergamon, Oxford; 1987. [Google Scholar]
  9. Ljung L, Glad T. On global identifiability for arbitrary model parametrizations. Automatica. 1994;30(2):265–276. doi: 10.1016/0005-1098(94)90029-9. [DOI] [Google Scholar]
  10. Audoly S, Bellu G, D'Angiò L, Saccomani MP, Cobelli C. Global identifiability of nonlinear models of biological systems. IEEE Trans Biomed Eng. 2001;48:55–65. doi: 10.1109/10.900248. [DOI] [PubMed] [Google Scholar]
  11. Walter E, Braems I, Jaulin L, Kieffer M. Guaranteed Numerical Computation as an Alternative to Computer Algebra for Testing Models for Identifiability. Lecture notes in computer science. 2004. pp. 124–131. full_text.
  12. Vajda S, Rabitz H. State isomorphism approach to global identifiability of nonlinear-systems. IEEE T Automat Contr. 1989;34(2):220–223. doi: 10.1109/9.21105. [DOI] [Google Scholar]
  13. Chappell M, Godfrey K. Structural identifiability of the parameters of a nonlinear batch reactor model. Math Biosci. 1992;108(2):241–51. doi: 10.1016/0025-5564(92)90058-5. [DOI] [PubMed] [Google Scholar]
  14. Peeters RLM, Hanzon B. Identifiability of homogeneous systems using the state isomorphism approach. AUTOMATICA. 2005;41(3):513–529. doi: 10.1016/j.automatica.2004.11.019. [DOI] [Google Scholar]
  15. Asprey SP, Macchietto S. Statistical tools for optimal dynamic model building. Comput Chem Eng. 2000;24(2-7):1261–1267. doi: 10.1016/S0098-1354(00)00328-8. [DOI] [Google Scholar]
  16. Asprey SP, Macchietto S. Designing robust optimal dynamic experiments. J Process Control. 2002;12(4):545–556. doi: 10.1016/S0959-1524(01)00020-8. [DOI] [Google Scholar]
  17. Miao H, Xia X, Perelson A, Wu H. On identifiability of nonlinear ODE models with applications in viral dynamics. SIAM Review: accepted. 2010. [DOI] [PMC free article] [PubMed]
  18. Brun R, Reichert P, Kunsch HR. Practical identifiability analysis of large environmental simulation models. Water Resour Res. 2001;37(4):1015–x1030. doi: 10.1029/2000WR900350. [DOI] [Google Scholar]
  19. Yao KZ, Shaw BM, Kou B, McAuley KB, Bacon DW. Modeling ethylene/butene copolymerization with multi-site catalysts: Parameter estimability and experimental design. Polym React Eng. 2003;11(3):563–588. doi: 10.1081/PRE-120024426. [DOI] [Google Scholar]
  20. Jacquez JA, Greif P. Numerical parameter identifiability and estimability: Integrating identifiability, estimability, and optimal sampling design. Math Biosci. 1985;77(1-2):201–227. doi: 10.1016/0025-5564(85)90098-7. [DOI] [Google Scholar]
  21. Degenring D, Froemel C, Dikta G, Takors R. Sensitivity analysis for the reduction of complex metabolism models. J Process Control. 2004;14(7):729–745. doi: 10.1016/j.jprocont.2003.12.008. [DOI] [Google Scholar]
  22. Froemel C. Master's thesis. Fachhochschule Aachen, Abteilung Juelich, Fachbereich Physikalische Technik, Studienrichtung Technomathematik; 2003. Parameterreduktion in Stoff wechselmodellen mit Methoden der Statistik. [Google Scholar]
  23. Surovtsova I, Simus N, Lorenz T, Konig A, Sahle S, Kummer U. Accessible methods for the dynamic time-scale decomposition of biochemical systems. Bioinformatics. 2009;25(21):2816. doi: 10.1093/bioinformatics/btp451. [DOI] [PubMed] [Google Scholar]
  24. Lam S, Goussis D. The CSP method for simplifying kinetics. International Journal of Chemical Kinetics. 1994;26(4):461–486. doi: 10.1002/kin.550260408. [DOI] [Google Scholar]
  25. Maas U, Pope S. Simplifying chemical kinetics: intrinsic low-dimensional manifolds in composition space. Combustion and Flame. 1992;88(3-4):239–264. doi: 10.1016/0010-2180(92)90034-M. [DOI] [Google Scholar]
  26. Huang Z, Chu Y, Hahn J. Model simplification procedure for signal transduction pathway models: An application to IL-6 signaling. Chemical engineering science. 2010;65(6):1964–1975. doi: 10.1016/j.ces.2009.11.035. [DOI] [Google Scholar]
  27. Smets I, Bernaerts K, Sun J, Marchal K, Vanderleyden J, Van Impe J. Sensitivity Function-Based Model Reduction. Biotechnology and bioengineering. 2002;80(2) doi: 10.1002/bit.10359. [DOI] [PubMed] [Google Scholar]
  28. Liebermeister W, Baur U, Klipp E. Biochemical network models simplified by balanced truncation. FEBS Journal. 2005;272(16):4034–4043. doi: 10.1111/j.1742-4658.2005.04780.x. [DOI] [PubMed] [Google Scholar]
  29. Hahn J, Edgar T. An improved method for nonlinear model reduction using balancing of empirical gramians. Computers & Chemical Engineering. 2002;26(10):1379–1397. [Google Scholar]
  30. Moore B. Principal component analysis in linear systems: Controllability, observability, and model reduction. IEEE Transactions on Automatic Control. 1981;26:17–32. doi: 10.1109/TAC.1981.1102568. [DOI] [Google Scholar]
  31. Schilling M, Maiwald T, Hengl S, Winter D, Kreutz C, Kolch W, Lehmann WD, Timmer J, Klingmueller U. Theoretical and experimental analysis links isoform-specific ERK signalling to cell fate decisions. Mol Syst Biol. 2009;5:334. doi: 10.1038/msb.2009.91. [DOI] [PMC free article] [PubMed] [Google Scholar]
  32. Schittkowski K. Experimental design tools for ordinary and algebraic differential equations. Ind Eng Chem Res. 2007;46(26):9137–9147. doi: 10.1021/ie0703742. [DOI] [Google Scholar]
  33. Lipniacki T, Paszek P, Brasier ARAR, Luxon B, Kimmel M. Mathematical model of NF-kappaB regulatory module. J Theor Biol. 2004;228(2):195–215. doi: 10.1016/j.jtbi.2004.01.001. [DOI] [PubMed] [Google Scholar]
  34. Chu Y, Hahn J. Parameter Set Selection via Clustering of Parameters into Pairwise Indistinguishable Groups of Parameters. Industrial & Engineering Chemistry Research. 2008;48(13):6000–6009. [Google Scholar]
  35. Grewal MS, Glover K. Identifiability of linear and nonlinear dynamical-systems. IEEE T Automat Contr. 1976;21(6):833–836. doi: 10.1109/TAC.1976.1101375. [DOI] [Google Scholar]
  36. Yamada S, Shiono S, Joo A, Yoshimura A. Control mechanism of JAK/STAT signal transduction pathway. FEBS Lett. 2003;534(1-3):190–196. doi: 10.1016/S0014-5793(02)03842-5. [DOI] [PubMed] [Google Scholar]
  37. Li S, Petzold L. Design of new DASPK for Sensitivity Analysis. Technical report. 1999. (TRCS99-28)
  38. Schittkowski K. Numerical data fitting in dynamical systems: a practical introduction with applications and software. Kluwer Academic Pub; 2002. [Google Scholar]
  39. Gill P, Murray W, Saunders M, Wright M. User's guide for NPSOL 5.0: A fortran package for nonlinear programming. Technical report. 1998. (TR SOL-86-1)
  40. McKay M, Connover W, Becknam R. A Comparison of Three Methods for Selecting Values of Input Variables in the Analysis of Output from a Computer Code. Technometrics. 1979;21(2):239–245. doi: 10.2307/1268522. [DOI] [Google Scholar]
  41. Press W, Flannery B, Teukolsky S, Vetterling W. Numerical Recipes in C: The Art of Scientific Computing. Cambridge Univ. Press, Cambridge; 1992. [Google Scholar]
  42. Abramowitz M, Stegun I. Handbook of mathematical functions with formulas, graphs, and mathematical tables. Dover publications, New York; 1964. [Google Scholar]
  43. Akaike H, Petrov B, F C. Second International Symposium of Information Theory. Akademiai Kiado. Budapest, Hungary; 1973. Information theory and an extension of the maximum likelihood principle; pp. 267–281. [Google Scholar]
  44. Burnham K, Anderson D. Model selection and multimodel inference: a practical information-theoretic approach. Springer, New York; 2002. [Google Scholar]
  45. Hurvich C, Tsai C. Regression and time series model selection in small samples. Biometrika. 1989;76(2):297. doi: 10.1093/biomet/76.2.297. [DOI] [Google Scholar]
  46. Greenlund AC, Morales MO, Viviano BL, Yan H, Krolewski J, Schreiber RD. Stat recruitment by tyrosine-phosphorylated cytokine receptors: an ordered reversible affinity-driven process. Immunity. 1995;2(6):677–687. doi: 10.1016/1074-7613(95)90012-8. [DOI] [PubMed] [Google Scholar]
  47. Igarashi K, Garotta G, Ozmen L, Ziemiecki A, Wilks AF, Harpur AG, Larner AC, Finbloom DS. Interferon-gamma induces tyrosine phosphorylation of interferon-gamma receptor and regulated association of protein tyrosine kinases, Jak1 and Jak2, with its receptor. J Biol Chem. 1994;269(20):14333–14336. [PubMed] [Google Scholar]
  48. Kaplan DH, Greenlund AC, Tanner JW, Shaw AS, Schreiber RD. Identification of an interferon-gamma receptor alpha chain sequence required for JAK-1 binding. J Biol Chem. 1996;271:9–12. doi: 10.1074/jbc.271.1.9. [DOI] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Additional file 1

The supplementary pdf file accompanying this article contains the supplementary tables S1-S8b and supplementary texts 1-3.

Click here for file (81.6KB, PDF)

Articles from BMC Systems Biology are provided here courtesy of BMC

RESOURCES