Skip to main content
PLOS Computational Biology logoLink to PLOS Computational Biology
. 2012 Mar 1;8(3):e1002401. doi: 10.1371/journal.pcbi.1002401

A Self-Organizing State-Space-Model Approach for Parameter Estimation in Hodgkin-Huxley-Type Models of Single Neurons

Dimitrios V Vavoulis 1,*, Volko A Straub 2, John A D Aston 3, Jianfeng Feng 1,4,*
Editor: Lyle J Graham5
PMCID: PMC3291554  PMID: 22396632

Abstract

Traditional approaches to the problem of parameter estimation in biophysical models of neurons and neural networks usually adopt a global search algorithm (for example, an evolutionary algorithm), often in combination with a local search method (such as gradient descent) in order to minimize the value of a cost function, which measures the discrepancy between various features of the available experimental data and model output. In this study, we approach the problem of parameter estimation in conductance-based models of single neurons from a different perspective. By adopting a hidden-dynamical-systems formalism, we expressed parameter estimation as an inference problem in these systems, which can then be tackled using a range of well-established statistical inference methods. The particular method we used was Kitagawa's self-organizing state-space model, which was applied on a number of Hodgkin-Huxley-type models using simulated or actual electrophysiological data. We showed that the algorithm can be used to estimate a large number of parameters, including maximal conductances, reversal potentials, kinetics of ionic currents, measurement and intrinsic noise, based on low-dimensional experimental data and sufficiently informative priors in the form of pre-defined constraints imposed on model parameters. The algorithm remained operational even when very noisy experimental data were used. Importantly, by combining the self-organizing state-space model with an adaptive sampling algorithm akin to the Covariance Matrix Adaptation Evolution Strategy, we achieved a significant reduction in the variance of parameter estimates. The algorithm did not require the explicit formulation of a cost function and it was straightforward to apply on compartmental models and multiple data sets. Overall, the proposed methodology is particularly suitable for resolving high-dimensional inference problems based on noisy electrophysiological data and, therefore, a potentially useful tool in the construction of biophysical neuron models.

Author Summary

Parameter estimation is a problem of central importance and, perhaps, the most laborious task in biophysical modeling of neurons and neural networks. An emerging trend is to treat parameter estimation in this context as yet another statistical inference problem, which can be tackled using well-established methods from Computational Statistics. Inspired by these recent advances, we adopted a self-organizing state-space-model approach augmented with an adaptive sampling algorithm akin to the Covariance Matrix Adaptation Evolution Strategy in order to estimate a large number of parameters in a number of Hodgkin-Huxley-type models of single neurons. Parameter estimation was based on noisy electrophysiological data and involved the maximal conductances, reversal potentials, levels of noise and, unlike most mainstream work, the kinetics of ionic currents in the examined models. Our main conclusion was that parameters in complex, conductance-based neuron models can be inferred using the aforementioned methodology, if sufficiently informative priors regarding the unknown model parameters are available. Importantly, the use of an adaptive algorithm for sampling new parameter vectors significantly reduced the variance of parameter estimates. Flexibility and scalability are additional advantages of the proposed method, which is particularly suited to resolve high-dimensional inference problems.

Introduction

Among several tools at the disposal of neuroscientists today, data-driven computational models have come to hold an eminent position for studying the electrical activity of single neurons and the significance of this activity for the operation of neural circuits [1][4]. Typically, these models depend on a large number of parameters, such as the maximal conductances and kinetics of gated ion channels. Estimating appropriate values for these parameters based on the available experimental data is an issue of central importance and, at the same time, the most laborious task in single-neuron and circuit modeling.

Ideally, all unknown parameters in a model should be determined directly from experimental data analysis. For example, based on a set of voltage-clamp recordings, the type, kinetics and maximal conductances of the voltage-gated ionic currents flowing through the cell membrane could be determined [5] and, then, combined in a conductance-based model, which replicates the activity of the biological neuron of interest under current-clamp conditions with sufficient accuracy. Unfortunately, this is not always possible, especially for complex compartmental models, which contain a large number of ionic currents.

A first problem arises from the fact that not all parameters can be estimated within an acceptable error margin, especially for small currents and large levels of noise. A second problem arises from the practice of estimating different sets of parameters based on data collected from different neurons of a particular type, instead of estimating all unknown parameters using data collected from a single neuron. Different neurons of the same type may have quite different compositions of ionic currents [6][9] (but, see also [10]). This implies that combining ionic currents measured from different neurons in the same model or even using the average of several parameters calculated over a population of neurons of the same type will not necessarily result in a model that expresses the experimentally recorded patterns of electrical activity under current-clamp conditions. Usually, only some parameters are well characterized, while others are difficult or impossible to measure directly. Thus, most modeling studies rely on a mixture of experimentally determined parameters and estimates of the remaining unknown ones using automated optimization methodology (see, for example, [11][22]). Typically, these methods require the construction of a cost function (for measuring the discrepancy between various features of the experimental data and the output of the model) and an automated parameter selection method, which iteratively generates new sets of parameters, such that the value of the cost function progressively decreases during the course of the simulation (see [23] for a review). Popular choices of such methods are evolutionary algorithms, simulated annealing and gradient descent methods. Often, a global search method (i.e. an evolutionary algorithm) is combined with local search (gradient descent) for locating multiple minima of the cost function with high precision. Since a poorly designed cost function (for example, one that merely matches model and experimental membrane potential trajectories) can seriously impede optimization, the construction of this function often requires particular attention (see, for example, [24]). Nevertheless, these computationally intensive methodologies have gained much popularity, particularly due to the availability of powerful personal computers at consumer-level prices and the development of specialized optimization software (e.g. [25]).

Alternative approaches also exist as, for example, methods based on the concept of synchronization between model dynamics and experimental data [26]. An emerging trend in parameter estimation methodologies for models in Computational Biology is to recast parameter estimation as an inference problem in hidden dynamical systems and then adopt standard Computational Statistics techniques to resolve it [27], [28]. For example, a particular study following this approach makes use of Sequential Monte Carlo methods (particle filters) embedded in an Expectation Maximization (EM) framework [28]. Given a set of electrophysiological recordings and a set of dynamic equations that govern the evolution of the hidden states, at each iteration of the algorithm the expected joint log-likelihood of the hidden states and the data is approximated using particle filters (Expectation Step). At a second stage during each iteration (Maximization Step), the log-likelihood is locally maximized with respect to the unknown parameters. The advantage of these methods, beyond the fact that they recast the estimation problem in a well-established statistical framework, is that they can handle various types of noisy biophysical data made available by recent advances in voltage and calcium imaging techniques.

Inspired by this emerging approach, we present a method for estimating a large number of parameters in Hodgkin-Huxley-type models of single neurons. The method is a version of Kitagawa's self-organizing state-space model [29] combined with an adaptive algorithm for selecting new sets of model parameters. The adaptive algorithm we have used is akin to the Covariance Matrix Adaption (CMA) Evolution Strategy [30], but other methods (e.g. Differential Evolution as described in [31]) may be used instead. We demonstrate the applicability of the algorithm on a range of models using simulated or actual electrophysiological data. We show that the algorithm can be used successfully with very noisy data and it is straightforward to apply on compartmental models and multiple datasets. An interesting result from this study is that by using the self-organizing state-space model in combination with a CMA-like algorithm, we managed to achieve a dramatic reduction in the variance of the inferred parameter values. Our main conclusion is that a large number of parameters in a conductance-based model of a neuron (including maximal conductances, reversal potentials and kinetics of gated ionic currents) can be inferred from low-dimensional experimental data (typically, a single or a few recordings of membrane potential activity) using the algorithm, if sufficiently informative priors are available, for example in the form of well-defined ranges of valid parameter values.

Methods

Modeling Framework

We begin by presenting the current conservation equation that describes the time evolution of the membrane potential for a single-compartment model neuron:

graphic file with name pcbi.1002401.e001.jpg (1)

where Inline graphic, Inline graphic and Inline graphic are all functions of time. In the above equation, Inline graphic is the membrane capacitance, Inline graphic is the membrane potential, Inline graphic is the externally applied (injected) current, Inline graphic and Inline graphic are the maximal conductance and reversal potential of the leakage current, respectively, and Inline graphic is the Inline graphic transmembrane ionic current. A voltage-gated current Inline graphic can be modeled according to the Hodgkin-Huxley formalism, as follows:

graphic file with name pcbi.1002401.e013.jpg (2)

where Inline graphic and Inline graphic are both functions of time. In the above expression, Inline graphic and Inline graphic are the maximal conductance and reversal potential of the Inline graphic ionic current, Inline graphic and Inline graphic are dynamic gating variables, which model the voltage-dependent activation and inactivation of the current, and Inline graphic is a small positive integer power (usually, not taking values larger than 4). The product Inline graphic is the proportion of open channels in the membrane that carry the Inline graphic current. The gating variables Inline graphic and Inline graphic obey first-order relaxation kinetics, as shown below:

graphic file with name pcbi.1002401.e026.jpg (3)

where the steady states (Inline graphic, Inline graphic) and relaxation times (Inline graphic, Inline graphic) are all functions of voltage.

Using vector notation, we can write the above system of Ordinary Differential Equations (ODEs) in more concise form:

graphic file with name pcbi.1002401.e031.jpg (4)

where the state vector Inline graphic is composed of the time-evolving state variables Inline graphic, Inline graphic and Inline graphic and the vector-valued function Inline graphic, which describes the evolution of Inline graphic in time, is formed by the right-hand sides of Eqs. 1 and 3. Notice that Inline graphic also depends on a parameter vector Inline graphic, which for now is dropped from Eq. 4 for notational clarity. Components of Inline graphic are the maximal conductances Inline graphic, the reversal potentials Inline graphic and the various parameters that control the voltage-dependence of the steady states and relaxation times in Eq. 3.

The above deterministic model does not capture the inherent variability in the electrical activity of neurons, but rather some average behavior of intrinsically stochastic events. In general, this variability originates from various sources, such as the random opening and shutting of transmembrane ion channels or the random bombardment of the neuron with external (e.g. synaptic) stimuli [32]. Here, we model the inherent variability in single-neuron activity by augmenting Eq. 4 with a noisy term and re-writing as follows:

graphic file with name pcbi.1002401.e043.jpg (5)

where Inline graphic is a covariance matrix and Inline graphic is a standard Wiener process over the state space of Inline graphic. Inline graphic may be a diagonal matrix of variances (Inline graphic, Inline graphic and Inline graphic) corresponding to each component of the state vector.

Typically, we assume that the above model is coupled to a measurement “device”, which permits indirect observations of the hidden state Inline graphic:

graphic file with name pcbi.1002401.e052.jpg (6)

where Inline graphic is an observation noise vector. In the simplest case, the vector of observations Inline graphic is one-dimensional and it may consist of noisy measurements of the membrane potential:

graphic file with name pcbi.1002401.e055.jpg (7)

where Inline graphic is the standard deviation of the observation noise and Inline graphic a random number sampled from a Gaussian distribution with zero mean and standard deviation equal to unity. More complicated non-linear, non-Gaussian observation functions may be used when, for example, the measurements are recordings of the intracellular calcium concentration, simultaneous recordings of the membrane potential and the intracellular calcium concentration or simultaneous recordings of the membrane potential from multiple sites (e.g. soma and dendrites) of a neuron.

Assuming that time Inline graphic is partitioned in a very large number Inline graphic of time steps Inline graphic, such that Inline graphic and the corresponding states are Inline graphic, we can approximate the solution to Eq. 5 using the following difference equation:

graphic file with name pcbi.1002401.e063.jpg (8)

where Inline graphic and Inline graphic is a random vector with components sampled from a normal distribution with zero mean and unit variance. The above expression implements a simple rule for computing the membrane potential, activation and inactivation variables at each point Inline graphic of the discretized time based on information at the previous time point Inline graphic and it can be considered as a specific instantiation of the Euler-Maruyama method for the numerical solution of Stochastic Differential Equations [33].

Then, the observation model becomes:

graphic file with name pcbi.1002401.e068.jpg (9)

In general, measurements do not take place at every point Inline graphic of the discretized time, but rather at intervals of Inline graphic time steps (depending on the resolution of the measurement device), thus generating a total of Inline graphic measurements. For simplicity in the above description, we have assumed that Inline graphic. However, all the models we consider in the Results section assume Inline graphic.

In terms of probability density functions, the non-linear state-space model defined by Eqs. 8 and 9 (known as the dynamics model) and the observation model, respectively) can be written as:

graphic file with name pcbi.1002401.e074.jpg (10)
graphic file with name pcbi.1002401.e075.jpg (11)

where the initial state Inline graphic is distributed according to a prior density Inline graphic. The above formulas are known as the state transition and observation densities, respectively [34].

Simulation-Based Filtering and Smoothing

In many inference problems involving state-space models, a primary concern is the sequential estimation of the following two conditional probability densities [29]: (a) Inline graphic and (b) Inline graphic, where Inline graphic, i.e. the set of observations (for example, a sequence of measurements of the membrane potential) up to the time point Inline graphic. Density (a), known as the filter density, models the distribution of state Inline graphic given all observations up to and including the time point Inline graphic, while density (b), known as the smoother density, models the distribution of state Inline graphic given the whole set of observations up to the final time point Inline graphic.

In principle, the filter density can be estimated recursively at each time point Inline graphic using Bayes' rule appropriately [29]:

graphic file with name pcbi.1002401.e087.jpg (12)

where Inline graphic and Inline graphic are the state transition and observation densities, respectively, and Inline graphic is the filter density at the previous time step Inline graphic.

Then, the smoother density can be obtained by using the following general recursive formula:

graphic file with name pcbi.1002401.e092.jpg (13)

which evolves backwards in time and makes use of the pre-calculated filter, Inline graphic. Given either of the above posterior densities, we can compute the expectation of any useful function of the hidden model state as:

graphic file with name pcbi.1002401.e094.jpg (14)

where Inline graphic is either the filter or the smoother density. Common examples of Inline graphic are Inline graphic itself (giving the mean Inline graphic) and the squared difference from the mean (giving the covariance of Inline graphic).

In practice, the computations defined by the above formulas can be performed analytically only for linear Gaussian models using the Kalman smoother/filter and for finite state-space hidden Markov models. For non-linear models, the extended Kalman filter is a popular approach, which however can fail when non-Gaussian or multimodal density functions are involved [34]. A more generally applicable, albeit computationally more intensive approach, approximates the filter and smoother densities using Sequential Monte Carlo (SMC) methods, also known as particle filters [34], [35]. Within the SMC framework, the filter density at each time point is approximated by a large number Inline graphic of discrete samples or particles, Inline graphic, and associated non-negative importance weights, Inline graphic:

graphic file with name pcbi.1002401.e103.jpg (15)

where Inline graphic is the Dirac delta function centered at the Inline graphic particle, Inline graphic.

Given an initial set of particles sampled from a prior distribution and their associated weights, a simple update rule involves the following steps [29]:

Step 1: For Inline graphic, sample a new set of particles from the proposal transition density function, Inline graphic. In general, one has enormous freedom in choosing the form of this density and even condition it on future observations, if these are available (see, for example, [36]). However, the simplest (and a quite common) choice is to use the transition density as the proposal, i.e. Inline graphic. This is the approach we follow in this paper.

Step 2: For each new particle Inline graphic, evaluate the importance weight:

graphic file with name pcbi.1002401.e111.jpg (16)

Notice that when Inline graphic, then the computation of the importance weights is significantly simplified, i.e. Inline graphic.

Step 3: Normalize the computed importance weights, by dividing each of them with their sum, i.e.

graphic file with name pcbi.1002401.e114.jpg (17)

The derived set of weighted samples Inline graphic is considered an approximation of the filter density Inline graphic.

In practice, the above algorithm is augmented with a re-sampling step (preceding Step 1), during which Inline graphic particles are sampled from the set of weighted particles computed at the previous iteration with probabilities proportional to their weights [34], [35]. All re-sampled particles are given weights equal to Inline graphic. This step results in discarding particles with small weights and multiplying particles with large weights, thus compensating for the gradual degeneration of the particle filter i.e. the situation where all particles but one have weights equal to zero. For performance reasons, the resampling step may be applied only when the effective number of particles drops below a threshold value, e.g. Inline graphic. An estimation of the effective number of particles is given by

graphic file with name pcbi.1002401.e120.jpg (18)

The above filter can be extended to a fixed-lag smoother, if instead of resampling just the particles at the current time step, we store and resample all particles up to Inline graphic time steps before the current time step, i.e. Inline graphic [29]. The resampled particles can be considered a realization from a posterior density Inline graphic, which is an approximation of the smoother density Inline graphic, for sufficiently large values of Inline graphic.

Within this Monte Carlo framework, the expectation in Eq. 14 can be approximated as:

graphic file with name pcbi.1002401.e126.jpg (19)

for a large number Inline graphic of weighted samples.

Simultaneous Estimation of Hidden States and Parameters

It is possible to apply the above standard filtering and smoothing techniques to parameter estimation problems involving state-space models. The key idea [29] is to define an extended state vector Inline graphic by augmenting the state vector Inline graphic with the model parameters, i.e. Inline graphic. Then, the time evolution of the extended state-space model becomes:

graphic file with name pcbi.1002401.e131.jpg (20)

while the observational model remains unaltered:

graphic file with name pcbi.1002401.e132.jpg (21)

The marginal posterior density of the parameter vector Inline graphic is given by:

graphic file with name pcbi.1002401.e134.jpg (22)

and, subsequently, the expectation of any function of Inline graphic can be computed as in Eq. 14:

graphic file with name pcbi.1002401.e136.jpg (23)

Furthermore, given a set of particles and associated weights, which approximate the smoother density Inline graphic as outlined in the previous section, i.e. Inline graphic for Inline graphic, the above expectation can be approximated as:

graphic file with name pcbi.1002401.e140.jpg (24)

for large Inline graphic.

Under this formulation, parameter estimation, which is traditionally treated as an optimization problem, is reduced to an integration problem, which can be tackled using filtering and smoothing methodologies for state-space models, a well-studied subject in the field of Computational Statistics.

Connection to Evolutionary Algorithms

It should be emphasized that although in Eq. 20 the parameter vector Inline graphic was assumed constant, i.e. Inline graphic, the same methodology applies in the case of parameters that are naturally evolving in time, such as a time-varying externally injected current Inline graphic. A particularly interesting case arises when an artificial evolution rule is imposed on a parameter vector, which is otherwise constant by definition. Such a rule allows sampling new parameter vectors based on samples at the previous time step, i.e. Inline graphic, and generating a sequence Inline graphic, which explores the parameter space and, ideally converges in a small optimal subset of it, after a sufficiently large number of iterations. It is at this point that the opportunity to use techniques borrowed from the domain of Evolutionary Algorithms arises. Here, we assume that the artificial evolution of the parameter vector Inline graphic is governed by a version of the Covariance Matrix Adaptation algorithm [30], a well-known Evolution Strategy, although the modeler is free to make other choices (e.g. Differential Evolution [31]). For the Inline graphic particle, we write:

graphic file with name pcbi.1002401.e149.jpg (25)

where Inline graphic is a random vector with elements sampled from a normal distribution with zero mean and unit variance. Inline graphic and Inline graphic are a mean vector and covariance matrix respectively, which are computed as follows:

graphic file with name pcbi.1002401.e153.jpg (26)
graphic file with name pcbi.1002401.e154.jpg (27)

In the above expressions, a and Inline graphic are small adaptation constants and Inline graphic and Inline graphic are the expectation and covariance of the weighted sample of Inline graphic, respectively. Inline graphic is a scale parameter that evolves according to a log-normal update rule:

graphic file with name pcbi.1002401.e160.jpg (28)

where Inline graphic is a small adaptation constant and Inline graphic is a normally distributed random number with zero mean and unit variance.

According to Eq. 25, the parameter vector Inline graphic is sampled at each iteration of the algorithm from a multivariate normal distribution, which is centered at Inline graphic and has a covariance matrix equal to Inline graphic:

graphic file with name pcbi.1002401.e166.jpg (29)

Both Inline graphic and Inline graphic are slowly adapting to the sample mean Inline graphic and covariance Inline graphic, with an adaptation rate determined by the constants Inline graphic and Inline graphic. Notice that by switching off the adaptation process (i.e. by setting Inline graphic), Inline graphic evolves according to a multivariate Gaussian distribution, which is centered at the previous parameter vector and has a covariance matrix equal to Inline graphic:

graphic file with name pcbi.1002401.e176.jpg (30)

Therefore, given an initial set of weighted particles Inline graphic sampled from some prior density function and an initial covariance matrix Inline graphic, which may be set equal to the identity matrix, the smoothing algorithm presented earlier becomes:

Step 1a: Compute the expectation Inline graphic and covariance Inline graphic of the weighted sample of Inline graphic

Step 1b: For Inline graphic, compute the scale factor Inline graphic according to Eq. 28. Notice that this scale factor is now part of the extended state Inline graphic for each particle

Step 1c: For Inline graphic, compute the mean vector Inline graphic, as shown in Eq. 26

Step 1d: Compute the covariance matrix Inline graphic, as shown in Eq. 27

Step 1e: For Inline graphic, sample Inline graphic, as shown in Eq. 25

Step 1f: For Inline graphic, sample a new set of state vectors from the proposal density Inline graphic, thus completing sampling the extended vectors Inline graphic. Notice that the proposal density Inline graphic is conditioned on the updated parameter vector Inline graphic.

Step 2–3: Execute steps 2 and 3 as described previously

Notice that in the algorithm outlined above, the order in which the components of Inline graphic are sampled is important. First, we sample the scaling factor Inline graphic. Then, we sample the parameter vector Inline graphic given the updated Inline graphic. Finally, we sample the state vector Inline graphic from a proposal, which is conditioned on the updated parameter vector Inline graphic. When resampling occurs, the state vectors Inline graphic with large importance weights are selected and multiplied with high probability along with their associated parameter vectors and scaling factors, thus resulting in a gradual self-adaptation process. This self-adaptation mechanism is very common in the Evolution Strategies literature.

Implementation

The algorithm described in the previous section was implemented in MATLAB and C (source code available as Supplementary Material; unmaintained FORTRAN code is also available upon request from the first author) and tested on parameter inference problems using simulated or actual electrophysiological data and a number of Hodgkin-Huxley-type models: (a) a single-compartment model (derived from the classic Hodgkin-Huxley model of neural excitability) containing a leakage, transient sodium and delayed rectifier potassium current, (b) a two-compartment model of a cat spinal motoneuron [37] and (c) a model of a B4 motoneuron in the Central Nervous System of the pond snail Lymnaea stagnalis [38], which was developed as part of this study. Each of these models is described in detail in the Results section. Models (a) and (b) were used for generating noisy voltage traces at a sampling rate of Inline graphic (one sample every Inline graphic). The simulated data was subsequently used as input to the algorithm in order to estimate a large number of parameters; typically, maximal conductances of ionic currents, reversal potentials, the parameters governing the activation and inactivation kinetics of ionic currents, as well as the levels of intrinsic and observation noise. Estimated parameter values were subsequently compared against the true parameter values in the model. The MATLAB environment was used for visualization and analysis of simulation results. For the estimation of the unknown parameters in model (c), actual electrophysiological data were used, as described in the next section.

Prior information was incorporated in the smoother by assuming that parameter values were not allowed to exceed well-defined upper or lower limits (see Tables 1, 2 and 3). For example, maximal conductances never received negative values, while time constants were always larger than zero. At the beginning of each simulation, the initial population of particles was uniformly sampled from within the acceptable range of parameter values and, during each simulation, parameters were forced to remain within their pre-defined limits.

Table 1. True and estimated values and prior intervals used during smoothing for all parameters in the single-compartment conductance-based model.

# Parameter Unit True Value Estimated Value1 Lower Bound Upper Bound
1 Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic
2 Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic1 Inline graphic
3 Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic
4 Inline graphic Inline graphic Inline graphic.0 34.3 Inline graphic Inline graphic
5 Inline graphic Inline graphic Inline graphic.0 Inline graphic Inline graphic Inline graphic
6 Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic
7 Inline graphic Inline graphic Inline graphic.0 Inline graphic Inline graphic Inline graphic
8 Inline graphic Inline graphic Inline graphic.0 Inline graphic Inline graphic Inline graphic
9 Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic (−45.0) 2 Inline graphic (−35.0)
10 Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic (−65.0) Inline graphic (−55.0)
11 Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic (−55.0) Inline graphic (−45.0)
12 Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic (5.0) Inline graphic (10.0)
13 Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic (−10.0) Inline graphic (−5.0)
14 Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic (10.0) Inline graphic (20.0)
15 Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic
16 Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic
17 Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic
18 Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic
19 Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic
20 Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic
21 Inline graphic - Inline graphic Inline graphic Inline graphic (0.0) Inline graphic (0.5)
22 Inline graphic - Inline graphic Inline graphic Inline graphic (0.0) Inline graphic (0.5)
23 Inline graphic - Inline graphic Inline graphic Inline graphic (0.5) Inline graphic (1.0)
1

These parameter values were estimated when we used the broad prior intervals (see Fig. 7Di).

2

Values in bold indicate the narrow prior intervals we used for generating Fig. 7Dii (and Supplementary Fig. S3).

Table 2. True and estimated values and prior intervals used during smoothing for all parameters in the two-compartment conductance-based model.

# Parameter Unit True Value Estimated Value1 Lower Bound Upper Bound
1 Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic
2 Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic
3 Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic
4 Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic
5 Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic
6 Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic
7 Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic
8 Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic
9 Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic
10 Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic
11 Inline graphic Inline graphic −35.0 Inline graphic −60.0 (−45.0) 2 Inline graphic (−25.0)
12 Inline graphic Inline graphic Inline graphic Inline graphic −60.0 (−65.0) Inline graphic (−45.0)
13 Inline graphic Inline graphic Inline graphic Inline graphic −60.0 (−40.0) Inline graphic (−20.0)
14 Inline graphic Inline graphic Inline graphic Inline graphic −60.0 (−40.0) Inline graphic (−20.0)
15 Inline graphic Inline graphic Inline graphic Inline graphic −60.0 (−55.0) Inline graphic (−35.0)
16 Inline graphic Inline graphic Inline graphic Inline graphic −60.0 (−50.0) Inline graphic (−30.0)
17 Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic (5.0) Inline graphic (10.0)
18 Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic (−10.0) Inline graphic (−5.0)
19 Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic (10.0) Inline graphic (20.0)
20 Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic (3.0) Inline graphic (8.0)
21 Inline graphic Inline graphic Inline graphic0 Inline graphic Inline graphic (−8.0) Inline graphic (−3.0)
22 Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic (5.0) Inline graphic (10.0)
23 Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic
24 Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic
25 Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic
26 Inline graphic - Inline graphic Inline graphic Inline graphic (0.5) Inline graphic (1.0)
27 Inline graphic - Inline graphic Inline graphic Inline graphic (0.5) Inline graphic (1.0)
28 Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic
29 Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic
30 Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic
1

These parameter values were estimated when we used the broad prior intervals (see Fig. 11Ai).

2

Values in bold indicate the narrow prior intervals we used for generating Figs. 11Aii, 11B, 11C (and Supplementary Figs. S4 and S5).

Table 3. Estimated mean values and prior limits used during smoothing for all parameters in the B4 model.

# Parameter Unit Estimated Mean Value1 , 2 Lower Bound Upper Bound
1 Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic
2 Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic
3 Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic
4 Inline graphic Inline graphic Inline graphic Inline graphic (−40.0) Inline graphic (−20.0)
5 Inline graphic Inline graphic Inline graphic Inline graphic (−40.0) Inline graphic (−20.0)
6 Inline graphic Inline graphic Inline graphic Inline graphic (−40.0) Inline graphic (−20.0)
7 Inline graphic Inline graphic Inline graphic Inline graphic (−20.0) Inline graphic (0.0)
8 Inline graphic Inline graphic Inline graphic Inline graphic (−70.0) Inline graphic (−40.0)
9 Inline graphic Inline graphic Inline graphic Inline graphic (5.0) Inline graphic (10.0)
10 Inline graphic Inline graphic Inline graphic Inline graphic (−10.0) Inline graphic (−5.0)
11 Inline graphic Inline graphic Inline graphic Inline graphic (10.0) Inline graphic (15.0)
12 Inline graphic Inline graphic Inline graphic Inline graphic (5.0) Inline graphic (10.0)
13 Inline graphic Inline graphic Inline graphic Inline graphic (−25.0) Inline graphic (−15.0)
14 Inline graphic Inline graphic Inline graphic Inline graphic (15.0) Inline graphic (25.0)
15 Inline graphic Inline graphic Inline graphic Inline graphic (25.0) Inline graphic (35.0)
16 Inline graphic Inline graphic Inline graphic Inline graphic (25.0) Inline graphic (35.0)
17 Inline graphic Inline graphic Inline graphic Inline graphic (35.0) Inline graphic (60.0)
1

These parameter values were estimated when we used the narrow prior intervals (in bold; see Fig. 12).

2

The parameter posteriors estimated when we used the broad prior intervals are illustrated in Supplementary Fig. S7.

All simulations were performed on an Intel dual-core i5 processor with 4 GB of memory running Ubuntu Linux. The number of particles used in each simulation was typically Inline graphic, where Inline graphic was the dimensionality of the extended state Inline graphic (equal to the number of free parameters and dynamic states in the model). The time step Inline graphic in the Euler-Maruyama method was set equal to Inline graphic. The parameter Inline graphic of the fixed-lag smoother was set equal to Inline graphic (unless stated otherwise), which is equivalent to a time window Inline graphic wide (since data were sampled every Inline graphic). The adaptation constants Inline graphic, Inline graphic and Inline graphic in Eqs. 26, 27 and 28 were all set equal to Inline graphic, unless stated otherwise. Depending on the size of Inline graphic, the complexity of the model and the length of the (actual or simulated) electrophysiological recordings, simulation times ranged from a few minutes up to more than Inline graphic hours.

Electrophysiology

As part of this study, we developed a single-compartment Hodgkin-Huxley-type model of a B4 neuron in the pond-snail Lymnaea stagnalis [38]. B4 neurons are part of the neural circuit that controls the rhythmic movements of the feeding muscles via which the animal captures and ingests its food. The Lymnaea central nervous system was dissected from adult animals (shell length Inline graphic) that were bred at the University of Leicester as described previously [39]. All dissections were carried out in Inline graphic-buffered saline containing (in Inline graphic) Inline graphic, Inline graphic, Inline graphic, Inline graphic, and Inline graphic, Inline graphic, in distilled water. All chemicals were purchased from Sigma. The buccal ganglia containing the B4 neurons were separated from the rest of the nervous system by cutting the cerebral buccal connectives and the buccal-buccal connective was crushed to eliminate electrical coupling between B4 neurons in the left and right buccal ganglion. Prior to recording, excess saline was removed from the dish and small crystals of protease type XIV were placed directly on top of the buccal ganglia to soften the connective tissue and aid the impalement of individual neurons. The protease crystals were washed of after about Inline graphic with multiple changes of Inline graphic-buffered saline. The B4 neuron was visually identified based on its size and position and impaled with two sharp intracellular electrodes filled with a mixture of Inline graphic potassium acetate and Inline graphic potassium chloride (resistance Inline graphic). During the recording, the preparation was bathed in Inline graphic-buffered saline plus Inline graphic hexamethonium chloride to block cholinergic synaptic inputs and suppress spontaneous fictive feeding activity.

The signals from the two intracellular electrodes were amplified using a Multiclamp 900A amplifier (Molecular Devices), digitized at a sampling frequency of Inline graphic using a CED1401plus A/D converter (Cambridge Electronic Devices) and recorded on a PC using Spike2 version 6 software (Cambridge Electronic Devices). A custom set of instructions using the Spike2 scripting language was used to generate sequences of current pulses consisting of individual random steps ranging in amplitude from Inline graphic to Inline graphic and a duration from Inline graphic to Inline graphic. The current signal was injected through one of the recording electrodes whilst the second electrode was used to measure the resulting changes in membrane potential.

Results

Hidden States, Intrinsic and Observational Noise are Simultaneously Estimated Using the Fixed-Lag Smoother

The applicability of the fixed-lag smoother presented above was demonstrated on a range of Hodgkin-Huxley-type models using simulated or actual electrophysiological data. The first model we examined consisted of a single compartment containing leakage, sodium and potassium currents, as shown below:

graphic file with name pcbi.1002401.e630.jpg
graphic file with name pcbi.1002401.e631.jpg (32)

where Inline graphic. Notice the absence of noise in the dynamics of Inline graphic, Inline graphic and Inline graphic, which is valid if we assume a very large number of channels (see Supplementary Material and Supplementary Figures S1 and S2 for the case were noise is present in the dynamics of these variables). The steady states and relaxation times of the activation and inactivation gating variables were voltage-dependent, as shown below (e.g. [5]):

graphic file with name pcbi.1002401.e636.jpg (33)

and

graphic file with name pcbi.1002401.e637.jpg (34)

where Inline graphic and Inline graphic. The parameters Inline graphic, Inline graphic, Inline graphic, Inline graphic and Inline graphic in Eqs. 33 and 34 were chosen such that Inline graphic and Inline graphic fit closely the corresponding steady-states and relaxation times of the classic Hodgkin-Huxley model of neural excitability in the giant squid axon [40]. Observations consisted of noisy measurements of the membrane potential, as shown in Eq. 7. The full set of parameter values in the above model is given in Table 1.

First, we used the fixed-lag smoother to simultaneously infer the hidden states (Inline graphic, Inline graphic, Inline graphic, Inline graphic) and standard deviations of the intrinsic (Inline graphic) and observation (Inline graphic) noise based on Inline graphic-long simulated recordings of the membrane potential Inline graphic. These recordings were generated by assuming a time-dependent Inline graphic in Eq. 31, which consisted of a sequence of current steps with amplitude randomly distributed between Inline graphic and Inline graphic and random duration up to a maximum of Inline graphic. Two simulated voltage recordings were generated corresponding to two different levels of observation noise, Inline graphic and Inline graphic, respectively. The second value (Inline graphic) was rather extreme and it was chosen in order to illustrate the applicability of the method even at very high levels of observation noise. Simulated data points were sampled every Inline graphic (Inline graphic). The standard deviation of the intrinsic noise was set at Inline graphic. The injected current Inline graphic and the induced voltage trace (for either value of Inline graphic) were then used as input to the smoother, during the inference phase. At this stage, all other parameters in the model (conductances, reversal potentials, and ionic current kinetics) were assumed known, thus the extended state vector took the form Inline graphic, where Inline graphic was a scale factor as in Eq. 25. New samples for Inline graphic were taken from a log-normal distribution (Eq. 28), while new samples for Inline graphic and Inline graphic were drawn from an adaptive bivariate Gaussian distribution at each iteration of the algorithm (Eq. 25). For each data set, smoothing was repeated for two different values of the smoothing lag, i.e. Inline graphic and Inline graphic. Inline graphic corresponds to filtering, while Inline graphic corresponds to smoothing with a fixed lag equal to Inline graphic. Our results from this set of simulations are summarized in Fig. 1.

Figure 1. Simultaneous estimation of hidden states, intrinsic and observation noise.

Figure 1

Estimation was based on a simulated recording of membrane potential with duration Inline graphic. For clarity, only Inline graphic of activity are shown in A and Bi,ii. (A) Smoothing of the membrane potential (the observed variable), when observation noise was low (Inline graphic). High-fidelity smoothing was achieved for either small (Inline graphic) or large (Inline graphic) values of the fixed smoothing lag Inline graphic. Simulated and smoothed data are difficult to distinguish due to their overlap. (Bi) Smoothing of the membrane potential at high levels of observation noise (Inline graphic). A large value of the smoothing lag (Inline graphic) was required for high-fidelity smoothing. (Bii) Inference of the unobserved activation (Inline graphic, Inline graphic) and inactivation (Inline graphic) variables for sodium and potassium currents as functions of time, during smoothing of the data shown in Bi for Inline graphic. (Ci) Inference of the standard deviations for the intrinsic and observation noise (Inline graphic and Inline graphic, respectively) during smoothing of the data shown in Bi for Inline graphic. Dashed lines indicate the true values of Inline graphic and Inline graphic. (Cii) Histograms of the time series for Inline graphic and Inline graphic in Ci. Again, dashed lines indicate the true values of the corresponding parameters. At this stage, maximal conductances, reversal potentials and kinetic parameters in the model were assumed known. The number of particles was Inline graphic. Also, Inline graphic. The scaling factors in Eq. 25 were all considered equal to Inline graphic.

We observed that at low levels of observation noise (Fig. 1A), the inferred expectation of the voltage (solid blue and red lines) closely matched the underlying (true) signal (solid black line). This was true for both values of the fixed lag Inline graphic used for smoothing. However, at high levels of observation noise (Fig. 1Bi), the true voltage was inferred with high fidelity when a large value of the fixed lag (Inline graphic) was used (solid red line), but not when Inline graphic (solid blue line). Furthermore, the inferred expectations of the unobserved dynamic variables Inline graphic, Inline graphic and Inline graphic (solid red lines in Fig. 1Bii) also matched the true hidden time series (solid black lines in the same figure) remarkably well, when Inline graphic.

We repeat that during these simulations an artificial update rule was imposed on the two free standard deviations Inline graphic and Inline graphic, as shown in Eq. 25. The artificial evolution of these parameters is illustrated in Fig. 1Ci, where the inferred expectations of Inline graphic and Inline graphic are presented as functions of time. These expectations converged immediately, fluctuating around the true values of Inline graphic and Inline graphic (dashed lines in Fig. 1Ci). This is also illustrated by the histograms in Fig. 1Cii, which were constructed from the data points in Fig. 1Ci. We observed that the peaks of these histograms were located quite closely to the true values of Inline graphic and Inline graphic (dashed lines in Fig. 1Cii).

In summary, the fixed-lag smoother was able to recover the hidden states and standard deviations of the intrinsic and observation noise in the model based on noisy observations of the membrane potential. This was true even at high levels of observation noise, subject to the condition that a sufficiently large smoothing lag Inline graphic was adopted during the simulation.

Adaptive Sampling Reduces the Variance of Inferred Parameter Distributions and Accelerates Convergence of the Algorithm

Next, we treated two more parameters in the model as unknown, i.e. the maximal conductances of the transient sodium (Inline graphic) and delayed rectifier potassium (Inline graphic) currents. The extended state vector, thus, took the form Inline graphic. As in the previous section, new samples for Inline graphic were drawn from a log-normal distribution (Eq. 28), while Inline graphic, Inline graphic, Inline graphic and Inline graphic were sampled by default from an adaptive multivariate Gaussian distribution at each iteration of the algorithm (Eq. 25).

In order to examine the effect of this adaptive sampling approach on the variance of the inferred parameter distributions, we repeated fixed-lag smoothing on Inline graphic-long simulated recordings of the membrane potential assuming each time that different aspects of the adaptive sampling process were switched off, as illustrated in Fig. 2. First, we assumed that no adaptation was imposed on Inline graphic or the “unknown” noise parameters and maximal conductances, i.e. the constants Inline graphic, Inline graphic and Inline graphic in Eqs. 26–28 were all set equal to zero. In this case, the multivariate Gaussian distribution from which new samples of Inline graphic, Inline graphic, Inline graphic and Inline graphic were drawn from reduced to Eq. 30. In addition, we assumed that Inline graphic in the same equation was equal to Inline graphic, for all samples Inline graphic. Under these conditions, the true values of the free parameters were correctly estimated through application of the fixed-lag smoother, as illustrated for the case of Inline graphic and Inline graphic in Figs. 2Ai and 2Aii.

Figure 2. The effect of adaptive parameter sampling on the variance of parameter estimates.

Figure 2

Merging the fixed-lag smoother with an adaptive sampling algorithm akin to the Covariance Matrix Adaptation Evolution Strategy reduced significantly the variance of parameter estimates. At this stage, the maximal conductances for the sodium (Inline graphic) and potassium (Inline graphic) currents were assumed unknown. Estimation was based on a simulated recording of membrane potential with duration Inline graphic and Inline graphic. (A) Inference of Inline graphic and Inline graphic during smoothing, when new parameter samples were drawn from a non-adaptive multi-variate normal distribution (Eq. 30). Dashed lines indicate the true parameter values. (B) Inference of Inline graphic and Inline graphic during smoothing, when new samples were drawn from a multi-variate normal distribution (Eq. 25) with an adaptive scaling factor Inline graphic (Inline graphic in Eq. 28). (C) Inference of Inline graphic and Inline graphic during smoothing, when new samples were drawn from a multi-variate normal distribution (Eq. 25) with adaptive scaling (as in B) and mean (Inline graphic in Eq. 26). (D) Inference of Inline graphic and Inline graphic during smoothing, when new samples were drawn from a multi-variate normal distribution with adaptive scaling (as in B), mean (as in C) and covariance (Inline graphic in Eq. 27). The histograms in the right plots were constructed from the time series in the left plots. Membrane potential, activation and inactivation variables, intrinsic and observation noise were also subject to estimation, as in Fig. 1. Smoothing lag and number of particles were Inline graphic and Inline graphic, respectively. The prior interval of the scaling factors Inline graphic was Inline graphic.

Subsequently, we repeated smoothing assuming that the scale factor Inline graphic evolved according to the log-normal update rule given by Eq. 28 with Inline graphic, while Inline graphic and Inline graphic were again set equal to Inline graphic. As illustrated in Figs. 2Bi and 2Bii for parameters Inline graphic and Inline graphic, by imposing this simple adaptation rule on the multivariate Gaussian distribution from which the free parameters in the model were sampled, we managed again to estimate correctly their values, but this time the variance of the inferred parameter distributions (the width of the histograms in Fig. 2Bii) was drastically reduced.

By further letting the mean and covariance of the proposal Gaussian distribution in Eq. 25 adapt (by setting Inline graphic in Eqs. 26 and 27), we achieved a further decrease in the spread of the inferred parameter distributions (Figs. 2C and 2D). Parameters Inline graphic and Inline graphic and the hidden states Inline graphic, Inline graphic, Inline graphic and Inline graphic were also inferred with very high fidelity in all cases (as in Fig. 1), but the variance of the estimated posteriors for Inline graphic and Inline graphic followed the same pattern as the variance of Inline graphic and Inline graphic.

It is worth observing that when all three adaptation processes were switched on (i.e. Inline graphic), the algorithm converged to a single point in parameter space within the first Inline graphic of simulation, which coincided with the true parameter values in the model (see Fig. 2D for the case of Inline graphic and Inline graphic). At this point, the covariance matrix Inline graphic became very small (i.e. all its elements were less than Inline graphic, although the matrix itself remained non-singular) and the mean Inline graphic was very close to the true parameter vector Inline graphic. We note that Inline graphic and Inline graphic, where Inline graphic stands for the expectation computed over the population of particles. In this case, it is not strictly correct to claim that the chains in Fig. 2Di approximate the posteriors of the unknown parameters Inline graphic and Inline graphic; since repeating the simulation many times would result in convergence at slightly different points clustered tightly around the true parameter values, it would be more reasonable to claim that these optimal points are random samples from the posterior parameter distribution and they can be treated as estimates of its mode.

Depending on the situation, one may wish to estimate the full posteriors of the unknown parameters or just an optimal set of parameter values, which can be used in a subsequent predictive simulation. In Fig. 3A, we examined in more detail how the scale factor Inline graphic affects the variance of the final estimates, assuming that Inline graphic. We repeat that each particle Inline graphic contains Inline graphic as a component of its extended state. Each scaling factor Inline graphic is updated at each iteration of the algorithm following a lognormal rule (Eq. 28, Step 1b of the algorithm in the Methods section). Sampling new parameter vectors is conditioned on these updated scaling factors (Eq. 25, Step 1e of the algorithm). When at a later stage weighting (and resampling) of the particles occurs, the scaling factors that are associated with high-weight parameters and hidden states are likely to survive into subsequent iterations (or “generations”) of the algorithm. During the course of this adaptive process, the scaling factors Inline graphic are allowed to fluctuate only within predefines limits, similarly to the other components of the extended state vector.

Figure 3. The effect of the size of the scaling factor .

Figure 3

Inline graphic and the number of particles Inline graphic on the variance of the estimates. Large minimal values of Inline graphic and small values of Inline graphic imply large variance of the estimates. (A) Resampling of particles (see Methods) implies adaptation of (among others) the scaling factors Inline graphic, which gradually approach the lower bound of their prior interval (red lines in Ai,ii). A prior interval with zero lower bound (i.e. Inline graphic) leads to estimates with negligible variance (Ai). A prior interval with relatively large lower bound (e.g. Inline graphic) leads to estimates with non-zero variance (Aii). Notice that the expectation Inline graphic in Ai does not actually take the value Inline graphic (instead it becomes approximately equal to Inline graphic). (B) A small number of particles (Bi, Inline graphic) implies estimates with large variance (compare to Bii, Inline graphic). Notice that the difference between Aii (Inline graphic) and Bii (Inline graphic) is negligible, implying the presence of a ceiling effect, when the number of particles becomes very large. In these simulations, Inline graphic and Inline graphic.

In Fig. 3Ai, we demonstrate the case where the scaling factors Inline graphic were allowed to take values from the prior interval Inline graphic. We observed that during the course of the simulation (which utilized Inline graphic-long simulated membrane potential recordings), the average value of the scaling factor, Inline graphic, decreased gradually towards Inline graphic and this was accompanied by a dramatic decrease in the variance of the inferred parameters Inline graphic and Inline graphic, which eventually “collapsed” to a point in parameter space located very close to their true values. This situation was the same as the one illustrated in Fig. 2D. Notice that although Inline graphic decreased towards zero, it never actually took this value; it merely became very small (Inline graphic). When we used a prior interval for Inline graphic with non-zero lower bound (i.e. Inline graphic]; see Fig. 3Aii), the final estimates had a larger variance, providing an approximation of the full posteriors of the “unknown” parameters Inline graphic and Inline graphic. Thus, controlling the lower bound of the prior interval for the scaling factors Inline graphic provides a simple method for controlling the variance of the final estimates. Notice that the variance of the final estimates also depends on the number of particles (Fig. 3B). A smaller number of particles resulted in a larger variance of the estimates (compare Fig. 3Bi to Fig. 3Bii). However, when a large number of particles was already in use, further increasing their number did not significantly affect the variance of the estimates or the rate of convergence (compare Fig. 3Bii to Fig. 3Aii), indicating the presence of a ceiling effect.

The adaptive sampling of the scaling factors Inline graphic further depends on parameter Inline graphic in Eq. 28, which determines the width of the lognormal distribution from which new samples are drawn. The value of this parameter provides a simple way to control the rate of convergence of the algorithm; larger values of Inline graphic resulted in faster convergence, when processing Inline graphic-long simulated recordings (compare Fig. 4A to Fig. 4B). The rate of convergence also depends on the number of particles in use (compare Fig. 4A to Fig. 4C), although it is more sensitive to changes in parameter Inline graphic; dividing the value of Inline graphic by Inline graphic (Fig. 4B) had a larger effect on the rate of convergence than dividing the number of particles by Inline graphic (Fig. 4C).

Figure 4. The effect of adaptation of the scaling factor .

Figure 4

Inline graphic and the number of particles Inline graphic on the speed of convergence. A slow rate of adaptation for Inline graphic and a small number of particles Inline graphic imply slow convergence of the algorithm. The rate at which Inline graphic adapts depends on the parameter Inline graphic in Eq. 28. Reducing Inline graphic in half results in a significant decrease in the rate of convergence (compare A to B). Also, reducing the number of particles by a factor of Inline graphic slows down the speed of convergence (compare A to C), but not as much as when parameter Inline graphic was adjusted. The plots on the right illustrate the profile of Inline graphic associated with the estimation of the parameters on the left plots. In these simulations, Inline graphic, Inline graphic and the prior interval for the scaling factors Inline graphic was Inline graphic.

In summary, by assuming an adaptive sampling process for the unknown parameters in the model, we managed to achieve a significant reduction in the spread of the inferred posterior distributions of these parameters. Furthermore, adjusting the prior interval and adaptation rate Inline graphic of the scaling factors Inline graphic provides a straightforward way to control the variance of the estimated posteriors and the rate of convergence of the algorithm. Alternatively, we could have set Inline graphic, i.e. set it to the same constant value for all particles Inline graphic and time steps Inline graphic (as in Fig. 2A). However, by permitting Inline graphic to adapt within a predefined interval, we potentially allow this parameter and, thus, the covariance matrices Inline graphic take large values, which in turn would permit the algorithm to escape local optima in the parameter space. For example, the time profiles of Inline graphic in Figs. 3 and 4 indicate that, early during the simulations, this quantity had relatively large values, which were associated with large variances of the posterior parameter estimates. During this initial period, the algorithm has the potential to “jump” away from local optima and towards more optimal regions of the parameter space. One may see, here, a distant analogy to simulated annealing, where a fictitious “temperature” control variable is gradually decreased, thus allowing the system to escape local minima and gradually settle to more optimal regions of the energy landscape.

Increasing Observation Noise Reduces the Accuracy and Precision of the Fixed-Lag Smoother

In a subsequent stage, we treated as unknown two more parameters in the model, i.e. the reversal potentials for the sodium and potassium currents, Inline graphic and Inline graphic, respectively. Thus, the extended state vector became Inline graphic. This time, we wanted to examine how increasing levels of observation noise (i.e. the value of parameter Inline graphic) affect the inference of unknown quantities in the model based on the fixed-lag smoother. For this reason, we repeated smoothing on four Inline graphic-long simulated data sets (i.e. recordings of membrane potential and the associated Inline graphic) corresponding to increasing values of the standard deviation of the observation noise Inline graphic, i.e. Inline graphic, Inline graphic, Inline graphic and Inline graphic.

The results from this set of simulations are summarized in Fig. 5. For Inline graphic, the expectations of the four parameters Inline graphic, Inline graphic, Inline graphic and Inline graphic (red solid lines in Figs. 5Ai–iv) eventually converged to their true values (dashed lines in the aforementioned figures). For Inline graphic, the expectations of these parameters (light red solid lines in Figs. 5Ai–iv) also converged, although the expectations for Inline graphic (Fig. 5Ai) and, to a lesser degree, Inline graphic (Fig. 5Aii) deviated noticeably from their true values. As expected, at higher levels of noise, the variance of the final estimates was larger, although the rate of convergence did not seem to be affected, due to the large number of particles we used (Inline graphic; see ceiling effect in Fig. 3Bii). The inferred parameters Inline graphic and Inline graphic (not illustrated for clarity) followed a similar convergence pattern.

Figure 5. The effect of observation noise on the accuracy and precision of parameter estimates.

Figure 5

Increasing observation noise decreases the accuracy and precision of the fixed-lag smoother. At this stage, the reversal potentials for the sodium and potassium currents (Inline graphic and Inline graphic, respectively) were also considered unknown. Estimation was based on a simulated recording of membrane potential with duration 2 s. The noise parameters were Inline graphic and Inline graphic, Inline graphic, Inline graphic or Inline graphic. (A) Inference of Inline graphic, Inline graphic, Inline graphic and Inline graphic during smoothing. The accuracy of the estimates decreases and their variance increases with increasing observation noise. (B) The box plot of the time series in A for Inline graphic. Data were first normalized according to Eq. 35. The reduction in the accuracy and precision at higher levels of observation noise were more prominent in the case of the maximal conductances (Inline graphic and Inline graphic) and less prominent in the case of reversal potentials (Inline graphic and, particularly Inline graphic). The membrane potential, activation and inactivation variables, intrinsic and observation noise were also subject to estimation, as in Fig. 1. In these simulations, Inline graphic, Inline graphic, Inline graphic and the prior interval of Inline graphic was Inline graphic.

In Fig. 5B, we show, for each tested value of Inline graphic, the box plots of the above four parameters, which were computed from the data points (as in Fig. 5A) corresponding to time Inline graphic. For each parameter and each value of Inline graphic, the data were first normalized as follows:

graphic file with name pcbi.1002401.e900.jpg (35)

where Inline graphic. The box plots in Fig. 5B were constructed from the normalized data points Inline graphic. The above normalization was necessary since it made possible the comparison between different data sets, each characterized by its own mean, variance and unit of measurement. In the box plots in Fig. 5B, zero (i.e. the dashed lines) corresponds to the true parameter values, while discrepancies from the true parameter values along the y-axis are given in relation to the average Inline graphic. We may observe that for very low levels of observation noise (Inline graphic), the posteriors of the four examined parameters were clustered tightly around their true values, but for larger levels of noise (Inline graphic, Inline graphic and Inline graphic), we observed larger discrepancies from the true parameter values and broader inferred posteriors. The parameters following more noticeably this trend were the conductances Inline graphic and Inline graphic, while Inline graphic and, particularly, Inline graphic were less affected. This indicates that smoothing is more sensitive to changes in some model parameters than others and this is why these parameters were tightly controlled. In summary, increasing the levels of measurement noise (i.e. the value of parameter Inline graphic) decreased the accuracy and precision of the algorithm, but it did not significantly affect the rate of convergence due to the large number of particles used during the simulations.

High-Dimensional Inference Problems are Resolved Given Sufficiently Informative Priors

At the next stage, we treated all parameters in the model (a total of 23 parameters; see Table 1) as unknown. Therefore, the extended state vector took the following (Inline graphic-dimensional) form:

graphic file with name pcbi.1002401.e914.jpg

where Inline graphic and Inline graphic. These parameters included the standard deviations of intrinsic and observation noise (Inline graphic and Inline graphic, respectively), the maximal conductances Inline graphic and reversal potentials Inline graphic of all currents in the model and the parameters controlling the steady-states and relaxation times of activation and inactivation for the sodium and potassium currents (Inline graphic, Inline graphic, Inline graphic, Inline graphic and Inline graphic). The results from this simulation are illustrated in Fig. 6.

Figure 6. Estimation of all parameters in a single-compartment conductance-based model using the fixed-lag smoother.

Figure 6

Estimation was based on a simulated recording of the membrane potential with duration Inline graphic. Noise parameters were Inline graphic. For clarity, only Inline graphic of activity are illustrated in Ai,ii. (A) Smoothing of the membrane potential (Ai) and the unobserved activation and inactivation variables for the sodium and potassium currents (Aii). (B, C) Estimated posteriors for Inline graphic (B), Inline graphic (Ci,ii) and Inline graphic (Ciii,iv). The histograms on the right were constructed form the data on left. (D) Box plot of the Inline graphic estimated parameter posteriors in the model. These included the standard deviations of intrinsic and observation noise, maximal conductances, reversal potentials and kinetics of all currents in the model (see Table 1). The estimates were first normalized according to Eq. 35. Parameter identification numbers are as in Table 1. In these simulations, Inline graphic, Inline graphic, Inline graphic and the prior interval for Inline graphic was Inline graphic.

We observed that the true signal (membrane potential) was inferred with very high fidelity (Fig. 6Ai). The sodium activation Inline graphic was also recovered with very high accuracy, while estimation of the hidden states Inline graphic and Inline graphic (sodium inactivation and potassium activation, respectively) was also satisfactory (despite significant deviations, the general form of the true hidden states was recovered without any observable impact on the dynamics of the membrane potential), as shown in Fig. 6Aii. Among the Inline graphic estimated parameters, we illustrate (in Figs. 6B and 6C) the estimated posteriors for the reversal potential of sodium Inline graphic (Fig. 6B) and for parameters Inline graphic (Figs. 6Ci,ii) and Inline graphic (Figs. 6Ciii,iv), which control the activation of sodium and potassium currents, respectively. We focus on these parameters, because they represent three different characteristic cases. The posteriors of parameters Inline graphic and Inline graphic are unimodal (see Figs. 6Bii and 6Cii) and they were estimated with relatively high accuracy. Particularly, the posterior for Inline graphic was estimated with very high precision and accuracy, despite its broad prior interval (the y-axis in Fig. 6Ci and the x-axis in Fig. 6Cii). On the other hand, the estimated posterior of Inline graphic covered a large part of its prior interval (the y-axis in Fig. 6Ciii and the x-axis in Fig. 6Civ), its main mode was located at a slightly larger value than the true parameter value, while at least two minor modes seem to be present near the upper bound of the prior interval (the arrow in Fig. 6Civ). These results reiterate our previous conclusion that smoothing may be particularly sensitive to some parameters, but not to others. The posteriors of parameters in the former category are very precise and narrow (as in the case of Inline graphic and, especially, Inline graphic), while the parameters in the latter category are characterized by broader posteriors. Also, we can observe that the fixed-lag smoother has the capability to provide a global approximation of the unknown posteriors, including their variance and the location of major and minor modes (i.e. global and local optima). An overview of all inferred posteriors is given by the box plot in Fig. 6D, which was constructed after all data (as in Figs. 6Bi, 6Ci and 6Ciii) were normalized according to Eq. 35. Again, it may be observed that while some of the estimated parameter posteriors are quite precise and accurate, such as Inline graphic (parameter Inline graphic), Inline graphic (parameter Inline graphic) and Inline graphic (parameter Inline graphic), others are less precise and accurate, such as the maximal conductances (parameters Inline graphic to Inline graphic), Inline graphic (parameter Inline graphic) and Inline graphic (parameter Inline graphic).

The simulation results presented above were obtained by assuming a prior interval for the scaling factors Inline graphic equal to Inline graphic. When we repeated the simulation using the prior interval Inline graphic, the true underlying membrane potential was again inferred with very high fidelity (Fig. 7Ai), while the hidden states Inline graphic, Inline graphic and Inline graphic were also estimated with sufficient accuracy (Fig. 7Aii). In this case, however, the estimates of the “unknown” parameters converged to single points in parameter space (as illustrated, for example, for parameters Inline graphic, Inline graphic and Inline graphic in Figs. 7Bi–ii), which fall within the support of the posteriors illustrated in Figs. 6B and 6C. The activation and inactivation steady states (Fig. 7Ci, red solid lines) and relaxation times (Fig. 7Cii, red solid lines) as functions of voltage, which were computed from these estimates, were also similar to their corresponding true functions, with the curves for Inline graphic and Inline graphic manifesting the largest deviation from truth (black solid lines in Figs. 7Ci,ii). An overview of the estimated parameter values (after normalizing using Eq. 35) is given in Fig. 7Di. As stated previously, some estimates were close to their true counterparts, while others were not. For example, the activation of the sodium current Inline graphic (Fig. 7Aii) and its steady state Inline graphic (Fig. 7Ci), which are important for the correct onset of the action potentials, were inferred with relatively high accuracy. On the other hand, larger errors were observed, for example, in the inference of sodium inactivation (Inline graphic; Fig. 7Aii) or in the estimation of Inline graphic (parameter Inline graphic; Fig. 7Di), the maximal conductance for the sodium current.

Figure 7. The effect of prior parameter intervals on the accuracy of the fixed-lag smoother.

Figure 7

Estimation was based on a simulated recording of the membrane potential with duration Inline graphic. Noise parameters were Inline graphic. For clarity, only Inline graphic of activity are illustrated in Ai,ii. Unlike Fig. 6, the prior interval for the scaling factors Inline graphic was now assumed equal to Inline graphic. (A) Smoothing of the membrane potential (Ai) and the unobserved activation and inactivation variables for the sodium and potassium currents (Aii). (B) Estimates for parameters Inline graphic (Bi), Inline graphic and Inline graphic (Bii). Convergence to an optimal parameter vector was achieved after approximately Inline graphic of activity. Notice that this optimal parameter vector falls within the support of the corresponding parameter posteriors (see Figs. 6Bii, 6Cii and 6Civ). (C) Inferred steady states (Ci) and relaxation times (Cii) for the activation and inactivation variables of sodium and potassium currents (red lines) against their true counterparts (black lines). (D) Inferred parameter values when broad (Di) or narrow (Dii) prior intervals were used for the parameters controlling the kinetics of sodium and potassium ionic currents (see Table 1). Plots A, B and C correspond to plot Di. In Dii, we also illustrate the estimated parameter values when very noisy data were used (see also Supplementary Fig. S3). In these simulations, Inline graphic, Inline graphic and Inline graphic.

Given the fact that the data on which inference was based (a single noisy recording of the membrane potential) was of much lower dimensionality than the extended state we aimed to infer, the observed discrepancies between inferred and true model quantities were unlikely to vanish unless we imposed more strict constraints on the model. When we repeated the previous simulation using more narrow prior intervals for some of the parameters controlling the kinetics of the sodium and potassium currents in the model (see red dashed boxes in Fig. 7Dii and bold intervals in Table 1), the estimated parameters settled closer to their true values (Fig. 7Dii). This was true even for parameters on which more narrow intervals were not directly applied, such as the maximal conductances (i.e. parameters Inline graphic to Inline graphic in Fig. 7Dii), and even when data with higher levels of observation noise were used (Fig. 7Dii, data points indicated with crosses; see also Fig. S3). It is important to mention that using more narrow prior constraints only affected the accuracy of the final estimates, not the quality of fitting the experimental data, which in all cases was of very high fidelity. Alternatively, we could have constrained the model by increasing the dimensionality of the observed signal, e.g. by using simultaneously more that one unique voltage traces (each generated under different conditions of injected current) during smoothing. We examine the use of multiple data sets simultaneously as input to the fixed-lag smoother later in the Results section.

In summary, the smoothing algorithm can be used to resolve high-dimensional inference problems. In combination with sufficient prior information (in the form of bounded regions within which parameters are allowed to fluctuate; see Table 1), the fixed-lag smoother can provide estimates of the intrinsic and observation noise, maximal conductances, reversal potentials and kinetics of ionic currents in a single-compartment Hodgkin-Huxley-type neuron model, based on low-dimensional noisy experimental data.

Parameter Estimation in Compartmental Models is Straightforward Using the Fixed-Lag Smoother

Next, we tested whether the fixed-lag smoother could be successfully applied on inference problems involving more complex models than the one we used in the previous sections. For this reason, we focused on a two-compartment model of a vertebrate motoneuron containing sodium, potassium and calcium currents and intracellular calcium dynamics, which were differentially distributed among a soma and a dendritic compartment [37]. The model (modified appropriately to include intrinsic noise terms) is summarized below:

graphic file with name pcbi.1002401.e993.jpg (36)
graphic file with name pcbi.1002401.e994.jpg (37)

where Inline graphic and Inline graphic is the membrane potential at the soma and dendritic compartments, respectively, and Inline graphic. The leakage conductance and reversal potential were Inline graphic and Inline graphic, respectively. The coupling conductance was Inline graphic and the ratio of the soma area to the total surface area of the cell was Inline graphic. The various ionic currents in the above model were as follows: (a) a transient sodium current, Inline graphic, (b) a delayed rectifier potassium current, Inline graphic, (c) a calcium-activated potassium current, Inline graphic, where Inline graphic and Inline graphic (the half-saturation constant), (d) an N-type calcium current, Inline graphic, where Inline graphic and (e) an L-type calcium current, Inline graphic. The various activation and inactivation dynamic variables in the above model were modeled using first-order relaxation kinetics (as in Eq. 32), where the various steady states were assumed to be sigmoid functions of voltage (Eq. 33). Notice, that the activation of Inline graphic was assumed instantaneous and therefore, it was given at any time by the voltage-dependent steady state Inline graphic. The relaxation times for sodium inactivation and potassium activation were also functions of voltage as in Eq. 34:

graphic file with name pcbi.1002401.e1012.jpg (38)
graphic file with name pcbi.1002401.e1013.jpg (39)

where the parameters Inline graphic, Inline graphic and Inline graphic (with Inline graphic and Inline graphic) were chosen by fitting the above expressions to the original model in [37]. The relaxation times for the remaining activation and inactivation variables were constant. All parameters values in the model are given in Table 2.

The intracellular calcium concentration at either the soma or the dendritic compartment was also modeled by a first-order differential equation, as follows:

graphic file with name pcbi.1002401.e1019.jpg (40)

where Inline graphic, Inline graphic and Inline graphic. The total calcium current is Inline graphic at the soma (Inline graphic) and Inline graphic at the dendritic compartment (Inline graphic).

The observation model assumed simultaneous noisy recordings of the membrane potential from both the soma and dendritic compartments, as follows:

graphic file with name pcbi.1002401.e1027.jpg (41)

where Inline graphic with Inline graphic. Notice that Inline graphic is the same for both compartments.

In the above model, the externally injected currents Inline graphic and Inline graphic were sequences of random current steps with duration up to Inline graphic (instead of Inline graphic as in the single-compartment model, due to the presence of slower currents in the two-compartment model) and magnitude between Inline graphic and Inline graphic. Current was injected in both the dendritic compartment and the soma (instead of just in the soma), because preliminary simulations indicated that this experimental setting facilitated parameter estimation, presumably due to the generation of a more variable (and, thus, information-rich) data set. The injected currents and the induced noisy voltage traces Inline graphic and Inline graphic comprised the simulated data on which parameter estimation was based.

First, we aimed to infer the noise parameters and maximal conductances of all voltage- and calcium-gated currents in the model, assuming that the kinetics of these currents were known. This implied an extended-state vector with Inline graphic components as shown below

graphic file with name pcbi.1002401.e1040.jpg

where Inline graphic. The results from this simulation are illustrated in Figs. 8 and 9. The fixed-lag smoother managed to recover the hidden dynamic states (including the time-evolution of the intracellular calcium; Fig. 8), the standard deviations of the intrinsic and observation noise (Figs. 9Ai,ii) and the true values of all the gated maximal conductances (Figs. 9Bi–iv) in the model using approximately Inline graphic of simulated data and Inline graphic particles. Notice that, in Figs. 8Ci–iv, the inferred hidden gating states (dashed red lines) coincide extremely well with the true ones (solid black lines), which is not surprising, since the voltage-dependent kinetics of these states were assumed known and the true membrane potential at the soma and dendritic compartment was recovered with very high fidelity (Figs. 8Ai,ii). Also, notice that, in Figs. 9Aii, 9Biii and 9Biv, the estimation of the standard deviation of the intrinsic noise, Inline graphic, and the maximal conductances of calcium and calcium-dependent currents in the dendritic compartment (Inline graphic, Inline graphic and Inline graphic) was improved after injecting current in both the soma and the dendritic compartment (compare the grey solid lines, which correspond to injection in the soma only, to the color ones in the aforementioned figures).

Figure 8. Simultaneous estimation of hidden model states (including intracellular calcium concentrations) and maximal conductances in a two-compartment model of a vertebrate motoneuron (I).

Figure 8

Estimation was based on two Inline graphic-long simulated recordings of the membrane potential, each recorded simultaneously from the soma and the dendritic compartment. Only part of the recorded activity is illustrated in A, B and C for clarity. Notice the different time scales between the right and left panels. (A) High-fidelity smoothing of the membrane potential at the soma (Ai) and the dendritic compartment (Aii). (B) Inference of the unobserved calcium concentrations at the soma (Bi) and the dendrite (Bii). (C) Inference of the unobserved activation and inactivation variables for the sodium and potassium currents (Ci) and the N-type calcium current (Ciii) at the soma and the N-type (Cii) and L-type (Civ) calcium currents at the dendritic compartment. Notice the almost complete overlap between true (black lines) and inferred (red lines) dynamic variables in Ci–iv. This was not surprising since we assumed, at this stage, that the kinetics of all gated currents were known. In these simulations, Inline graphic, Inline graphic, Inline graphic and the prior interval for Inline graphic was Inline graphic.

Figure 9. Simultaneous estimation of hidden model states (including intracellular calcium concentrations) and maximal conductances in a two-compartment model of a vertebrate motoneuron (II).

Figure 9

Inference of maximal conductances and noise parameters during fixed-lag smoothing. (A) The standard deviations of the observation (Ai) and the intrinsic (Aii) noise at the soma and the dendrite. (B) Inferred maximal conductances of the sodium and potassium currents at the soma (Bi), of the N-type calcium current and the calcium-activated potassium current at the soma (Bii), of the calcium-activated potassium current at the dendrite (Biii) and of the N-type and L-type calcium currents at the dendrite (Biv). In all cases, parameter expectations gradually converged towards the true parameter values (dashed lines) after less than Inline graphic. The grey lines in Aii, Biii and Biv correspond to estimated parameters, when current was injected in the soma only. In these simulations, Inline graphic, Inline graphic, Inline graphic and the prior interval for Inline graphic was Inline graphic.

In a second stage, we assumed that the kinetics of all voltage-gated ionic currents were also unknown, implying an extended state vector with Inline graphic components, as follows:

graphic file with name pcbi.1002401.e1061.jpg

where Inline graphic, Inline graphic and Inline graphic. Our results from this simulation are summarized in Figs. 10 and 11. Again, the membrane potential at the soma and the dendrite were inferred with very high fidelity (Fig. 10Ai,ii). However, the estimated hidden dynamics of most ionic currents and intracellular calcium concentrations in the model deviated significantly from their true counterparts (Fig. 10B,C). The expectations of all estimated parameters are illustrated in Fig. 11Ai. As in the case of the single-compartment model, by imposing tighter prior constraints on some of the parameters controlling the kinetics of ionic currents in the model (see red dashed box in Fig. 11Aii and Table 2), we managed to reduce the discrepancies of the estimates from their true values (Fig. 11Aii and Supplementary Fig. S4). This was true even for parameters on which stricter priors were not directly applied. The inference was completed after processing almost Inline graphic of data, as shown in Fig. 11B for the maximal conductances of sodium and potassium currents at the soma. Interestingly, the algorithm seems to temporarily settle at local optima (see arrows in Fig. 11B) before “jumping” away and, eventually, converge at the final estimates. The inferred voltage-dependent steady-states of the sodium, potassium and calcium currents (Figs. 11Ci,ii) and the relaxation times for the sodium inactivation and potassium activation (Fig. 11Ciii) were also very similar to their true corresponding functions. The algorithm remained operational when more noisy data were used, as illustrated in Fig. 11Aii and in Supplementary Fig. S5.

Figure 10. Simultaneous estimation of hidden model states, maximal conductances and kinetic parameters in a two-compartment model of a vertebrate motoneuron (I).

Figure 10

Estimation was based on two simulated Inline graphic-long simultaneous recordings of the membrane potential from the soma and dendritic compartment. Only part of this data is illustrated for clarity. Notice the different time scales between the left and right panels. (A) High-fidelity smoothing of the observed voltage at the soma (Ai) and the dendrite (Aii). (B) Inference of unobserved calcium concentrations at the soma (Bi) and dendritic compartment (Bii). (C) Inference of the unobserved activation and inactivation variables for all voltage-gated currents at the soma and the dendrite. Since the kinetics of voltage-gated currents were assumed unknown, the difference between true (black lines) and inferred (red lines) dynamic variables was significant (compare to Fig. 8). The inferred parameters are shown in Fig. 7Ai. In these simulations, Inline graphic, Inline graphic, Inline graphic and the prior interval for Inline graphic was Inline graphic.

Figure 11. Simultaneous estimation of hidden model states, maximal conductances and kinetic parameters in a two-compartment model of a vertebrate motoneuron (II).

Figure 11

Inference of maximal conductances, noise and kinetic parameters during smoothing. (A) Inferred parameters in the model using broad or narrow prior intervals and high or low levels of observation noise. Estimates were normalized according to Eq. 35. Parameter identification numbers are as in Table 2. The estimates in Ai were obtained using broad prior intervals (see Table 2). The maximal conductance Inline graphic (parameter #7) converged to zero and, for this reason, it is indicated with a red square. These estimates correspond to the results shown in Fig. 10. Estimates in Aii were obtained using narrow prior intervals for some of the parameters controlling the kinetics of ionic currents (see red dashed boxes) at either low (Inline graphic) or high (Inline graphic) levels of observation noise (see also Supplementary Figs. S4 and S5). (B) Inferred maximal conductances for sodium (Inline graphic) and potassium (Inline graphic) when narrow prior intervals and low levels of observation noise were used (circles in Aii). Notice the temporary convergence of the estimates (arrows) before jumping away towards their final values. (C) True (black lines) and inferred (red lines) activation and inactivation steady-states for the sodium and potassium currents (Ci) and the N-type and L-type calcium currents (Cii) and for the relaxation times for sodium inactivation and potassium activation (Ciii), when narrow prior intervals and low levels of observation noise were used (circles in Aii). In these simulations, Inline graphic, Inline graphic and the prior interval for Inline graphic was Inline graphic. The number of particles was Inline graphic in Ai and Inline graphic in Aii, B and C (see main text for further comments).

An interesting fact regarding the simulation results presented in Figs. 10 and 11Ai was that, in order to obtain high-fidelity estimates of the true membrane potential at the soma and dendritic compartment (as shown in Figs. 10Ai,ii) we had to use more than Inline graphic particles, the number calculated by the Inline graphic rule (see Methods). In particular, we used Inline graphic particles, although we cannot exclude that a smaller number may have sufficed. After applying more narrow prior constraints (Figs. 11Aii, B, C, S4 and S5), using the number of particles calculated by the above simple heuristic (Inline graphic in this case) was again sufficient for accurately inferring the true membrane potential (see Fig. S4Ai,ii and S5Ai,ii). This implies that as the complexity (and dimensionality) of the estimation problem increases, a non-linearly growing number of particles may be required in order to obtain acceptable results, but this situation may be compensated for by providing highly informative priors. It should be mentioned that the two-compartment model allows for the physical separation of currents and as such it is a slightly better approximation of a real neuron with differential expression of individual currents in different cellular compartments. However, in no way does it capture the full morphological complexity of a real neuron. As such, current injection into the dendritic compartment can not be replicated accurately in a real neuron, as current injection in the model will have a uniform effect on all currents in that compartment, whilst current injection into the dendrite of a neuron would have far more complex effects on dendritic currents, which potentially would be dependent on the distance from the injection site. Thus, whilst it would be possible, albeit challenging, to carry out dual recordings from the soma and dendrites in a real neuron this would not be the same as the dual current injection in the model. In this case, application of the fixed-lag smoother on a more spatially detailed model would be necessary (and feasible). In principle, the method can also assimilate other types of spatial data, such as calcium imaging data, in case recordings from multiple neuron locations are not available (although we do not examine this case in detail in this paper).

Given the large number of unknown parameters and hidden states in combination with the low dimensionality of the data (notice that the intracellular calcium concentration was assumed unobserved), it was truly remarkable that the algorithm managed to recover much of the extended state vector with relatively satisfactory accuracy. However, it should be noted that in our simulations we assumed knowledge of important information, such as the passive conductances Inline graphic and Inline graphic and the reversal potentials of sodium, potassium and calcium currents. This and the fact that the availability of prior information in the form of more narrow parameter boundaries improved significantly the accuracy of the final estimates emphasizes our previous conclusion that prior information is important for the successful inference of unknown model parameters and hidden model states using the fixed-lag smoother. Given such information, inference in complex compartmental models based on simultaneous recordings from several neuron locations and, possibly, measurements of intracellular calcium, can be naturally achieved via appropriate formulation of the extended state vector and application of the fixed-lag smoother.

Parameters in a Model of an Invertebrate Motoneuron were Inferred from Actual Electrophysiological Data Using the Fixed-Lag Smoother

In a final set of simulations, we applied the smoother on actual electrophysiological data in order to estimate the unknown parameters in a single-compartment model of the B4 motoneuron from the nervous system of the pond snail, Lymnaea stagnalis [38]. This neuron is part of a population of motoneurons, which receive rhythmic electrical input from upstream Central Pattern Generator interneurons and in turn innervate and control the movements of the feeding muscles via which the animal captures and ingests its food. Previous studies in these neurons have demonstrated the presence of a transient inward sodium current Inline graphic, a delayed outward potassium current Inline graphic and a transient outward potassium current Inline graphic [41]. A hyperpolarization-activated current Inline graphic was conditional on the presence of serotonin in the solution [38] and, therefore, this current was not included in this instance of the B4 model. Thus, the current conservation equation for a single-compartment model of the B4 motoneuron (appropriately modified to include an intrinsic noise term) took the following form:

graphic file with name pcbi.1002401.e1093.jpg (42)

where the leakage conductance, leakage reversal potential and membrane capacitance in the above model were estimated a priori based on neuron responses to negative current pulses (Inline graphic, Inline graphic and Inline graphic, respectively). The voltage-activated currents that appear in the above expression were modeled as follows: (a) Inline graphic, (b) Inline graphic and (c) Inline graphic, where Inline graphic and Inline graphic as in [41]. The dynamic activation and inactivation variables of these currents (Inline graphic, Inline graphic, Inline graphic and Inline graphic) obeyed first-order relaxation kinetics (as in Eq. 32) with voltage-dependent steady-states (Eq. 33) and relaxation times (Eq. 34 with Inline graphic and Inline graphic), similarly to previously published neuron models in the central nervous system of Lymnaea [42]. The observation model was as in Eq. 7.

The raw data we used for inferring the parameters in the above model took the form of four independent Inline graphic-long recordings of the membrane potential from the same B4 motoneuron. Each recording was taken while injecting an external current in the neuron consisting of a sequence of random steps ranging in amplitude between Inline graphic and Inline graphic and with duration between Inline graphic and Inline graphic. A particular characteristic of the data generated under these conditions was the presence of brief bursts of spikes, which were interrupted by relatively long intervals of non-activity (corresponding to sub-threshold excitatory and inhibitory current injections, respectively; see Figs. 12Ai–iv). These long intervals of inactivity were not informative and they negatively affected the performance of the smoother by permitting the random drift of particles towards non-optimal regions of the parameter space (see Supplementary Fig. S6). However, when the four recordings are considered together, the intervals of inactivity at any single voltage trace overlap with intervals of activity at the remaining three voltage traces, resulting in a four-dimensional data set, where the overall intervals of inactivity were minimized. This four-dimensional data set was used as input to the smoother during the inference phase.

Figure 12. Parameter estimation in a model invertebrate motoneuron based on actual electrophysiological data.

Figure 12

Estimation was based on four independent Inline graphic-long recordings of the membrane potential from the same B4 motoneuron. (A) Simultaneous, high-fidelity smoothing of the four membrane potential recordings. (B) A total of Inline graphic free parameters in the model were inferred during smoothing (see Table 3), including the maximal conductances of the transient sodium and potassium and persistent potassium currents (Bi), the half steady-state activation values (Bii) and the relaxation times for the activation of the potassium currents (Biii). The remaining inferred parameters are not illustrated for clarity, but they follow a similar convergence pattern. (C) Box plot of all inferred parameters in the model. Parameter identification numbers are as in Table 3. Estimates were normalized as explained in the main text (the non-normalized mean parameter values are given in Table 3). In this simulation, Inline graphic, Inline graphic, Inline graphic and the prior interval for Inline graphic was Inline graphic.

Thus, the Inline graphic-dimensional extended state vector became:

graphic file with name pcbi.1002401.e1121.jpg

where Inline graphic, Inline graphic and Inline graphic. Notice the presence of four groups of hidden dynamic states, {Inline graphic, Inline graphic, Inline graphic, Inline graphic, Inline graphic, Inline graphic}, where each group corresponds to a different voltage trace (and associated externally injected current, Inline graphic). The evolution of all four groups of dynamic variables was governed by a common (shared) set of parameters. In total, we had to estimate Inline graphic unknown parameters. The boundaries within which the values of these parameters were allowed to fluctuate are given in Table 3 (indicated in bold) and they were chosen from within the support of the posteriors in Supplementary Fig. S7 (after a few trial-and-error simulations), which were obtained by using the broader prior intervals given in Table 3. Notice that the marginal distributions illustrated in Fig. S7 have large variance and multiple modes and, although they provide a global view of the structure of the parameter space, they cannot be used to identify a single combination of optimal parameters values, since they do not include any information regarding correlations between parameters. Using the major modes of the inferred posteriors did not lead to an accurate (or even spiking) predictive model. Thus, the estimation was based on using more narrow prior intervals, which helped us estimate unimodal posteriors with small variance (see Fig. 12C) and, thus, identify a single combination of optimal parameters that could be used in predictive simulations. We cannot prove that other optimal combinations of parameters do not exist, but we were not able to find any (i.e. by choosing different narrow prior intervals) after a reasonable amount of time. Also, notice that the standard deviations of the intrinsic and observation noise were not subject to estimation, but instead they were given (through trial and error) the minimal fixed values Inline graphic and Inline graphic, respectively. If left free during smoothing, the values of these parameters fluctuated uncontrollably, masking the contribution of the remaining parameters in the model and, thus, achieving an almost perfect (but meaningless) smoothing of the experimental data. This is an indication that the B4 model we used may be missing one or more relevant components, such as additional currents and compartments (see below for further analysis of this point). We did not observe this effect in the cases examined in the previous sections, where simulated data was used, because the models responsible for the generation of this data were, by definition, precisely known.

Our results from this set of simulations are illustrated in Fig. 12. Simultaneous smoothing of all four data sets was again accomplished with high fidelity, as illustrated in Figs. 12Ai–iv. The artificial evolution of the expectations of the conductances for the transient sodium, persistent potassium and transient potassium currents, as well as of some of the kinetic parameters that were estimated in the model is illustrated in Figs. 12Bi–iii. The distributions of all inferred parameters (normalized after replacing Inline graphic in Eq. 35 with Inline graphic, for each tested parameter) are also illustrated in Fig. 12C. The inferred expectations of all parameters are given in Table 3.

In order to examine the predictive value of the model given the estimated parameter expectations in Table 3, we compared its activity to that of the biological B4 neuron, when both were injected with a Inline graphic-long random current consisting of a sequence of current pulses with amplitude ranging from Inline graphic to Inline graphic and duration from Inline graphic to Inline graphic. Our results from this simulation are illustrated in Fig. 13. We observed that the overall pattern of activity of the model was similar to that of the biological neuron (Fig. 13A). Whilst the model overall generated more action potentials, some individual spikes were absent in the simulated data. A more detailed examination of our data revealed specific differences between the biological and model neurons, which explain the differences in the overall activity between the two (Fig. 13B, C). The spike shape of the model neuron was quite similar to that of its biological counterpart (Fig. 13Bi), including spike threshold, peak, trough and height (i.e. trough-to-peak amplitude; Fig. 13Biii), but the simulated spike had a slightly longer duration than the biological one (half-width: Inline graphic vs Inline graphic; Fig. 13Bii).

Figure 13. Comparison between B4 model activity and the biological neuron.

Figure 13

(A) Response of the model and the biological B4 motoneuron to a sequence of current steps with random amplitude and duration. Current step amplitudes were from Inline graphic to Inline graphic and current step durations from Inline graphic to Inline graphic. Intrinsic and observation noise in the model were Inline graphic and Inline graphic, respectively. (B) Comparison between model and biological B4 action potentials. The width of the spikes was measured at half their peak amplitude. (C) Current-Voltage (IV) and Current-Frequency (IF) relations for the model and biological B4 neurons. In order to construct these relations both the model and biological neurons were injected with Inline graphic-long current pulses with amplitude between Inline graphic and Inline graphic.

In a second set of experiments, both the biological and model neurons were injected with Inline graphic-long current pulses ranging from Inline graphic to Inline graphic and their current-voltage (IV) and current-frequency (IF) relations were constructed (Fig. 13C). The IV plot showed some non-linear behavior in response to negative current pulses in the experimental data (probably due to the presence of a residual Inline graphic current), which was not present in the simulations (Fig. 13Ci). As a result, the slope of the part of the IV curve corresponding to Inline graphic was more shallow in the simulations than in the experimental data. Moreover, the rheobase was lower in the experimental data than in the model, but the slope of the IF curve was steeper in the simulated data, which resulted in higher firing rates for the model at injected currents larger than approximately Inline graphic (Fig. 13Cii). This feature can account for the overall level of spiking in the model neuron when compared to the biological one (Fig. 13A).

Overall, this analysis illustrates that the assumed B4 model did not capture all the aspects of the real neuron. However, this does not mean that our estimation method is flawed. It just shows that the model is actually missing some relevant components, such as additional ionic currents or compartments, which would be necessary for approximating more accurately the spatial structure and biophysical properties of the biological neuron. In the first part of the manuscript we have demonstrated that if the underlying model is complete, then our method produced accurate estimates of the true parameter values, given sufficient informative priors. Thus, it is safe to assume that the observed differences between the biological and model neurons can be minimized, if the fixed-lag smoother is applied on a more complex model of the B4 motoneuron.

In summary, we used the fixed-lag smoother to estimate the unknown parameters in a single-compartment model of an invertebrate motoneuron based on actual electrophysiological data. The model, although a simplification of the actual biological system, was still quite complex containing a number of non-linearly interacting components and a total of Inline graphic unknown parameters. By using the methodologies outlined in the previous sections, we managed to estimate the values of these parameters, such that the resulting model mimicked with satisfactory accuracy the overall activity of its biological counterpart. Furthermore, we demonstrated the flexibility of the fixed-lag smoother by showing how it can be used to process simultaneously multiple data sets, given an appropriate formulation of the extended state vector.

Discussion

Parameter estimation in conductance-based neuron models traditionally involves a global optimization algorithm (for example, an evolutionary algorithm), usually in combination with a local search method (such as gradient descent), in order to find combinations of model parameters that minimize a pre-defined cost function. In this paper, we have addressed the problem of parameter estimation in Hodgkin-Huxley-type models of single neurons from a different perspective. By adopting a hidden-dynamical-systems formalism and expressing parameter estimation as an inference problem in these systems, we made possible the application of a range of well-established inference methods from the field of Computational Statistics. Although it is usually assumed that the kinetics of ionic currents in a conductance-based model are known a priori, here we assumed that this was not the case and, typically, we estimated kinetic parameters, along with the maximal conductances and reversal potentials of ionic currents in the models we examined.

The particular method we used was Kitagawa's self-organizing state-space model, which was implemented as a fixed-lag smoother. The smoother was combined with an adaptive algorithm for sampling new sets of parameters akin to the Covariance Matrix Adaptation Evolution Strategy. Alternatively, we could have approximated the smoother distribution (Eq. 13) with a two-pass algorithm, consisting of a forward filter followed by a backward smoothing phase, which would make use of the precomputed filter [34]. This would require storing the filter for the whole duration of the smoothed data, which in turn would have very high memory requirements when large numbers of particles or high-dimensional problems are considered. In contrast, the fixed-lag smoother has the advantage that only the particles up to Inline graphic time steps in the past need to be stored, which is less demanding in memory size and computationally more efficient. Moreover, the fixed-lag smoother, being a single-pass algorithm, was more natural to use in the context of on-line parameter estimation.

The applicability of the algorithm was demonstrated on a number of conductance-based models using noisy simulated or actual electrophysiological data. In a recent study, it was found that increasing observation noise led to an increase in the variance of parameter estimates and a decrease in the rate of convergence of the algorithm [28]. Similarly, we observed that at high levels of observation noise, although the algorithm remained functional, its accuracy and precision were reduced (Fig. 5). It is emphasized that, at a particular level of observation noise, the outcome of the algorithm is an approximation of the posterior distributions of hidden states and unknown parameters in the model, given the available experimental data and prior information. In general, these approximate posteriors provide an overview of the structure of the parameter space and they potentially have multiple modes (or local optima). By taking advantage of the adaptive nature of the fixed-lag smoother (and, in particular, by controlling the scaling factor that determines the width of the proposal distribution in Eq. 25), we managed to reduce the variance of these posteriors and, in the limit case, we could force the algorithm to converge to a single optimal point (belonging to the support of the parameter posteriors), which could subsequently be used in predictive simulations (e.g. see Figs. 7D and 11A). Unlike the study in [28], we did not observe any significant reduction in the rate of convergence of the algorithm at high levels of observation noise, which was attributed to a ceiling effect due to the large number of particles we used in our simulations (typically, Inline graphic, where Inline graphic was the dimensionality of the estimation problem; see Figs. 3B and 4C). Thus, we cannot exclude observing such a reduction in the rate of convergence, if a smaller number of particles is used and/or problems of higher dimension are examined. Furthermore, the proposed method requires only a single forward pass of the experimental data, instead of multiple passes, as in the case of off-line estimation methods, including the Expectation Maximization (EM) algorithm. On the other hand, this means that, in general, the proposed algorithm requires processing longer data time series in order to converge. In addition, unlike off-line estimation methods, it does not take into account the complete data trace at each iteration, but at most Inline graphic past data points (but, also, see [36] for a partial “remedy” of this situation). In principle, it would be possible to combine previous work on parameter estimation (e.g. [43], [44]) within an EM inference framework in order to estimate various types of parameters (including maximal conductances and channel kinetics) in conductance-based neuron models. This could be an interesting topic for further research.

Our main conclusion was that, using this algorithm and a set of low-dimensional experimental data (typically, one or more traces of membrane potential activity), it was possible to fit complex compartmental models to this data with high fidelity and, simultaneously, estimate the hidden dynamic states and optimal values of a large number of parameters in these models. Based on simulation experiments using simulated data, we found that the estimated optimal parameter values and hidden states were close to their true counterparts, as long as sufficient prior information was made available to the algorithm. This information took the form of knowledge of the values of particular parameters (for example, the passive properties of the membrane) or of relatively narrow ranges of permissible parameter values. Such prior information could have included the kinetics of the ion currents that flow through the membrane or the spatial distribution of various parameter values along different neuron compartments (e.g. the ratio of maximal conductance A between compartment 1 and compartment 2). In real-life situations, such information may become available through current- or voltage-clamp experiments. For example, the passive properties in the B4 model (membrane capacitance, leakage maximal conductance and reversal potential) were inferred from current-clamp data and, thus, they were fixed during the subsequent smoothing phase.

It has been demonstrated that this requirement for prior information may be relaxed, if the data set used as input was sufficiently variable to tease apart the relative contribution of different parameters in a model [15]. A well-established result in conductance-based modeling is that the same pattern of electrical activity may be produced by different parameter configurations of the same model [6][9]. This implies that it is impossible to identify, during the course of an optimization procedure, a unique set of parameters using just this single pattern of activity as input to the method. For example, as we observed in the case of the B4 model, the posteriors of the estimated parameters may be characterized by multiple modes (i.e. local optima) or quite large variances, which makes identification of a unique set of optimal parameter values for use in predictive simulations rather difficult (Supplementary Fig. S7). A more variable data set would be necessary in order to constrain the model under study, thus forcing the optimization process to converge towards a unique solution. It should be noted that this conclusion was reached by treating as unknown only the maximal conductances in a conductance-based model [15]. Although it is reasonable to assume that this holds true when the kinetics of ion channels are also treated as unknown, it still needs to be demonstrated whether the generation of a data set sufficiently variable to constrain both the maximal conductances and kinetics of ion channels in a complex conductance-based model is practical or even feasible. A more pragmatic approach would be to rely on a mixture of prior information and one or more sufficiently variable electrophysiological recordings as input to the optimization algorithm. It was shown in this study that both the injection of prior information (in the form mentioned above) and the simultaneous assimilation of multiple data sets is straightforward using the proposed algorithm.

It is important to notice that, unlike more traditional approaches, explicitly defining a cost or fitness function was not required by the fixed-lag smoother. Given the fact that the efficiency of any optimizer can be seriously impeded by a poorly designed cost function, bypassing the need to define such a function may be viewed as an advantage of the proposed method. As in previous studies [43], [44], here lies the implicit assumption that by fitting (or smoothing) with high fidelity the raw experimental data (for example, one or more recordings of the membrane potential), the estimated model would capture a whole range of features embedded in this data, such as the current-frequency response of the neuron. Although this is a reasonable assumption, we found that it did not hold completely true, when our knowledge of the form of the underlying model was not exact, as in the case of the B4 neuron. In this case, although we could achieve a very good smoothing of the experimental data, subsequent predictive simulations using the inferred model parameters revealed discrepancies between simulation output and experimental data. It is likely that these discrepancies will be minimized, if important missing components are added to the model, such as additional ionic currents or, importantly, an approximation of the spatial structure of the biological neuron.

An important outcome of this study was to demonstrate the intimate relation between the self-organizing state-space model and evolutionary algorithms. When used for parameter estimation, the self-organizing state-space model undergoes at each iteration a process of new particle (individual) generation (mutation/recombination) and resampling (selection and multiplication), which parallels similar processes in evolutionary algorithms. At the root of this parallelism is the fact that we need to impose an artificial evolution on model parameters as part of the formulation of the self-organizing state-space model (see Methods), thus providing a unique opportunity to merge the two classes of algorithms. Here, we decided to combine the self-organizing state-space model with an adaptive algorithm similar to the Covariance Matrix Adaptation Evolution Strategy [30] and by following this adaptive strategy, we managed to achieve a dramatic reduction in the variance of parameter estimates. However, this choice is by no means exclusive and other evolutionary algorithms may be chosen instead, e.g. the Differential Evolution algorithm [31]. This is a topic open to further exploration. Notice that, similarly to Evolutionary Algorithms, the proposed method has, in principle, the ability to estimate the possibly multi-modal posterior distribution of the unknown parameters in the examined model, i.e. it is a global estimation method (for example, see Fig. 6C, 11B and S7). At each iteration, the algorithm retains a population of particles, which are characterized by a degree of variability and, thus, give the algorithm the opportunity to randomly explore a wide range of the parameter space, spending on average more time in the vicinity of optimal regions. By imposing narrow prior constraints on some of the unknown parameters, we are effectively reducing the dimensionality of the problem and we force the algorithm to converge towards a particular optimum, which can be later used in predictive simulations.

A point of potential improvement concerns our choice of the proposal density, Inline graphic. Here, we made the common and straightforward choice to use the transition density Inline graphic as our proposal. However, the modeler is free to make other choices. For example, a recent study demonstrated that the efficiency of particle filters can be significantly increased by conditioning the proposal density on future observations [36].

An important practical aspect of the proposed algorithm was its high computational cost. This cost increased as a function of the number Inline graphic of particles used during smoothing, the length of the fixed smoothing lag Inline graphic, the complexity of the model and the number of unknown parameters in the model. Our simulations on an Intel dual-core i5 processor with four gigabytes of memory took from a few minutes to more than 12 hours to complete. An emerging trend in Scientific Computing is the use of modern massively parallel Graphics Processing Units (GPUs) in order to accelerate general purpose computations, as those presented in this paper. The utility of this approach in achieving significant accelerations of Monte Carlo simulations has been recently demonstrated [45] and it has even been applied recently on parameter estimation problems in conductance-based models of single neurons [46]. Preliminary results using a GPU-accelerated version of the fixed-lag smoother (data not shown) have indeed demonstrated reduced simulation times, but the accelerations we observed were not as dramatic as those reported in the literature [45], [46]. This can always be attributed to the fact that our implementation of the algorithm was not optimized. On the other hand, we observed significant accelerations in our simulations involving the serial implementation of the fixed-lag smoother, just by switching from an open-source compiler (GNU) to a commercial one (Intel), which presumably emitted better optimized machine code for the underlying hardware. Nevertheless, the use of GPUs for general purpose computing is becoming common and it is likely to become quite popular with the advent of cheaper hardware and, importantly, more flexible and programmer-friendly Application Programming Interfaces (APIs).

Overall, our results point towards a generic four-stage heuristic for parameter estimation in conductance-based models of single neurons: (a) First, the general structure of the model is decided, such as the number of ionic currents and compartments it should include. (b) Second, prior information is exploited in order to fix as many parameters as possible in the model and tightly constrain the remaining ones. For example, the capacitance, reversal potentials and leakage conductance in the model may be fixed to values estimated from current-clamp data. By further exploiting current– and voltage-clamp data, narrow constraints may be imposed on the remaining free (e.g. kinetic) parameters in the model. (c) At a third stage, more precise parameter value distributions are estimated by applying the fixed-lag smoother on current-clamp data, such as one or more recordings of the electrical activity of the membrane induced by random current injections. (d) Finally, the predictive value of the model is assessed through comparison to independent data sets and the model is modified, if necessary. It is important to notice that the techniques outlined in this paper are applicable on a wide range of research domains and that they provide a disciplined way to merge complex stochastic dynamic models, noisy data and prior information under a common inference framework.

In conclusion, the class of statistical estimation methods, which the algorithm presented in this paper belongs to, in combination with Monte Carlo approximation techniques are particularly suitable to address high-dimensional inference problems in a disciplined manner. This makes them potentially useful tools at the disposal of biophysical modelers of neurons and neural networks and it is predicted that these methodologies will become more popular in the future among this research community.

Supporting Information

Figure S1

Simultaneous estimation of hidden states and channel noise in a stochastic single-compartment model. Estimation was based on a simulated Inline graphic-long recording of membrane potential generated by Supplementary Eqs. S1, S2 and S4. For clarity, only Inline graphic of activity are shown in Figs. Ai,ii. Notice that in these simulations, we assumed the absence of synaptic input (i.e. Inline graphic). Activity in the model neuron was driven by a random sequence of current steps Inline graphic with amplitude between Inline graphic and Inline graphic and duration up to Inline graphic. (A) Simultaneous inference of the observed membrane potential (Ai) and the hidden activation (Inline graphic, Inline graphic) and inactivation (Inline graphic) gating variables for the sodium and potassium currents. (B) Inference of the standard deviation of the observation noise Inline graphic (Bi) and the parameters Inline graphic and Inline graphic, which control the variance of the sodium and potassium channel noise (Bii). Estimates converged to their final values after approximately Inline graphic. The dashed lines indicate the true values of these parameters. The y-axes in Bi,ii indicate the width of the prior intervals imposed on the corresponding parameters. Simulation parameters were: Inline graphic, Inline graphic and Inline graphic. The prior interval for the scaling factors Inline graphic was Inline graphic.

(TIFF)

Figure S2

Simultaneous estimation of hidden states, channel noise and presynaptic firing rates in a stochastic single-compartment model. Estimation was based on a simulated Inline graphic-long recording of membrane potential generated by Eqs. S1, S2 and S4 with Inline graphic. For clarity, only Inline graphic of activity are shown in Figs. Ai,ii. (A) Simultaneous inference of the observed membrane potential (Ai) and the hidden activation (Inline graphic, Inline graphic) and inactivation (Inline graphic) gating variables for the sodium and potassium currents (Aii). (B) Inference of the standard deviation of the observation noise Inline graphic (Bi), parameters Inline graphic and Inline graphic, which control the variance of the sodium and potassium channel noise (Bii) and the presynaptic firing rates Inline graphic and Inline graphic (Biii). Estimates converged to their final values after approximately Inline graphic of activity. The dashed lines indicate the true values of the parameters. The y-axes in B and C indicate the width of the prior intervals imposed on the corresponding parameters. Discrepancies from the true values in B are due to the overlapping effects of different parameters controlling observation, channel and synaptic noise. Simulation parameters were: Inline graphic, Inline graphic and Inline graphic. The prior interval for the scaling factors Inline graphic was Inline graphic.

(TIFF)

Figure S3

Simultaneous inference of hidden states and unknown parameters in the single compartment model (see main text) at high levels of observation noise. This figure corresponds to Fig. 7Dii in the main text for Inline graphic. (A) Inferred membrane potential (Ai) and unobserved gating variables (Aii). (B) Examples of simultaneously inferred parameters, such as maximal conductances (Bi,ii) and reversal potentials (Biii,iv). The y-axes in Bi–iv indicate the prior intervals of the corresponding parameters. Simulation details are as in Fig. 7 in the main text.

(TIFF)

Figure S4

Simultaneous inference of hidden states in the two-compartment model (see main text) at low levels of observation noise. This figure corresponds to Figs. 11Aii for Inline graphic, 11B and 11C in the main text. (A) Inference of the membrane potential at the soma (Ai) and the dendritic compartment (Aii). (B) Inference of the unobserved concentration of intracellular calcium at the soma (Bi) and the dendritic compartment (Bii). (C) Inference of the unobserved gating variables for the sodium and potassium currents at the soma (Ci), the N-type calcium current at the soma (Ciii), the N-type calcium current at the dendritic compartment (Cii) and the L-type calcium current at the dendritic compartment (Civ). Simulation details are as in Fig. 11 in the main text.

(TIFF)

Figure S5

Simultaneous inference of hidden states in the two-compartment model (see main text) at high levels of observation noise. This figure corresponds to Fig. 11Aii for Inline graphic. (A) Inference of the membrane potential at the soma (Ai) and the dendritic compartment (Aii). (B) Inference of the unobserved concentration of intracellular calcium at the soma (Bi) and the dendritic compartment (Bii). (C) Inference of the unobserved gating variables for the sodium and potassium currents at the soma (Ci), the N-type calcium current at the soma (Ciii), the N-type calcium current at the dendritic compartment (Cii) and the L-type calcium current at the dendritic compartment (Civ). Simulation details are as in Fig. 11 in the main text.

(TIFF)

Figure S6

Inference in the B4 model using a single recording of the membrane potential. A single Inline graphic-long recording of B4 activity induced by injecting a sequence of random current steps in the neuron was used during smoothing. Random current amplitude was between Inline graphic and Inline graphic and random step duration was between Inline graphic and Inline graphic. (A) Inference of the membrane potential (Ai) and the unobserved gating variables for the sodium and potassium currents in the model (Aii). (B) Examples of simultaneously inferred model parameters: maximal conductances of all currents (Bi), half steady-state activation voltages for all currents (Bii) and maximal relaxation times for the activation of the potassium currents in the model (Biii). Notice that in all cases the parameter estimates converge exactly to the middle of their prior intervals (indicated by the y-axes in Bi–iii). This convergence takes place while the algorithm processes the “inactive” region of the data (approximately, from second Inline graphic to second 3 in Ai). Based on these converged estimates, the model incorrectly emits spikes later during smoothing (see arrows in Ai), indicating that the estimated parameters are not optimal for smoothing during the whole duration of experimental data. Simulation parameters were as follows: Inline graphic, Inline graphic and Inline graphic. The prior interval for the scaling factors Inline graphic was Inline graphic. For the Inline graphic free parameters in the model, we used the narrow prior intervals in Table 3.

(TIFF)

Figure S7

Inferred posterior distributions of all unknown parameters in the B4 model using the broad prior intervals in Table 3 . Inference was based on simultaneously smoothing four Inline graphic-long voltage recordings from the B4 neuron as in Fig. 12A in the main text. As in that case, data smoothing was accomplished with very high fidelity, as illustrated in Fig. 12A. (A) Inferred maximal conductances. (B) Inferred half steady-state activation and inactivation voltages. (C) Inferred activation (Ci) and inactivation (Cii) voltage sensitivities (parameters Inline graphic in the model). (D) Activation and inactivation relaxation times. The x-axes in all plots indicate the prior parameter intervals we used (Table 3). Notice that most posteriors are very broad (covering a large portion of the prior interval) and not unimodal. Simulation parameters were as described in Fig. 12 of the main text.

(TIFF)

Text S1

Supplementary text analysing in more detail several points in the manuscript. To be read in conjuction with the accompanying supplementary figures.

(PDF)

Text S2

The MATLAB/C99 source code we used in this study.

(BZ2)

Acknowledgments

The authors would like to thank four anonymous reviewers, whose constructive criticism helped substantially improve this paper.

Footnotes

The authors have declared that no competing interests exist.

This study was supported by the Future and Emerging Technologies (FET) European Commission program BION (number 213219). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1.Herz AVM, Gollisch T, Machens CK, Jaeger D. Modeling single-neuron dynamics and computations: a balance of detail and abstraction. Science. 2006;314:80–5. doi: 10.1126/science.1127240. [DOI] [PubMed] [Google Scholar]
  • 2.De Schutter E, Ekeberg O, Kotaleski JH, Achard P, Lansner A. Biophysically detailed modelling of microcircuits and beyond. Trends Neurosci. 2005;28:562–9. doi: 10.1016/j.tins.2005.08.002. [DOI] [PubMed] [Google Scholar]
  • 3.Grillner S. The motor infrastructure: from ion channels to neuronal networks. Nat Rev Neurosci. 2003;4:573–86. doi: 10.1038/nrn1137. [DOI] [PubMed] [Google Scholar]
  • 4.Marder E, Calabrese RL. Principles of rhythmic motor pattern generation. Physiol Rev. 1996;76:687–717. doi: 10.1152/physrev.1996.76.3.687. [DOI] [PubMed] [Google Scholar]
  • 5.Willms AR, Baro DJ, Harris-Warrick RM, Guckenheimer J. An improved parameter estimation method for hodgkin-huxley models. J Comput Neurosci. 1999;6:145–68. doi: 10.1023/a:1008880518515. [DOI] [PubMed] [Google Scholar]
  • 6.Goldman MS, Golowasch J, Marder E, Abbott LF. Global structure, robustness, and modulation of neuronal models. J Neurosci. 2001;21:5229–38. doi: 10.1523/JNEUROSCI.21-14-05229.2001. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Golowasch J, Goldman MS, Abbott LF, Marder E. Failure of averaging in the construction of a conductance-based neuron model. J Neurophysiol. 2002;87:1129–31. doi: 10.1152/jn.00412.2001. [DOI] [PubMed] [Google Scholar]
  • 8.Prinz AA, Billimoria CP, Marder E. Alternative to hand-tuning conductance-based models: construction and analysis of databases of model neurons. J Neurophysiol. 2003;90:3998–4015. doi: 10.1152/jn.00641.2003. [DOI] [PubMed] [Google Scholar]
  • 9.Prinz AA, Bucher D, Marder E. Similar network activity from disparate circuit parameters. Nat Neurosci. 2004;7:1345–52. doi: 10.1038/nn1352. [DOI] [PubMed] [Google Scholar]
  • 10.Nowotny T, Szücs A, Levi R, Selverston AI. Models wagging the dog: are circuits constructed with disparate parameters? Neural Comput. 2007;19:1985–2003. doi: 10.1162/neco.2007.19.8.1985. [DOI] [PubMed] [Google Scholar]
  • 11.Lepora NF, Overton PG, Gurney K. Efficient fitting of conductance-based model neurons from somatic current clamp. J Comput Neurosci. 2011 doi: 10.1007/s10827-011-0331-2. E-pub ahead of print. [DOI] [PubMed] [Google Scholar]
  • 12.Hendrickson EB, Edgerton JR, Jaeger D. The use of automated parameter searches to improve ion channel kinetics for neural modeling. J Comput Neurosci. 2011;31:329–46. doi: 10.1007/s10827-010-0312-x. [DOI] [PubMed] [Google Scholar]
  • 13.Keren N, Bar-Yehuda D, Korngreen A. Experimentally guided modelling of dendritic excitability in rat neocortical pyramidal neurones. J Physiol. 2009;587:1413–37. doi: 10.1113/jphysiol.2008.167130. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Pospischil M, Toledo-Rodriguez M, Monier C, Piwkowska Z, Bal T, et al. Minimal hodgkinhuxley type models for different classes of cortical and thalamic neurons. Biol Cybern. 2008;99:427–41. doi: 10.1007/s00422-008-0263-8. [DOI] [PubMed] [Google Scholar]
  • 15.Hobbs KH, Hooper SL. Using complicated, wide dynamic range driving to develop models of single neurons in single recording sessions. J Neurophysiol. 2008;99:1871–83. doi: 10.1152/jn.00032.2008. [DOI] [PubMed] [Google Scholar]
  • 16.Nowotny T, Levi R, Selverston AI. Probing the dynamics of identified neurons with a datadriven modeling approach. PLoS One. 2008;3:e2627. doi: 10.1371/journal.pone.0002627. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Reid MS, Brown EA, DeWeerth SP. A parameter-space search algorithm tested on a hodgkinhuxley model. Biol Cybern. 2007;96:625–634. doi: 10.1007/s00422-007-0156-2. [DOI] [PubMed] [Google Scholar]
  • 18.Druckmann S, Banitt Y, Gidon A, Schürmann F, Markram H, et al. A novel multiple objective optimization framework for constraining conductance-based neuron models by experimental data. Front Neurosci. 2007;1:7–18. doi: 10.3389/neuro.01.1.1.001.2007. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Achard P, De Schutter E. Complex parameter landscape for a complex neuron model. PLoS Comput Biol. 2006;2:e94. doi: 10.1371/journal.pcbi.0020094. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Keren N, Peled N, Korngreen A. Constraining compartmental models using multiple voltage recordings and genetic algorithms. J Neurophysiol. 2005;94:3730–42. doi: 10.1152/jn.00408.2005. [DOI] [PubMed] [Google Scholar]
  • 21.Tabak J, Murphey CR, Moore LE. Parameter estimation methods for single neuron models. J Comput Neurosci. 2000;9:215–236. doi: 10.1023/a:1026531603628. [DOI] [PubMed] [Google Scholar]
  • 22.Vanier MC, Bower JM. A comparative survey of automated parameter-search methods for compartmental neural models. J Comput Neurosci. 1999;7:149–71. doi: 10.1023/a:1008972005316. [DOI] [PubMed] [Google Scholar]
  • 23.Van Geit W, De Schutter E, Achard P. Automated neuron model optimization techniques: a review. Biol Cybern. 2008;99:241–51. doi: 10.1007/s00422-008-0257-6. [DOI] [PubMed] [Google Scholar]
  • 24.LeMasson G, Maex R. Computational Neuroscience: realistic modelling for experimentalists. CRC Press; 2001. pp. 1–25. [Google Scholar]
  • 25.Van Geit W, Achard P, De Schutter E. Neurofitter: a parameter tuning package for a wide range of electrophysiological neuron models. Front Neuroinform. 2007;1:1. doi: 10.3389/neuro.11.001.2007. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Tokuda I, Parlitz U, Illing L, Kennel M, Abarbanel HD. Parameter estimation for neuron models. 2003. pp. 251–256. In: Experimental Chaos: 7th Experimental Chaos Conference. volume 676.
  • 27.Lillacci G, Khammash M. Parameter estimation and model selection in computational biology. PLoS Comput Biol. 2010;6:e1000696. doi: 10.1371/journal.pcbi.1000696. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Huys QJM, Paninski L. Smoothing of, and parameter estimation from, noisy biophysical recordings. PLoS Comput Biol. 2009;5:e1000379. doi: 10.1371/journal.pcbi.1000379. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.Kitagawa G. A self-organizing state-space model. J Am Stat Assoc. 1998;93:1203–1215. [Google Scholar]
  • 30.Igel C, Hansen N, Roth S. Covariance matrix adaptation for multi-objective optimization. Evol Comput. 2007;15:1–28. doi: 10.1162/evco.2007.15.1.1. [DOI] [PubMed] [Google Scholar]
  • 31.Price KV, Storn RM, Lampinen JA. Differential evolution: a practical approach to global optimization. Springer; 2005. p. 558 p. [Google Scholar]
  • 32.Faisal AA, Selen LPJ, Wolpert DM. Noise in the nervous system. Nat Rev Neurosci. 2008;9:292–303. doi: 10.1038/nrn2258. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Kloeden PE, Platen E. Numerical Solution of Stochastic Differential Equations. Springer; 1999. [Google Scholar]
  • 34.Cappe O, Godsill SJ, Moulines E. An overview of existing methods and recent advances in sequential monte carlo. P IEEE. 2007;95:899–924. [Google Scholar]
  • 35.Doucet A, de Freitas N, Gordon N, editors. Sequential Monte Carlo Methods in Practice. Springer; 2001. [Google Scholar]
  • 36.van Leeuwen PJ. Nonlinear data assimilation in geosciences: an extremely efficient particle filter. Q J R Meteorol Soc. 2010;136:1991–1999. [Google Scholar]
  • 37.Booth V, Rinzel J, Kiehn O. Compartmental model of vertebrate motoneurons for ca2+- dependent spiking and plateau potentials under pharmacological treatment. J Neurophysiol. 1997;78:3371–85. doi: 10.1152/jn.1997.78.6.3371. [DOI] [PubMed] [Google Scholar]
  • 38.Straub VA, Benjamin PR. Extrinsic modulation and motor pattern generation in a feeding network: a cellular study. J Neurosci. 2001;21:1767–78. doi: 10.1523/JNEUROSCI.21-05-01767.2001. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39.Straub VA, Grant J, O'Shea M, Benjamin PR. Modulation of serotonergic neurotransmission by nitric oxide. J Neurophysiol. 2007;97:1088–99. doi: 10.1152/jn.01048.2006. [DOI] [PubMed] [Google Scholar]
  • 40.Koch C. Biophysics of computation: information processing in single neurons. Oxford University Press; 2005. p. 588 p. [Google Scholar]
  • 41.Vehovszky A, Szabó H, Elliott CJH. Octopamine increases the excitability of neurons in the snail feeding system by modulation of inward sodium current but not outward potassium currents. BMC Neurosci. 2005;6:70. doi: 10.1186/1471-2202-6-70. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42.Vavoulis DV, Nikitin ES, Kemenes I, Marra V, Feng J, et al. Balanced plasticity and stability of the electrical properties of a molluscan modulatory interneuron after classical conditioning: a computational study. Front Behav Neurosci. 2010;4:19. doi: 10.3389/fnbeh.2010.00019. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43.Haufler D, Morin F, C LJ, K SF. Parameter estimation in single-compartment neuron models using a synchronization-based method. Neurocomputing. 2007;70:1605–1610. [Google Scholar]
  • 44.Huys QJM, Ahrens MB, Paninski L. Efficient estimation of detailed single-neuron models. J Neurophysiol. 2006;96:872–90. doi: 10.1152/jn.00079.2006. [DOI] [PubMed] [Google Scholar]
  • 45.Lee A, Yau C, Giles MB, Doucet A, Holmes CC. On the utility of graphics cards to perform massively parallel simulation of advanced monte carlo methods. J Comput Graph Stat. 2010;19:769–789. doi: 10.1198/jcgs.2010.10039. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 46.Quinn JC, Abarbanel HD. Data assimilation using a gpu accelerated path integral monte carlo approach. 2011. arXiv : http://arxiv.org/abs/1103.4887.

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Figure S1

Simultaneous estimation of hidden states and channel noise in a stochastic single-compartment model. Estimation was based on a simulated Inline graphic-long recording of membrane potential generated by Supplementary Eqs. S1, S2 and S4. For clarity, only Inline graphic of activity are shown in Figs. Ai,ii. Notice that in these simulations, we assumed the absence of synaptic input (i.e. Inline graphic). Activity in the model neuron was driven by a random sequence of current steps Inline graphic with amplitude between Inline graphic and Inline graphic and duration up to Inline graphic. (A) Simultaneous inference of the observed membrane potential (Ai) and the hidden activation (Inline graphic, Inline graphic) and inactivation (Inline graphic) gating variables for the sodium and potassium currents. (B) Inference of the standard deviation of the observation noise Inline graphic (Bi) and the parameters Inline graphic and Inline graphic, which control the variance of the sodium and potassium channel noise (Bii). Estimates converged to their final values after approximately Inline graphic. The dashed lines indicate the true values of these parameters. The y-axes in Bi,ii indicate the width of the prior intervals imposed on the corresponding parameters. Simulation parameters were: Inline graphic, Inline graphic and Inline graphic. The prior interval for the scaling factors Inline graphic was Inline graphic.

(TIFF)

Figure S2

Simultaneous estimation of hidden states, channel noise and presynaptic firing rates in a stochastic single-compartment model. Estimation was based on a simulated Inline graphic-long recording of membrane potential generated by Eqs. S1, S2 and S4 with Inline graphic. For clarity, only Inline graphic of activity are shown in Figs. Ai,ii. (A) Simultaneous inference of the observed membrane potential (Ai) and the hidden activation (Inline graphic, Inline graphic) and inactivation (Inline graphic) gating variables for the sodium and potassium currents (Aii). (B) Inference of the standard deviation of the observation noise Inline graphic (Bi), parameters Inline graphic and Inline graphic, which control the variance of the sodium and potassium channel noise (Bii) and the presynaptic firing rates Inline graphic and Inline graphic (Biii). Estimates converged to their final values after approximately Inline graphic of activity. The dashed lines indicate the true values of the parameters. The y-axes in B and C indicate the width of the prior intervals imposed on the corresponding parameters. Discrepancies from the true values in B are due to the overlapping effects of different parameters controlling observation, channel and synaptic noise. Simulation parameters were: Inline graphic, Inline graphic and Inline graphic. The prior interval for the scaling factors Inline graphic was Inline graphic.

(TIFF)

Figure S3

Simultaneous inference of hidden states and unknown parameters in the single compartment model (see main text) at high levels of observation noise. This figure corresponds to Fig. 7Dii in the main text for Inline graphic. (A) Inferred membrane potential (Ai) and unobserved gating variables (Aii). (B) Examples of simultaneously inferred parameters, such as maximal conductances (Bi,ii) and reversal potentials (Biii,iv). The y-axes in Bi–iv indicate the prior intervals of the corresponding parameters. Simulation details are as in Fig. 7 in the main text.

(TIFF)

Figure S4

Simultaneous inference of hidden states in the two-compartment model (see main text) at low levels of observation noise. This figure corresponds to Figs. 11Aii for Inline graphic, 11B and 11C in the main text. (A) Inference of the membrane potential at the soma (Ai) and the dendritic compartment (Aii). (B) Inference of the unobserved concentration of intracellular calcium at the soma (Bi) and the dendritic compartment (Bii). (C) Inference of the unobserved gating variables for the sodium and potassium currents at the soma (Ci), the N-type calcium current at the soma (Ciii), the N-type calcium current at the dendritic compartment (Cii) and the L-type calcium current at the dendritic compartment (Civ). Simulation details are as in Fig. 11 in the main text.

(TIFF)

Figure S5

Simultaneous inference of hidden states in the two-compartment model (see main text) at high levels of observation noise. This figure corresponds to Fig. 11Aii for Inline graphic. (A) Inference of the membrane potential at the soma (Ai) and the dendritic compartment (Aii). (B) Inference of the unobserved concentration of intracellular calcium at the soma (Bi) and the dendritic compartment (Bii). (C) Inference of the unobserved gating variables for the sodium and potassium currents at the soma (Ci), the N-type calcium current at the soma (Ciii), the N-type calcium current at the dendritic compartment (Cii) and the L-type calcium current at the dendritic compartment (Civ). Simulation details are as in Fig. 11 in the main text.

(TIFF)

Figure S6

Inference in the B4 model using a single recording of the membrane potential. A single Inline graphic-long recording of B4 activity induced by injecting a sequence of random current steps in the neuron was used during smoothing. Random current amplitude was between Inline graphic and Inline graphic and random step duration was between Inline graphic and Inline graphic. (A) Inference of the membrane potential (Ai) and the unobserved gating variables for the sodium and potassium currents in the model (Aii). (B) Examples of simultaneously inferred model parameters: maximal conductances of all currents (Bi), half steady-state activation voltages for all currents (Bii) and maximal relaxation times for the activation of the potassium currents in the model (Biii). Notice that in all cases the parameter estimates converge exactly to the middle of their prior intervals (indicated by the y-axes in Bi–iii). This convergence takes place while the algorithm processes the “inactive” region of the data (approximately, from second Inline graphic to second 3 in Ai). Based on these converged estimates, the model incorrectly emits spikes later during smoothing (see arrows in Ai), indicating that the estimated parameters are not optimal for smoothing during the whole duration of experimental data. Simulation parameters were as follows: Inline graphic, Inline graphic and Inline graphic. The prior interval for the scaling factors Inline graphic was Inline graphic. For the Inline graphic free parameters in the model, we used the narrow prior intervals in Table 3.

(TIFF)

Figure S7

Inferred posterior distributions of all unknown parameters in the B4 model using the broad prior intervals in Table 3 . Inference was based on simultaneously smoothing four Inline graphic-long voltage recordings from the B4 neuron as in Fig. 12A in the main text. As in that case, data smoothing was accomplished with very high fidelity, as illustrated in Fig. 12A. (A) Inferred maximal conductances. (B) Inferred half steady-state activation and inactivation voltages. (C) Inferred activation (Ci) and inactivation (Cii) voltage sensitivities (parameters Inline graphic in the model). (D) Activation and inactivation relaxation times. The x-axes in all plots indicate the prior parameter intervals we used (Table 3). Notice that most posteriors are very broad (covering a large portion of the prior interval) and not unimodal. Simulation parameters were as described in Fig. 12 of the main text.

(TIFF)

Text S1

Supplementary text analysing in more detail several points in the manuscript. To be read in conjuction with the accompanying supplementary figures.

(PDF)

Text S2

The MATLAB/C99 source code we used in this study.

(BZ2)


Articles from PLoS Computational Biology are provided here courtesy of PLOS

RESOURCES