Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2019 Oct 11.
Published in final edited form as: Neurocomputing (Amst). 2002 Jun;44-46:1071–1076. doi: 10.1016/s0925-2312(02)00418-6

A general framework for neurobiological modeling: an application to the vestibular system

Chris Eliasmith a,*, MB Westover b, CH Anderson b
PMCID: PMC6788744  NIHMSID: NIHMS1053520  PMID: 12744262

Abstract

The otolith organs in the vestibular system are excellent detectors of linear accelerations. However, any measurement of linear acceleration is ambiguous between a tilt in a gravitational field and an inertial acceleration. Angelaki et al. have put forward a general hypothesis about how inertial accelerations can be computed based on vestibular signals (J. Neurosci. 19 (1999) 316). We have constructed a realistic, detailed model of the relevant systems to test this hypothesis. The model produces useful predictions about what kinds of neurons should be found in the vestibular nucleus if such a computation is actually performed in the vestibular system. The model is constructed using general principles of neurobiological simulation (J. Neurophys. 84 (2000) 2113).

Keywords: Vestibular system, Neural representation, Large-scale models

1. Introduction

Given any accelerometer, it is not possible to distinguish inertial acceleration due to translation from gravitational acceleration during tilting. In the mammalian vestibular system, the otoliths serve as an accelerometer. Thus, the otolith signals are ambiguous between these two kinds of acceleration. However, animals can successfully distinguish such motions. Precisely, how this distinction is made is a matter of some debate in contemporary neuroscience [2,3,6].

In a series of experiments on macaque monkeys, Angelaki et al. show that inactivation of the semicircular canals results in precisely the kinds of eye movement errors we would expect if only the otolith were used to distinguish inertial from tilt accelerations [2]. Similarly, Merfeld et al. show that reflexive eye movements in humans are most consistent with a combination of canal and otolith information [6]. This kind of evidence strongly supports the contention that there is an integration of sensory signals from both the otolith and semicircular canals used to determine appropriate motor responses (see also [7,8]). Angelaki et al. have suggested a series of transformations that could be used to determine an unambiguous inertial acceleration given the otolith and semicircular canal signals [2]. In this paper, we present a detailed model that implements these transformations using our framework for constructing realistic neurobiological models [5,4,1].

2. A quantitative hypothesis about vestibular function

Angelaki et al. suggest that the following computation is implemented in the brain to determine inertial acceleration, I(t), provided the otolith signal A(t) and the canal signal (t) [2]:

I(t)=Ω(t)×I(t)+A(t)Ω(t)×A(t) (1)
=A(t)+Ω(t)×[A(t)I(t)]. (2)

The well-established assumptions underlying this hypothesis are: (1) that the three semicircular canals generally operate as orthogonal integrating accelerometers which together provide a good estimate of angular velocity, (t); and (2) that the two otolith organs encode accelerations in the horizontal and vertical planes, providing a good estimate of linear acceleration A(t) [8]. In order to characterize the dynamics for a neural population that represents this inertial acceleration, we use a first-order Taylor expansion:

I(t+τ)I(t)+τI(t), (3)
=I(t)+τ[A(t)+Ω(t)×[A(t)I(t)]]. (4)

Eqs. (2) and (4), along with neurophysiological considerations suggest a neural circuit as that shown in Fig. 1. This diagram provides a particular modularization of the computation needed to determine the true inertial acceleration.

Fig. 1.

Fig. 1.

A high-level modularization of a vestibular circuit that can compute inertial acceleration from the labyrinth inputs.

We think that this modularization is a reasonable one given that there are strong convergent projections of both the otolith and canal systems within the vestibular nucleus [8, p. 166]. In our model, we use leaky integrate-and-fire neurons tuned to mimic the properties of neurons found in the labyrinth and relevant parts of the vestibular nucleus [8].1

3. The model and results

In order to construct a model of this system, we embed each of the higher-level representations and transformations identified in Fig. 1 into the relevant neural populations. The general principles behind this approach are articulated in [1]. The representations and transformations up to the vestibular nucleus are relatively simple, so we will focus instead on the representation and transformations of the vestibular neurons that perform the cross-product, a highly non-linear operation.

We take the first population of neurons involved in the cross-product transformation, ck(t), to construct an integrated representation of the input space, C, by encoding (t) and A(t) − I(t) simultaneously using a six-dimensional encoding vector, ϕ˜C:

ck(C=[Ω,AI])=Gk[αkϕ˜kCC+Jkb]=Gk[αkϕ˜kΩΩ+ϕ˜kΔ(AI)+Jkb]=Gk[kωkjbj+iωkiai+lωkldl+Jkb],

where ωkj=αkϕ˜kΩϕ˜jΩ, ωki=αkϕ˜kΔϕ˜iA, and ωkl=αkϕ˜kΔϕ˜lI.

In order to use this encoding to compute the cross-product of elements in the C space, we need to determine a decoding for the function P(C), which finds the cross-product of the first and last three elements of any vector C. This decoding will be P(C)=kϕkpck. We find the decoding vectors by solving the appropriate least-squares minimization problem.

This decoding can be used in a description of the dynamics of the subsequent population. Eq. (3) gives an expression for neurons encoding the inertial acceleration:dl(t+τ)=Gl[αlϕ˜lII(t+τ)+Jlb]

Substituting from (4):

dl(t+τ)=Gl[αlϕ˜lI(I(t)+τ[A(t)+Ω(t)×[A(t)I(t)]])+Jlb]=Gl[lωlldl(t)+τkωlkck(t)+τiωliai(t)+Jlb],

where ωll=αlϕ˜lIϕlI, and ωlk=αlϕ˜lIϕkp. A(t) requires separate (though similar) treatment. However, we do not discuss it here.

In order to model these populations with spiking neurons, we use a linear decoding filter on the spike trains to give neuron activity variables, ai(t), bj(t), etc. Thus, for each population we write an expression of the form: ai(t)= Σn hi(t)*δ(ttn)= Σn hi(ttn), where the linear filters, hi(t) are taken to be the postsynaptic currents produced in the postsynaptic neuron given the incoming spike train, δ(ttn). The hi(t) responses are appropriately weighted by the relevant weights, ω, determined as above.

We can now test our model by comparing it to experimental results from macaques.2 In order to show that the vestibular system is able to estimate true inertial acceleration, despite the ambiguity in the otolith signal, the following three experiments were performed. First, the animal was subjected only to translational acceleration. Second, the animal was subjected to a translational acceleration coupled with an angular rotation chosen to double the otolith signal. Third, the animal was subjected to a translation acceleration coupled with an angular rotation chosen to cancel the otolith signal. Any estimate of translational acceleration based solely on otolith information would be severely adversely affected in the last two of these cases. Nevertheless, eye position changes resulting from the translational vestibular ocular reflex (tVOR) are very similar. This suggests that the otolith signal is not solely responsible for the tVOR.

Fig. 2 shows the results of the model compared to both a direct solution of Eq. (2) and the data, for the case when the otolith signal is canceled. Similar agreement between experimental and model results was obtained for the other conditions. Thus, the model verifies that the hypothesis is a reasonable one, and that it can be implemented in a biologically plausible, spiking network.

Fig. 2.

Fig. 2.

Experimental data, equation solution, and model solution. The model consists of a total of 5000 LIF neurons. Comparisons incorporate a gain reduction which has been experimentally verified for tVOR, and ranges between about 0.1 and 0.25.

4. Conclusion

One of the reasons for constructing this model is to determine what the tuning properties of certain populations in the model are expected to look like. Of particular interest are those neurons performing the cross product. On the basis of this model, we expect these neurons to have preferred direction vectors in a six-dimensional space, which are sensitive to both angular velocity and linear acceleration. However, this expectation is based mainly on our chosen means of implementing the equation. It is possible, for instance, to perform the cross product as a series of multiplications between components of these vectors. In this case, neuron preferred direction vectors would lie in two-dimensional spaces. Nevertheless, however we choose to implement the cross-product, there must be neurons whose preferred direction vectors mix the otolith and vestibular signals. Furthermore, given the fact that there is a non-linear transformation being performed, we expect neurons in this population to have fairly low background firing rates and more often than not require significant input to begin firing. This is due to the fact that for non-linear transformations, a greater response range is required as we move away from the resting state compared to linear transformations.

In addition, the dynamics of the network provide predictions regarding the kinds of receptors present on neurons in the various populations. Although we have not discussed the dynamics in detail here, based on this model we expect the neurons participating in the I population to have recurrent connections to NMDA-type receptors because the recurrence requires slow synaptic decays to function well. The remaining connections can be fast, AMPA-type receptors. This suggests that some cells in the I population will have both kinds of receptors on the same cell. Notably, this is consistent with the general properties of neurons in the vestibular nucleus [9].

Acknowledgments

This research has been supported by the McDonnell Center for Higher Brain Function at Washington University and the Mathers Foundation.

Footnotes

1

We can replace the LIF neurons with more complex single cell models. But, the computational cost for this increase in detail is unacceptable in this case.

2

Thanks are due to Dora Angelaki for providing the experimental data.

References

  • [1].Angelaki DE, Dickman JD, Spatiotemporal processing of linear acceleration : primary afferent and central vestibular neuron responses, J. Neurophys 84 (2000) 2113–2132. [DOI] [PubMed] [Google Scholar]
  • [2].Angelaki DE, McHenry MQ, Dickman JD, Newlands SD, Hess BJM, Computation of inertial motion: neural strategies to resolve ambiguous otolith information, J. Neurosci 19 (1999) 316–327. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [3].Chen LW, Yung KKL, Chan YS, Co-localization of NMDA receptors and AMPA receptors in neurons, Brain Res. 884 (2000) 87–97. [DOI] [PubMed] [Google Scholar]
  • [4].Eliasmith C, Anderson CH, Developing and applying a toolkit from a general neurocomputational framework, Neurocomputing 26 (1999) 1013–1018. [Google Scholar]
  • [5].Eliasmith C, Anderson CH , Neurobiological simulation: principles and methods, Neurocomputing 38 (2001) 581–586. [Google Scholar]
  • [6].Eliasmith C, Anderson CH, Neural Engineering: The Principles of Neurobiological Simulation, MIT Press, Cambridge, in press. [Google Scholar]
  • [7].Merfeld DM, Zupan L, Peterka RJ, Humans use internal models to estimate gravity and linear acceleration, Nature 398 (1999) 615–618. [DOI] [PubMed] [Google Scholar]
  • [8].Telford L, Seidman SH, Paige GD, Dynamics of squirrel monkey linear vestibuloocular reflex and interactions with fixation distance, J. Neurophys 78 (1997) 1775–1790. [DOI] [PubMed] [Google Scholar]
  • [9].Wilson VJ, Jones GM, Mammalian Vestibular Physiology, Plenum Press, New York, 1979. [Google Scholar]

RESOURCES