Abstract
When navigating through the environment, our brain needs to infer how far we move and in which direction we are heading. In this estimation process, the brain may rely on multiple sensory modalities, including the visual and vestibular systems. Previous research has mainly focused on heading estimation, showing that sensory cues are combined by weighting them in proportion to their reliability, consistent with statistically optimal integration. But while heading estimation could improve with the ongoing motion, due to the constant flow of information, the estimate of how far we move requires the integration of sensory information across the whole displacement. In this study, we investigate whether the brain optimally combines visual and vestibular information during a displacement estimation task, even if their reliability varies from trial to trial. Participants were seated on a linear sled, immersed in a stereoscopic virtual reality environment. They were subjected to a passive linear motion involving visual and vestibular cues with different levels of visual coherence to change relative cue reliability and with cue discrepancies to test relative cue weighting. Participants performed a two-interval two-alternative forced-choice task, indicating which of two sequentially perceived displacements was larger. Our results show that humans adapt their weighting of visual and vestibular information from trial to trial in proportion to their reliability. These results provide evidence that humans optimally integrate visual and vestibular information in order to estimate their body displacement.
Introduction
To accurately navigate through the environment, we not only need to know the direction in which we move, called heading, but we also need to keep track of our displacement. In inferring these components of self-motion, the brain may rely on various sensory modalities, especially visual (e.g., optic flow) [1] and vestibular information [2–9].
Heading estimation has received much attention, particularly with regard to how the brain combines overlapping but noisy information from different modalities. In recent years this problem has typically been approached from a Bayesian inference perspective. In this framework an optimal observer reduces the uncertainty in the internal estimate of heading by weighting each signal in proportion to its reliability [3, 10–17]. This framework still holds when visual cue reliability varies unpredictably across trials. This notion of dynamic cue reweighting in multisensory heading perception has been validated in recent years, both at the behavioral [2] and the neural level [18].
Would the same computational principle also hold up fort the estimation of travelled distance? For several reasons, displacement estimation may not rely on the same neural computations as heading estimation. First, heading estimation can be derived from the instantaneous direction of translation and may improve through evidence accumulation with the ongoing motion [19]. The estimate of displacement, in contrast, requires the combining of sensory information and (double) integration of the acceleration and velocity information across the whole displacement, i.e., from start to end of the motion. Second, geometrically, while optic flow signals can readily be used to estimate heading [1], they require depth scaling to provide unambiguous cues to travelled distance [20]. Third, heading and displacement detection thresholds appear to be susceptible to motion directions in different reference frames [9].
Thus far, most studies on displacement estimation have been limited to testing the contribution of single cues, such as vision [20–23] or vestibular input [6, 8, 24]. Only a few studies reported on the interplay between visual and vestibular information during displacement estimation. In the study by Harris et al. (2000) subjects had to indicate the position of a previously perceived visual or physical stop target by a button-press when “passing” that target during passive translation with either ‘only visual’, ‘only vestibular’ or ‘visual and vestibular’ information being present [4]. They observed differences in the accuracies when comparing displacement estimates from different modalities, i.e., visual or vestibular. Another study tested absolute displacement estimation in an oscillatory up-down movement [5]. They showed that a model of gain-dependent linear weighting of visual and vestibular cues could explain their results for in-phase, but not out-of-phase, visual and vestibular motion. Along the same line, Campos et al. (2012) used a task in which subjects had to match a previous displacement by adjusting the distance of a virtual target [25]. Using the accuracies of displacement estimates in single and combined cue conditions, they were able to predict the observed linear weighting of discrepant visual and vestibular cues. The latter two studies [5, 25]explained their data using a linear integration model. However, because they did not experimentally vary cue reliability, it is unknown whether this integration was performed optimally by combining displacement cues relative to their reliability. Therefore, the objective of the present study is to examine whether visual and vestibular cues are dynamically weighted in displacement estimation when cue reliability is varied randomly across trials.
Here, using a psychometric approach, we tested whether observers show optimal visual and vestibular cue integration in displacement estimation for different levels of cue reliability, analogous to previous experiments on heading estimation in humans and macaque monkeys[2]. By manipulating cue reliability randomly across trials, we were able to test whether cue weights are dynamically updated at the single trial level. We found that human subjects optimally integrate visual and vestibular cues by dynamically adjusting cue weights depending on their relative reliability from trial to trial.
Materials and Methods
Participants
Six healthy subjects (4 female), 19–25 years old participated in the study. All subjects had normal or corrected to normal vision, including normal stereovision (tested using the Randot Stereo test (Stereo Optical Inc., Chicago, USA)) and no known history of neurological, visual, or vestibular sensory disorders. Informed written consent was obtained from all subjects prior to the experiment and the experiment was approved by the Ethics Committee Faculty of Social Sciences (ECSS).
Equipment
Vestibular stimuli were administered using a linear sled on an 800-mm track with which subjects were laterally translated. The sled, powered by a linear motor (TB15N, Technotion, Almelo, The Netherlands), was controlled by a Kollmorgen S700 (Danaher, Washington, DC) drive. The kinematics of the trajectory of the subjects were accurately controlled at 1000 Hz with accuracy better than 34 μm, 2 mm/s, and 150 mm/s2. Subjects were seated on the sled with their interaural axis aligned with the sled’s motion axis and were restrained using a five point seat belt and a chin rest (Fig 1). The head was firmly held in place using a chin rest and an ear-fixed mold. Cushioning was provided at the back and sides of the head for comfort and to prevent confounding cues related motion related vibrations. White noise was administered through headphones to prevent the subjects from using motion related noise of the sled to be used as a cue.
Visual stimuli were projected using two digital stereo DLP®-rear projection cubes (EC-3D-67-SXT+ -CP, Eyevis GmbH, Reutlingen, Germany) on a 2.83 X 1.05 m surface with a resolution of 2800 by 1050 pixels. Subjects were seated 1.2 m in front of the screens, which thus subtended 97.3° X 45.4° of visual angle. Vertical retraces of the screens were synchronized using an Nvidea Quadro K5000 graphics card. The visual display was updated at 60 Hz. Stereoscopic images were generated using channel separation, based on interference filter technology (INFITEC® GmbH, Ulm, Germany). Images for the left and right eye were projected at different wavelengths. Subjects wore a pair of glasses with selective interference filters for each eye. Visual stimuli depicted optic flow that accurately simulated movement of the observer through a world stable cloud of random stars in a virtual space of 3m wide, 2m high and 1.2m deep (0.6 m in front and 0.6 m behind the screen). Star density was 0.01/cm³, with each star being a 0.15 cm X 0.15 cm triangle. The visual stimulus also contained a white head-fixed fixation point at screen depth at which the subjects had to fixate during the trial. The visual scene provided multiple depth cues such as relative size, motion parallax and binocular disparity.
Projection of perspective changes due to motion was controlled by custom developed python software, using OpenGL. The sled position was sampled at 1000Hz and by using an extrapolation paradigm on the current and previous sled positions we predicted the position on t + 66.67ms to overcome the latency of the projection system. This predicted position was used as observer position in our OpenGL software to render the 3D environment. In the case of only visual cues and no self-displacement, a sled-movement was simulated to administer the same visual cues to the observer as during a condition in which visual and vestibular (i.e., self-motion) cues were presented.
Procedure
The task consisted of a two-interval two-alternative forced-choice (2AFC) task in which subjects had to judge whether the second of two consecutive movements was shorter or longer than the first. One of the intervals contained the “reference” movement the other imposed the “probe” movement. The reference movement consisted of a rightward translation of 20 cm and the probe movement was also rightward but varied in fine steps around the reference amplitude using the adaptive Ψ-method [26], but was limited between 5cm and 35 cm. The order of reference and probe movement was counterbalanced across trials. Subjects were unaware of the order of the reference and probe movements. At the end of the second movement subjects were presented an auditory cue indicating they could respond. After the response was given, subjects were translated leftward to the starting position. In each trial subjects were presented with two 1.5s translational motion stimuli with a bell-shaped velocity profile (peak velocity = 0.25 m/s, peak acceleration = 0.98 m/s² for the reference movement). Both motion profiles were delivered in one of three randomly interleaved stimulus modalities: vestibular only (inertial movement without optic flow in complete darkness), visual only (optic flow without inertial movement) and combined (synchronous inertial movement and optic flow). Within a trial the motion profiles only differed in amplitude and the stimulus properties remained the same (modality and coherence). We note that during inertial movement, also extra-vestibular cues were available to the subjects such as somatosensation and proprioception; however, we refer to this condition as vestibular because both behavioral and electrophysiological performance has been shown to strongly depend on intact vestibular labyrinths [16].
Combined condition trials were randomly assigned one of three conflict amplitudes: +Δ,—Δ, or 0 (no conflict), with Δ being 5cm. Conflict amplitudes only existed in the reference movement (Fig 2). Positive Δ indicates smaller movement amplitude for vestibular relative to visual cues (vice versa for negative Δ). When Δ was nonzero, the vestibular displacement was 17.5 cm or 22.5 cm and visual displacement was 22.5 cm or 17.5 cm for Δ = 5 cm and Δ = -5 cm, respectively.
During the experiment, visual cue reliability was manipulated by changing the motion coherence of the optic flow pattern by randomly replacing a fixed percentage of the stars in each frame. This visual motion coherence was manipulated in three steps, i.e. 20%, 60% and 100%, randomly applied across trials. Vestibular cue reliability was held constant. All stimulus triangles had a limited lifetime of 30 frames i.e., approximately 500ms. Prior to the movement the stars were randomly divided into 30 triangle-sets. Every frame one of these sets was re-initialized at a new random location in the world and this location remained stationary for the next 30 frames, i.e. was drawn in accordance with the observers’ displacement.
Each subject was presented with 13 different conditions, see Table 1, divided over 5 sessions of one hour each. All visual only conditions were randomly interleaved as well as the combined conditions, though both sets of conditions were acquired separately. Also the vestibular only condition was acquired in a separate session. Prior to each set of conditions, subjects conducted 25 practice trials leading to a total of 2025 trials (13 conditions * 150 trials + 3 practice blocks * 25 trials).
Table 1. Conditions with corresponding cues, motion coherence and discrepancy.
Condition | Cue | Motion Coherence (%) | Δ (cm) |
---|---|---|---|
1 | Vestibular | NA | NA |
2 | Visual | 100 | NA |
3 | Visual | 60 | NA |
4 | Visual | 20 | NA |
5 | Combined | 100 | -5 |
6 | Combined | 60 | -5 |
7 | Combined | 20 | -5 |
8 | Combined | 100 | 0 |
9 | Combined | 60 | 0 |
10 | Combined | 20 | 0 |
11 | Combined | 100 | +5 |
12 | Combined | 60 | +5 |
13 | Combined | 20 | +5 |
Theory and model predictions
Given the visual S visand vestibular S ves sensory estimates of an environmental propertyS, the joint likelihood of these estimates can be described as P(S|S vis,S ves). Using Bayes’ rule, this probability can be re-written as
(1) |
Assuming a uniform prior and independent noise for the single cues, Eq 1 can be rewritten as
(2) |
When the likelihood functions of the two single cues are considered to be Gaussian, the posterior probability will also be Gaussian. This results in a simple statistical optimal estimation model in which the means and , and variances and of the single cue distributions can be used to obtain the combined mean and variance . We find:
(3) |
(4) |
with
(5) |
Thus, in this optimal integration model the combined mean is the weighted average of the individual mean estimates, with weights proportional to the reliability (i.e. the inverse of the variance) of the corresponding cues, and the estimated combined variance is smaller than either of the unimodal variances, with the largest improvement (by a factor of 2) when the single cue variances are equal.
The predicted combined cue variance and optimal weighting for the combined mean based on the individual cues’ variances are compared to the observed combined cue variance and the observed weights in the data. Assuming that both the vestibular and visual sensors are accurate, experimental information about the weighting of the single cues is only available in the combined conditions that have a discrepancy between the visual and vestibular cues, i.e. the Δ = -5 and Δ = +5 conditions [2]. We used these data as follows. Consider the case of positive Δ, i.e., with the visual cue amplitude smaller than the vestibular cue. The weighting can be estimated by measuring the shift of the point of subjective equality (PSE) relative to the zero conflict combined condition. When the PSE is shifted by +Δ/2, the subject estimated 50% of the trials in which the probe was as large as the vestibular reference cue (cue conflict was only applied to the reference) to be larger than the reference. This can only be the case when the subject completely relies upon the vestibular cue, i.e., with a vestibular weight of 1. Following this analysis, the observed weights were estimated by rewriting Eq 3, again assuming accurate sensors:
Substituting the discrepancy terms (-Δ/2 and +Δ/2 for vestibular and visual cues, respectively) we obtain:
resulting in:
(6) |
Data analysis
Analyses were performed using Matlab R2012a (MathWorks) and SPSS Statistics 19 (IBM). Response data was summarized by fitting a cumulative Gaussian to the proportion of “probe amplitude larger than the reference” responses as function of probe amplitude. The psychometric threshold and point of subjective equality (PSE) were taken as the SD (σ) and mean (μ), respectively. Furthermore, the lapse rate (typically represented by λ) was set at 0.04. Psychometric curve parameters (μ, σ) were fitted for each subject, stimulus condition, visual coherence level, and discrepancy using Maximum Likelihood Estimation (MLE). These parameters were inserted into Eqs 4–6 to predict the combined thresholds and weights.
Optimal integration predicts the combined threshold to be smaller than that of either single cue condition, although the size of this effect will depend on the discrepancy between the two single cue thresholds. That is, with large differences between the single cues in the low coherence levels the combined threshold will not decrease much with respect to the lowest single cue threshold, see Eq 4. We performed a direct comparison between the lowest single cue and combined thresholds for each coherence level separately using single tailed paired t-tests. We also tested whether the predicted and observed thresholds differ with a two-tailed paired t-test for each coherence level. In addition, we tested the effect of adding a discrepancy versus no discrepancy by using two-tailed paired t-tests, one for each coherence level. To test the effect of coherence level on dynamic weighting we tested whether the predicted and observed weights differ in a 2x3 analysis of variance (ANOVA) with the following design: 2 levels of outcome (predicted, observed) and three levels of visual motion coherence level (20%, 60%, 100%). Furthermore, to provide insight in individual differences we compared observed and predicted response patterns per subject for the thresholds and weights. To this end we computed the 95% confidence intervals (CIs) for the unimodal and combined cue thresholds using a bootstrap procedure. We randomly sampled 150 data points per condition with replacement. We then calculated the predicted combined variance using Eq 4, the predicted single cue weight using Eq 5 and the observed single cue weights using Eq 6. By repeating this procedure 999 times we constructed the upper and lower bounds of the 95% CI of the mean. These 95% confidence intervals were used to detect individual straying from optimality.
Results
Fig 3 shows the single and combined cue results of a typical subject. These psychometric curves illustrate the proportion of ‘larger’ responses as a function of probe amplitude. The different slopes of the psychometric functions in Fig 3A, quantified by their threshold, indicate that visual cue reliability depends on visual coherence levels. From the single cue thresholds, we predicted the weights that the subject should use when combining both cues in a statistically optimal manner. Each coherence level has different predicted weights, computed from each pairing of vestibular and visual single cue thresholds using Eq 5.
Single cue and combined thresholds
Within Fig 3B–3D the predicted (dashed) and actual results are shown for all combined cue conditions; different panels are organized by the visual motion coherence levels. When visual cue reliability was low (i.e., coherence 20%, see Fig 3B), the combined cue psychometric functions during cue conflict (i.e., Δ = -5cm or Δ = +5cm) shifted according to vestibular dominance, that is, low visual weights (negatively biased red curves and positively biased blue curves). In contrast, when visual reliability was high (i.e., coherence 100%), the curves shifted based on visual dominance (i.e., higher visual weights), see Fig 3D. Overall, the observed results match the optimal predictions for all three coherence levels quite well.
Fig 4A–4F shows the individual subjects’ actual and predicted thresholds for the single cue and combined conditions with Δ = 0. Fig 4G shows the average result across subjects. As mentioned, optimal cue integration models predicts the combined threshold to be lower than the lowest single cue threshold (Eq 4). We validated this prediction: we found a significant reduction in the combined threshold as compared to the lowest single cue threshold for 100% (t(5) = 2.23, p < 0.05) and 60% (t(5) = 2.68, p < 0.05), but not for 20% (p = 0.47) of visual motion coherence. In comparing the actual and predicted thresholds in the combined condition we found no significant differences (all p > 0.3). Notably, subject 2 shows large variability in the visual threshold in the 20% visual motion coherence condition. This can possible be explained by the large variability in the responses resulting in a relatively high visual threshold and a large variance in thresholds obtained via the bootstrap procedure.
We also tested the effect of whether adding a discrepancy on the combined cue thresholds. The results show a higher threshold for the nonzero- than for the zero delta conditions for 100% (t(5) = 2.673, p < 0.05) and 60% (t(5) = 6.208, p < 0.005) of visual motion coherence, but not for 20% (p = 0.12).
In Fig 5, individual differences between the predicted (Eq 4) and observed (Eq 6) combined cue thresholds can be observed. Shown are the estimates with their 95% confidence intervals (CI’s). As can be seen, the diagonal line, representing optimality, falls within CI’s of all but two subjects in the 60% and 100% coherence conditions. Interestingly, larger deviations from optimality can be observed for the 20% coherence condition, with the CI’s of 5 subjects falling outside the diagonal.
Reliability based cue weighting
Predicted and observed weights as obtained from Eqs 5 and 6, respectively, are summarized in Fig 6. The general result is that all subjects show robust changes both in actual and predicted weights as function of visual motion coherence. Statistical test results show no significant difference between predicted and observed visual weights (p > 0.41; η² = 0.136), or interaction with motion coherence (p > 0.08; η² = 0.397). The main effect of motion coherence was significant [F(2,10) = 35.406; p < 0.001; η² = 0.867] and reflects the reliability-dependent weighting, with higher visual weighting for reliable visual stimuli and lower visual weighting for less reliable visual cues.
In analogy to the individual threshold differences, individual differences in visual weighting can be observed, as thresholds and weights are closely related, see Eq 5. In order to gain more insight in these differences we plotted the observed and predicted weights in a scatterplot, as shown by Fig 7. For the 100% and 60% visual coherence conditions the observed and predicted visual weights match closely. In analogy to the combined threshold values, for the 20% coherence condition some deviations from optimality can be observed.
Discussion
We studied the integration of visual and vestibular cues in a whole body lateral displacement discrimination task. Visual cues were presented with different levels of visual coherence (20%, 60% and 100%) to change relative cue reliability and discrepancies between visual and vestibular cues were imposed to test relative cue weighting. Our results are generally in line with two primary predictions of optimal integration schemes as derived from Bayesian inference theory: 1) human subjects reduce the variance of the combined cue percept and 2) they dynamically weight individual cues based on their reliabilities. Our results are therefore complementary to studies on multisensory integration during heading perception [2, 3, 11].
Visual-vestibular interaction
Within the literature on the integration of visual and vestibular cues during self-motion a distinction can be made between studies addressing visual-vestibular integration from a reliability-based or gain-based cue combination perspective. Studies on reliability based cue weighting use probabilistic inference to determine the relative (weighted) contribution of single cues based on their reliability. In contrast, studies on gain-based cue weighting used differences in displacement estimates to determine cue weighting.
Several heading estimation studies showed cue reliability dependent weighting of cues (e.g., [2, 3]). Our study shows dynamic cue weighting for displacement estimation. The interaction of visual and vestibular cues during heading perception is consistently characterized by an overestimation of vestibular cues [2, 11, 13]. More specifically, the observed vestibular weighting is consistently higher than the predicted vestibular weighting based on single cue reliability. In contrast, our study does not show a significant vestibular overestimation. This overrating during heading estimation has been attributed to the use of virtual reality systems [13] or explained as the result of training effects [2]. Another, but yet to be explored possibility is the involvement of causal inference, in which multisensory integration is performed in two steps [27]. The first step determines the information provided by each cue and in the second step the likelihood of two cues coming from the same source or from multiple sources is estimated. When visual and vestibular displacement cues provide conflicting information, causal inference decides whether the displacement was indicated by a single cue or both cues. In the case of too large conflicts this may lead to the observer using only one cue to perform the task. Although the involvement of causal inference has been proposed as a possible explanation for the vestibular overweighting during heading discrimination [2, 13], it is still a speculative account [28]. As our results show no differences between the predicted and observed vestibular weights, we have no reason to assume that causal inference would have resulted in cue segregation.
Studies on gain based cue integration show controversy with regard to the relative weighting of visual and vestibular cues. In an experiment in which subjects had to match a predefined walked distance by adjusting a target in the depth dimension [25] showed nearly equal weighting of visual and vestibular cues, 0.54 and 0.46, respectively. Similarly, [5] showed nearly equal weighting when asking for absolute displacement estimates in an oscillatory up-down movement with 0.35–0.42 and 0.58–0.65 for the visual and vestibular weights, respectively, depending on motion amplitude. Despite the fact that our study uses a reliability based approach for the weighting of visual and vestibular cues, our results show remarkable agreement with the findings by [25] and [5]. They exhibit, on average, a nearly equal weighting for the single cues (visual: 0.46, vestibular: 0.54) with 100% visual coherence. In contrast, the study by [4] reports a dominance of non-visual cues in displacement estimation. They used a fore-after movement and had people indicate the position of a previously perceived visual or physical stop target by a button-press when “passing” that target during passive motion. There are several possible reasons for the inconsistency between the results of [4] and those of the current study. First, it has been argued that the relative importance of visual and vestibular cues is task dependent, with neither of the cues covering all aspects of self-motion perception [29]. Differences might already occur at the sensory level as vestibular detection thresholds and neural firing patterns of primary otolith afferents differ between motion directions [30, 31]. Second, in the study by [4] the visual stimulus was presented without stereoscopic vision. It is believed that adding stereoscopic vision facilitates the integration of visual and vestibular cues by disambiguating depth percept (but see [2, 12]). Third, the distances used by [4] are many times larger (1.5 m up to 10 m) than those used by us (0.05 m up to 0.35 m). According to Weber’s Law, increases in discrimination thresholds are proportional to the pedestal and this dependency might be very different for visual and vestibular cues, which in turn may result in pedestal dependent differing cue reliabilities.
Differences between heading- and displacement discrimination
Recent literature shows that during heading perception one receives accumulating evidence for direction of travel (i.e., leftward or rightward) during self-motion and hence heading estimates’ certainty will increase over time [19]. For displacement discrimination, however, it is unlikely that evidence will increase over time. In order to estimate the travelled distance one needs to take the double integral of the acceleration profile of the entire movement. Furthermore, in a comparison between heading-discrimination- and a displacement-discrimination task it was shown that heading discrimination dependent on movement directions relative to the head, but not movement directions in world coordinates. In contrast, displacement discrimination depended on movements in world coordinates, but not movements relative to the head [9]. Therefore, the processing of sensory information for displacement discrimination during self-motion might be very different. Up to a certain level, our results are in line with the findings of reliability based weighting during heading estimation. However, during heading estimation a vestibular overweighting as been reported consistently [2, 10, 11], whereas we did not observe such an overweighting with respect to optimal integration schemes. How these differences are manifested at the neuronal level is unknown. It is likely that sets of neurons can be found that specifically code for displacement as there are neurons specifically tuned to coding heading angles [3, 17, 32].
In most experiments regarding relative cue weighting in self-motion perception there is an inherent concern that the visual only condition introduces a sensory conflict between the visual and vestibular information. That is, while observing optic flow, the visual system detects movement while the vestibular system specifies a stationary position. Consequently, the visual only condition might not be a purely unisensory condition. We expect this influence to be minor in our task, since [16] showed that in non-human primates there was no difference in visual heading estimation between labyrinthectomized- and control animals.
The results on the thresholds show that there is a significant reduction of the combined threshold with respect to the lowest single cue thresholds, but only for the 100% and 60% visual motion coherence conditions. This is in line with the theory on optimal integration in which the reduction of the combined cue threshold is strongest when the two single cues have equal thresholds. An increasing difference between the single cue thresholds, as is the case in the 20% coherence condition, will henceforth result in only a small reduction of the combined threshold with respect to the lowest single cue threshold.
Finally, our data show that adding discrepancies between the single cues increases the combined thresholds in the 100% and 60% visual motion conditions compared to the combined thresholds in congruent conditions. This is consistent with results on heading perception, showing that adding discrepancies affects the integration process of visual and vestibular cues, also when subject were not aware of any discrepancies [11].
Conclusion
In this study, we showed that the integration of visual and vestibular cues during displacement estimation is in line with two predictions of probabilistic inference theory. We showed that perceptual precision is increased when both cues are combined with respect to either of the single cues and that cue weighting is dynamically based on relative cue reliability.
Data Availability
All data are contained within the paper. Raw data are available at Dryad: doi:10.5061/dryad.v2k7s.
Funding Statement
This work is supported by grants from the European Research Council (EU-ERC 283567) and the Netherlands Organisation for Scientific Research (NWO-VICI: 453-11-001; NWOVENI: 451-10-017). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
References
- 1. Gibson JJ. The perception of visual surfaces. The American journal of psychology. 1950;63(3):367–84. Epub 1950/07/01. . [PubMed] [Google Scholar]
- 2. Fetsch CR, Turner AH, DeAngelis GC, Angelaki DE. Dynamic reweighting of visual and vestibular cues during self-motion perception. The Journal of neuroscience: the official journal of the Society for Neuroscience. 2009;29(49):15601–12. Epub 2009/12/17. 10.1523/JNEUROSCI.2574-09.2009 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3. Gu Y, Angelaki DE, Deangelis GC. Neural correlates of multisensory cue integration in macaque MSTd. Nature neuroscience. 2008;11(10):1201–10. Epub 2008/09/09. 10.1038/nn.2191 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4. Harris LR, Jenkin M, Zikovitz DC. Visual and non-visual cues in the perception of linear self-motion. Experimental brain research Experimentelle Hirnforschung Experimentation cerebrale. 2000;135(1):12–21. Epub 2000/12/05. . [DOI] [PubMed] [Google Scholar]
- 5. Wright WG, DiZio P, Lackner JR. Vertical linear self-motion perception during visual and inertial motion: more than weighted summation of sensory inputs. Journal of vestibular research: equilibrium & orientation. 2005;15(4):185–95. Epub 2005/11/16. . [PubMed] [Google Scholar]
- 6. Israel I, Berthoz A. Contribution of the otoliths to the calculation of linear displacement. Journal of neurophysiology. 1989;62(1):247–63. Epub 1989/07/01. . [DOI] [PubMed] [Google Scholar]
- 7. Israel I, Chapuis N, Glasauer S, Charade O, Berthoz A. Estimation of passive horizontal linear whole-body displacement in humans. Journal of neurophysiology. 1993;70(3):1270–3. Epub 1993/09/01. . [DOI] [PubMed] [Google Scholar]
- 8. Israel I, Grasso R, Georges-Francois P, Tsuzuku T, Berthoz A. Spatial memory and path integration studied by self-driven passive linear displacement. I. Basic properties. Journal of neurophysiology. 1997;77(6):3180–92. Epub 1997/06/01. . [DOI] [PubMed] [Google Scholar]
- 9. MacNeilage PR, Banks MS, DeAngelis GC, Angelaki DE. Vestibular Heading Discrimination and Sensitivity to Linear Acceleration in Head and World Coordinates. Journal of Neuroscience. 2010;30(27):9084–94. 10.1523/Jneurosci.1304-10.2010 ISI:000279581700007. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10. Gu Y, Watkins PV, Angelaki DE, DeAngelis GC. Visual and nonvisual contributions to three-dimensional heading selectivity in the medial superior temporal area. The Journal of neuroscience: the official journal of the Society for Neuroscience. 2006;26(1):73–85. Epub 2006/01/10. 10.1523/JNEUROSCI.2356-05.2006 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11. Butler JS, Smith ST, Campos JL, Bulthoff HH. Bayesian integration of visual and vestibular signals for heading. Journal of vision. 2010;10(11):23 Epub 2010/10/05. 10.1167/10.11.23 . [DOI] [PubMed] [Google Scholar]
- 12. Butler JS, Campos JL, Bulthoff HH, Smith ST. The role of stereo vision in visual-vestibular integration. Seeing and perceiving. 2011;24(5):453–70. Epub 2011/09/06. 10.1163/187847511X588070 . [DOI] [PubMed] [Google Scholar]
- 13. de Winkel KN, Weesie J, Werkhoven PJ, Groen EL. Integration of visual and inertial cues in perceived heading of self-motion. Journal of vision. 2010;10(12):1 Epub 2010/11/05. 10.1167/10.12.1 . [DOI] [PubMed] [Google Scholar]
- 14. Angelaki DE, Gu Y, DeAngelis GC. Multisensory integration: psychophysics, neurophysiology, and computation. Current opinion in neurobiology. 2009;19(4):452–8. Epub 2009/07/21. 10.1016/j.conb.2009.06.008 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15. Angelaki DE, Gu Y, Deangelis GC. Visual and vestibular cue integration for heading perception in extrastriate visual cortex. The Journal of physiology. 2011;589(Pt 4):825–33. Epub 2010/08/04. 10.1113/jphysiol.2010.194720 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16. Gu Y, DeAngelis GC, Angelaki DE. A functional link between area MSTd and heading perception based on vestibular signals. Nature neuroscience. 2007;10(8):1038–47. Epub 2007/07/10. 10.1038/nn1935 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17. Gu Y, Fetsch CR, Adeyemo B, Deangelis GC, Angelaki DE. Decoding of MSTd population activity accounts for variations in the precision of heading perception. Neuron. 2010;66(4):596–609. Epub 2010/06/01. 10.1016/j.neuron.2010.04.026 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18. Fetsch CR, Pouget A, DeAngelis GC, Angelaki DE. Neural correlates of reliability-based cue weighting during multisensory integration. Nature neuroscience. 2012;15(1):146–54. Epub 2011/11/22. 10.1038/nn.2983 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19. Drugowitsch J, DeAngelis GC, Klier EM, Angelaki DE, Pouget A. Optimal multisensory decision-making in a reaction-time task. eLife. 2014;3 Epub 2014/06/16. 10.7554/eLife.03005 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20. Frenz H, Lappe M. Absolute travel distance from optic flow. Vision research. 2005;45(13):1679–92. Epub 2005/03/29. 10.1016/j.visres.2004.12.019 . [DOI] [PubMed] [Google Scholar]
- 21. Bremmer F, Lappe M. The use of optical velocities for distance discrimination and reproduction during visually simulated self motion. Experimental Brain Research. 1999;127(1):33–42. 10.1007/s002210050771 ISI:000081095400004. [DOI] [PubMed] [Google Scholar]
- 22. Redlick FP, Jenkin M, Harris LR. Humans can use optic flow to estimate distance of travel. Vision research. 2001;41(2):213–9. Epub 2001/02/13. . [DOI] [PubMed] [Google Scholar]
- 23. Sun HJ, Campos JL, Chan GS. Multisensory integration in the estimation of relative path length. Experimental brain research Experimentelle Hirnforschung Experimentation cerebrale. 2004;154(2):246–54. Epub 2003/12/20. 10.1007/s00221-003-1652-9 . [DOI] [PubMed] [Google Scholar]
- 24. Berthoz A, Israel I, Georges-Francois P, Grasso R, Tsuzuku T. Spatial memory of body linear displacement: what is being stored? Science. 1995;269(5220):95–8. Epub 1995/07/07. . [DOI] [PubMed] [Google Scholar]
- 25. Campos JL, Butler JS, Bulthoff HH. Multisensory integration in the estimation of walked distances. Experimental brain research Experimentelle Hirnforschung Experimentation cerebrale. 2012;218(4):551–65. Epub 2012/03/14. 10.1007/s00221-012-3048-1 . [DOI] [PubMed] [Google Scholar]
- 26. Kontsevich LL, Tyler CW. Bayesian adaptive estimation of psychometric slope and threshold. Vision research. 1999;39(16):2729–37. Epub 1999/09/24. . [DOI] [PubMed] [Google Scholar]
- 27. Kording KP, Beierholm U, Ma WJ, Quartz S, Tenenbaum JB, Shams L. Causal Inference in Multisensory Perception. PloS one. 2007;2(9). ARTN e943 10.1371/journal.pone.0000943 ISI:000207455800022. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28. Fetsch CR, Gu Y, DeAngelis GC, Angelaki DE. Self-Motion Perception: Multisensory Integration in Extrastriate Visual Cortex In: Trommershäuser J, Körding K, Landy MS, editors. Sensory Cue Integration. Computational Neuroscience. New York: Oxford University Press; 2011. [Google Scholar]
- 29. Campos JL, Bulthoff HH. Multimodal integration during Self-Motion in Virtual Reality In: Simon SA, Nicolelis MAL, editors. The neural bases of multisensory processes. Frontiers in neuroscience. Boca Raton: CRC Press; 2012. [Google Scholar]
- 30. Angelaki DE, Dickman JD. Spatiotemporal processing of linear acceleration: primary afferent and central vestibular neuron responses. Journal of neurophysiology. 2000;84(4):2113–32. Epub 2000/10/12. . [DOI] [PubMed] [Google Scholar]
- 31. Kingma H. Thresholds for perception of direction of linear acceleration as a possible evaluation of the otolith function. BMC ear, nose, and throat disorders. 2005;5(1):5 Epub 2005/06/24. 10.1186/1472-6815-5-5 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 32. Angelaki DE, Cullen KE. Vestibular system: the many facets of a multimodal sense. Annual review of neuroscience. 2008;31:125–50. Epub 2008/03/15. 10.1146/annurev.neuro.31.060407.125555 . [DOI] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
All data are contained within the paper. Raw data are available at Dryad: doi:10.5061/dryad.v2k7s.