Abstract
Neurological evidence suggests that disturbed vestibular processing may play a key role in triggering out-of-body experiences (OBEs). Little is known about the brain mechanisms during such pathological conditions, despite recent experimental evidence that the scientific study of such experiences may facilitate the development of neurobiological models of a crucial aspect of self-consciousness: embodied self-location. Here we apply Bayesian modeling to vestibular processing and show that OBEs and the reported illusory changes of self-location and translation can be explained as the result of a mislead Bayesian inference, in the sense that ambiguous bottom-up signals from the vestibular otholiths in the supine body position are integrated with a top-down prior for the upright body position, which we measure during natural head movements. Our findings have relevance for self-location and translation under normal conditions and suggest novel ways to induce and study experimentally both aspects of bodily self-consciousness in healthy subjects.
Keywords: self, body, out-of-body experience, vestibular, Bayesian model, self-motion, uncertainty, illusion
Introduction
Out-of-body experiences (OBEs) are illusions, where people experience themselves as being located outside their physical body (disembodied self-location) and often report sensations of flying and to see the world from an elevated perspective. Investigating such neurological conditions is a promising approach to study the neuronal basis of the bodily self and might facilitate the development of neurobiological models of bodily self-consciousness (Ehrsson, 2007; Lenggenhager et al., 2007; Metzinger, 2008; Vogeley and Fink, 2003; Vogeley et al., 2004). This has been demonstrated in recent experiments where key aspects of the bodily self (self-location and self-identification) have been manipulated experimentally (Ehrsson, 2007; Lenggenhager et al., 2007; Mizumoto and Ishikawa, 2005), but further conceptual advances depend on explicitly modeling aspects of the bodily self.
For almost every sensory system, the bottom-up signals from the sensory periphery are ambiguous and need to be disambiguated based on previous experience (see Poggio et al., 1985 for a prominent example in vision). Such a disambiguation could be done in computationally different ways. For example, it could be computed close to the sensory periphery, where previous experience may have shaped neuronal circuits such that the interpretation of the ambiguous sensory signals is always biased towards the same most plausible interpretation. Other disambiguations have to be flexible if they depend on the particular context via, for example, signals from other modalities or the recent history of sensory stimulation. Top-down signals, which could change more rapidly than less adaptive processes closer to the sensory periphery, are a possible mechanism for the latter kind of computation.
The signals from the otholithic vestibular system are inherently ambiguous, because the otholiths cannot distinguish between gravity and acceleration. Since subjects are usually moving in their environment, the brain has to continuously disambiguate these signals. The way the brain performs this computation is currently not completely known. In order to estimate self-motion, however, the brain probably integrates vestibular bottom-up and top-down signals derived from, for example, other gravitoreceptors in the body, from the visual system (MacNeilage et al., 2007), and at a subcortical level probably also from the semicircular canals (Angelaki et al., 2001). The Bayesian approach is a natural framework to model this integration and has become a widely accepted analogy to information processing in the brain (Friston, 2005; Knill and Richards, 1996; Pouget et al., 2003; Rao and Ballard, 1999) as it is supported by experimental (Ernst and Banks, 2002; Körding and Wolpert, 2004; MacNeilage et al., 2007; van Beers et al., 1996) and theoretical studies relating it to neuronal activations (Denève et al., 2007; Huys et al., 2007; Ma et al., 2006; Rao, 2004; Sahani and Dayan, 2003).
Here we apply the Bayesian approach to model the brain's processing of vestibular signals (see also Laurens and Droulez, 2007) as such processing has been proposed to encode not just location and movements of one's body, but also aspects of the self such as self-location and the first-person perspective. Neurological evidence showed that the occurrence of OBEs in neurological patients is associated with disturbed otholithic processing (Blanke and Mohr, 2005). The experienced spatial location of the self (self-location) is a key aspect of (self) consciousness (Blanke et al., 2004), and it likely depends on self-motion signals like the vestibular signals via, for example, path integration mechanisms (McNaughton et al., 2006). Recent behavioural and neuroimaging research has utilized the illusory characteristics of disembodied self-location during OBE to design experiments for the study of embodied self-location under normal conditions (Arzy et al., 2006; Blanke et al., 2005; Ehrsson, 2007; Lenggenhager et al., 2007). Using virtual reality, these studies showed that self-location can be manipulated systematically by multisensory mismatch of bodily information (Ehrsson, 2007; Lenggenhager et al., 2007) and that self-location depends on temporoparietal activity (Arzy et al., 2006) compatible with neurological data (Blanke et al., 2004; De Ridder et al., 2007; Devinsky et al., 1989; Maillard et al., 2004). Here, we propose that OBEs and associated illusory changes in self-location can be explained as the result of a mislead Bayesian inference, in the sense that the ambiguous bottom-up signals from the vestibular otholiths in the supine body position are integrated with a top-down prior for the upright body position, which is not appropriate for the current (supine) body position.
We then show that the variances of the top-down priors for the head acceleration and pitch rotation are crucially affecting an inferred illusory self-translation, which here we identify with pathological self-location. We measure these variances via analyzing natural head movements and find that illusory self-translations are predicted especially for subjects engaged in rapid sport-like activities, but less so for slowly walking or stationary subjects. Collectively, our results suggest a solid interpretation for pathological self-location and other vestibular illusions during OBEs. These findings also have relevance for embodied self-location under normal conditions and suggest novel ways to induce and study experimentally this crucial aspect of bodily self-consciousness in healthy subjects.
Materials and Methods
In order to obtain quantitative estimates of the top-down prior information for the upright position, we instructed nine subjects (seven male, age 29 ± 4) to perform three kinds of natural movements for a duration of 4–5 min each. In the standing condition, the subjects were standing and moved only their head and partly the upper torso in order to inspect all locations surrounding them. In the waiting condition, subjects walked continuously around the capture area (3 m × 3 m) as if they were waiting and looking for someone on a plaza. In the action condition, subjects were instructed to move as if they were playing tennis, such that they serve, play with the forehand and backhand, make fast movements to the net, etc.
Subjects were wearing a headband, on which we mounted n = 4 infrared markers. Each condition started with a period of 5 s, during which the subjects were standing and looking straight ahead. The time-averaged three-dimensional coordinates of the markers recorded during this period were then used as the base pose xi, i = 1,…, n. The base pose was translated such that where we set h = 15 cm for all subjects. In other words, the base pose was shifted to the origin in the horizontal plane and 15 cm along the vertical axis. This way, rotations and translations of the base pose are relative to the approximate center of the head. We also used larger and smaller values of h, but the measured statistics did not depend strongly on these choices of h. As a recording device we used a ReActor 2 motion capture system (Ascension Technology) operating at a sampling rate of 30 Hz.
Using this system, we measured for each time the translation and rotation of the recorded marker position relative to the base pose. While a few commercial products are available for this computations, we decided to implement our own postprocessing, since the implementation details are often only poorly documented. Details of our prostprocessing, which intentionally does not apply any temporal smoothing, are documented in the Supplementary Material.
Results
The otholiths as gravitoinertial sensors
We model the afferent signals as if coming from a single device sensing the gravitoinertial acceleration f = g −a , where g = (0,−9.81 m/s2,0)T is the gravitational vector and a is the acceleration of the head, i.e. −a is the inertial acceleration. Each of the vectors f, g and a are given in world-coordinates. In order to obtain the gravitoinertial acceleration x sensed by the otholiths, f needs to be transformed into the coordinate system of the head, which we assume to be a rotated version of the world-coordinate system. This transformation depends on the orientation of the head, which we always describe in terms of yaw, pitch and roll angles (see Materials and Methods for details). Let the rotation of the head be given by a pure rotation matrix R, such that R−1 = RT. Then, with
(1) |
one obtains the sensed gravitoinertial acceleration x in head-coordinates. With ahd = RTa we refer to the acceleration of the head in head-centered coordinates. The conventions for the coordinate system and rotation angles are shown in Figure 1A.
The same sensed gravitoinertial acceleration could be produced by a multitude of head rotations and accelerations. Figures 1B–D illustrate such ambiguities. For example, based only on the otholith signals the brain cannot distinguish between a stationary backward tilt and an upright body position with a forward acceleration (Figure 1B). This ambiguity plays a key role in our explanation of illusory self-motion as reported in OBEs. If subjects are in a supine body position, but they internally assume to be upright, then the brain's “explanation” for the sensory-bottom up signals consistent with the upright body position is a forward acceleration. Figures 1C,D illustrate two other ambiguities for the stationary upright body position.
The effect of a non-appropriate upright prior – model
We consider the acceleration of the head, ahd, in the head-centered coordinate system and the rotation of the head, R, relative to gravity as the two relevant so-called hidden variables, which the brain may use to account for the sensed gravitoinertial acceleration. Given a particular sensed gravitoinertial acceleration x, the posterior for the two hidden variables is obtained by Bayes' rule as
(2) |
where
(3) |
is the likelihood, which describes the way the two hidden variables give rise to a sensed gravitoinertial acceleration x. In other words, it is a forward model of the sensory signals, where we assume that these signals are perturbed by additive Gaussian noise. Here, N(μ, σ2)(x) denotes a Gaussian probability density with mean μ and covariance matrix σ2I evaluated at x. P(ahd, R) is the top-down prior assumed before any sensory signals are observed, and P(x) is the overall probability of x. When comparing different possible explanations of a particular value of x (in terms of values for ahd and R), P(x) becomes a constant factor not affecting such a comparison.
We now consider the case, where ahd and x are vectors in the xy-plane (see Figure 1A), and the rotation matrix implementing only pitch rotations. We first set to the stimulation corresponding to a stationary upright position. In this case, the otholith signals do not distinguish between the correct stationary upright position and other alternatives (Figures 1C,D).
For such a sensory stimulus, the likelihood for the accelerations in the x- and y-direction in head-centered coordinates are shown in the first two panels of Figure 2A. In the first panel, the likelihood is plotted for no acceleration in the y-direction. The combination of the true values with no pitch, ϕ = 0, and no acceleration, , has the highest likelihood. However, high values are also assigned to combinations with ϕ < 0 and (forward tilt and forward acceleration) and ϕ > 0 and (backward tilt and backward acceleration). The second panel shows the likelihood with and varying acceleration in the y-direction. Again, high values are assigned to the true parameter combination with ϕ = 0 and . However, an upside-down rotation of the head, ϕ = ±π, with an acceleration of is also consistent with the stationary upright position and no forward acceleration.
The last three panels in Figure 2A show the posterior for the two hidden variables, where we have integrated over , and ϕ, respectively. In other words, the last three panels are different views of the information one obtains about the state of the head in terms of the pitch ϕ and acceleration ahd, if the sensory data x is combined with the top-down prior. Here, the posterior assigns a high value to the true parameter combination. This is the expected result, because we set the sensory data x to a value corresponding to a stationary upright position, and the top-down prior implemented the assumption of a stationary upright position. In this scenario, there is no conflict between the sensory data and the top-down prior.
Let us now consider the case of a sensory stimulation corresponding to a supine position, where we set with Figure 2B shows the likelihoods, priors and posteriors. Again, the highest values for the likelihoods (first two panels in Figure 2B) are for the parameter combination we actually used to compute the sensory stimulation x. However, due to the otholith ambiguity, other parameter combinations are also consistent with this sensory stimulation. With no acceleration in the y-direction, the face-down position with a forward acceleration of is consistent with the stationary supine position (first panel in Figure 2B). For no acceleration in the x-direction, positions with an even further tilt backwards and positive accelerations as well as less tilt backwards and negative accelerations are also consistent with the stationary supine position (second panel in Figure 2B). Now, however, there is a conflict between this bottom-up sensory information and the top-down prior, which we assume to implement the assumption of an upright position. How do they interact according to Bayesian inference and what are the predicted perceptual consequences in the sense of the posteriors?
The last three panels in Figure 2B show the obtained posteriors. Two features are worth noting. First, the pitch rotations with the highest posterior probabilities are closer to the value ϕ = 0 for the upright position. Second, non-vanishing accelerations in the x- and y-directions are predicted. In other words, combining the ambiguous otholith signals in a supine position with a prior for the upright position leads to an inferred state of the head with a forward and downward acceleration in head-coordinates and less tilt backward than the supine position.
In our model, this prediction is a direct consequence of the multiplication of two Gaussians. Let and be two Gaussians for some random variable x with diagonal covariance matrices and . Then, their product is given as
(4) |
which is also a Gaussian, but with the new mean computed as the weighted sum of the means of the individual Gaussians. In particular, the variance of the first Gaussian, normalized by becomes the weight for the mean μ2 of the second Gaussian and vice versa. This basic relation is the key property of Bayesian models of multisensory integration (Alais and Burr, 2004; Battaglia et al., 2003; Ernst and Banks, 2002). In such models, Eq. 4 describes the posterior (assuming a non-informative prior) over a state variable x of interest (like the location of a target), and N1 and N2 describe how x is inferred via the individual modalities (like vision and audition). Equation 4 shows how the reliability of the individual modalities, as quantified by their variances, affects the mean of the posterior distribution. If one modality becomes less reliable (the variance increases), then the mean of the posterior is shifted towards the mean predicted by the other modality. The same relation, however, also dictates how a Gaussian likelihood is combined with a Gaussian prior within a single modality, which is the case we are considering.
For the sake of illustration, let us consider the Bayesian inference of only the head acceleration ahd with the true pitch ϕ* used to compute a sensory stimulation and the internally assumed pitch ϕ as parameters, and let us assume that the mean of the acceleration prior is zero, μacc = 0. Then, the posterior for the head acceleration ahd is given as
(5) |
If ϕ* = ϕ, the mean of the head acceleration posterior is zero, independent of the values of and . In other words, if the internally assumed pitch ϕ and the true pitch ϕ* coincide, then the inferred head acceleration in terms of the mean of the posterior is zero (Figure 2A). If ϕ ≠ ϕ*, however, then the mean of the posterior is a weighted sum (see Eq. 4) of μacc = 0 and Here, large values of bias the solution of the otholithic ambiguity towards the zero head acceleration with μacc = 0. Conversely, large values of bias the solution of this ambiguity towards the non-zero head acceleration which can account for the mismatch between the internally assumed and the true pitch.
This non-zero “compensatory acceleration” is shown in Figure 3A in the xy-plane for the supine body position (the true pitch ϕ*), but different internally assumed pitch rotations. If ϕ* = ϕ (“face up” in Figure 3A), then no compensatory acceleration is needed to account for the sensory signals, but if the internally assumed pitch is the upright position (“upright” in Figure 3A), the brain has to assume a non-zero acceleration in order to account for the bottom-up signals in the supine body position. The absolute value of this compensatory acceleration is shown in Figure 3B for different internally assumed pitch rotations.
In order to compute the full Bayesian acceleration posteriors (last panels in Figures 2A,B) with the internally assumed pitch ϕ being also a random variable, one evaluates
(6) |
Now both the acceleration and the pitch are the hidden state variables of interest, which we assumed to be independent, i.e. P(ahd, ϕ) = P(ahd)P(ϕ). Interestingly, we find that the variance σpitch of the pitch prior P(ϕ) affects the strength of the inferred illusory self-translation, because the absolute value of the compensatory self-translation (Figure 3B) depends in a non-linear way on the internally assumed pitch. In other words, P(ϕ) averages over a non-linear function (the compensatory acceleration for a given sensory stimulation x), and the range over which this averaging is performed affects its result. Figure 3C shows that for large σpitch the absolute value of the illusory self-translation becomes smaller, because at the internally assumed upright position, ϕ = 0, the absolute value of the compensatory acceleration decays more rapidly towards the supine position than it increases towards further tilts forward. Figure 3C also shows that for larger σacc, or conversely for smaller σn, the inferred illusory self-translation increases.
The effect of a non-appropriate upright prior – predictions
In the last section we have demonstrated, using a toy model of otholithic processing, that an otherwise optimal processing strategy can lead to illusory perceptions, when the internally assumed upright body position differs from the actual body position. In particular, if the actual body position is supine, but the internally assumed position is upright, then the brain may explain this discrepancy in terms of an illusory self-translation, which we predict to be forward and downward in head-centered coordinates. Moreover, we have derived two additional testable dependencies of the magnitude of the illusion. We predict that the illusion is stronger, if the subjects are more uncertain about their acceleration (larger values of σacc, Figure 3C), and that the uncertainty regarding the pitch of the head should be minimized (small values of σpitch, Figure 3C) for stronger illusions. While testing these predictions is a methodological challenge (see Discussion), a valuable first step is a characterization of the range these two values take in everyday life. We did this by measuring the head movements of subjects using a motion capture system.
The effect of the measured upright priors
We assume that the brain may have inferred the priors based on stimuli it is exposed to during everyday life. These priors certainly change on different time-scales and depend on the actual context. To the best of our knowledge, however, no statistics of natural head movements are reported in the literature. In order to obtain a first approximation to these statistics, we measured the head movements of subjects in the three different conditions standing, waiting and action (see Materials and Methods). The results are shown in Figures 4A–D as histograms and Gaussian fits to the histograms. While the measured σpitch is similar for the waiting and action conditions, it is larger for the standing condition (Figure 4A), where standing subjects moved their head in order to inspect their surrounding. The measured σacc is largest for the action, smaller for the waiting and even smaller for the standing condition (Figures 4B–D). Hence, given the dependence of the illusory self-translation on σpitch and σacc, it will be largest for the action condition. Figures 4E–G show that this is indeed the case. Interestingly, Figures 4E–G show that in the waiting condition almost no illusory self-motion is predicted whereas in the action condition the larger values of σacc lead to a notable self-translation of approx. 1/2 g in magnitude. This dependence of the magnitude of the illusory self-translation in the supine position on the variance of the acceleration prior may also underlie the everyday experience of some vestibular aftereffects like, for example, illusory self-motion when lying in bed after a day in the theme park or after a day of skiing.
Discussion
The main result of this paper is the interpretation of the vestibular component of OBEs as a mislead Bayesian inference and that this accounts for disembodied and elevated self-location. We have shown that a mismatch between the actual body position and an internally assumed upright body position can lead to illusory self-translations due to the otholithic ambiguity. While this ambiguity has been the subject of previous studies (see Angelaki et al., 2001 for a review), we have shown that illusory self-motions are also crucially affected by the variances of the accelerations and pitch priors. Since most previous studies have not considered these variances as candidates for the independent variable in experimental designs, our model makes strong testable predictions for future behavioral experiments with healthy subjects. Moreover, we have measured the top-down priors by analyzing natural head movements and determined the approximate values of these variances. Together with our model this data suggests that subjects who use a top-down acceleration prior corresponding to, for example, rapid sport-like movements may experience illusory self-translation in a supine body position.
Based on neurological observations, it has been proposed that the vestibular component of OBEs may both induce and allow distinguishing OBEs from related bodily illusions. OBEs have been induced by electrical stimulation of vestibular cortex (Blanke et al., 2002; Penfield et al., 1955) and share many similarities with bodily illusions of healthy subjects in microgravity conditions such as orbital and parabolic flights (Kornilova, 1997; Lackner, 1992). Hence, in terms of aspects relevant to the study of the bodily self, the present study suggests that using a top-down prior reflecting an upright body position in order to disambiguate otholith signals in a supine body position may account for the following key elements of an OBE. Disembodied self-location would be due to illusory self-translation affecting self-location by means of, for example, path integration mechanisms (McNaughton et al., 2006), while the actual body position is static. Elevation of self-location would be due to the illusory upright body position while the actual body is in a supine position. However, it could also be a consequence of the illusory self-translation in the forward direction. The vestibular sensations of floating and flying would be a direct consequence of the predicted illusory self-translation.
We argue that the concept of combining empirically measured statistics of vestibular (and other sensory) signals with computational modeling is an important step towards the systematic and detailed description of how the brain computes a phenomenon like self-location by the integration of basic bodily sensory signals.
Signals from the semicircular canals and other modalities
An apparently strong limitation of our model is its focus on only the otholithic signals. We do not model the information from the semicircular canals, which signal angular acceleration and help to disambiguate the otholithic signals (Angelaki et al., 2001). However, we are considering a scenario with a mismatch between the actual static supine position and an internally assumed upright position. The inferred orientation as described by the Bayesian posterior is almost upright as well. Hence, the semicircular canals are neither actually contributing sensory signals, nor would there be any further discrepancy between the inferred (almost) upright position and the upright position imposed by the prior in terms of canal signals. However, a more complete model would certainly have to account for the signals form the semicircular canals and how they interact with the otholithic signals. Moreover, an apparently arbitrary choice we made is setting the noise for the otholithic signals to This sensory noise can be measured experimentally, but estimates for the detection threshold of movement onset vary widely depending on the method of measurement (Korhuber, 1974). If our explanation of the vestibular component in OBEs is correct, we have to assume low sensory noise. Hence, we selected a value for the sensory noise, which corresponds to the upper end of measured range of detection threshold and is a conservative estimate.
In normal life, however, being in a supine position and imagining oneself standing does not induce OBE-like experiences. This is certainly due to the fact that the top-down priors used to disambiguate the otholithic signals (beside their subcortical interactions with the canal signals) change on many different time-scales and depend on the context. For example, visual signals (like optic flow) and proprioceptive signals (like neck muscle activity to sustain the head against gravity) are certainly used by the brain in order to infer self-location and self-translation. Moreover, when healthy subjects are lying awake in a supine position, they usually know that they are lying, which affects their interpretation of the sensory signals. Hence, our assumption of an upright prior in a supine position is a simplification, which may only be appropriate when signals from other modalities are attenuated. We hypothesize that this is the case in some neurological conditions leading to OBEs, but it may also be the case in dreams or transitions from wakefulness to sleep. In other words, in OBEs the actual processing of sensory and top-down information would still be intact (or even optimal in the Bayesian sense), but “wrong” inputs to an otherwise optimal processing could lead to illusions such as OBEs as suggested by the induction of OBE-like experiences via multisensory stimulation (Ehrsson, 2007; Lenggenhager et al., 2007). It is a methodological challenge to render other modalities non-informative when experimentally testing our model with healthy subjects.
The analogy of bayesian inference
Bayesian inference is a mathematical theory rooted in logic and statistics (Cox, 1961; Jaynes, 2003), which deserves justification when used to model brain functions (Knill and Richards, 1996). We selected it, because the Bayesian approach formalizes a task the brain has to solve during processing of sensory information and it provides a solution to it. The task is to infer the state of the body (or other objects in the world) based on ambiguous and noisy sensory signals in terms of values of hidden variables like the orientation of the head relative to gravity, which cannot be measured directly. Bayesian inference is the solution to this task, because within such a setting, where information can have different degrees of certainty, it can be derived as the optimal way of combining prior information with new (sensory) information (Cox, 1961). Behavioral experiments have shown that in some sensory (Ernst and Banks, 2002; MacNeilage et al., 2007) and sensory-motor (Körding and Wolpert, 2004; van Beers et al., 1999) tasks the human performance is well described by the Bayesian posterior. In order to link this approach to OBEs we assumed that the posterior distributions are a proper description of subjects' experiences. Sharing this assumption, a recent Bayesian model of vestibular processing (Laurens and Droulez, 2007) explains a set of other vestibular illusions.
Linking our approach to brain activations is also possible, but at this point it is hindered for two reasons. First, although neuronal implementations of Bayesian computations have been suggested (Denève et al., 2007; Huys et al., 2007; Ma et al., 2006; Rao, 2004; Sahani and Dayan, 2003), the mere similarity of the performance or perception of subjects to Bayesian inference does not imply that the brain indeed performs computations analogous to Bayesian inference. Second, the vestibular system is a distributed processing network (Guldin and Grüsser, 1998) whose functions and patterns of activation have not been described in as much detail as those of other sensory systems (Berthoz, 1996; Brandt and Dieterich, 1999). This would complicate the interpretation of neuroimaging results using Bayesian modeling. The Bayesian approach, however, leads to clearly testable predictions for behavioral experiments. For example, illusory translations should occur if the top-down prior is manipulated such that it reflects an upright position (i.e. by manipulating visual inputs) and while subjects are in a supine body position, but such experiments may turn out to be challenging methodologically (see above).
Previous work on OBEs and self-location
Strictly speaking, our model may be described only as a model of a vestibular illusion. But as its predictions are consistent with reported vestibular sensations during OBEs we expect merit for studying the bodily self under normal conditions. Possible conceptual advances could derive from our explicit postulate of identifying the Bayesian posterior with the experience of the subject as compared to a description of subjects' behavioral performance. Consequently, here we suggest that the bodily self is a statistical model of body-related multisensory signals with the vestibular modality playing a key role, because it delivers information about the whole body in space. Does this perspective help to interpret previous empirical findings regarding the computational role of particular brain regions? Direct evidence for the causal implication of temporal-parietal cortex in OBEs came from the induction of OBEs by electrical stimulation of this area (Blanke et al., 2002; De Ridder et al., 2007; Penfield et al., 1955) as well as brain damage to this area (Blanke et al., 2004). These neurologically induced OBEs were characterized by disembodied and elevated self-location and first-person perspective as well as vestibular sensations. It is also activated when employing mental imagery using self-location (Arzy et al., 2006; Blanke et al., 2005) and mental perspective taking (Vogeley et al., 2004). Hence, this brain region could be part of a (statistical) model of the bodily self. In other words, it may serve as an area representing aspects of the bodily self like self-location and self-translation relative to gravity, and its involvement in visuo-spatial perspective taking could reflect its role as the source for imagined self-motion signals.
Recently, it has been shown that self-location can be manipulated experimentally in healthy subjects by creating conflicts between multisensory bodily signals using virtual reality (Ehrsson, 2007; Lenggenhager et al., 2007). Extending a similar paradigm about the perception and integration of multisensory arm signals, the so-called “rubber hand illusion” (Botvinick and Cohen, 1998; Tsakiris and Haggard, 2005), to the entire body, changes in self-location and self-identification could be induced. In Lenggenhager et al. (2007) subjects saw their own body (virtual body) in 3D through a head-mounted display (HMD) as if standing 2 m in front of them. They saw their virtual body being stroked synchronously or asynchronously with respect to a felt stroking on their back. This stimulation led to visual capture and systematic errors in self-location as the subjects localized themselves as drifted towards the virtual body in the synchronous condition but not in the asynchronous condition. Although, none of the subjects reported disembodied self-location (as in OBEs), these data suggest that participants localized their bodily self outside their corporeal borders. This was corroborated by participants' heightened self-identification with the virtual body and self-attribution of the visual stimuli applied to the “skin” of the virtual body. These behavioural data as well as the present data suggest that self-location can be manipulated relatively easily using conflicting sensory stimulation. Hence, online processing of body-related multisensory information in the brain is more like ongoing puzzle solving of which the normally experienced embodied self-location is just a fragile and only temporarily stable solution, which is a setting that is naturally suited for the Bayesian approach to sensory information processing. However, in order to model the experimentally identified aspects of self-location one has to refer to particular coordinate systems. In our model, we assumed an body-centered coordinate system and identified the illusory self-motion with altered self-location. More precisely, we suggest that the illusory body-centered self-motion signals are used in predictive models (Kilner et al., 2007; Rao and Ballard, 1999; Wolpert et al., 1995) for body- and world-related sensory signals. Since the physical body is not actually moving, this leads to prediction errors in other modalities, and those may correspond to other aspects of multisensory OBEs. By explaining disembodied self-location as illusory self-translation while the actual body position is static and elevated self-location during OBEs due to the illusory upright body position while the actual body is in a supine position our results suggest a solid interpretation for pathological self-location during OBEs. These findings are likely to have relevance for the continuously calculated embodied self-location and suggest novel ways to induce and study experimentally this crucial aspect of bodily self-consciousness under normal conditions.
Conflict of Interest Statement
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Supplementary Material
Pose determination
We have computed the statistics of the natural head movements in terms of motion parameters that the brain may use in order to estimate the orientation and movement of the body in space. In particular, we were interested in the statistics of the accelerations in head-centered coordinates, ahd, and the angles characterizing the rotation matrix R. However, these motion parameters could not be measured directly, but had to be estimated based on the three-dimensional Cartesian coordinates of the markers we attached to the head of our subjects. In order to estimate the translation and rotation of the head relative to the base pose, we solved the so-called pose determination problem by means of a least-squares estimate: Given a base pose xi, i = 1,…,n, and another set of points yi, i = 1,…,n, for the locations of the n markers at some other time during the movement, we seek a rotation matrix R and a translation vector r such that
(7) |
is minimized. Let and It has been shown (Umeyama, 1991) that
(8) |
is a unique solution to Eq. 7, with being a singular value decomposition of the covariance matrix
(9) |
The diagonal matrix S in Eq. 8 is defined as
(10) |
for rank(Σxy) = m − 1 and
(11) |
for rank(Σxy) ≥ m − 1, where m = 3 is the number of dimensions of the points xi and yi. In this way, we obtained for each time t during the movement a rotation matrix Rt and a translation rt by transforming the base pose to the measured coordinates of the markers at time t.
Head acceleration and orientation
Given a sequence of rotation matrices Rt and translation vectors rt, we computed (1) the accelerations of the head in head-centered coordinates, ahd, and (2) the orientation of the head in terms of yaw, pitch, and roll angles. The accelerations are obtained by first computing the accelerations in world-coordinates based on the translation vectors rt. Then, these accelerations were transformed into the head-centered coordinate system.
The head orientations were characterized using yaw, pitch and roll rotations as follows: We denote the angles for the yaw, pitch and roll rotations as θ, ϕ and ψ, respectively. The corresponding rotation matrices for these individual rotations are given as
where we used shorthand notations cθ for cos (θ), sθ for sin (θ), etc. A rotation matrix R = (RθRϕ)Rψ given as
(12) |
corresponds to a sequence of rotations, where we first carry out the yaw rotation in world coordinates, then the pitch rotation in the rotated coordinate system, and finally the roll rotation in the coordinate system already rotated by yaw and pitch rotations.
Now, for each rotation matrix R estimated using the least-squares estimate, we computed the values of θ, ϕ, and ψ. For |1 − R21| < 0.001 they are given as θ = arctan2(R13, R33), and ψ = 0. For |−1 − R21| < 0.001 the values are θ = arctan2(R13, R33), and ψ = 0. Otherwise, they are given by
(13) |
Here, arctan2 (y, x) is the four-quadrant inverse tangents. The case distinction was introduced in order to account for the singularities at the north and south pole. In other words, head orientations with a pitch of are described as only yaw and pitch, but no roll rotations.
Gaussian fits to the histograms
As we were mainly interested in the variances of the priors, we subtracted the mean for the pitch angles and accelerations for each subjects before combining the histograms. Given histograms computed over all subjects, we derived Gaussian fits to these histograms by minimizing the Kullback–Leibler divergence (Cover and Thomas, 2006) between the Gaussians fits and the histograms. The results of these fits are given in Table S1.
Table S1.
Condition | Standing | Waiting | Action |
---|---|---|---|
μϕ | <0.01 | <0.01 | 0.02 |
σϕ | 0.41 | 0.28 | 0.21 |
μx | <0.01 | <0.01 | 0.11 |
σx | 0.39 | 1.03 | 2.25 |
μy | <0.01 | −0.03 | 0.01 |
σy | 0.16 | 0.6 | 2.4 |
μz | <0.01 | <0.01 | <0.01 |
σz | 0.35 | 1.08 | 2.27 |
Acknowledgments
Supported by the Fondation de Famille Sandoz and DFG (803371). The authors like to thank Henning Sprekeler for carefully reading the manuscript and helpful comments.
References
- Alais D., Burr D. (2004). The ventriloquist effect results from near-optimal bimodal integration. Curr. Biol. 14, 257–262 [DOI] [PubMed] [Google Scholar]
- Angelaki D. E., Wei M., Merfeld D. M. (2001). Vestibular discrimination of gravity and translational acceleration. Ann. N. Y. Acad. Sci. 942, 114–127 [DOI] [PubMed] [Google Scholar]
- Arzy S., Thut G., Mohr C., Michel C., Blanke O. (2006). Neural basis of embodiment: distinct contributions of temporoparietal junction and extrastriate body area. J. Neurosci. 26, 8074–8081 10.1523/JNEUROSCI.0745-06.2006 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Battaglia P. W., Jacobs R. A., Aslin R. N. (2003). Bayesian integration of visual and auditory signals for spatial localization. J. Opt. Soc. Am. A Opt. Image Sci. Vis. 20, 1391–1397 10.1364/JOSAA.20.001391 [DOI] [PubMed] [Google Scholar]
- Berthoz A. (1996). How does the cerebral cortex process and utilize vestibular signals? In Disorders of the Vestibular System, Baloh R., Halmagyi G. M., eds (New York, Oxford University Press; ), pp. 113–125 [Google Scholar]
- Blanke O., Landis T., Spinelli L., Seeck M. (2004). Out-of-body experience and autoscopy of neurological origin. Brain 127, 243–258 10.1093/brain/awh040 [DOI] [PubMed] [Google Scholar]
- Blanke O., Mohr C. (2005). Out-of-body experience, heautoscopy, and autoscopic hallucination of neurological origin Implications for neurocognitive mechanisms of corporeal awareness and self-consciousness. Brain Res. Brain Res. Rev. 50, 184–199 10.1016/j.brainresrev.2005.05.008 [DOI] [PubMed] [Google Scholar]
- Blanke O., Mohr C., Michel C., Pascual-Leone A., Brugger P., Seeck M., Landis T., Thut G. (2005). Linking out-of-body experience and self processing to mental own-body imagery at the temporoparietal junction. J. Neurosci. 25, 550–557 10.1523/JNEUROSCI.2612-04.2005 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Blanke O., Ortigue S., Landis T., Seeck M. (2002). Stimulating illusory own-body perceptions. Nature 419, 269–270 10.1038/419269a [DOI] [PubMed] [Google Scholar]
- Botvinick M., Cohen J. D. (1998). Rubber hands ‘feel’ touch that eyes see. Nature 391, 756. 10.1038/35784 [DOI] [PubMed] [Google Scholar]
- Brandt T., Dieterich M. (1999). The vestibular cortex. Its locations, functions, and disorders. Ann. N. Y. Acad. Sci. 871, 293–312 10.1111/j.1749-6632.1999.tb09193.x [DOI] [PubMed] [Google Scholar]
- Cover T. M., Thomas J. A. (2006). Elements of Information Theory. New Jersey, Wiley and Sons [Google Scholar]
- Cox R. T. (1961). The Algebra of Probable Inference. Baltimore, The Johns Hopkins Press [Google Scholar]
- De Ridder D., Van Laere K., Dupont P., Menovsky T., Van de Heyning P. (2007). Visualizing out-of-body experience in the brain. N. Engl. J. Med. 357, 1829–1833 10.1056/NEJMoa070010 [DOI] [PubMed] [Google Scholar]
- Denève S., Duhamel J. R., Pouget A. (2007). Optimal sensorimotor integration in recurrent cortical networks: a neural implementation of Kalman filters. J. Neurosci. 27, 5744–5756 10.1523/JNEUROSCI.3985-06.2007 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Devinsky O., Feldmann E., Burrowes K., Bromfield E. (1989). Autoscopic phenomena with seizures. Arch. Neurol. 46, 1080–1088 [DOI] [PubMed] [Google Scholar]
- Ehrsson H. H. (2007). The experimental induction of out-of-body experiences. Science 317, 1048. 10.1126/science.1142175 [DOI] [PubMed] [Google Scholar]
- Ernst M. O., Banks M. S. (2002). Humans integrate visual and haptic information in a statistically optimal fashion. Nature 415, 429–433 10.1038/415429a [DOI] [PubMed] [Google Scholar]
- Friston K. J. (2005). A theory of cortical responses. Philos. Trans. R. Soc. Lond., B, Biol. Sci. 360, 815–836 10.1098/rstb.2005.1622 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Guldin W. O., Grüsser O. J. (1998). Is there a vestibular cortex? Trends Neurosci. 21, 254–259 10.1016/S0166-2236(97)01211-3 [DOI] [PubMed] [Google Scholar]
- Huys Q. J., Zemel R. S., Natarajan R., Dayan P. (2007). Fast population coding. Neural Comput. 19, 404–441 10.1162/neco.2007.19.2.404 [DOI] [PubMed] [Google Scholar]
- Jaynes E. T. (2003). Probability theory: the logic of science. Cambridge, Cambridge University Press [Google Scholar]
- Kilner J. M., Friston K. J., Frith C. D. (2007). The mirror-neuron system: a Bayesian perspective. Neuroreport 18, 619–623 10.1097/WNR.0b013e3281139ed0 [DOI] [PubMed] [Google Scholar]
- Knill D. C., Richards W. (eds) (1996). Perception as Bayesian Inference. New York, Cambridge University Press [Google Scholar]
- Körding K. P., Wolpert D. M. (2004). Bayesian integration in sensorimotor learning. Nature 427, 244–247 10.1038/nature02169 [DOI] [PubMed] [Google Scholar]
- Korhuber H. H. (1974). Vestibular system. Part 1: basic mechanisms. In Handbook of Sensory Physiology, Kornhuber H. H., ed. (Berlin: Springer; ), 676 p. [Google Scholar]
- Kornilova L. N. (1997). Orientation illusions in spaceflight. J. Vestib. Res., 7, 429–439 10.1016/S0957-4271(96)00184-X [DOI] [PubMed] [Google Scholar]
- Lackner J. R. (1992). Spatial orientation in weightless environments. Perception 21, 803–812 10.1068/p210803 [DOI] [PubMed] [Google Scholar]
- Laurens J., Droulez J. (2007). Bayesian processing of vestibular information. Biol. Cybern. 96, 389–404 10.1007/s00422-006-0133-1 [DOI] [PubMed] [Google Scholar]
- Lenggenhager B., Tadi T., Metzinger T., Blanke O. (2007). Video ergo sum: manipulating bodily self-consciousness. Science 317, 1096–1099 10.1126/science.1143439 [DOI] [PubMed] [Google Scholar]
- Ma W. J., Beck J. M., Latham P. E., Pouget A. (2006). Bayesian inference with probabilistic population codes. Nat. Neurosci. 9, 1432–1438 10.1038/nn1790 [DOI] [PubMed] [Google Scholar]
- MacNeilage P. R., Banks M. S., Berger D. R., Bülthoff H. H. (2007). A Bayesian model of the disambiguation of gravitoinertial force by visual cues. Exp. Brain Res. 179, 263–290 10.1007/s00221-006-0792-0 [DOI] [PubMed] [Google Scholar]
- Maillard L., Vignal J. P., Anxionnat R., TaillandierVespignani L. (2004). Semiologic value of ictal autoscopy. Epilepsia 45, 391–394 10.1111/j.0013-9580.2004.39103.x [DOI] [PubMed] [Google Scholar]
- McNaughton B. L., Battaglia F. P., Jensen O., Moser E. I., Moser M. B. (2006). Path integration and the neural basis of the ‘cognitive map’. Nat. Rev. Neurosci. 7, 663–678 10.1038/nrn1932 [DOI] [PubMed] [Google Scholar]
- Metzinger T. (2008). Empirical perspectives from the self-model theory of subjectivity: a brief summary with examples. Prog. Brain Res. 168, 215–245 10.1016/S0079-6123(07)68018-2 [DOI] [PubMed] [Google Scholar]
- Mizumoto M., Ishikawa M. (2005). Immunity to error through misidentification and the bodily illusion experiment. J. Conscious. Stud. 12, 3–19 [Google Scholar]
- Penfield W. (1955). The twenty-ninth Maudsley lecture: the role of the temporal cortex in certain psychical phenomena. J. Ment. Sci. 101, 451–465 [DOI] [PubMed] [Google Scholar]
- Poggio T., Torre V., Koch C. (1985). Computational vision and regularization theory. Nature 317, 314–319 10.1038/317314a0 [DOI] [PubMed] [Google Scholar]
- Pouget A., Dayan P., Zemel R. S. (2003). Inference and computation with population codes. Annu. Rev. Neurosci. 26, 381–410 10.1146/annurev.neuro.26.041002.131112 [DOI] [PubMed] [Google Scholar]
- Rao R. P., Ballard D. H. (1999). Predictive coding in the visual cortex: a functional interpretation of some extra-classical receptive-field effects. Nat. Neurosci. 2, 79–87 10.1038/4580 [DOI] [PubMed] [Google Scholar]
- Rao R. P. N. (2004). Bayesian computation in recurrent neural circuits. Neural Comput. 16, 1–38 10.1162/08997660460733976 [DOI] [PubMed] [Google Scholar]
- Sahani M., Dayan P. (2003). Doubly distributional population codes: simultaneous representation of uncertainty and multiplicity. Neural Comput. 15, 2255–2279 10.1162/089976603322362356 [DOI] [PubMed] [Google Scholar]
- Tsakiris M., Haggard P. (2005). The rubber hand illusion revisited: visuotactile integration and self-attribution. J. Exp. Psychol. Hum. Percept. Perform. 31, 80–91 10.1037/0096-1523.31.1.80 [DOI] [PubMed] [Google Scholar]
- Umeyama S. (1991). Least squares estimation of transformation parameters between two point patterns. IEEE Trans. Pattern Anal. Mach. Intell. 13, 376–380 10.1109/34.88573 [DOI] [Google Scholar]
- van Beers R. J., Sittig A. C., Denier van der Gon J. J. (1996). How humans combine simultaneous proprioceptive and visual position information. Exp. Brain Res. 111, 253–261 10.1016/S0079-6123(08)60413-6 [DOI] [PubMed] [Google Scholar]
- van Beers R. J., Sittig A. C., Gon J. J. (1999). Integration of proprioceptive and visual position-information: an experimentally supported model. J. Neurophysiol. 81, 1355–1364 [DOI] [PubMed] [Google Scholar]
- Vogeley K., Fink G. R. (2003). Neural correlates of the first-person-perspective. Trends Cogn. Sci. 7, 38–42 10.1016/S1364-6613(02)00003-7 [DOI] [PubMed] [Google Scholar]
- Vogeley K., May M., Ritzl A., Falkai P., Zilles K., Fink G. R. (2004). Neural correlates of first-person perspective as one constituent of human self-consciousness. J. Cogn. Neurosci. 16, 817–827 10.1162/089892904970799 [DOI] [PubMed] [Google Scholar]
- Wolpert D. M., Ghahramani Z., Jordan M. I. (1995). An internal model for sensorimotor integration. Science 269, 1880–1882 10.1126/science.7569931 [DOI] [PubMed] [Google Scholar]