Abstract
Incorporating the fact that the senses are embodied is necessary for an organism to interpret sensory information. Before a unified perception of the world can be formed, sensory signals must be processed with reference to body representation. The various attributes of the body such as shape, proportion, posture, and movement can be both derived from the various sensory systems and can affect perception of the world (including the body itself). In this review we examine the relationships between sensory and motor information, body representations, and perceptions of the world and the body. We provide several examples of how the body affects perception (including but not limited to body perception). First we show that body orientation effects visual distance perception and object orientation. Also, visual-auditory crossmodal-correspondences depend on the orientation of the body: audio “high” frequencies correspond to a visual “up” defined by both gravity and body coordinates. Next, we show that perceived locations of touch is affected by the orientation of the head and eyes on the body, suggesting a visual component to coding body locations. Additionally, the reference-frame used for coding touch locations seems to depend on whether gaze is static or moved relative to the body during the tactile task. The perceived attributes of the body such as body size, affect tactile perception even at the level of detection thresholds and two-point discrimination. Next, long-range tactile masking provides clues to the posture of the body in a canonical body schema. Finally, ownership of seen body parts depends on the orientation and perspective of the body part in view. Together, all of these findings demonstrate how sensory and motor information, body representations, and perceptions (of the body and the world) are interdependent.
Keywords: body representation, distance, gravity, auditory, crossmodal, tactile, self-perception
Introduction
Since the pioneering philosophical approach of Merleau-Ponty (1945), it has been acknowledged that the senses are embodied. The implication of this approach is that the senses can only be understood by acknowledging the attributes of the body in which they are necessarily situated. In vision, it is obvious that the eyes are in the head and that their viewpoints will be affected by the head’s position and orientation. What is perhaps less obvious is that these properties of the eyes’ vehicle contribute to processing such “visual” judgments as the orientation of the ground plane (Schreiber et al., 2008) and, as we will see, perceived distance. Head position influences the three-dimensional position of the eyes by means of static and dynamic three-dimensional vestibulo-ocular reflexes and through eye height. Information concerning head position is therefore critical to “externalize” the information in the retinal images: that is in creating a representation of the external world. Similar arguments apply to the ears, which are also passengers on the head. Head motion can even help to correctly scale the representation of external space, e.g., the distance between objects, which is notoriously hard to extract from static auditory or visual information alone (Gogel, 1963; Philbeck and Loomis, 1997). Information about the body is also needed to interpret tactile information about the world. When the hands explore and interact with objects in the world, it is necessary to take into account the arrangements of the hands and fingers in order to interpret the patterns of pressure sensed by the fingertips. The representation of the body is also needed to interpret the pressure and location of even simple touches on the skin in order to take into account the uneven density of tactile receptors over the surface of the body in the same way as the visual system must take into account the uneven density of photoreceptors in the retina. In this review we will outline some of the interesting, unexpected and fundamental roles that the body plays in determining our perception of the world.
The Effect of Body Orientation on Perceived Distance
Things look different when viewed with the head in an unusual orientation. It is amusing, for example, to look out of a tall building and watch people walking on the street below. Their legs seem to move in a strange way and they often look too small, “like ants,” suggesting a failure in size-distance constancy when looking straight downward which also extends to the perception of speed (Owens et al., 1990). It has long been suspected that body orientation or perceived body orientation may be connected to perhaps the most famous distance-related illusion in psychology: the moon illusion (Rock and Kaufman, 1962). Casual observation shows that the moon appears smaller when it is in the zenith and viewed by looking straight up than when it is close to the horizon and viewed straight ahead. Although the illusion continues to defy complete explanation (Hershenson, 1989; Ross and Plug, 2002; Weidner et al., 2014), it is usually explained with reference to changes in the moon’s perceived distance. We (Harris and Mander, 2014) were the first to measure the effect of posture (and perceived posture) on the perceived distance of objects at biologically significant distances (Cutting and Vishton, 1995), as opposed to the unknowable distance of celestial bodies. We used the York University Tumbling Room facility (Howard and Hu, 2001) in which the orientation of an observer and the surrounding room can be independently varied (Figures 1A,B). We showed that lying supine causes the opposite wall of the room to appear closer than when viewed from an upright position (Figures 1C,D). Rotating the room around an upright observer (Figure 1E) produces an illusion of lying supine (Howard and Hu, 2001). Just feeling supine due to this illusion turned out to be sufficient to create this shortening of perceived distance (Harris and Mander, 2014). Thus, it is the perceived orientation of the body that is important in interpreting visual cues to distance. This may be related to the geometrical requirement of taking eye orientation—itself dependent on head orientation—into account in order to interpret binocular cues correctly (Blohm et al., 2008). This unexpected involvement of the body in visual distance perception underscores the importance of the body in interpreting sensory information—it is not a raw sensory signal that leads to perception, but rather the representation in the brain of world features (including the body itself) that is modified in response to sensory input and that determines perception.
The Effect of Body Orientation on the Perceived Orientation of Objects
In the previous section we showed that the body’s orientation in pitch (head over heels) affected distance and size perception. Other changes in self-orientation can also lead to errors in perceptual judgments. When the body is rolled to one side (Figure 2), individuals systematically misperceive the orientation of an object relative to gravity. For example, when judging the orientation of a visual line with gravity vertical, estimates are biased toward the body midline (Aubert, 1886; Mittelstaedt, 1983). In contrast, when setting a bar to gravity vertical using only touch, there is a bias in the opposite direction, away from the body midline (Bauermeister et al., 1964). We (Fraser et al., 2014; Harris et al., 2014) compared visual and manual, touch-based estimates of gravity vertical while the body and head were tilted relative to each other (Figure 2).
Touch-based orientation judgments were affected more by the orientation of the body (Figure 2B), whereas visual errors were largely driven by head tilt (Figure 2C; Guerraz et al., 1998; Tarnutzer et al., 2009). Together, these results show that it is over simplistic to refer to the representation of the body as a single unit when considering the effect of self-orientation on perception. Changes in orientation of the head and body can have different effects on different sensory inputs and so they should be taken into account separately: posture is an important factor.
The Effect of Body Orientation on Auditory Localization
The ability to localize sound in elevation is tricky. What ability we have depends largely on reflections within the external pinna (Batteau, 1967; Fisher and Freedman, 1968; Makous and Middlebrooks, 1990; Blauert, 1996) and is thus bound to the head. Deducing where sounds are in the external world therefore requires taking into account the position and orientation of the head. Errors in sound localization when the head and body are tilted show that head orientation is only partially taken into account (Goossens and van Opstal, 1999; Parise et al., 2014). In fact, the perceived elevation of a sound, like the perceived orientation of a line we described above, depends on the perceived orientation of the head, which is determined by several factors.
Sounds that are played through headphones with no intrinsic location at all can nevertheless be perceived as having an elevation by virtue of their frequency content. This is an example of a cross-modal correspondence (Spence, 2011), in this case between pitch and perceived elevation, in which “higher” frequencies are perceived as coming from “higher” in space. But is this elevation defined in head or space coordinates? We showed that such sounds were perceived as lying on an axis defined neither by the head nor gravity but rather that lined up with the perceptual upright (Carnevale and Harris, 2013). Non-spatial sounds (tones played through headphones) that differed only in their frequency content (either rising or falling frequencies) were presented while observers viewed ambiguous visual motion in either the horizontal or vertical directions created by superimposing two gratings moving in opposite directions (left and right or up and down; Figure 3). Observers were tested lying on their sides to separate body and gravitational uprights. A disambiguating effect of sound was found in both directions (up relative to the head and up relative to gravity), suggesting that an auditory upright exists in between the head and gravitational reference frames—a direction very similar to the perceptual upright. The perceptual upright is the orientation in which objects are best identified and represents the brain’s best guess of the direction of up derived from a combination of visual and gravity cues (Dyde et al., 2006) and a tendency to revert to the body midline (Mittelstaedt, 1983). As we showed above for the influence of the body in determining perceived distance and orientation of objects, the perceived orientation of the body also determines the layout of auditory space (see also Parise et al., 2014, who used external sounds). So both visual and auditory perceptions depend on body orientation. What about the perception of touch?
Tactile Responses Depend on the Direction of Gaze
The orientation of the eyes and head are also involved in determining the perceived location of a touch such that the perceived location of a touch is shifted depending on gaze position (Harrar and Harris, 2009; Pritchett and Harris, 2011). Of course the direction of gaze is usually also the point to which attention is directed and attention is known to affect some aspects of tactile perception (Michie et al., 1987) in a way that depends on eye position (Gherri and Forster, 2014). However, Harrar and Harris (2009) found, by overtly orienting attention away from eye position, that attention could account for only about 17% of the effect. Even actions toward a touch are directed toward the shifted perceived position (Harrar and Harris, 2010). The effect appears to be equally affected by either eye or head displacement and is therefore best described as relating to gaze, the sum of eye and head position (Pritchett and Harris, 2011). The perceived location of touch also depends on whether a participant moves their gaze between the presentation of the touch and reporting its perceived location (Pritchett et al., 2012; Mueller and Fiehler, 2014). The perceived location shifts in the same direction as gaze if a gaze change occurs before the report (Harrar and Harris, 2009; Pritchett and Harris, 2011; Harrar et al., 2013), but in the opposite direction if the person does not move before making their report (Ho and Spence, 2007; Pritchett et al., 2012). What do these strange reversals tell us about the involvement of the body in the coding of touch? We can partially explain these gaze-related shifts in terms of the frame of reference in which touch location is coded. The direction of gaze and the direction in which the body is facing are misperceived toward one another when gaze is held eccentrically: the perceived straight ahead of the body is shifted in the direction of gaze, and the perceived direction of gaze is underestimated and perceived as closer to the body’s “straight ahead” (Hill, 1972; Morgan, 1978; Yamaguchi and Kaneko, 2007; Harris and Smith, 2008). Figure 4 shows how the direction in which the perceived location of touch shifts, may depend on whether it is coded relative to one or other of these misperceived reference directions. Displacements in a gaze-centered frame might also be evoked if the location of touch were attracted toward the direction of gaze. We can therefore conclude that touch is initially coded relative to the body midline but, if the location needs to be remembered during a gaze movement, it is switched to a gaze-based reference frame. Touch localization therefore depends on the orientation of the body and gaze. In next section we consider the effect of body size on the perception of touch.
Tactile Responses Depend on the Perceived Size of the Body
In order to identify the size of an object held against the skin it is necessary to correct for the variation in the density of tactile receptors on that part of the body surface. The object will stretch over an array of receptors on the body. The same size object will extend over a different number of receptors depending on the density of receptors in that area of skin. Receptor density must therefore be taken into account if an object’s felt size and proportions are to be accurately determined. In fact, small errors in the perceived size of felt objects are found in which an object felt on an area with a high density of receptors (e.g., the hand) is judged as slightly larger than when the same object is felt on an area with a low density of receptors (e.g., the back). This phenomenon, known as the Weber Illusion (Longo and Haggard, 2011), suggests incomplete compensation for the variation in receptor density and the associated distortions of the homunculus found in the primary somatosensory cortex (Penfield and Boldrey, 1937). The perceived size of the body even in adults is rather plastic and can be altered not only in response to normal growth but also in response to altered feedback concerning body size. For example, the perceived position of a limb can be manipulated by applying vibration to the associated tendon organs. If the affected limb is in contact with another body part, for example the tip of the nose, its perceived location in space will be inferred from the distorted position of the limb. Thus, the nose can appear lengthened: the aptly-named Pinocchio Illusion (Lackner, 1988). Such distortions in the size of a body part are passed on to objects felt on the skin (de Vignemont et al., 2005). If the body part is extended, an object pressed onto the skin is felt as correspondingly longer. Curiously, when perceived body size is distorted in either direction (made either larger or smaller) tactile sensitivity and acuity are both reduced compared to control conditions with non-tendon vibration and attention maintained constant throughout (Figure 5; D’Amour et al., 2015; but cf. Volcic et al., 2013). Distorting the perceived size of the body represents a major change in the critical, universal reference system of the brain: the body. Disrupting the body reference system has multiple fundamental consequences. But what might a reliable body reference look like?
The Body Reference System
Tactile sensitivity depends on many things. We have demonstrated that it depends on the body representation (Figure 5; D’Amour et al., 2015), and it is very likely that cognitive factors such as attention are also involved (Michie et al., 1987; Spence, 2002; Gherri and Forster, 2012, 2014). An additional factor is that tactile sensitivity can be influenced by simultaneous tactile stimulation on remote areas of the body. This is known as long-range tactile masking (Sherrick, 1964; Braun et al., 2005; Tamè et al., 2011) and seems to indicate a precise connection between the representations of certain patches of skin. For example, the sensitivity to touch on one arm can be influenced by long-range masking only by touch on the corresponding point on the other arm (Figure 6A; D’Amour and Harris, 2014a). Likewise touches on the stomach can be affected by simultaneous touch on the corresponding part of the back (Figure 6B; D’Amour and Harris, 2014b). These effects are quantified relative to when the masking stimulus is positioned at another point on the body so that any attentional effects caused by the presence of a second tactile stimulus were controlled. The question then becomes, how are the “corresponding points” defined and what can they tell us about the nature of the brain’s body schema?
In Head and Holmes’ (1911) original description of the representation of the body in the brain, they postulate a body schema in a “canonical posture” to which the actual posture is later added. The nature of this canonical representation can only be inferred but is presumably based on statistical probabilities of where the various body parts are likely to be (Bremner et al., 2012), that is a prior with the left arm and leg on the left and vice versa. This might correspond to the “position of orthopedic rest” (Bromage and Melzack, 1974), the position that astronauts adopt when relaxed in zero gravity1 although the detailed layout is hard to access. The prior is likely to rely on visual information about the body (Röder et al., 2004), which might provide a representation of the type shown in Figure 6C although the existence of phantom limbs in people born without arms or legs (Ramachandran and Hirstein, 1998; Brugger et al., 2000) indicates a genetic component to the body schema. Positioning the limbs in a non-canonical position (e.g., crossed) can provide hints about the canonical arrangement. If a touch is applied to the left hand while it is positioned on the right side of the body, saccades toward the touch will often start off directed toward its expected position on the left side (Groh and Sparks, 1996) and reaction times to the touch will be speeded by a visual cue on the left side (Azañón and Soto-Faraco, 2008). More detailed work is required testing many parts of the body (such as the hands, and the upper and lower sections of the limbs) to obtain a more precise impression of the canonical representation. Further, there are likely to be multiple schemas each adapted to a particular aspect of perception (de Vignemont, 2007; Longo et al., 2010).
Obviously the relationship between the front and back of the torso is fixed in all frames of reference, but for the limbs this is not the case. By varying the position of the limbs relative to each other, we have demonstrated that long-range tactile masking also depends on the position of the limbs in space (D’Amour and Harris, 2014a). Such modulation by posture suggests that long-range tactile masking is a phenomenon at or beyond the point at which the postural body schema is derived rather than at or before the level of the primary somatosensory cortex. The connections between the sides of the body has a neurophysiological correlate in which many somatosensory cells with receptive fields on the arms and hands are responsive to stimuli from either side of the body (Iwamura et al., 1994, 1993; Taoka et al., 1998). Such cells thus provide a signal that an arm was touched but do not distinguish which arm: at some level the postural schema seems to have only one arm! There is some indication that cross-body connections might also occur between the legs (Gilson, 1969; Iwamura et al., 2002, 2001) which suggests that this bilateral representation may include the whole body (Figure 6D). We refer to this as a Nasnas body after the monster in the Book of 1001 Nights. The Nasnas body may be a somatosensory equivalent to the way that vision is referred to a single cyclopean eye (Mapp and Ono, 1999).
The Representation of the Body in Defining the Self
The ability to move one’s own body and see that it behaves in the expected way is an important aspect of determining agency (Gallagher, 2000; Tsakiris et al., 2007a,b) and thus in deriving, establishing, and maintaining a sense of ownership of our own body. We established that sensitivity for detecting delay between initiating and seeing a movement was enhanced if the moving body part were seen in its natural orientation (the first-person perspective) as opposed to if it were seen as if it were someone else’s hand (from a third-person perspective, Hoover and Harris, 2012, see Figure 7). This variation with perspective gives us an objective measure of what the brain regards as the body’s first person view. Hand and head movements that are seen from a natural first-person perspective (looking down at the hand or seeing the hand or head in a mirror) are associated with a strong self-advantage, but views of the body from behind or of an arm stretched out toward us in a third-person perspective are not (Hoover and Harris, 2015). This suggests that body parts that can be seen in a first-person perspective are preferentially treated as belonging to us (Petkova et al., 2011b). Parts of the body that cannot be seen directly and thus have no representation from a first-person perspective, such as the back of the body, may not be regarded as parts of the self in the same way as parts of the body that can be seen directly. However, this can altered by providing an unusual first-person visual view of the back (Ehrsson, 2007; Lenggenhager et al., 2007; Spapé et al., 2015) demonstrating the role of learning and experience in forming our perception of our “self.” The suggestion that vision determines what is regarded as self either directly or from the view in a mirror, is compatible with our representation of the body in the brain as having only a two-dimensional representation of the torso as shown in Figure 6D.
Discussion
This review emphasizes the reciprocal nature of the perception of our bodies in the world and the world that we perceive. Multisensory integration operates not only at the level of integrating redundant cues about object properties—such as when auditory and visual cues signal the location of an event (Alais and Burr, 2004; Burr and Alais, 2006) or when cues about the size of an object are conveyed by both vision and touch (Ernst and Banks, 2002). Multisensory integration also determines the representation of the body in the brain (Maravita et al., 2003; Petkova et al., 2011a), and this representation in turn is fundamental in interpreting all sensory information.
The Body in the Brain
What is the nature of the body’s representation in the brain? Here we are not considering the consciously accessible representation of the body which may be divided into parts known as body mereology (de Vignemont et al., 2006) with their various cultural associations and accessible to consciousness (Longo, 2014). That is better referred to as a body image. Instead we are attempting to access the internal, possibly monstrous, representation(s) to which all sensory information is related at a neurophysiological level. This representation may be fragmented (Coslett and Lie, 2004; Kammers et al., 2009; Mancini et al., 2011) and apparently illogical in its arrangement. Many converging studies (e.g., Driver and Grossenbacher, 1996; Röder et al., 2004; Soto-Faraco et al., 2004; Longo et al., 2008) suggest that, counterintuitive to the idea of a fragmented, distorted representation, there might be a strong visual component to this representation, at least in normally sighted individuals (see Figure 6C). However, the view of the body is limited in the sense that only some parts can be seen at all and mostly from what we might paradoxically think of as an “odd angle” (see Figure 6C). In which case it is not surprising that there is reduced ownership of the back, which is not directly visible (Hoover and Harris, 2015), and that perception of the back may be closely linked to the more visible front (D’Amour and Harris, 2014b). Representing the three-dimensional body using the two-dimensional flat mapping process that seems to be so common in the brain (Chklovskii and Koulakov, 2004) clearly requires some transformations. It is necessary that sensory inputs are connected appropriately so that for example, a stimulus drawn across the body’s surface is perceived as moving continuously at a constant speed and without discontinuities as it moves from one side to the other or between regions of high and low acuity. That is, it is necessary that the unconscious, distorted body schema be related to the consciously accessible, three-dimensional body image in some.
The processes involved in creating and using a representation of the body in the brain are summarized in Figure 8. The body schema, in some canonical posture, has posture added to it, using information from proprioception and vision. This representation is then situated in space using proprioceptive vision (vision about the body and its relationship with space) and vestibular cues concerning the direction of up (Harris, 2009). The movement of the body, obtained also from visual and vestibular cues also needs to be taken into account, so that the position of earth-fixed features can be appropriately updated to register their new positions relative to the body both during the movement itself and following repositioning in space.
To consider sensory functioning in isolation of the multisensory context provided by the other senses and without regard to the body of which they are a part has to be regarded as being artificial. It is now the turn of our own bodies to take central stage if we are to understand how we are able to construct our perception of the external world and interact with it.
Conflict of Interest Statement
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Acknowledgments
The core funding for the experiments described in this paper was provided by the Natural Sciences and Engineering Council of Canada (NSERC) to LH. MC, SD, LF, and LP were partly supported from the NSERC CREATE program. LF, VH, and AH were supported by NSERC post-graduate scholarships. MC, SD, and LF received Ontario Graduate Scholarships.
Footnotes
JSC-09551, Skylab Experience Bulletin No. 17—Neutral Body Posture in Zero G, NASA-JSC, 7–75 cited in http://msis.jsc.nasa.gov/sections/section03.htm
References
- Alais D., Burr D. C. (2004). The ventriloquist effect results from near-optimal bimodal integration. Curr. Biol. 14, 257–262. 10.1016/j.cub.2004.01.029 [DOI] [PubMed] [Google Scholar]
- Aubert H. (1886). Die bewegungsempfindung. Pflugers Arch. J. Physiol. 39, 347–370. 10.1007/BF01612166 [DOI] [Google Scholar]
- Azañón E., Soto-Faraco S. (2008). Changing reference frames during the encoding of tactile events. Curr. Biol. 18, 1044–1049. 10.1016/j.cub.2008.06.045 [DOI] [PubMed] [Google Scholar]
- Batteau D. W. (1967). The role of pinna in human localization. Proc. R. Soc. Lond. B Biol. Sci. 168, 158–180. 10.1098/rspb.1967.0058 [DOI] [PubMed] [Google Scholar]
- Bauermeister M., Werner H., Wapner S. (1964). The effect of body tilt on tactual-kinesthetic perception of verticality. Am. J. Psychol. 77, 451–456. 10.2307/1421016 [DOI] [PubMed] [Google Scholar]
- Blauert J. (1996). Spatial Hearing. The Psychophysics of Human Sound Localization. Cambridge, MA: MIT Press. [Google Scholar]
- Blohm G., Khan A. Z., Ren L., Schreiber K. M., Crawford J. D. (2008). Depth estimation from retinal disparity requires eye and head orientation signals. J. Vis. 8, 3.1–3.23. 10.1167/8.16.3 [DOI] [PubMed] [Google Scholar]
- Braun C., Hess H., Burkhardt M., Wuhle A., Preissl H. (2005). The right hand knows what the left hand is feeling. Exp. Brain Res. 162, 366–373. 10.1007/s00221-004-2187-4 [DOI] [PubMed] [Google Scholar]
- Bremner A. J., Holmes N. P., Spence C. (2012). “The development of multisensory representations of the body and of the space around the body,” in Multisensory Development, eds Bremner A. J., Lewkowicz D. J., Spence C. (Oxford: Oxford University Press; ), 113–136. [Google Scholar]
- Bromage P. R., Melzack R. (1974). Phantom limbs and the body schema. Can. Anaesth. Soc. J. 21, 267–274. 10.1007/BF03005731 [DOI] [PubMed] [Google Scholar]
- Brugger P., Kollias S. S., Müri R. M., Crelier G., Hepp-Reymond M. C., Regard M. (2000). Beyond re-membering: phantom sensations of congenitally absent limbs. Proc. Natl. Acad. Sci. U.S.A. 97, 6167–6172. 10.1073/pnas.100510697 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Burr D., Alais D. (2006). Combining visual and auditory information. Prog. Brain Res. 155, 243–258. 10.1016/S0079-6123(06)55014-9 [DOI] [PubMed] [Google Scholar]
- Carnevale M. J., Harris L. R. (2013). The contribution of sound in determining the perceptual upright. Multisens. Res. 26, 125 10.1163/22134808-000S0091 [DOI] [Google Scholar]
- Chklovskii D. B., Koulakov A. A. (2004). Maps in the brain: what can we learn from them? Annu. Rev. Neurosci. 27, 369–392. 10.1146/annurev.neuro.27.070203.144226 [DOI] [PubMed] [Google Scholar]
- Coslett H. B., Lie E. (2004). Bare hands and attention: evidence for a tactile representation of the human body. Neuropsychologia 42, 1865–1876. 10.1016/j.neuropsychologia.2004.06.002 [DOI] [PubMed] [Google Scholar]
- Cutting J. E., Vishton P. M. (1995). “Perceiving layout and knowing distances: the interaction, relative potency, and contextual use of different information about depth,” in Perception of Space and Motion, eds Epstein W., Rogers S. (San Diego, CA: Academic Press; ), 66–117. [Google Scholar]
- D’Amour S., Harris L. R. (2014a). Contralateral tactile masking between forearms. Exp. Brain Res. 232, 821–826. 10.1007/s00221-013-3791-y [DOI] [PubMed] [Google Scholar]
- D’Amour S., Harris L. R. (2014b). Vibrotactile masking through the body. Exp. Brain Res. 232, 2859–2863. 10.1007/s00221-014-3955-4 [DOI] [PubMed] [Google Scholar]
- D’Amour S. D., Pritchett L. M., Harris L. R. (2015). Bodily illusions disrupt tactile sensations. J. Exp. Psychol. Hum. Percept. Perform. 41, 42–49. 10.1037/a0038514 [DOI] [PubMed] [Google Scholar]
- de Vignemont F. (2007). How many representations of the body? Behav. Brain Sci. 30, 204–205. 10.1017/S0140525X07001434 [DOI] [Google Scholar]
- de Vignemont F., Ehrsson H. H., Haggard P. (2005). Bodily illusions modulate tactile perception. Curr. Biol. 15, 1286–1290. 10.1016/j.cub.2005.06.067 [DOI] [PubMed] [Google Scholar]
- de Vignemont F., Tsakiris M., Haggard P. (2006). “Body mereology,” in Human Body Perception from the Inside Out, eds Knoblich G., Thornton I. M., Grosjean M., Shiffrar M. (Oxford: Oxford University Press; ), 147–170. [Google Scholar]
- Driver J., Grossenbacher P. G. (1996). “Multimodal spatial constraints on tactile selective attention,” in Attention and Performance XVI: Information Integration in Perception and Communication, eds Toshio I., McClelland J. L. (Cambridge, MA: MIT Press; ), 209–236. [Google Scholar]
- Dyde R. T., Jenkin M., Harris L. R. (2006). The subjective visual vertical and the perceptual upright. Exp. Brain Res. 173, 612–622. 10.1007/s00221-006-0405-y [DOI] [PubMed] [Google Scholar]
- Ehrsson H. H. (2007). The experimental induction of out-of-body experiences. Science 317, 1048. 10.1126/science.1142175 [DOI] [PubMed] [Google Scholar]
- Ernst M. O., Banks M. S. (2002). Humans integrate visual and haptic information in a statistically optimal fashion. Nature 415, 429–433. 10.1038/415429a [DOI] [PubMed] [Google Scholar]
- Fisher H. G., Freedman S. J. (1968). The role of the pinna in auditory localization. J. Aud. Res. 8, 15–26. [Google Scholar]
- Fraser L., Makooie B., Harris L. R. (2014). Contributions of the body and head to perceived vertical: cross-modal differences. J. Vis. 14, 1097 10.1167/14.10.1097 [DOI] [Google Scholar]
- Gallagher I. (2000). Philosophical conceptions of the self: implications for cognitive science. Trends Cogn. Sci. 4, 14–21. 10.1016/S1364-6613(99)01417-5 [DOI] [PubMed] [Google Scholar]
- Gherri E., Forster B. (2012). The orienting of attention during eye and hand movements: ERP evidence for similar frame of reference but different spatially specific modulations of tactile processing. Biol. Psychol. 91, 172–184. 10.1016/j.biopsycho.2012.06.007 [DOI] [PubMed] [Google Scholar]
- Gherri E., Forster B. (2014). Attention to the body depends on eye-in-orbit position. Front. Psychol. 5:683. 10.3389/fpsyg.2014.00683 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Gilson R. D. (1969). Vibrotactile masking: some spatial and temporal aspects. Percept. Psychophys. 5, 176–180. 10.3758/BF03209553 [DOI] [Google Scholar]
- Gogel W. (1963). The visual perception of size and distance. Vision Res. 3, 101–120. 10.1016/0042-6989(63)90033-6 [DOI] [PubMed] [Google Scholar]
- Goossens H. H., van Opstal A. J. (1999). Influence of head position on the spatial representation of acoustic targets. J. Neurophysiol. 81, 2720–2736. [DOI] [PubMed] [Google Scholar]
- Groh J. M., Sparks D. L. (1996). Saccades to somatosensory targets. 1. behavioral-characteristics. J. Neurophysiol. 75, 412–427. [DOI] [PubMed] [Google Scholar]
- Guerraz M., Poquin D., Luyat M., Ohlmann T. (1998). Head orientation involvement in assessment of the subjective vertical during whole body tilt. Percept. Mot. Skills 87, 643–648. [DOI] [PubMed] [Google Scholar]
- Harrar V., Harris L. R. (2009). Eye position affects the perceived location of touch. Exp. Brain Res. 198, 403–410. 10.1007/s00221-009-1884-4 [DOI] [PubMed] [Google Scholar]
- Harrar V., Harris L. R. (2010). Touch used to guide action is partially coded in a visual reference frame. Exp. Brain Res. 203, 615–620. 10.1007/s00221-010-2252-0 [DOI] [PubMed] [Google Scholar]
- Harrar V., Pritchett L. M., Harris L. R. (2013). Segmented space: measuring tactile localisation in body coordinates. Multisens. Res. 26, 3–18. 10.1163/22134808-00002399 [DOI] [PubMed] [Google Scholar]
- Harris L. R., Mander C. (2014). Perceived distance depends on the orientation of both the body and the visual environment. J. Vis. 14, 1–8. 10.1167/14.12.17 [DOI] [PubMed] [Google Scholar]
- Harris L. R., Smith A. (2008). The coding of perceived eye position. Exp. Brain Res. 187, 429–437. 10.1007/s00221-008-1313-0 [DOI] [PubMed] [Google Scholar]
- Harris L. R. (2009). “Visual-vestibular interactions,” in The International Encyclopedia of Neuroscience, ed. Squire L. R. (North Holland: Elsevier; ), 381–387. [Google Scholar]
- Harris L. R., Fraser L., Makooie B. (2014). Vestibular and neck proprioceptive contributions to the perceived vertical. Iperception 5, 441. [Google Scholar]
- Head H., Holmes G. (1911). Sensory disturbances from cerebral lesions. Brain 34, 102–254. 10.1093/brain/34.2-3.102 [DOI] [Google Scholar]
- Hershenson M. (1989). The Moon Illusion. New Jersey, NJ: Lawrence Erlbaum Associates. [Google Scholar]
- Hill A. L. (1972). Directional constancy. Percept. Psychophys. 11, 175–178. 10.3758/BF03210370 [DOI] [Google Scholar]
- Ho C., Spence C. (2007). Head orientation biases tactile localization. Brain Res. 1144, 136–141. 10.1016/j.brainres.2007.01.091 [DOI] [PubMed] [Google Scholar]
- Hoover A. E. N., Harris L. R. (2012). Detecting delay in visual feedback of an action as a monitor of self recognition. Exp. Brain Res. 222, 389–97. 10.1007/s00221-012-3224-3 [DOI] [PubMed] [Google Scholar]
- Hoover A. E. N., Harris L. R. (2015). The role of the viewpoint on body ownership. Exp. Brain Res. 233, 1053–1060. 10.1007/s00221-014-4181-9 [DOI] [PubMed] [Google Scholar]
- Howard I. P., Hu G. (2001). Visually induced reorientation illusions. Perception 30, 583–600. 10.1068/p3106 [DOI] [PubMed] [Google Scholar]
- Iwamura Y., Iriki A., Tanaka M. (1994). Bilateral hand representation in the postcentral somatosensory cortex. Nature 369, 554–556. 10.1038/369554a0 [DOI] [PubMed] [Google Scholar]
- Iwamura Y., Tanaka M., Iriki A., Taoka M., Toda T. (2002). Processing of tactile and kinesthetic signals from bilateral sides of the body in the postcentral gyrus of awake monkeys. Behav. Brain Res. 135, 185–190. 10.1016/S0166-4328(02)00164-X [DOI] [PubMed] [Google Scholar]
- Iwamura Y., Tanaka M., Sakamoto M., Hikosaka O. (1993). Rostrocaudal gradients in the neuronal receptive field complexity in the finger region of the alert monkey’s postcentral gyrus. Exp. Brain Res. 92, 360–368. 10.1007/BF00229023 [DOI] [PubMed] [Google Scholar]
- Iwamura Y., Taoka M., Iriki A. (2001). Bilateral activity and callosal connections in the somatosensory cortex. Neuroscientist 7, 419–429. 10.1177/107385840100700511 [DOI] [PubMed] [Google Scholar]
- Kammers M. P., Longo M. R., Tsakiris M., Dijkerman H. C., Haggard P. (2009). Specificity and coherence of body representations. Perception 38, 1804–1820. 10.1068/p6389 [DOI] [PubMed] [Google Scholar]
- Lackner J. R. (1988). Some Proprioceptive influences on the perceptual representation of body shape and orientation. Brain 111, 281–297. 10.1093/brain/111.2.281 [DOI] [PubMed] [Google Scholar]
- Lenggenhager B., Tadi T., Metzinger T., Blanke O. (2007). Video ergo sum: manipulating bodily self-consciousness. Science 317, 1096–1099. 10.1126/science.1143439 [DOI] [PubMed] [Google Scholar]
- Longo M. R. (2014). Three-dimensional coherence of the conscious body image. Q. J. Exp. Psychol. 68, 1116–1123. 10.1080/17470218.2014.975731 [DOI] [PubMed] [Google Scholar]
- Longo M. R., Azañón E., Haggard P. (2010). More than skin deep: body representation beyond primary somatosensory cortex. Neuropsychologia 48, 655–668. 10.1016/j.neuropsychologia.2009.08.022 [DOI] [PubMed] [Google Scholar]
- Longo M. R., Cardozo S., Haggard P. (2008). Visual enhancement of touch and the bodily self. Conscious. Cogn. 17, 1181–1191. 10.1016/j.concog.2008.01.001 [DOI] [PubMed] [Google Scholar]
- Longo M. R., Haggard P. (2011). Weber’s illusion and body shape: anisotropy of tactile size perception on the hand. J. Exp. Psychol. Hum. Percept. Perform. 37, 720–726. 10.1037/a0021921 [DOI] [PubMed] [Google Scholar]
- Makous J. C., Middlebrooks J. C. (1990). Two-dimensional sound localization by human listeners. J. Acoust. Soc. Am. 87, 2188–2200. 10.1121/1.399186 [DOI] [PubMed] [Google Scholar]
- Mancini F., Longo M. R., Iannetti G. D., Haggard P. (2011). A supramodal representation of the body surface. Neuropsychologia 49, 1194–1201. 10.1016/j.neuropsychologia.2010.12.040 [DOI] [PubMed] [Google Scholar]
- Mapp A., Ono H. (1999). Wondering about the wandering cyclopean eye. Vision Res. 39, 2381–2386. 10.1016/S0042-6989(98)00278-8 [DOI] [PubMed] [Google Scholar]
- Maravita A., Spence C., Driver J. (2003). Multisensory integration and the body schema: close to hand and within reach. Curr. Biol. 13, R531–R539. 10.1016/S0960-9822(03)00449-4 [DOI] [PubMed] [Google Scholar]
- Merleau-Ponty M. (1945). Phenomenology of Perception. Paris: Gallimard. [Google Scholar]
- Michie P. T., Bearpark H. M., Crawford J. M., Glue L. C. (1987). The effects of spatial selective attention on the somatosensory event-related potential. Psychophysiology 24, 449–463. 10.1111/j.1469-8986.1987.tb00316.x [DOI] [PubMed] [Google Scholar]
- Mittelstaedt H. (1983). A new solution to the problem of the subjective vertical. Naturwissenschaften 70, 272–281. 10.1007/BF00404833 [DOI] [PubMed] [Google Scholar]
- Morgan C. L. (1978). Constancy of egocentric visual direction. Percept. Psychophys. 23, 61–63. 10.3758/BF03214296 [DOI] [PubMed] [Google Scholar]
- Mueller S., Fiehler K. (2014). Effector movement triggers gaze-dependent spatial coding of tactile and proprioceptive-tactile reach targets. Neuropsychologia 62, 184–193. 10.1016/j.neuropsychologia.2014.07.025 [DOI] [PubMed] [Google Scholar]
- Owens D. A., Hahn J. P., Francis E. L., Wist E. R. (1990). Binocular vergence and perceived object velocity: a new illusion. Investig. Ophthalmol. Vis. Sci. 31(Suppl.), 93. [Google Scholar]
- Parise C. V., Knorre K., Ernst M. O. (2014). Natural auditory scene statistics shapes human spatial hearing. Proc. Natl. Acad. Sci. U. S. A. 111, 6104–6108. 10.1073/pnas.1322705111 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Penfield W., Boldrey E. (1937). Somatic motor and sensory representation in the cerebral cortex of man as studied by electrical stimulation. Brain 60, 389–443. 10.1093/brain/60.4.389 [DOI] [Google Scholar]
- Petkova V. I., Bjornsdotter M., Gentile G., Jonsson T., Li T. Q., Ehrsson H. H. (2011a). From part- to whole-body ownership in the multisensory brain. Curr. Biol. 21, 1118–1122. 10.1016/j.cub.2011.05.022 [DOI] [PubMed] [Google Scholar]
- Petkova V. I., Khoshnevis M., Ehrsson H. H. (2011b). The perspective matters! Multisensory integration in ego-centric reference frames determines full-body ownership. Front. Psychol. 2:35. 10.3389/fpsyg.2011.00035 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Philbeck J. W., Loomis J. M. (1997). Comparison of two indicators of perceived egocentric distance under full-cue and reduced-cue conditions. J. Exp. Psychol. Hum. Percept. Perform. 23, 72–85. 10.1037/0096-1523.23.1.72 [DOI] [PubMed] [Google Scholar]
- Pritchett L. M., Carnevale M. J., Harris L. R. (2012). Reference frames for coding touch location depend on the task. Exp. Brain Res. 222, 437–445. 10.1007/s00221-012-3231-4 [DOI] [PubMed] [Google Scholar]
- Pritchett L. M., Harris L. R. (2011). Perceived touch location is coded using a gaze signal. Exp. Brain Res. 213, 229–234. 10.1007/s00221-011-2713-0 [DOI] [PubMed] [Google Scholar]
- Ramachandran V. S., Hirstein W. (1998). The perception of phantom limbs. The D. O. Hebb lecture. Brain 121, 1603–1630. 10.1093/brain/121.9.1603 [DOI] [PubMed] [Google Scholar]
- Rock I., Kaufman L. (1962). The Moon Illusion II: the moon’s apparent size is a function of the presence or absence of terrain. Science 136, 1023–1031. 10.1126/science.136.3521.1023 [DOI] [PubMed] [Google Scholar]
- Röder B., Rosler F., Spence C. (2004). Early vision impairs tactile perception in the blind. Curr. Biol. 14, 121–124. 10.1016/j.cub.2003.12.054 [DOI] [PubMed] [Google Scholar]
- Ross H. E., Plug C. (2002). The Mystery of the Moon Illusion: Exploring Size Perception. Oxford: Oxford University Press. [Google Scholar]
- Schreiber K. M., Hillis J. M., Filippini H. R., Schor C. M., Banks M. S. (2008). The surface of the empirical horopter. J. Vis. 8, 7.1–7.20. 10.1167/8.3.7 [DOI] [PubMed] [Google Scholar]
- Sherrick C. (1964). Effects of double simultaneous stimulation of the skin. Am. J. Psychol. 77, 42–53. 10.2307/1419270 [DOI] [PubMed] [Google Scholar]
- Soto-Faraco S., Ronald A., Spence C. (2004). Tactile selective attention and body posture: assessing the multisensory contributions of vision and proprioception. Percept. Psychophys. 66, 1077–1094. 10.3758/BF03196837 [DOI] [PubMed] [Google Scholar]
- Spapé M. M., Ahmed I., Jacucci G., Ravaja N. (2015). The self in conflict: actors and agency in the mediated sequential Simon task. Front. Psychol. 6:304. 10.3389/fpsyg.2015.00304 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Spence C. (2002). Multisensory attention and tactile information-processing. Behav. Brain Res. 135, 57–64. 10.1016/S0166-4328(02)00155-9 [DOI] [PubMed] [Google Scholar]
- Spence C. (2011). Crossmodal correspondences: a tutorial review. Atten. Percept. Psychophys. 73, 971–995. 10.3758/s13414-010-0073-7 [DOI] [PubMed] [Google Scholar]
- Tamè L., Farnè A., Pavani F. (2011). Spatial coding of touch at the fingers: insights from double simultaneous stimulation within and between hands. Neurosci. Lett. 487, 78–82. 10.1016/j.neulet.2010.09.078 [DOI] [PubMed] [Google Scholar]
- Taoka M., Toda T., Iwamura Y. (1998). Representation of the midline trunk, bilateral arms, and shoulders in the monkey postcentral somatosensory cortex. Exp. Brain Res. 123, 315–322. 10.1007/s002210050574 [DOI] [PubMed] [Google Scholar]
- Tarnutzer A. A., Bockisch C. J., Straumann D. (2009). Head roll dependent variability of subjective visual vertical and ocular counterroll. Exp. Brain Res. 195, 621–626. 10.1007/s00221-009-1823-4 [DOI] [PubMed] [Google Scholar]
- Tsakiris M., Hesse M. D., Boy C., Haggard P., Fink G. R. (2007a). Neural signatures of body ownership: a sensory network for bodily self-consciousness. Cereb. Cortex 17, 2235–2244. 10.1093/cercor/bhl131 [DOI] [PubMed] [Google Scholar]
- Tsakiris M., Schutz-Bosbach S., Gallagher S. (2007b). On agency and body-ownership: phenomenological and neurocognitive reflections. Conscious. Cogn. 16, 645–660. 10.1016/j.concog.2007.05.012 [DOI] [PubMed] [Google Scholar]
- Volcic R., Fantoni C., Caudek C., Assad J. A., Domini F. (2013). Visuomotor adaptation changes stereoscopic depth perception and tactile discrimination. J. Neurosci. 33, 17081–17088. 10.1523/JNEUROSCI.2936-13.2013 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Weidner R., Plewan T., Chen Q., Buchner A., Weiss P. H., Fink G. R. (2014). The moon illusion and size-distance scaling—evidence for shared neural patterns. J. Cogn. Neurosci. 26, 1871–1882. 10.1162/jocn_a_00590 [DOI] [PubMed] [Google Scholar]
- Yamaguchi M., Kaneko H. (2007). Integration system of head, eye, and retinal position signals for perceptual direction. Opt. Rev. 14, 411–415. 10.1007/s10043-007-0411-8 [DOI] [Google Scholar]