Skip to main content
PLOS ONE logoLink to PLOS ONE
. 2013 Jul 5;8(7):e68438. doi: 10.1371/journal.pone.0068438

When Kinesthesia Becomes Visual: A Theoretical Justification for Executing Motor Tasks in Visual Space

Michele Tagliabue 1,*, Joseph McIntyre 1
Editor: Robert J van Beers2
PMCID: PMC3702599  PMID: 23861903

Abstract

Several experimental studies in the literature have shown that even when performing purely kinesthetic tasks, such as reaching for a kinesthetically felt target with a hidden hand, the brain reconstructs a visual representation of the movement. In our previous studies, however, we did not observe any role of a visual representation of the movement in a purely kinesthetic task. This apparent contradiction could be related to a fundamental difference between the studied tasks. In our study subjects used the same hand to both feel the target and to perform the movement, whereas in most other studies, pointing to a kinesthetic target consisted of pointing with one hand to the finger of the other, or to some other body part. We hypothesize, therefore, that it is the necessity of performing inter-limb transformations that induces a visual representation of purely kinesthetic tasks. To test this hypothesis we asked subjects to perform the same purely kinesthetic task in two conditions: INTRA and INTER. In the former they used the right hand to both perceive the target and to reproduce its orientation. In the latter, subjects perceived the target with the left hand and responded with the right. To quantify the use of a visual representation of the movement we measured deviations induced by an imperceptible conflict that was generated between visual and kinesthetic reference frames. Our hypothesis was confirmed by the observed deviations of responses due to the conflict in the INTER, but not in the INTRA, condition. To reconcile these observations with recent theories of sensori-motor integration based on maximum likelihood estimation, we propose here a new model formulation that explicitly considers the effects of covariance between sensory signals that are directly available and internal representations that are ‘reconstructed’ from those inputs through sensori-motor transformations.

Introduction

A number of previous studies have suggested that the CNS plans and executes targeted movements of the hand using a visual representation of the movement even when the target is presented kinesthetically (e.g. pointing with one hand to the other) and even when no visual feedback about the hand is allowed [1][5]. This is in apparent contrast with our own studies on human sensori-motor integration [6], [7] in which we observed that if subjects were asked to align their hidden hand to the orientation of a kinesthetically felt target, they completely ignored the information related to the visual scene, indicating that the brain executes purely kinesthetic tasks (K-K: kinesthetic target and kinesthetic response) without using a visual representation of the movement. This apparent contradiction, however, could be related to a fundamental difference between the motor tasks that the subjects were asked to perform in these different sets of studies. Whilst in our study subjects felt and reproduced the target position/orientation with the same hand, participants in the other aforementioned studies had to sense the target with the left hand or with a foot, hidden under a table, and reproduce its position with the right hand. Based on this observation, we postulated that the use of a visual representation of the movement could be related to the necessity of performing an inter-limb transformation of the kinesthetic information.

We set out to test this hypothesis by asking healthy volunteers to reproduce, by pronating/supinating their unseen hand, a kinesthetically memorized orientation (i.e. in a replication of the K-K condition of our previous study [6]) and we compared their responses in two different conditions (Figure 1). In the first, called INTRA-manual, subjects memorized and reproduced the orientation with the same hand. In the second, called INTER-manual, subjects memorized the orientation with one hand and responded with the other. Note that in the INTER-manual task, subjects could have simultaneously sensed the target with one hand and reproduced the target orientation with the other. We chose instead to have subjects reproduce the target from memory even in the bilateral task, so that any differences observed between INTER and INTRA could not be attributed to differing memory requirements of the tasks.

Figure 1. Experimental conditions.

Figure 1

In both INTRA- and INTER-manual condition subjects felt the orientation of a kinesthetic target and reproduced it by prono-supinating the hand (blue arrows). Subjects never saw their hand, thanks to a virtual reality system. Whilst in the INTRA condition the subject sensed the target and reproduced the orientation with the same (right) hand, in the INTER condition targets were sensed with the left and reproduced with the right. To detect the use of visual representation of the task, in some of the trials the virtual scene imperceptibly rotated with respect to gravity (red arrows).

In both INTER and INTRA, we exploited the peculiarity of our reach-to-grab paradigm in virtual reality that allows us to assess the weight given to visual information even when the subject had only kinesthetic sensory inputs to control the movement [6], [7]. Specifically, subjects were asked to tilt the head laterally after the target memorization and during the head movement of half of the trials we introduced an imperceptible sensory conflict between the orientation of the visual scene and the direction of gravity. Under these circumstances, deviations of the final hand orientation should be proportional to the relative weight given to the visual-scene versus gravito-kinesthetic information used to control the movement. If our hypothesis is correct, we expect visual encoding to play a role – and hence we expect to see a deviation of the responses due to the visual scene tilt – in the INTER-manual, but not in the INTRA-manual task.

In an effort to explain our results, we applied models of optimal sensory-motor integration based on the maximum likelihood principle (MLP), which falls in the domain of Bayesian optimal estimation [8][16]. Recent models assume that to control goal-directed upper-limb movements the brain evaluates the difference between target and hand concurrently in visual and kinesthetic reference frames [4], [6], [7], [17]. When direct sensory information is not available in one or more sensory modalities, internal representations of the target or hand may be ‘reconstructed’ via coordinate transformations, through recurrent neural networks in the parietal cortex (for review see [18]). Thus, when reaching a visual target with an unseen hand, the CNS might, for instance, reconstruct a visual representation of the hand to be compared with the available sensory information about the target. The relative weight given to the comparisons performed in visual or kinesthetic space would be determined by the expected variability within each representation, taking into account the variability added by any sensori-motor transformations that may be required [14], [19][21].

In the context of this model, we asked whether kinesthetic information from the two hands is encoded in a single, kinesthetic, perhaps trunk-centered, reference frame, allowing them to be directly compared, or whether the kinesthetic information from the right and left arms is encoded in two different, arm-specific reference frames, perhaps centered on each shoulder [22], and thus requiring sensory transformations between them. For each of these two alternatives we proposed the corresponding mathematical formulation of the model (Figure 2), and we tested their ability to predict subjects' performance on each task. We then considered the effects of co-variance between transformed sensory signals as a means of reconciling the predictions of MLP with our experimental observations and with our hypothesis that the brain reconstructs concurrent movement representations only if a direct target-hand comparison is not possible [6].

Figure 2. Models of sensory-motor integration.

Figure 2

Both models A and B assume that the brain can perform concurrent visual and kinesthetic comparisons (Inline graphic, Inline graphic) between the target (T) and the hand (H) and optimally combine the results of these comparisons to estimate the target-hand distance (Inline graphic). Both models also include the possibility of performing sensory reconstructions of the information about the target and the effector, as represented by the green arrows. Model A does not distinguish between the kinesthetic information from left and right limb and they can be compared directly, therefore two concurrent comparisons are possible (Inline graphic, Inline graphic). For model B kinesthetic information about the target and the response can be directly compared only if they come from the same limb. Therefore this model explicitly differentiates three concurrent references frames for the comparisons: visual (Inline graphic), kinesthetic linked to the right limb (Inline graphic) and kinesthetic linked to the left limb (Inline graphic).

Results

Figure 3 shows for the INTRA and INTER-manual conditions the average responses to each target orientation in trials without conflict, which do not differ appreciably between the two experimental conditions: statistical analyses on the aligning errors showed no significant effect of the experimental condition (ANOVA F(1,15) = 2.13, p = 0.17) or interactions between condition and target orientation (ANOVA F(6,90) = 0.92, p = 0.49). On the other hand, clear differences can be observed in Figure 4 for the responses in the trials with conflict: in the INTRA condition the responses were not consistently deviated by the conflict, in the INTER condition conflict caused the responses to all target orientations to be deviated in the same direction. These latter observations were confirmed by the statistical analysis of the global deviation of the responses. Figure 5A shows that in the INTER-manual condition responses were significantly deviated by the inclination of the visual surround (one-tailed t-test with respect to 0Inline graphic: t(15) = 2.27, p = 0.02), whilst no statistical difference from the null deviation could be detected in the INTRA-manual condition (one-tailed t-test with respect to 0Inline graphic: t(15) = −0.37, pInline graphic0.25). The differing effects between the two experimental conditions was confirmed by a significant difference between the INTRA- and INTER-manual conditions (ANOVA F(1,15) = 8.81, p = 0.009). The comparison of the response variability between the two experimental conditions, reported in Figure 5B, shows that subjects were more precise when they had to reproduce an orientation felt with the same hand than with the other hand (ANOVA F(1,15) = 10.38, p = 0.005).

Figure 3. Subject responses to each target orientation in the INTRA- and INTER-manual conditions for the trials without conflict.

Figure 3

Thick lines are the average responses (transparent areas’ width represent the standard error) of all subjects, combining trials with left and right head tilt as explained in the methods.

Figure 4. Subject responses to each target orientation in the INTRA- and INTER-manual conditions.

Figure 4

The responses for the trials with and without conflict are represented separately. Thick lines are the average responses (transparent areas’ width represent the standard error) of all subjects, combining trials with left and right head tilt as explained in the methods. Green arrows represent the measured responses’ deviations due to the tilt of the visual scene.

Figure 5.

Figure 5

A. Experimental results for the deviation of responses induced by the imperceptible tilt of the visual scene and the variability of responses in the INTRA- and INTER-manual conditions. Results are expressed as a percentage of the theoretical deviation expected if subjects aligned the response with respect to the visual scene. B. Average within-subject variability (standard deviation) of the response without conflict. In both panels vertical whiskers represent the 0.95 confidence intervals. Stars represent the significance of the main effects in the ANOVA and the results of the t-test comparison with the nominal 0Inline graphic value. (** Inline graphic; * Inline graphic).

Theoretical Modeling

Given the clear empirical observations shown above, we then undertook a mathematical analysis to better understand the implications of these results for recent theories of sensorimotor integration. To represent the performance of this task, we considered the two candidate models shown in Fig. 2, both of which assume that the brain performs concurrent target-hand comparisons in different sensory modalities and then optimally combines the result of these comparisons to create the movement vector (Inline graphic). For model A, Inline graphic. For model B Inline graphic (for details about the notations, see the caption to Figure 2). Both models also assume that the brain can reconstruct missing information from available signals through recurrent neural networks (for review see [18]). To make predictions using these two models, we hypothesized that the weighting (Inline graphic) of each individual comparison would be determined by the maximum likelihood principle, which states that the two or three quantities computed in each case will be combined according to the relative variance of each signal (see Methods). We assumed that sensorimotor transformations add variability [19][21], such that reconstructed signals will have greater variance than the source signal. Note that this is a slight simplification. A transformation might literally add stochastic noise if the transformation involves the integration of the primary signal with other, noisy sensory inputs. But transformations might also amplify the variability of the primary signal if they included distortions, perceptual biases or other non-linearities. We assume here, however, that on a target-by-target basis the effects of such distortions will be small, such that the effects of a sensorimotor transformation on the variability of the transformed output can be adequately represented by the simple addition of variance and that the MLP equations can be applied directly to the sum, as was done in other sensori-motor integration models [4], [14]. We hence computed relative weights between directly sensed and reconstructed comparisons based on the assumption that transformations add variance [6].

With few exceptions [23], all studies in Neuroscience to date have, to our knowledge, applied the maximum likelihood equations as if the original and reconstructed signals are independent (i.e. uncorrelated). To demonstrate the limitations of that approach, and to gain insight into how covariance might affect the results, we first computed the model predictions without taking into account the covariance between the original sensory inputs and internal representations that are computed through sensorimotor transformations of those signals. Applying Eqs. 4 and 5 from the Methods for model A we have:

graphic file with name pone.0068438.e019.jpg (A1)
graphic file with name pone.0068438.e020.jpg (A2)

where Inline graphic and Inline graphic are the variance associated to the kinesthetic and visual comparison respectively. For model B, applying Eqs 6–8 from the Methods gives:

graphic file with name pone.0068438.e023.jpg (B1)
graphic file with name pone.0068438.e024.jpg (B2)
graphic file with name pone.0068438.e025.jpg (B3)

where Inline graphic, Inline graphic, and Inline graphicare the variances associated to the left-hand kinesthetic, right-hand kinesthetic and visual comparison respectively. The net variance of each of the comparisons, Inline graphic, Inline graphic, Inline graphic and Inline graphic, depends on the variance of the input signals (e.g. Inline graphic, the variance of target orientation as measured via kinesthesia) and the variance added by any sensorimotor transformations required to reconstruct an internal representation (e.g. Inline graphic, representing the variance added when transforming target information from the kinesthetic to visual domains, including the variability of the sensory information that allows one to measure the orientation of the head, such as the visual scene, vestibular signals and neck proprioception). For the detailed equations defining the variance of each comparison, depending on the experimental conditions and hypotheses, see Figure 6. In order to reduce the number of parameters of the model and allow a meaningful fitting to the experimental data, the following assumptions were made:

Figure 6. Model predictions when MLP is applied while ignoring the covariance between direct and reconstructed sensory signals.

Figure 6

On the right, graphical representations of the information flow for model A and B and for the INTRA and INTER conditions, together with the equations of the variance corresponding to each concurrent target-hand comparison. The best-fit Model A predicts equal weight given to visual versus kinesthetic information between the INTRA and INTER tasks, and the same response variability in both cases. Model B predicts a greater weighting of visual information and higher response variability for the INTER task, qualitatively similar to the empirical data, but does not correctly predict a zero weighting of visual information in the INTRA-manual task.

  • The variability of the kinesthetic information from the left and right arm was similar [24]: Inline graphic.

  • The variance associated to the inter-limb transformation of kinesthetic information about the target and about the response was the same (Inline graphic).

  • The variance associated to the reconstruction of visual information from kinesthesia from either the right or left arm was the same (Inline graphic).

Note that asymmetries in errors when matching kinesthetic or visual targets with the right or left hand [25] mean that the last of these three assumptions may not be entirely true. We nevertheless adopted these assumptions so as to reduce the number of free parameters to three (variance of the kinesthetic signals Inline graphic, variance due to inter-limbs transformations Inline graphic and variance due to visuo-kinesthetic transformations Inline graphic) whose values could be determined to best fit the experimentally observed values of response deviation and response variability that are reported in Figure 5.

Figure 6 shows the model predictions based on equations A1–A2 and B1–B3, i.e. when co-variation between sensory signals and reconstructed internal representations is ignored. To the left we show a 2×2 grid, corresponding to the predictions of each of the two models (A and B) for each of the two experimental conditions (INTRA and INTER). Within each model we show the sources of variance associated with each individual comparison necessary for equations A1–A2 and B1–B3, including the noise associated with input signals that are present and variance added by any sensory transformations that may be required. Sensory inputs that are not available directly are grayed out in the schematic diagrams. To the right we superimpose on the experimental data the predictions of each of the two models with weights calculated from equations A1–A2 and B1–B3 that best fit all of the data (deviations and variability). The estimated variability associated with the kinesthetic information and each transformation as a result of the best-fit procedure are shown in Table 1.

Table 1. Values (expressed in °) of the predicted variability (±95% confidence interval) associated with the kinesthetic information (Inline graphic), inter-limb kinesthetic transformations (Inline graphic) and visuo-kinesthetic reconstructions (Inline graphic) that best fit the experimental data for the models in Figure 2.

Inline graphic Inline graphic Inline graphic
Model A 8.5Inline graphic0.3 22.8Inline graphic14.5
Model A′ 8.1Inline graphic0.3
Model B 8.1Inline graphic0.5 16.0Inline graphic3.5 20.2Inline graphic14.7
Model B′ 7.2Inline graphic0.3 17.2Inline graphic1.6 17.2Inline graphic1.8

From this analysis one can eliminate Model A as an explanation of the empirical data. According to this model, there is no difference between comparing the posture of the right hand to the left hand versus comparing the right hand’s posture with itself. Thus, this model predicts precisely the same weight given to the visual comparison, and thus precisely the same deviations of the response due to tilt of the visual scene between the INTER and INTRA tasks. Model A also predicts that the variability of responses will be equal between the INTER and INTRA tasks. We observed, however, a statistically-significant difference in both deviation and variability between the INTER and INTRA conditions, inconsistent with these predictions (two-tailed t-test between Model A’s prediction and the experimentally observed difference for response deviations between the INTER and INTRA conditions: t(15) = 3.0, pInline graphic0.01).

Model B does a better job of capturing the qualitative features of the measured data. According to this model, comparing the posture of the left hand to the right hand requires inter-limb sensory transformations. Thus, the kinesthetic comparisons will be more variable in the INTER task compared to the INTRA task. This allows the model to predict a different weighting between visual and kinesthetic information for the INTER vs. INTRA tasks, meaning that the deviations induced by rotations of the visual scene and the overall variability of responses are expected to be greater in the INTER task than in the INTRA task. Nevertheless, Model B, with weights computed to best fit the data, predicts a smaller differential of the response deviations between the INTRA and INTER tasks than what we actually observed (two-tailed t-test between Model B's prediction and the experimentally observed difference in response deviations between the INTER and INTRA conditions: t(15) = 2.2, pInline graphic0.05). Indeed, using reasonable assumptions about the relative amounts of variability in each sensory signal and in each sensory reconstruction, one cannot expect to see zero weight given to a visual comparison in the INTRA condition using Model B, even though that is what we observed experimentally.

This inability to reproduce the deviation data with Model B was insensitive to the precise values of response variability used to perform the fitting. The only way that one could expect to see the observed difference in deviations between INTER and INTRA with Model B would be if the difference in response variability between INTRA and INTER was much greater than what was observed. But to achieve the better than 7∶1 ratio that would be required to reproduce the deviation data with Model B, either the overall response variability would have to be much smaller (Inline graphic) in the INTRA condition or much larger (Inline graphic) in the INTER condition than the actually observed values (Inline graphic 1.9 for INTRA and Inline graphic for INTER).

To adequately fit both the deviation and variability data we needed to take into account the covariance between a reconstructed signal and its source when computing the MLP weights. As explained in the Methods section, this means that the relative weighting of the parallel comparisons is based on the variance of the independent components of each comparison, neglecting the variance of the common components. In Figure 7 we therefore report the variance associated to each possible comparison with the component of the variance common to all branches grayed out.

Figure 7. Model predictions when covariance is taken into account in the calculation of the optimal weights for each individual comparison.

Figure 7

On the right, graphical representations of the information flow for Model A’ and B’ and for the INTRA and INTER conditions, together with the equations of the variance corresponding to each concurrent target-hand comparison. Components of variance common to all branches, and hence not used to define the relative weights for Inline graphic, are grayed out. Model A’ predicts that there will be no reconstruction of the task in visual space for either INTRA or INTRA, because there is no variance in the Inline graphic comparison that is not included in the Inline graphic comparison. Deviation of the response with visual scene tilt should be 0 and the variance of the response should be the same for both conditions. Model B’ predicts that the task will be carried out only in kinesthetic space for the INTRA condition, but that both kinesthetic and visual comparisons will be made in the INTER condition. Only Model B’ can accurately fit the data.

For model A we note that there is no component of variance in kinesthetic comparison Inline graphic that is not also included in the visual comparison Inline graphic. Applying Eqs. 16–17 of the Methods, the predicted weight is zero for the visual comparison and one for the kinesthetic comparison for both INTER and INTRA conditions:

graphic file with name pone.0068438.e067.jpg (A'1)
graphic file with name pone.0068438.e068.jpg (A'2)

Because both conditions rely on Inline graphic only, one should also observe equal variance of responses between the two conditions. The statistically significant differences of deviations and variability between INTER and INTRA, and the statistically non-zero weight given in to visual information in the INTER condition mean that Model A should be rejected even if covariance is taken into account (two-tailed t-test between the predictions of A’1–A’2 and the experimentally observed difference for response deviations between the INTER and INTRA conditions: t(15) = 3.0, pInline graphic0.01).

On the other hand applying the same concepts for Model B does predict different results between the INTER and INTRA conditions. As shown in Figure 7, for the INTRA situation, there is no variance associated with the direct intra-manual comparison Inline graphic that is not also included in Inline graphic and Inline graphic. Applying Eqs. 18–20 to this situation gives the following weights for the three concurrent comparisons:

graphic file with name pone.0068438.e074.jpg
graphic file with name pone.0068438.e075.jpg
graphic file with name pone.0068438.e076.jpg
graphic file with name pone.0068438.e077.jpg
graphic file with name pone.0068438.e078.jpg (B'1)
graphic file with name pone.0068438.e079.jpg (B'2)
graphic file with name pone.0068438.e080.jpg (B'3)

No weight will be given to either Inline graphic or Inline graphic and thus, the deviation of the responses in the INTRA condition is predicted by Model B to be strictly zero. On the contrary, Figure 6 shows that in the INTER condition each of the comparisons Inline graphic, Inline graphic and Inline graphic contains a component of variance that is not included in the two others, because each comparison requires at least one transformation, the transformations are all different, and each of these transformation adds variability that is independent from the others. In this case applying Eqs. 18–20 predicts that some non-zero weight will be given to each of the three comparisons to determine the optimal outcome.

graphic file with name pone.0068438.e086.jpg (B'4)
graphic file with name pone.0068438.e087.jpg (B'5)
graphic file with name pone.0068438.e088.jpg (B'6)

The estimated variances associated with the kinesthetic information and sensory transformations as a result of the best-fit procedure are shown in Table 1. The non-zero weight given to Inline graphic means that Model B does predict a deviation of responses due to tilt of the visual scene. Unlike for the predictions of Eqs. B1–B3, predictions made by taking into account the covariance when applying MLP (Eqs. B’1–B’6) are not statistically different from the empirically estimated difference in weights between the INTER and INTRA conditions (two-tailed t-test between the predictions of Model B’ and the experimentally observed difference for response deviations between the INTER and INTRA conditions: t(15) = 0.4, pInline graphic0.70). The variability added by the different transformations also means that the final variance of responses is expected to be greater for the INTER condition, compared to the INTRA condition, which is what we observed. Equations B’1–B’6 are therefore able to fit the salient features of the empirical data (deviation of responses in INTER but not INTRA, different variability between INTER and INTRA).

Discussion

The results show that, for a purely kinesthetic task of reproducing a kinesthetically sensed orientation with an unseen hand (K-K condition), the brain gives significant weight to visual information when the task requires an inter-limb information transmission (INTER condition), but not when subjects memorized and responded with the same hand (INTRA condition). The lack of a significant effect of visual information in the INTRA condition matches our previous results [6] and is coherent with studies of reaching movements in which subjects also used the same arm to feel the target and to reproduce its remembered position [26][29]. The use of visual encoding in otherwise purely kinesthetic pointing tasks has nevertheless been observed in a number of studies that have required comparing one limb to another [1], [4], [5]. Our results suggest that the use of visual encoding in these studies was most likely due to the bilateral nature of the task and that responses would have been different if the subjects had used the same limb to feel the target and to reproduce its position.

In our previous work [6] we concluded that the CNS avoids reconstructing sensory information whenever a direct target-effector comparison is feasible and that if a sensorimotor transformation cannot be avoided, the CNS creates concurrent representations of the task in multiple reference frames potentially tied to other sensory modalities. Based on the reconstruction of a visual representation in the INTER condition observed here, we therefore postulate that inter-manual comparisons require sensorimotor transformations and are therefore not ‘direct’. Independent support for this hypothesis can be found in the literature. Baud-Bovy and colleagues [27] showed that variable errors are oriented toward the left or the right shoulder depending on the arm used to memorize the kinesthetically-presented target, suggesting the existence of distinct reference frames tied to each arm. A transformation would therefore be required to compare the position of the two hands. Evidence for arm-specific reference frames for the control of pointing movements to visual targets has also been reported [19], [22], [30], [31]. The recreation of the task in multiple, alternative reference frames has been reported for other tasks as well. Just as we saw a reconstruction of the task in an alternative (visual) reference frame in the INTER, but not in the INTRA condition, Rao and colleagues [29] reasoned that tactile information plays more of a role in a bilateral reaching task, as compared to a unilateral one, because the bilateral tasks requires a transformation anyway, and so it made sense to transform the task into tactile space as well. Subjects did use an eye-centered reference frame to encode the kinesthetic information even in a unilateral kinesthetic matching task [32], but those subjects verbally reported whether one passive movement imposed by a robot was more to the left or to the right of another. Given that ‘left’ and ‘right’ refer to directions that are not intrinsically defined by the kinesthetic receptors of the arm, sensory transformations would nevertheless be required to verbalize the response. Thus, the need for at least one transformation, or not, appears to be the key factor in determining if additional representations of the task (visual or otherwise) are constructed. This concept explains, in a parsimonious way, the difference between INTER and INTRA reported here and a wide variety of findings reported in the literature.

Avoidance of sensory transformations, including inter-manual transformations, can also explain how the CNS chooses one sensory input over another when sensory information is available simultaneously in more than one modality. In an orientation-matching paradigm [6], [7], where we compared movements to a visual target with only visual feedback (V-V) or with both visual and kinesthetic feedback (V-VK) of the hand, we observed very similar results between these conditions both in terms of global variability and strength of the oblique effect (for review about oblique effect see [33]), suggesting a similar importance given to visual information in both conditions. Helms Tillery also observed similar variable error between a VK-V and V-V condition in a 3D pointing task [20]. Similarly, compared to a unilateral kinesthetic-only task (K-K), we saw no evidence for a reconstruction in visual space when visual information about the hand was added (K-VK). But other studies have reported the use of visual information in K-VK tasks [2][4], [8], [9], [34]. In those studies, however, subjects did not use the same arm to sense the target and produce the movement. Transforming the kinesthetic target to be compared with visual feedback about the hand would, in the bilateral case, appear to be advantageous. Conversely, adding kinesthetic information about a visual target had little effect in a study by Sabes that compared V-VK and bilateral VK-VK conditions [17]. Generally speaking, eye-centered representation of the movement dominate over kinesthetic cues when subjects reach for visual targets [35], [36]. Indeed, neural activation in the posterior parietal cortex of monkeys during reaching toward visual targets with the hand in view suggests that a target-hand comparison is performed in retinal space, without integration of kinesthetic information about the limb [37].

MLP can explain why the CNS gives greater weight to direct versus reconstructed comparisons if one takes into account the additional variability inherent to transformed signals [14], [19][21]. As a direct evidence supporting this idea, Burns and collegues [14] showed that for reaching to a visual target (V-VK) increasing the variability of cross-modal transformations by tilting the head [6], [38] increased the weight given to the direct visual comparison. It is worth noticing, therefore, that subjects in our experiments showed greater variability for INTER than for INTRA (Fig. 5B). A similar increase in variability was observed when contrasting cross-modal conditions (V-K or K-V) to conditions allowing a direct visual or unilateral kinesthetic comparison (V-V and K-K, respectively) [6]. The greater variability in the INTER condition observed here and elsewhere [39], can therefore most likely be ascribed to the variability added by inter-limb transformations, be it through added noise from sensory signals required for the transformation, or due to distortions or non-linearities in the transformation itself. In this vein, the loss of precision when tactile information was compared between limbs [40], [41] has been been attributed to the inter-hemispheric relay [42][44] or to the coordinate transformations that would be required to compare stimuli to different fingers within the same hand [45]. In sum, sensorimotor transformations in general, and inter-limb transformations in particular, appear to add variability. According to MLP, performing additional comparisons based on transformed information would do little to improve performance when a direct comparison is possible.

Simply considering the variability added by inter-limb comparisons was not sufficient, however, to properly predict the experimental results within the context of our models. We could only predict the total lack of effect of scene tilt in the INTRA condition by specifically considering the correlation between reconstructed and direct sensory information. More precisely, we showed that if a direct target-response comparison was possible (Inline graphic in the INTRA-manual condition), combining it with other comparisons (e.g. Inline graphic) reconstructed from the available kinesthetic information could not reduce response variability. On the other hand, in our INTER condition, where a direct target-effector comparison was not possible, all possible comparisons required some sort of sensory transformation. If the variability of each transformation is independent from the others, combining them can result in a decrease of the overall response variability. Accordingly, subjects reconstructed a visual representation of the task (Inline graphic), even though just one of the kinesthetic comparisons (Inline graphic or Inline graphic),would have been sufficient. These results are in line with our working premise that once a transformation becomes inevitable, a broader slate of redundant comparisons are automatically performed [6]. Furthermore, explicitly taking into account the covariance of transformed signals in the application of MLP provides a firm theoretical basis that explains not only when, but also why the CNS would reconstruct a visual representation of a kinesthetic task [1][5].

In conclusion, the fact that in the intra-manual task no role of visual information could be detected in our experiments demonstrates that the brain prefers direct comparisons whenever possible. We have shown that this is because additional reconstructed representations would strongly correlate to the direct comparison and hence would not reduce movement variability. On the other hand, when a sensory transformation is necessary to compare the hand and target position, even if it is just the transformations required to compare one arm to the other, the brain reconstructs the movement in multiple reference frames, thus creating a visual representation of a purely kinesthetic task.

Methods

Ethics Statement

The experimental protocol was approved by the IRB of University Paris Descartes (Comité de Protection des Personnes Ile-de-France II, IRB registration number 00001073, protocol 20121300001072) and all participants gave written informed consent in line with the Declaration of Helsinki.

Experimental Setup

The experimental setup used here to test our hypothesis was a modified version of the one employed in our previous studies [6], [7]. The system consisted of the following elements: a motion-analysis system with active markers (CODA; Charnwood Dynamics), that was used to measure the three-dimensional position of 27 infrared LEDs in real time (submillimeter accuracy, 200 Hz sampling frequency). Eight markers were distributed ∼10 cm apart on the surface of stereo virtual-reality goggles (nVisor sx60, NVIS) worn by the subjects, eight on the surface of tools (350 g, isotropic inertial moment around the roll axis) that were attached to each of the subjects’ hands and three attached to a fixed reference frame placed in the laboratory. For the goggles and the tools, a numerical model of the relative positions of the LEDs was implemented in advance, so that an optimal matching algorithm could be used to effectively and robustly estimate the position and the orientation of the object, even in case of partially hidden markers. We exploited the redundancy of the high number of markers on the helmet and on the tools to reduce errors in the position and orientation estimation, resulting in a standard error in the measured viewpoint orientation below the visual resolution of the goggles (0.078°). To minimize the effect of the noise and computational delays of the system, a predictive Kalman filter was applied to the angular coordinates of the objects.

The virtual environment consisted of a cylindrical horizontal tunnel whose walls were characterized by longitudinal marks parallel to the tunnel axis (Figure 1). These marks helped the subjects to perceive their own spatial orientation in the virtual world. Identification of the visual vertical was facilitated by the fact that the marks went from white on the ceiling to black on the floor. The real-time position and orientation of the goggles were then used to update, at 50 Hz, the visual scene viewed by the subject in the virtual environment. Data from markers on the tools attached to the hands were used when necessary (i.e. during training) to place a representation of the hand in the scene and to record the subjects’ movements.

Experimental Procedure

In both INTRA and INTER experimental conditions the trial started with the subject’s head upright. Then automatic auditory commands asked the subject to raise one hand in front of him (the left or the right hand in the INTER or INTRA condition respectively). At this point the walls of the tunnel started changing color depending on the hand prono-supination angle. The color went from red to green as the hand approached the orientation that had to be memorized. Once the subject achieved the hand desired orientation, he or she had 2.5 sec to memorize it, after which the walls became insensitive to the hand orientation and the subject was instructed to lower the hand. After the target was acquired, subjects had 5 sec to tilt their head by 15°, to the right or to the left, depending on the trial. To guide subjects to the desired inclination of the head, audio feedback was provided: a sound with a left-right balance corresponding to the direction of the desired head inclination decreased in volume as the head approached 15°. If the subject was not able to turn off the sound within 5 sec, the trial was interrupted and was repeated later on. If he/she was able to reach the desired head inclination, after the 5 sec delay period that included the head roll movement, a signal was given to the subject to align the unseen right hand with the remembered target orientation and to validate the response by pressing a pedal with a foot.

Sensory Conflict During Tilting of the Head: Tracking the virtual-reality goggles was normally used to hold the visual scene stable with respect to the real world during movements of the head. But in half of the trials we generated a gradual, imperceptible conflict such that when the head tilted 15°, the subject received visual information corresponding to a rotation of 24°. The amplitude of the angle between the visual vertical and gravity varied proportionally (0.6 times) with the actual head tilt with respect to gravity, so that when the head was straight there was no conflict, and when the head was tilted to 15°, it was about 9°. At the end of the experiment the experimenter explicitly interviewed the subjects about the conflict perception. None of the subjects in this experiment reported to have noticed the tilt of the visual scene.

Participants

After giving written informed consent, 16 subjects participated in this study: 8 male and 8 female, age 23±3 years (mean±standard deviation). All subjects performed both experimental conditions, and to compensate for possible order effects, half of them started with the INTRA condition and the other half with the INTER condition. For each of the two experimental conditions the subjects performed 56 trials: two for each combination of seven target orientations (−45°, −30°, −15°, 0°, +15°, +30°, +45°), two head inclinations (±15°), and two levels of conflict (no and yes).

Data Analysis

We analyzed the recorded data in terms of errors (Inline graphic) made at the moment of the response in aligning the hand with respect to the memorized target orientation. To quantify the specific effect of the sensory conflict in each condition, we first corrected for any global rotation of responses that might occur, for instance, due to possible Muller or Aubert effects (for review see [46]), independent from the tilt of the visual scene in the conflict situation. To do so, we subtracted from all values of Inline graphic involving a head tilt to the right mean of such values obtained in the absence of sensory conflict and we did the same for leftward tilt, on a subject-by-subject basis. This allowed combining the trials with right and left head tilts. Next, for each value of Inline graphic obtained with conflict we computed the relative deviation (Inline graphic) from the mean of all responses without conflict expressed as percentage of the expected deviation if only visual information was used, taking into account the actual amount of head tilt measured during the response phase of each trial.

graphic file with name pone.0068438.e100.jpg (1)

Finally, we computed the mean value of these relative deviations, which is a direct measure of the overall weight given to the visual versus other sources of information, for each subject and for each condition.

graphic file with name pone.0068438.e101.jpg (2)

The variability of the performance of each individual subject in the trials without conflict was also evaluated. To robustly estimate each subject’s precision, despite the fact that only two responses could be used for each combination of target orientation and head inclination, the following procedure was used. First, the responses with the head tilted to the right and to the left were combined by compensating for possible Aubert-Muller effect, as reported above. Then, the variance, Inline graphic, of the responses obtained for each of the 14 combinations of target and head orientation were combined as reported in the following equation to obtain the global standard deviation [47].

graphic file with name pone.0068438.e103.jpg (3)

where Inline graphic is the number of responses of the Inline graphic combination of target inclination and head orientation.

We used ANOVA for repeated measures on the values of response deviations and variability obtained with the methods described above to compare the INTRA and INTER conditions. The variability data, Inline graphic, were transformed by the function Inline graphic, before performing the ANOVA [48]. To test whether the response deviations induced by the conflict were significantly different from the purely kinesthetic response (0Inline graphic), one-tailed Student’s t-test was used. To test whether the difference for response deviation between INTER and INTRA condition predicted by the models and experimentally observed differ significantly, two-tailed Student’s t-test was used.

Mathematical Modeling

We evaluated our experimental results in the context of recent models by which sensory signals are combined and compared based on the principles of maximum likelihood. According to MLP, two signals (Inline graphic and Inline graphic) that are statistically independent will be optimally combined (Inline graphic) by assigning weights to each signal based on the relative variance between them:

graphic file with name pone.0068438.e112.jpg (4)
graphic file with name pone.0068438.e113.jpg (5)

Similarly if three independent signals (Inline graphic, Inline graphic and Inline graphic) have to be combined (Inline graphic) the optimal weights are given by the following equations:

graphic file with name pone.0068438.e118.jpg (6)
graphic file with name pone.0068438.e119.jpg (7)
graphic file with name pone.0068438.e120.jpg (8)

But if one signal is reconstructed from another, the two signals will not be independent. In this case the MLP equations used to determine the relative weights must be modified to take into account the covariance between signals, as follows. Let Inline graphic and Inline graphic be two variables where each is the sum of a independent components Inline graphic and Inline graphic, and a common component Inline graphic.

graphic file with name pone.0068438.e126.jpg (9)
graphic file with name pone.0068438.e127.jpg (10)

This additive formulation is representative of the computation required to shift two signals into a common reference frame. The variance of each variable is simply the sum of the variances of each component:

graphic file with name pone.0068438.e128.jpg (11)
graphic file with name pone.0068438.e129.jpg (12)

while the covariance between x and y is simply equal to the variance of the common component c:

graphic file with name pone.0068438.e130.jpg (13)

It can be easily demonstrated from basic principles (see Equations S1–S16 in online Supporting Information and [23]) that the relative weight, Inline graphic and Inline graphic, which minimizes the variance of the combination of Inline graphic and Inline graphic: Inline graphic is:

graphic file with name pone.0068438.e136.jpg (14)
graphic file with name pone.0068438.e137.jpg (15)

and substituting Eqs. 11, 12 and 13:

graphic file with name pone.0068438.e138.jpg (16)
graphic file with name pone.0068438.e139.jpg (17)

To optimally weight three quantities Inline graphic, Inline graphic and Inline graphic, each of which is the sum of an independent component Inline graphic, Inline graphic and Inline graphic, respectively, and a common component Inline graphic, one finds a similar result:

graphic file with name pone.0068438.e147.jpg (18)
graphic file with name pone.0068438.e148.jpg (19)
graphic file with name pone.0068438.e149.jpg (20)

One can see from this derivation that computing the optimum weighting of Inline graphic and Inline graphic, or of Inline graphic, Inline graphic and Inline graphic, consists of computing the relative weights based on the variance of the independent components of each variable (Inline graphic, Inline graphic, Inline graphic), leaving out the variance of the common component Inline graphic. The standard deviation of the responses expected of each model computed is.

graphic file with name pone.0068438.e159.jpg (21)

or

graphic file with name pone.0068438.e160.jpg (22)

depending on the structure of the model.

To illustrate the importance of taking into account co-variation of signals, and to compare with previous studies, we applied the MLP equations both ways, deliberately ignoring the co-variation between signals using Eqs. 4–5 for model A and Eqs. 6–8 for model B, and by correctly applying MLP in the case of co-variance using Eqs. 16–17 for model A’ and Eqs. 18–20 for model B’. In each case we attempted to fit the model parameters to the experimental data in order to estimate the ability of the different model formulations to predict the experimental results. Specifically, we searched for the set of free parameters, Inline graphic, Inline graphic, Inline graphic for each model (see Table 1), that, in conjunction with the modeling assumptions listed in the section entitled ‘theoretical modeling’ above, would minimize the difference between the actual and predicted responses deviations (Eq. 2 and Eq. 23 respectively), where the predicted deviation expressed as a percentage is given by:

graphic file with name pone.0068438.e164.jpg (23)

and would minimize the difference between the actual and predicted response standard deviations (Eq. 3 and Eq. 21–22 respectively), simultaneously in both conditions (INTER and INTRA). The four data points were fit simultaneously by minimizing the weighted sum of the square of the differences between the experimental data and the model predictions for each data point, with the respective weights corresponding to the inverse of the squared confidence interval for each data point.

Supporting Information

Supporting Information S1

Equations for optimal sensory weighting in case of correlated signals.

(PDF)

Acknowledgments

We gratefully acknowledge the support of the Paris Descartes Platform for Sensorimotor Studies (Université Paris Descartes, CNRS, INSERM, Région Île-de-France) for the performance of the experimental work.

Funding Statement

This work was supported by the French space agency (Centre national d'études spatiales). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1. Pouget A, Ducom JC, Torri J, Bavelier D (2002) Multisensory spatial representations in eye-centered coordinates for reaching. Cognition 83: B1–11. [DOI] [PubMed] [Google Scholar]
  • 2. Sober SJ, Sabes PN (2005) Flexible strategies for sensory integration during motor planning. Nat Neurosci 8: 490–497. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3. Sarlegna FR, Sainburg RL (2007) The effect of target modality on visual and proprioceptive contributions to the control of movement distance. Exp Brain Res 176: 267–280. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4. McGuire LMM, Sabes PN (2009) Sensory transformations and the use of multiple reference frames for reach planning. Nat Neurosci 12: 1056–1061. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5. Jones SAH, Cressman EK, Henriques DYP (2010) Proprioceptive localization of the left and right hands. Exp Brain Res 204: 373–383. [DOI] [PubMed] [Google Scholar]
  • 6. Tagliabue M, McIntyre J (2011) Necessity is the mother of invention: Reconstructing missing sensory information in multiple, concurrent reference frames for eye-hand coordination. J Neurosci 31: 1397–1409. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7. Tagliabue M, McIntyre J (2012) Eye-hand coordination when the body moves: Dynamic egocentric and exocentric sensory encoding. Neurosci Lett 513: 78–83. [DOI] [PubMed] [Google Scholar]
  • 8. van Beers RJ, Sittig AC, Denier van der Gon JJ (1996) How humans combine simultaneous proprioceptive and visual position information. Exp Brain Res 111: 253–261. [DOI] [PubMed] [Google Scholar]
  • 9. van Beers RJ, Sittig AC, Gon JJ (1999) Integration of proprioceptive and visual position-information: An experimentally supported model. J Neurophysiol 81: 1355–1364. [DOI] [PubMed] [Google Scholar]
  • 10.Ghahramani Z, Wolpert DM, Jordan MI (1997) Computational models of sensorimotor integration. In: Morasso P, Sanguineti V, editors. Self-organization, Computational Maps, and Motor Control. Volume 119 of Advances in Psychology. Oxford: Elsevier. 117–147. doi:10.1016/S0166–4115(97)80006–4.
  • 11. Ernst MO, Banks MS (2002) Humans integrate visual and haptic information in a statistically optimal fashion. Nature 415: 429–433. [DOI] [PubMed] [Google Scholar]
  • 12. Kording KP, Wolpert DM (2004) Bayesian integration in sensorimotor learning. Nature 427: 244–247. [DOI] [PubMed] [Google Scholar]
  • 13. Kersten D, Mamassian P, Yuille A (2004) Object perception as bayesian inference. Annu Rev Psychol 55: 271–304. [DOI] [PubMed] [Google Scholar]
  • 14. Burns JK, Blohm G (2010) Multi-sensory weights depend on contextual noise in reference frame transformations. Front Hum Neurosci 4: 221. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15. Bays PM, Wolpert DM (2007) Computational principles of sensorimotor control that minimize uncertainty and variability. J Physiol 578: 387–396. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16. O’Reilly JX, Jbabdi S, Behrens TEJ (2012) How can a bayesian approach inform neuroscience? Eur J Neurosci 35: 1169–1179. [DOI] [PubMed] [Google Scholar]
  • 17. Sabes PN (2011) Sensory integration for reaching: models of optimality in the context of behavior and the underlying neural circuits. Prog Brain Res 191: 195–209. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18. Pouget A, Deneve S, Duhamel JR (2002) A computational perspective on the neural basis of multisensory spatial representations. Nat Rev Neurosci 3: 741–747. [DOI] [PubMed] [Google Scholar]
  • 19. Soechting JF, Flanders M (1989) Errors in pointing are due to approximations in sensorimotor transformations. J Neurophysiol 62: 595–608. [DOI] [PubMed] [Google Scholar]
  • 20. Tillery SI, Flanders M, Soechting JF (1991) A coordinate system for the synthesis of visual and kinesthetic information. J Neurosci 11: 770–778. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21. Schlicht EJ, Schrater PR (2007) Impact of coordinate transformation uncertainty on human sensorimotor control. J Neurophysiol 97: 4203–4214. [DOI] [PubMed] [Google Scholar]
  • 22. McIntyre J, Stratta F, Lacquaniti F (1998) Short-term memory for reaching to visual targets: psychophysical evidence for body-centered reference frames. J Neurosci 18: 8423–8435. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23. Oruc I, Maloney LT, Landy MS (2003) Weighted linear cue combination with possibly correlated error. Vision Res 43: 2451–2468. [DOI] [PubMed] [Google Scholar]
  • 24. van Beers RJ, Sittig AC, van der Gon JJD (1998) The precision of proprioceptive position sense. Exp Brain Res 122: 367–377. [DOI] [PubMed] [Google Scholar]
  • 25. Goble DJ, Brown SH (2008) Upper limb asymmetries in the matching of proprioceptive versus visual targets. J Neurophysiol 99: 3063–3074. [DOI] [PubMed] [Google Scholar]
  • 26. Adamovich S, Berkinblit M, Fookson O, Poizner H (1998) Pointing in 3d space to remembered targets. i. kinesthetic versus visual target presentation. J Neurophysiol 79: 2833–2846. [DOI] [PubMed] [Google Scholar]
  • 27. Baud-Bovy G, Viviani P (1998) Pointing to kinesthetic targets in space. J Neurosci 18: 1528–1545. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28. Baud-Bovy G, Viviani P (2004) Amplitude and direction errors in kinesthetic pointing. Exp Brain Res 157: 197–214. [DOI] [PubMed] [Google Scholar]
  • 29. Rao AK, Gordon AM (2001) Contribution of tactile information to accuracy in pointing movements. Exp Brain Res 138: 438–445. [DOI] [PubMed] [Google Scholar]
  • 30. Soechting JF, Tillery SIH, Flanders M (1990) Transformation from head-to shoulder-centered representation of target direction in arm movements. J Cognitive Neuroscience 2: 32–43. [DOI] [PubMed] [Google Scholar]
  • 31. Flanders M, Tillery SIH, Soechting JF (1992) Early stages in a sensorimotor transformation. Behavioral and Brain Sciences 15: 309–320. [Google Scholar]
  • 32. Fiehler K, Rsler F, Henriques DYP (2010) Interaction between gaze and visual and proprioceptive position judgements. Exp Brain Res 203: 485–498. [DOI] [PubMed] [Google Scholar]
  • 33. Appelle S (1972) Perception and discrimination as a function of stimulus orientation: the “oblique effect” in man and animals. Psychol Bull 78: 266–278. [DOI] [PubMed] [Google Scholar]
  • 34. Mon-Williams M, Wann JP, Jenkinson M, Rushton K (1997) Synaesthesia in the normal limb. Proc Biol Sci 264: 1007–1010. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35. Henriques DY, Klier EM, Smith MA, Lowy D, Crawford JD (1998) Gaze-centered remapping of remembered visual space in an open-loop pointing task. J Neurosci 18: 1583–1594. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36. Beurze SM, Pelt SV, Medendorp WP (2006) Behavioral reference frames for planning human reaching movements. J Neurophysiol 96: 352–362. [DOI] [PubMed] [Google Scholar]
  • 37. Buneo CA, Jarvis MR, Batista AP, Andersen RA (2002) Direct visuomotor transformations for reaching. Nature 416: 632–636. [DOI] [PubMed] [Google Scholar]
  • 38. McIntyre J, Lipshits M (2008) Central processes amplify and transform anisotropies of the visual system in a test of visual-haptic coordination. J Neurosci 28: 1246–1261. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39. Jones SAH, Fiehler K, Henriques DYP (2012) A task-dependent effect of memory and hand-target on proprioceptive localization. Neuropsychologia 50: 1462–1470. [DOI] [PubMed] [Google Scholar]
  • 40. Charron JF, Collin I, Braun CM (1996) Intermanual transfer of somaesthetic information: a two-point discrimination experiment. Neuropsychologia 34: 873–877. [DOI] [PubMed] [Google Scholar]
  • 41. Nefs HT, Kappers AML, Koenderink JJ (2005) Intermanual and intramanual tactual grating discrimination. Exp Brain Res 163: 123–127. [DOI] [PubMed] [Google Scholar]
  • 42. Fabri M, Polonara G, Del Pesce M, Quattrini A, Salvolini U, et al. (2001) Posterior corpus callosum and interhemispheric transfer of somatosensory information: an fmri and neuropsychological study of a partially callosotomized patient. J Cogn Neurosci 13: 1071–1079. [DOI] [PubMed] [Google Scholar]
  • 43. Fabri M, Del Pesce M, Paggi A, Polonara G, Bartolini M, et al. (2005) Contribution of posterior corpus callosum to the interhemispheric transfer of tactile information. Cogn Brain Res 24: 73–80. [DOI] [PubMed] [Google Scholar]
  • 44. Sternad D, Wei K, Diedrichsen J, Ivry RB (2007) Intermanual interactions during initiation and production of rhythmic and discrete movements in individuals lacking a corpus callosum. Exp Brain Res 176: 559–574. [DOI] [PubMed] [Google Scholar]
  • 45. van der Horst BJ, Duijndam MJA, Ketels MFM, Wilbers MTJM, Zwijsen SA, et al. (2008) Intra-manual and intermanual transfer of the curvature aftereffect. Exp Brain Res 187: 491–496. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 46.Howard I (1982) Human visual orientation. New York: Wiley.
  • 47. McIntyre J, Stratta F, Lacquaniti F (1997) Viewer-centered frame of reference for pointing to memorized targets in three-dimensional space. J Neurophysiol 78: 1601–1618. [DOI] [PubMed] [Google Scholar]
  • 48. Luyat M, Mobarek S, Leconte C, Gentaz E (2005) The plasticity of gravitational reference frame and the subjective vertical: peripheral visual information affects the oblique effect. Neurosci Lett 385: 215–219. [DOI] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Supporting Information S1

Equations for optimal sensory weighting in case of correlated signals.

(PDF)


Articles from PLoS ONE are provided here courtesy of PLOS

RESOURCES