Skip to main content
Proceedings of the Royal Society B: Biological Sciences logoLink to Proceedings of the Royal Society B: Biological Sciences
. 2018 Aug 8;285(1884):20180638. doi: 10.1098/rspb.2018.0638

Perceptual teleology: expectations of action efficiency bias social perception

Matthew Hudson 1,†,‡,, Katrina L McDonough 1,, Rhys Edwards 1, Patric Bach 1,
PMCID: PMC6111183  PMID: 30089623

Abstract

Primates interpret conspecific behaviour as goal-directed and expect others to achieve goals by the most efficient means possible. While this teleological stance is prominent in evolutionary and developmental theories of social cognition, little is known about the underlying mechanisms. In predictive models of social cognition, a perceptual prediction of an ideal efficient trajectory would be generated from prior knowledge against which the observed action is evaluated, distorting the perception of unexpected inefficient actions. To test this, participants observed an actor reach for an object with a straight or arched trajectory on a touch screen. The actions were made efficient or inefficient by adding or removing an obstructing object. The action disappeared mid-trajectory and participants touched the last seen screen position of the hand. Judgements of inefficient actions were biased towards the efficient prediction (straight trajectories upward to avoid the obstruction, arched trajectories downward towards the target). These corrections increased when the obstruction's presence/absence was explicitly acknowledged, and when the efficient trajectory was explicitly predicted. Additional supplementary experiments demonstrated that these biases occur during ongoing visual perception and/or immediately after motion offset. The teleological stance is at least partly perceptual, providing an ideal reference trajectory against which actual behaviour is evaluated.

Keywords: representational momentum, action prediction, prediction errors, teleological reasoning, motion perception, social perception

1. Introduction

Human and non-human primates take the ‘intentional stance’ when watching conspecifics [1], interpreting their behaviour as purposeful and goal-directed [26]. Crucial to this is the understanding that others' actions are optimized to achieve their goals in the most efficient and rational way, minimizing time and energy expenditure, given the environmental constraints. Both human infants and macaque monkeys, for example, show surprise when intentional agents do not attempt to avoid an obstacle, or take an unnecessary long way to reach their goal [5,7]. This simple efficient action heuristic can provide a foundation for the development of sophisticated capacities for mentalizing and theory of mind in adult humans (e.g. [5,8]). For example, seeing a seemingly inefficient action (e.g. a reach straight for an object despite an obstacle in the way) can prompt the insight that others may act according to beliefs that differ from one's own (i.e. they may not have seen the obstacle). Indeed, seeing such actions captures attention [9] and alters activity in brain areas implicated in action perception and mentalizing (e.g. [10,11]).

Yet, despite the crucial role of teleological/intentional reasoning in human and animal social cognition, little is known about the underlying processes. The currently dominant view sees social perception as a bottom-up ‘resonance’ of one's own motor apparatus with others' actions, which allows the associated goals and internal states (sensations and emotions) to be derived (e.g. [12,13]). Action efficiency would, in such a model, be conceptualized as a post hoc motoric signal of effort or energy expenditure, which can be compared with a reference value for this type of action (e.g. [14]). However, such models are challenged by findings that children make efficiency judgements for movements of biomechanically impossible actions for which motor resonance is unlikely [15], that they can process efficiency before acquiring competence in the seen actions (e.g. [16,17]) or that, in adults, eye movements indicate expectations of efficient action before action onset, when such kinematic information is not yet available [18].

An alternative is that teleological reasoning might not emerge from a ‘late’ motoric signal, but from earlier perceptual signals (e.g. [6]). Recent predictive coding frameworks argue that perception in general—and social perception in particular—is informed by prior expectations, derived from one's knowledge about the world and other people, and that these expectations guide processing of the perceptual input [1924]. Predictive influences have been demonstrated in a diverse range of perceptual abilities, including the perception of ‘true’ colour from surrounding illumination [25], anticipated effects of physical dynamics on motion perception [26] and three-dimensional (3D) concave/convexity from the presumed location of light sources [27]. In a similar way, the environment provides all the necessary information to generate an ideal reference trajectory that a fully rational, intentional actor would take to achieve their goal (i.e. location of goal objects and possible obstructions), and which would provide a comparison to immediately flag observed actions as being efficient or not, confirming prior attributions of goals and intentionality.

Here, we provide a first test of (i) whether human observers make such predictions of how rational actors who are aware of all environmental constraints efficiently traverse the given action space, (ii) whether these predictions are realized in a perceptual format that can serve as a reference image for the observed action, and (iii) whether this format biases the perceptual representation of the observed action. We rely on the well-established phenomenon that when a moving stimulus suddenly disappears, participants' estimations of its last seen position show robust distortions towards the expected path (i.e. representational momentum [28]), in line with the notion that the considerable uncertainty during motion perception is sharpened by top-down information (e.g. [29,30]), or that predicted paths are perceptually ‘filled in’ after the sudden offset [31]. Importantly, these distortions rely on changes to lower-level visual representations (e.g. [32,33]), occur even when participants are warned against them [34,35] and can integrate higher-level information such as the physical forces acting on the objects (e.g. momentum, friction and gravity; for a review, see [28]) or prior action expectations [3638].

Here, we use this paradigm to reveal the expectations of efficient action that guide the perception of others' actions. In three studies, participants watched an actor reach towards an object. The action disappeared mid-trajectory, and participants indicated the perceived disappearance point on a touch screen. In two conditions, the actions were efficient, showing a reach either straight towards the object or arched over an obstacle placed in between. In two other conditions, the actions were made inefficient by either adding an obstacle to the path of the straight reaches (such that the actor would knock into the obstacle) or removing the obstacle for the arched reaches (such that the actor reached over empty space). If others’ behaviour is perceived relative to what would be expected under the implicit assumption of efficient action, then the perceived kinematics should be displaced along the trajectory that an intentional, rational actor might take. Unexpected inefficient actions should be ‘corrected’ towards the predicted efficient action trajectory: straight reaches would be perceived upward if approaching an obstacle where an avoidance movement would be predicted, while an arched reach would be displaced downwards if made over empty space as this energy expenditure is unnecessary. Moreover, such distortions should be observed spontaneously when participants passively observe these actions, but should increase the more that attention is drawn to the environmental constraints and the behaviour of a rational actor. As an additional between-subjects task manipulation, we therefore varied whether the actions were viewed under no additional instructions (no task), or whether participants were asked to report ‘yes’ or ‘no’ in response to the presence of an obstacle prior to action onset (report obstacle) or to predict whether a rational actor would ideally have to reach ‘straight’ or ‘over’ an obstacle before the action started (predict trajectory). Further experiments showed that these deviations are not observed when static non-action stimuli have to be judged in the same scenes (electronic supplementary material, Experiment 1), that they generalize to probe judgements tasks that do not require working memory or touch screen responses (electronic supplementary material, Experiment 2) and that they are substantially reduced through dynamic visual noise masks that disrupt recurrent interactions between early visual areas and top-down information (electronic supplementary material, Experiment 3).

2. Material and methods

(a). Participants

Eighty-five participants took part (mean age = 24 years, s.d. = 7.7, 62 females; no task: n = 30, report obstacle: n = 27, predict trajectory: n = 28). Seven additional participants were excluded due to performance (see Results). All participants were right-handed, had normal/corrected vision, were recruited from Plymouth University and wider community, and received course credit or payment. The study received ethical approval from the University of Plymouth's ethics board, in accordance with those of the ESRC and the Declaration of Helsinki. A priori power analyses of previous experiments investigating similar effects with the same method ([36], Experiment 3) revealed that a sample size of 14 is required to achieve power of 0.95.

(b). Apparatus

Stimuli were filmed with a Sony HD video camera at 50 f.p.s. with a widescreen aspect ratio (16 : 9) and a resolution of 1920 × 1080 (2.1 megapixels) and edited with Adobe Photoshop. The experiment was delivered using Presentation (NeuroBS) via an NEC Multisync P221w LCD touch screen monitor (1680 × 1050). Verbal responses for the report obstacle and predict trajectory conditions were recorded using Presentation's sound threshold logic via a Logitech PC120 combined microphone and headphone set.

(c). Stimuli

Example stimuli can be seen in figure 1a. Videos were filmed of an arm starting in a rest position at the right of the screen and reaching to grasp a target object on the left (either an apple, bottle, crisps, glue stick or stapler). In the original set of videos, the actor's reach was either (i) unobstructed and the trajectory of the arm was straight towards the target object (straight/efficient), (ii) obstructed by one of four objects (iPad, lamp, pencil holder or photo-frame) and the trajectory of the arm was arched over the obstruction (arched/efficient). From each video, 19 frames were extracted for the experimental stimuli, beginning with the onset of movement (frame 1) to mid-way through the action (frame 19). Inefficient action sequences were created by digitally removing the obstructing objects in the arched/efficient videos (arched/inefficient). For each of the straight/efficient actions, a new set of videos were created by adding each of the obstructing objects to show the actor was reaching straight for the target, despite the obstruction (straight/inefficient). This created a set of inefficient actions that were identical to the efficient actions in terms of movement kinematics, and differed only by the presence/absence of the obstructing object. Finally, for each action, a single frame was created in which the hand was digitally removed. This served as a response stimulus at the end of each trial where participants estimated the disappearance point of the action.

Figure 1.

Figure 1.

Stimulus conditions and trial sequence. The stimulus conditions are depicted in (a). The action trajectory was either straight (i,ii) or arched over (iii,iv). The presence or absence of an obstructing object made the action trajectory either efficient (i,iii) or inefficient (ii,iv). In all examples, the hand is in the initial start position, and the white markers depict the final four frames of the trajectory of the index finger tip. The action sequence disappeared at one of these four points. An example trial sequence is depicted in (b), depicting an efficient arched trajectory over an obstruction. (Online version in colour.)

(d). Procedure

Participants completed two blocks of 80 trials in which each combination of action trajectory (straight and arched) and efficiency (efficient and inefficient) was represented by 20 trials. Participants were instructed that, on each trial, they would see an actor reach from the right of the screen for a target object on the far left, but that sometimes there would be a second object in between.

An example trial sequence can be seen in figure 1b. At the start of each trial, participants were instructed to ‘hold the spacebar’ and to keep it depressed to prevent them from tracking the observed action with their finger to improve performance. They then saw the first frame of the action sequence as a static image. In the no task condition, the action sequence began after a randomly chosen delay of between 1000 and 3000 ms. In the report obstacle condition and predict trajectory conditions, it began 1000 ms after the participant's verbal response had been detected. In the report obstacle condition, participants said ‘no’ if there was no obstruction and ‘yes’ if there was an obstruction. In the predict trajectory condition, participants said ‘straight’ if there was no obstruction and ‘over’ if there was an obstruction. The action depicted the frame order progressing at three frame intervals for a total sequence of between four and seven frames for 80 ms each (e.g. frames 1–4–7–10–13–16–19). Starting frames and sequence length were randomly chosen on each trial to prevent memorization of the final position from the starting frame. The final frame was immediately replaced by the response stimulus, creating the impression that the hand simply disappeared from the scene. Participants released the spacebar and, with their right hand, touched the screen where they thought the final seen position of the tip of the index finger was. As soon as a response was registered, the next trial began. Debriefing during pilot testing in the report obstacle condition has established that this experimental set-up is relatively opaque to participants and therefore unlikely to capture demand effects (e.g. [39,40]). Participants believed the aim to be investigating speed of motor responses and/or sequence learning, but none of them mentioned action efficiency/rationality or the hypothesized effect of perceptual mis-localizations.

3. Results

Participants were excluded if the distance between the real and selected positions exceeded 3 s.d. of the sample mean (mean = 49.2 pixels, s.d. = 12.4, no exclusions), or if the correlation between the real and selected positions was more than 3 s.d. below the median r value (X-axis: median = 0.914, s.d. = 0.055; Y-axis: median = 0.901, s.d. = 0.077, four participants excluded). For each participant, individual trials were excluded if the response procedure was incorrect (spacebar released before the action offset, 4.1%), or if response initiation or execution times were less than 200 ms or more than 3 s.d. above the sample mean (5.1%, initiation: mean = 443.5 ms, s.d. = 84.4; execution: mean = 817.2 ms, s.d. = 230.4). Three additional participants were excluded for having an excessive number of trial exclusions (greater than 50%).

The real final screen coordinate of the tip of the index finger was subtracted from participants' selected screen coordinate on each trial. Analysis was conducted on this residual localization error, which provided a directional measure of how far, in pixels (px), participants' responses were displaced along the X- and Y-axes. An accurate response would produce a value of 0 on both axes. On the X-axis, positive values denote a rightward displacement (against the direction of motion) and negative values a leftward displacement. On the Y-axis, positive and negative values denote upward and downward displacements, respectively.

Overall, there was a significant leftward bias (X-axis: mean = −8.4 px, s.d. = 19.2, t84 = −4.01, p < 0.001, d = 0.61, 95% CI (−4.3, −12.5)), and a significant downward bias (Y-axis: mean = −15.1 px, s.d. = 15.0, t84 = −9.27, p < 0.001, d = 1.35, 95% CI (−11.9, −18.3)). The differences along the X- and Y-axes and the result of the one-sample t-test, for each experimental condition across all tasks and for each task individually, can be seen in electronic supplementary material, table S1 (figure 2a–d). These difference values were entered into a 2 × 2 × 3 mixed-measures ANOVA for the X- and Y-axes separately, with trajectory (arched and straight) and efficiency (efficient and inefficient) as within-subjects factors, and task (no task, report obstacle and predict trajectory) as a between-subjects factor.

Figure 2.

Figure 2.

The trajectory × efficiency interactions for each task condition are depicted in (a) no task, (b) report obstacle and (c) predict trajectory. The difference between the real final position and the selected final position is plotted for the X-axis and Y-axis. The centre of each plot represents the real final position on any given trial (0 px difference on each axis). (d) A descriptive representation of the data in real screen coordinates (collapsed across task conditions). The solid lines represent the mean real final position of the arched (white) and straight (black) trajectories for the four possible disappearance points. The selected screen coordinates for each trajectory are plotted for the efficient (filled line) and inefficient (dashed line) conditions. The data are placed over a spatially aligned backdrop of a representative stimulus image of the action start point with an obstructing object to provide a reference of how the data relate to the stimuli. (e) A comparison of the size of the Y-axis interaction in pixels, equivalent to the total amount by which inefficient actions were corrected towards a more efficient trajectory. Error bars depict 95% confidence intervals.

(a). Y-axis

The main prediction is that perceptual judgements of inefficient actions would be displaced towards the expected trajectory, that is, downwards for inefficient arched trajectories and upwards for inefficient straight trajectories. Indeed, the analysis revealed a main effect of efficiency (F1,82 = 12.04, p = 0.001, Inline graphic) that was qualified by an interaction of efficiency and trajectory (F1,82 = 136.2, p < 0.001, Inline graphic). As predicted, inefficient arched trajectories (−19.7 px) were displaced below efficient arched trajectories (−11.0 px, t84 = −9.33, p < 0.001, d = 0.47), and inefficient straight actions (−12.0 px) were displaced above efficient straight actions (−17.7 px, t84 = 8.51, p < 0.001, d = 0.44), despite the actual hand disappearance points being identical within each trajectory. Importantly, there was a three-way interaction between trajectory, efficiency and task (F2,82 = 10.6, p < 0.001, Inline graphic). The interaction effect was re-quantified as a single value for each participant ((arched/efficient − arched/inefficient) − (straight/efficient − straight/inefficient)) to reveal the total amount in pixels by which inefficient actions were corrected towards a more efficient trajectory for each task (figure 2e). Between-subjects t-tests on this interaction value—mathematically equivalent to the pairwise three-way interactions of trajectory, efficiency and task—show that the interaction was marginally larger in the predict trajectory condition than the report obstacle condition (t53 = −1.95, p = 0.057, d = 0.53), which in turn was significantly larger than in the no task condition (t55 = −2.81, p = 0.007, d = 0.73). Demonstrating the robustness of the interaction, exploratory analysis showed that the interaction of trajectory and efficiency was evident in all conditions with a corrected α-level of p = 0.017 (no task: F1,29 = 35.3, p < 0.001, Inline graphic; report obstacle: F1,26 = 44.5, p < 0.001, Inline graphic; predict trajectory: F1,27 = 57.7, p < 0.001, Inline graphic; figure 2a–c). There were no further main effects or interactions.

(b). X-axis

We did not have specific predictions about how action rationality would affect perceptual displacements on the X-axis, and the analysis indeed did not reveal either a main effect of efficiency (F1,82 = 0.837, p = 0.363, Inline graphic) or an interaction of efficiency and trajectory (F1,82 = 1.39, p = 0.242, Inline graphic). We report the remaining effects of no interest below, but due to α-inflation of unpredicted effects in an ANOVA [41], they should be considered exploratory and interpreted with caution. A main effect of task (F2,82 = 8.81, p < 0.001, Inline graphic) revealed a general leftward displacement in the no task (−16.8 px, t29 = −5.14, p < 0.001, d = 1.35) and predict trajectory conditions (−10.0 px, t27 = −3.43, p = 0.002. d = 0.86), but not in the report obstacle condition (9 px, t26 = 0.720, p = 0.478, d = 0.19). A main effect of trajectory (F1,82 = 1231.4, p < 0.001, Inline graphic) showed a leftward displacement for arched trajectories (−24.4 px, t84 = −11.0, p < 0.001, d = 1.7) and a rightward displacement for straight trajectories (7.7 px, t84 = 3.74, p < 0.001, d = 0.53), most likely reflecting the further right displaced centre of gravity of the straight arm configurations ([42]; see also [36]). Finally, an interaction between trajectory and task (F1,82 = 9.88, p < 0.001, Inline graphic) revealed that the trajectory effect was larger in the no task than in the report obstacle (t55 = 4.15, p < 0.001, d = 1.1) and predict trajectory conditions (t56 = 3.25, p = 0.002, d = 0.86), which did not differ from each other (t53 = 1.18, p = 0.243, d = 0.32). There were no further interactions (all values of p > 0.351).

(c). Non-biological stationary stimuli (see electronic supplementary material, Experiment 1)

An alternative explanation for the Y-axis interaction between trajectory and efficiency on perceptual judgements is that the presence of an obstacle reduced the amount of general downward displacement. In electronic supplementary material, Experiment 1 (reported in full in the electronic supplementary material), the hands were therefore replaced by a static geometric shape in the same position as the disappearance points of the task relevant index finger, which crucially could not be interpreted in terms of intentionality nor efficiency. However, here, the presence/absence of an obstacle had no effect on these perceptual non-action judgements.

(d). Replication with probe judgements (see electronic supplementary material, Experiment 2)

Electronic supplementary material, Experiment 2 (reported in full in the electronic supplementary material) confirmed that the perceptual shifts towards efficient actions could also be observed in a psychophysical task without motor or working memory components. Here, participants simply reported—with a press of a button—whether the index finger's seen disappearance was identical to a probe stimulus presented directly after action offset (250 ms gap to prevent masking), which could be displaced either subtly upwards or downwards from the real disappearance point. Mirroring the results of the main experiments, participants were more likely to misidentify probes displaced towards the predicted ‘ideal’ trajectory with the actually perceived disappearance point (F1,36 = 11.39, p = 0.002, Inline graphic). They more readily accepted downwards probes as the same as the last seen position of inefficient arched reaches, and upwards probes for inefficient straight reaches. This replication rules out that the effects emerge from perceptual changes to the action's representation in later working memory or motor control stages, and instead reveals a contribution to immediate perceptual processing, either during ongoing motion perception (e.g. [43,44]) or to perceptual ‘filling in’ in the brief interval directly after its sudden offset (e.g. [31]).

(e). Masking with dynamic visual noise (see electronic supplementary material, Experiment 3)

Electronic supplementary material, Experiment 3 (reported in full in the electronic supplementary material) tested whether the perceptual shifts towards the efficient trajectories can be disrupted with a short (560 ms) dynamic visual noise mask directly after action offset that was presented in 50% of trials, at the same frequency as the prior motion stimuli (80 ms per frame), while maintaining equivalent task demands as the main experiment. Such masks reliably disrupt lower-level perceptual processes [45,46], eliciting similar effects to transcranial magnetic stimulation of occipital cortices (e.g. [47]). They specifically interrupt re-entrant interactions between V1 and higher visual areas that are crucial for conscious access to a stimulus during actual perception (e.g. backwards masking [48,49]) or visual imagery (e.g. [50,51]). Indeed, while the non-mask trials fully replicated the perceptual shifts towards efficient actions, they were reduced to about one-third of their size in the masked trials (F1,26 = 8.89, p = 0.006, Inline graphic). These masking effects therefore further confirm that the perceptual biases in the main experiments reflect changes to early visual stimulus representation, specifically tying them to the re-entrant feedback required for stabilizing percepts in perception or visual imagery (e.g. [48,51]).

4. Discussion

The present study showed for the first time that the teleological interpretation humans have of others' behaviour is perceptually instantiated and provides a visual reference signal for an expected ‘ideal’ trajectory during action observation. Participants watched a hand reach for objects with either efficient or inefficient kinematics and reported its last position after it had suddenly disappeared. Across several samples, perceptual reports were consistently biased towards the ideal reference kinematics. Straight reaches were reported higher if there was an obstacle in the way, as if lifted to avoid it. Conversely, reaches with a high arched trajectory were reported lower if the path was clear. These biases were evident automatically, but became more pronounced when observers explicitly processed the potential obstacles that could constrain the action and particularly when they predicted the most efficient action kinematics through the scene.

Further experiments showed that the biases in perceptual judgements were action-specific and not elicited when locations of briefly presented non-biological stationary objects were judged in the same scenes (electronic supplementary material, Experiment 1). They could also be observed in a probe judgement task without working memory or motor components already at 250 ms after action offset (electronic supplementary material, Experiment 2), did not increase with longer response times (electronic supplementary material, response time (RT) analysis) and were effectively disrupted by dynamic visual noise masks (electronic supplementary material, Experiment 3), which interfere with the recurrent (top-down) feedback to early visual cortex [48], preventing its use in awareness (i.e. backwards masking, [46,47]) or visual imagery (e.g. [50,51]). The observed biases in perceptual judgements are therefore unlikely to stem from unspecific perceptual changes in memory or motor control (e.g. [40]; see [52] for an example for perceptual changes in action memory). Instead, they support a role in ongoing motion perception (e.g. [43]) or occurring directly after its sudden offset, when the predicted future trajectory is visually ‘filled in’ (e.g. [31]).

The results are also unlikely to reflect the cognitive/social demands of the experimental set-up (e.g. [39,40]). The rapid nature of action sequences and resulting touch responses precluded considered decision-making that takes into account the required pattern across experimental conditions. Moreover, the effect was replicated in a more complex and cognitively opaque probe judgement task (electronic supplementary material, Experiment 2), which is largely unaffected by conscious strategic manipulation [34,35]. Finally, even though cues to the research question were equivalent, the perceptual biases were successfully reduced by brief dynamic visual noise masks directly after stimulus offset (electronic supplementary material, Experiment 3), which are known to disrupt perceptual processes specifically [48,49].

Together, these results reveal that, during social perception, the principle of efficient action provides a similar perceptual reference signal to the assumption that light comes from above [27] or that gravity pulls objects downwards [53], constantly pushing the perceptual representation of inefficient actions towards a more rational path. The resulting biases in perceptual judgements cannot be accounted for by an abstract awareness of the action's goal, such as when eye movements jump towards an action's target [54] or perceptual judgements are biased towards them [3638]. Instead, they reveal concrete expectations of the specific trajectory that the action will take through the scene. Moreover, making this awareness explicit prior to action observation increased the perceptual bias. Action efficiency is therefore not only evaluated after an action has been completed [5,7,9,55], but constantly updated, at every step in the trajectory, by predictions that can be derived from contextual cues prior to the motor behaviour.

Our results support predictive coding frameworks of social perception [1924], which argue that social perception, like perception in general, is hypothesis-driven and guided by top-down expectations. In such models, observers constantly test their inferences about others' goals and beliefs by predicting how they would behave, and matching this prediction to—and integrating it with—the actual perceptual input. In such a view, predictions of efficient action can contribute to the perceptual sharpening of the visual uncertainty during action perception (i.e. motion blurring, [29]), or after its offset, constantly biasing perception towards the expected avoidance or straightening movement, with the amount of bias constrained by the visual uncertainty. In addition, they allow humans to rapidly confirm the intentionality of others' behaviour, only requiring a match of the observed actions to the ideal kinematics that would follow from these goals. Prediction errors, by contrast, would signal inefficient actions, triggering more sophisticated mentalizing processes to re-evaluate the actor's goal or how their beliefs may differ from one's own [5,10,11]. In this way, the relatively simple perceptual process of prediction and prediction error would not only support perception, but also provide a foundation for higher-level judgements about others' beliefs or intentions, even in cases in which motor experience is unlikely [1517].

An important question is at which level the present predictive biases on social perception arise. While attribution of goals is often seen as a higher-level process (e.g. [19,23,24]), the detection of intentionality has been argued to be a feature of perceptual processing itself, based on specific stimulus features that signal intentionality and which allow humans to ‘see’ the agency behind others' movements (for a review, see [6]). A perceptual expectation of efficient action could emerge directly from such low-level unconscious perceptual inferences. Indeed, a series of follow-up studies ([56]; pre-print at: https://doi.org/10.17605/OSF.IO/QWJTF) has linked the perceptual displacements to stimulus features that imply intentional action, being almost completely eliminated for moving stimuli that are not intentional (e.g. a ball) and do not follow the characteristic motion profile of intentional action towards objects (e.g. [57]). Such a perceptual origin is also consistent with the finding that, in humans, sensitivity to kinematic efficiency emerges early in development, that it is present in other primates [7] and that it is spared in individuals with autism spectrum conditions, for whom only more advanced mental state reasoning proves problematic [9,58]. The key leap to the sophisticated socio-cognitive abilities of humans may therefore lie in the abstraction of these lower-level content-based representations of goals, environment and action, to higher-order representations of desires, beliefs and intentions, respectively [5]. For example, it is clear that human infants from the age of 4 onwards are able to predict what others will do not based on the actual environmental constraints (toy is in box A), but in terms of what the actor believes the state of the environment to be (they believe it is in box B), suggesting that the bottleneck emerges at this later state that requires sophisticated coordination of representation, such as others’ beliefs that differ from one's own beliefs or objective reality [5961].

5. Conclusion

The principle of efficient action allows observers to perceive others' actions relative to ideal reference actions, thereby confirming prior goal attributions or revising them in the case of a conflict. Such perceptual mechanisms for efficiency perception support rapid attribution of intentionality and facilitate the perception of others’ behaviour and our interactions with them. The burden of social cognition is placed on mechanisms that account for unexpected behaviour through a re-evaluation of their beliefs, desire and intentions, so that our model of the social world can be refined, and predictions of other's behaviour can be made more accurately.

Supplementary Material

Supplementary Analysis, Experiments, and Tables
rspb20180638supp1.pdf (458.9KB, pdf)

Data accessibility

A summary of the data for each experiment is provided in an Excel file as the electronic supplementary material and available in the Dryad Digital Repository: https://doi.org/10.5061/dryad.m9j93jg [62].

Authors' contributions

All authors devised the experiment. Stimuli were created by K.L.M. and R.E. Data were collected by M.H., K.L.M. and R.E. Data were analysed by M.H., K.L.M. and R.E. The manuscript was written by M.H., K.L.M. and P.B. Supplementary Experiments 2 and 3 design, data collection, analysis and write up were by K.L.M. and P.B. All authors gave final approval for publication.

Competing interests

We declare we have no competing interests.

Funding

This work was supported by the Economic and Social Research Council (grant no. ES/J019178/1) awarded to P.B. and a Plymouth University PhD grant to K.L.M.

References

  • 1.Dennett DC. 1987. The intentional stance. Cambridge, MA: MIT Press. [Google Scholar]
  • 2.Baillargeon R, Scott RM, Bian L. 2016. Psychological reasoning in infancy. Annu. Rev. Psychol. 67, 159–186. ( 10.1146/annurev-psych-010213-115033) [DOI] [PubMed] [Google Scholar]
  • 3.Baker CL, Saxe R, Tenenbaum JB. 2009. Action understanding as inverse planning. Cognition 113, 329–349. ( 10.1016/j.cognition.2009.07.005) [DOI] [PubMed] [Google Scholar]
  • 4.Csibra G, Gergely G. 2007. Obsessed with goals: functions and mechanisms of teleological interpretation of actions in humans. Acta Psychol. 124, 60–78. ( 10.1016/j.actpsy.2006.09.007) [DOI] [PubMed] [Google Scholar]
  • 5.Gergely G, Csibra G. 2003. Teleological reasoning in infancy: the naïve theory of rational action. Trends Cogn. Sci. 7, 287–292. ( 10.1016/S1364-6613(03)00128-1) [DOI] [PubMed] [Google Scholar]
  • 6.Scholl BJ, Gao T. 2013. Perceiving animacy and intentionality: visual processing or higher-level judgment? In Social perception: detection and interpretation of animacy, agency, and intention (eds Rutherford MD, Kuhlmeier VA), pp. 197–230. Cambridge, MA: MIT Press. [Google Scholar]
  • 7.Rochat MJ, Serra E, Fadiga L, Gallese V. 2008. The evolution of social cognition: goal familiarity shapes monkeys' action understanding. Curr. Biol. 18, 227–232. ( 10.1016/j.cub.2007.12.021) [DOI] [PubMed] [Google Scholar]
  • 8.Wellman HM, Brandone AC. 2009. Early intention understandings that are common to primates predict children's later theory of mind. Curr. Opin Neurobiol. 19, 57–62. ( 10.1016/j.conb.2009.02.004) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Vivanti G, McCormick C, Young GS, Abucayan F, Hatt N, Nadig A, Ozonoff S, Rogers SJ. 2011. Intact and impaired mechanisms of action understanding in autism. Dev. Psychol. 47, 841–856. ( 10.1037/a0023105) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Desmet C, Brass M. 2015. Observing accidental and intentional unusual actions is associated with different subregions of the medial frontal cortex. Neuroimage 122, 195–202. ( 10.1016/j.neuroimage.2015.08.018.) [DOI] [PubMed] [Google Scholar]
  • 11.Marsh LE, Mullett TE, Ropar D, Hamilton AFC. 2014. Responses to irrational actions in action observation and mentalising networks of the human brain. Neuroimage 103, 81–90. ( 10.1016/j.neuroimage.2014.09.020) [DOI] [PubMed] [Google Scholar]
  • 12.Iacoboni M. 2009. Imitation, empathy, and mirror neurons. Annu. Rev. Psychol. 60, 653–670. ( 10.1146/annurev.psych.60.110707.163604) [DOI] [PubMed] [Google Scholar]
  • 13.Rizzolatti G, Sinigaglia C. 2010. The functional role of the parieto-frontal mirror circuit: interpretations and misinterpretations. Nat. Rev. Neurosci. 11, 264–274. ( 10.1038/nrn2805.) [DOI] [PubMed] [Google Scholar]
  • 14.Jara-Ettinger J, Gweon H, Tenenbaum JB, Schulz LE. 2015. Children's understanding of the costs and rewards underlying rational action. Cognition 140, 14–23. ( 10.1016/j.cognition.2015.03.006) [DOI] [PubMed] [Google Scholar]
  • 15.Southgate V, Johnson MH, Csibra G. 2008. Infants attribute goals even to biomechanically impossible actions. Cognition 107, 1059–1069. ( 10.1016/j.cognition.2007.10.002) [DOI] [PubMed] [Google Scholar]
  • 16.Gredeback G, Melinder A. 2010. Infants understanding of everyday social interactions: a dual process account. Cognition 114, 197–206. ( 10.1016/j.cognition.2009.09.004) [DOI] [PubMed] [Google Scholar]
  • 17.Sodian B, Schoeppner B, Metz U. 2004. Do infants apply the principle of rational action to human agents? Infant Behav. Dev. 27, 31–41. ( 10.1016/j.infbeh.2003.05.006) [DOI] [Google Scholar]
  • 18.Fischer MH, Prinz J, Lotz K. 2008. Grasp cueing shows obligatory attention to action goals. Q. J. Exp. Psychol. 61, 860–868. ( 10.1080/17470210701623738) [DOI] [PubMed] [Google Scholar]
  • 19.Bach P, Nicholson T, Hudson M. 2014. The affordance-matching hypothesis: how objects guide action understanding and prediction. Front. Hum. Neurosci. 8, 254 ( 10.3389/fnhum.2014.00254) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Bach P, Nicholson T, Hudson M. 2015. Pattern completion does not negate matching: a response to Uithol and Maranesi. Front. Hum. Neurosci. 9, 685 ( 10.3389/fnhum.2015.00685) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Bach P, Schenke KC. 2017. Predictive social perception: towards a unifying framework from action observation to person knowledge. Soc. Personal Psychol. Compass 11, e12312, 1–17. ( 10.1111/spc3.12312) [DOI] [Google Scholar]
  • 22.Csibra G. 2008. Action mirroring and action understanding: an alternative account. In Sensory motor foundations of higher cognition. Attention and performance XXII (eds Haggard P, Rosetti Y, Kawato M), pp. 435–459. ( 10.1093/acprof:oso/9780199231447.003.0020) [DOI] [Google Scholar]
  • 23.Kilner JM, Friston KJ, Frith CD. 2007. The mirror-neuron system: a Bayesian perspective. Neuroreport 18, 619–623. ( 10.1097/WNR.0b013e3281139ed000001756-200704160-00018) [DOI] [PubMed] [Google Scholar]
  • 24.Kilner JM, Friston KJ, Frith CD. 2007. Predictive coding: an account of the mirror neuron system. Cogn. Process 8, 159–166. ( 10.1007/s10339-007-0170-2) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Bloj MG, Kersten D, Hurlbert AC. 1999. Perception of three-dimensional shape influences colour perception through mutual illumination. Nature 402, 877–879. ( 10.1038/47245) [DOI] [PubMed] [Google Scholar]
  • 26.Sotiropoulos G, Seitz AR, Series P. 2011. Changing expectations about speed alters perceived motion direction. Curr. Biol. 21, R883–R884. ( 10.1016/j.cub.2011.09.013) [DOI] [PubMed] [Google Scholar]
  • 27.Adams WJ, Graf EW, Ernst MO. 2004. Experience can change the ‘light-from-above’ prior. Nat. Neurosci. 7, 1057–1058. ( 10.1038/nn1312) [DOI] [PubMed] [Google Scholar]
  • 28.Hubbard TL. 2005. Representational momentum and related displacements in spatial memory: a review of the findings. Psychon. Bull. Rev. 12, 822–851. ( 10.3758/BF03196775) [DOI] [PubMed] [Google Scholar]
  • 29.Hammett ST. 1997. Motion blur and motion sharpening in the human visual system. Vision Res. 37, 2505–2510. ( 10.1016/S0042-6989(97)00059-X) [DOI] [PubMed] [Google Scholar]
  • 30.Bex PJ, Edgar GK, Smith AT. 1995. Sharpening of drifting, blurred images. Vision Res. 35, 2539–2546. ( 10.1016/0042-6989(95)00060-D) [DOI] [PubMed] [Google Scholar]
  • 31.Ekman M, Kok P, de Lange FP. 2017. Time-compressed preplay of anticipated events in human primary visual cortex. Nat. Commun. 8, 15276 ( 10.1038/ncomms15276) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Senior C, Barnes J, Giampietroc V, Simmons A, Bullmore ET, Brammer M, David AS. 2000. The functional neuroanatomy of implicit-motion perception or ‘representational momentum’. Curr. Biol. 10, 16–22. ( 10.1016/S0960-9822(99)00259-6) [DOI] [PubMed] [Google Scholar]
  • 33.Senior C, Ward J, David AS. 2002. Representational momentum and the brain: an investigation into the functional necessity of V5/MT. Vis. Cogn. 9, 81–92. ( 10.1080/13506280143000331) [DOI] [Google Scholar]
  • 34.Courtney JR, Hubbard TL. 2008. Spatial memory and explicit knowledge: an effect of instruction on representational momentum. Q. J. Exp. Psychol. 61, 1778–1784. ( 10.1080/17470210802194217) [DOI] [PubMed] [Google Scholar]
  • 35.Ruppel SE, Fleming CN, Hubbard TL. 2009. Representational momentum is not (totally) impervious to error feedback. Can. J. Exp. Psychol./Rev. Can. Psychol. Exp. 63, 49–58. ( 10.1037/a0013980) [DOI] [PubMed] [Google Scholar]
  • 36.Hudson M, Bach P, Nicholson T. 2017. You said you would! The predictability of other's behavior from their intentions determines predictive biases in action perception. J. Exp. Psychol. Hum. Percept. Perform. 44, 320–335. ( 10.1037/xhp0000451) [DOI] [PubMed] [Google Scholar]
  • 37.Hudson M, Nicholson T, Ellis R, Bach P. 2016. I see what you say: prior knowledge of other's goals automatically biases the perception of their actions. Cognition 146, 245–250. ( 10.1016/j.cognition.2015.09.021) [DOI] [PubMed] [Google Scholar]
  • 38.Hudson M, Nicholson T, Simpson WA, Ellis R, Bach P. 2016. One step ahead: the perceived kinematics of others' actions are biased toward expected goals. J. Exp. Psychol. Gen. 145, 1–7. ( 10.1037/xge0000126) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39.Durgin FH, Baird JA, Greenburg M, Russell R, Shaughnessy K, Waymouth S. 2009. Who is being deceived? The experimental demands of wearing a backpack. Psychon. Bull. Rev. 16, 964–969. ( 10.3758/PBR.16.5.964) [DOI] [PubMed] [Google Scholar]
  • 40.Firestone C, Scholl BJ. 2016. Cognition does not affect perception: evaluating the evidence for ‘top-down’ effects. Behav. Brain Sci. 39, e229 ( 10.1017/S0140525X15000965) [DOI] [PubMed] [Google Scholar]
  • 41.Cramer AO, et al. 2016. Hidden multiplicity in exploratory multiway ANOVA: prevalence and remedies. Psychon. Bull. Rev. 23, 640–647. ( 10.3758/s13423-015-0913-5) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42.Coren S, Hoenig P. 1972. Effect of non-target stimuli upon length of voluntary saccades. Percept. Mot. Skills 34, 499–508. ( 10.2466/pms.1972.34.2.499) [DOI] [PubMed] [Google Scholar]
  • 43.Muckli L, Kohler A, Kriegeskorte N, Singer W. 2005. Primary visual cortex activity along the apparent-motion trace reflects illusory perception. PLoS Biol. 3, e265 ( 10.1371/journal.pbio.0030265) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 44.Yantis S, Nakama T. 1998. Visual interactions in the path of apparent motion. Nat. Neurosci. 1, 508–512. ( 10.1038/2226) [DOI] [PubMed] [Google Scholar]
  • 45.Kinsbourne M, Warrington EK. 1962. The effect of an after-coming random pattern on the perception of brief visual stimuli. Q. J. Exp. Psychol. 14, 223–234. ( 10.1080/17470216208416540) [DOI] [Google Scholar]
  • 46.Breitmeyer B, Öğmen H. 2006. Visual masking: time slices through conscious and unconscious vision. Oxford, UK: Oxford University Press; ( 10.1093/acprof:oso/9780198530671.001.0001) [DOI] [Google Scholar]
  • 47.Tapia E, Beck DM. 2014. Probing feedforward and feedback contributions to awareness with visual masking and transcranial magnetic stimulation. Front. Psychol. 5, 1173 ( 10.3389/fpsyg.2014.01173) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 48.Fahrenfort JJ, Scholte HS, Lamme VA. 2007. Masking disrupts reentrant processing in human visual cortex. J. Cogn. Neurosci. 19, 1488–1497. ( 10.1162/jocn.2007.19.9.1488) [DOI] [PubMed] [Google Scholar]
  • 49.Lamme VA, Zipser K, Spekreijse H. 2002. Masking interrupts figure-ground signals in V1. J. Cogn. Neurosci. 14, 1044–1053. ( 10.1162/089892902320474490) [DOI] [PubMed] [Google Scholar]
  • 50.Andrade J, Kemps E, Werniers Y, May J, Szmalec A. 2002. Insensitivity of visual short-term memory to irrelevant visual information. Q. J. Exp. Psychol. A 55, 753–774. ( 10.1080/02724980143000541) [DOI] [PubMed] [Google Scholar]
  • 51.Borst G, Ganis G, Thompson WL, Kosslyn SM. 2012. Representations in mental imagery and working memory: evidence from different types of visual masks. Mem. Cogn. 40, 204–217. ( 10.3758/s13421-011-0143-7) [DOI] [PubMed] [Google Scholar]
  • 52.Ianì F, Mazzoni G, Bucciarelli M. 2018. The role of kinematic mental simulation in creating false memories. J. Cogn. Psychol. 30, 1–15. ( 10.1080/20445911.2018.1426588) [DOI] [Google Scholar]
  • 53.Hubbard TL, Bharucha JJ. 1988. Judged displacement in apparent vertical and horizontal motion. Percept. Psychophys. 44, 211–221. ( 10.3758/BF03206290) [DOI] [PubMed] [Google Scholar]
  • 54.Eshuis R, Coventry KR, Vulchanova M. 2009. Predictive eye movements are driven by goals, not by the mirror neuron system. Psychol. Sci. 20, 438–440. ( 10.1111/j.1467-9280.2009.02317.x) [DOI] [PubMed] [Google Scholar]
  • 55.Scott RM, Baillargeon R. 2013. Do infants really expect agents to act efficiently? A critical test of the rationality principle. Psychol. Sci. 24, 466–474. ( 10.1177/0956797612457395) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 56.McDonough KL, Hudson M, Bach P. 2018. Seeing the best way forward: intentionality cues bias action perception toward the most efficient trajectory (doi:10.17605/OSF.IO/QWJTF)
  • 57.Beggs WDA, Howarth CI. 1972. The movement of the hand toward a target. Q J. Exp. Psychol. 24, 448–453. ( 10.1080/14640747208400304) [DOI] [PubMed] [Google Scholar]
  • 58.Marsh LE, Pearson A, Ropar D, Hamilton AFC. 2015. Predictive gaze during observation of irrational actions in adults with autism spectrum conditions. J. Autism Dev. Disord. 45, 245–261. ( 10.1007/s10803-014-2215-6) [DOI] [PubMed] [Google Scholar]
  • 59.Samson D, Apperly IA, Kathirgamanathan U, Humphreys GW. 2005. Seeing it my way: a case of a selective deficit in inhibiting self-perspective. Brain 128, 1102–1111. ( 10.1093/brain/awh464) [DOI] [PubMed] [Google Scholar]
  • 60.Santiesteban I, White S, Cook J, Gilbert SJ, Heyes C, Bird G. 2012. Training social cognition: from imitation to theory of mind. Cognition 122, 228–235. ( 10.1016/j.cognition.2011.11.004) [DOI] [PubMed] [Google Scholar]
  • 61.Steinbeis N. 2016. The role of self–other distinction in understanding others' mental and emotional states: neurocognitive mechanisms in children and adults. Phil. Trans. R. Soc. B 371, 20150074 ( 10.1098/rstb.2015.0074) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 62.Hudson M, McDonough KL, Edwards R, Bach P. 2018. Data from: Perceptual teleology: expectations of action efficiency bias social perception Dryad Digital Repository. ( 10.5061/dryad.m9j93jg) [DOI] [PMC free article] [PubMed]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Citations

  1. Hudson M, McDonough KL, Edwards R, Bach P. 2018. Data from: Perceptual teleology: expectations of action efficiency bias social perception Dryad Digital Repository. ( 10.5061/dryad.m9j93jg) [DOI] [PMC free article] [PubMed]

Supplementary Materials

Supplementary Analysis, Experiments, and Tables
rspb20180638supp1.pdf (458.9KB, pdf)

Data Availability Statement

A summary of the data for each experiment is provided in an Excel file as the electronic supplementary material and available in the Dryad Digital Repository: https://doi.org/10.5061/dryad.m9j93jg [62].


Articles from Proceedings of the Royal Society B: Biological Sciences are provided here courtesy of The Royal Society

RESOURCES