Abstract
Facial expressions are vital for social communication, yet the underlying mechanisms are still being discovered. Illusory faces perceived in objects (face pareidolia) are errors of face detection that share some neural mechanisms with human face processing. However, it is unknown whether expression in illusory faces engages the same mechanisms as human faces. Here, using a serial dependence paradigm, we investigated whether illusory and human faces share a common expression mechanism. First, we found that images of face pareidolia are reliably rated for expression, within and between observers, despite varying greatly in visual features. Second, they exhibit positive serial dependence for perceived facial expression, meaning an illusory face (happy or angry) is perceived as more similar in expression to the preceding one, just as seen for human faces. This suggests illusory and human faces engage similar mechanisms of temporal continuity. Third, we found robust cross-domain serial dependence of perceived expression between illusory and human faces when they were interleaved, with serial effects larger when illusory faces preceded human faces than the reverse. Together, the results support a shared mechanism for facial expression between human faces and illusory faces and suggest that expression processing is not tightly bound to human facial features.
Keywords: facial expression, emotion processing, serial dependence, pareidolia, rapid adaptation
1. Introduction
Facial expressions are one of the most powerful and universal methods we have for social communication [1–3]. Our ability to recognize facial expressions in others and understand the emotions they signify involves both affective and perceptual components which are still not wholly understood [1,4,5]. Faces capture our attention automatically [6], and emotional faces have been shown to have priority over neutral faces in numerous behavioural tasks [7–11]. Consistent with the prioritization of faces by the visual system, facial expression recognition is thought to be supported by specialized brain regions that respond to dynamic facial cues [12]. However, very little is understood about the tuning properties of the brain regions engaged in facial expression recognition, including which visual features are critical for this complex psychological judgement.
One significant challenge in understanding what drives the processing of facial expressions is in disambiguating which visual features determine expression recognition. As visual stimuli, facial expressions are incredibly rich, originating from the many intricate muscle movements which convey our internal emotional states [4,13]. There is evidence from several behavioural tasks that different visual features are relevant for recognizing different emotions [14–17] (e.g. eyebrows for ‘sad’, and the mouth for ‘happy’ [16]). What these approaches have in common is a goal of characterizing the key visual features that differentiate human expressions and relating them to expression recognition. In most experimental paradigms, this is achieved via manipulation of human faces, for example, by removing facial features [15,16,18] or morphing different facial expressions together into the same face [14]. Although this strategy has proven fruitful in revealing key differences between expressions that impact behavioural performance, these approaches are united in assuming that the local processing of visual features is fundamental in expression recognition. Further, it is not clear to what degree expression recognition is bound to the specific low-level visual features that define facial features and their associated muscle movements [13] in human faces.
Here, we take a complementary approach and examine global facial expression processing in non-human faces. Specifically, we examine expression in cases of face pareidolia, the perception of illusory facial features in inanimate objects. Face pareidolia is a spontaneous error of face detection which we share with other primates [19,20]. Human neuroimaging [21] and behavioural studies [22,23] have revealed that illusory faces share some mechanisms with human faces. However, there are notable differences in the neural representation of illusory and real faces, with the initial ‘face-like’ response to illusory faces resolving after only one-quarter of a second [21]. One behavioural manifestation of this difference has been observed in visual search—although observers are faster to find a target object in search displays when it contains an illusory face, they are even faster to find a human face [22]. Understanding the processing of expression in errors of face detection (pareidolia) is important because examples of pareidolia are an intriguing case in which the facial ‘expression’ occurs in the absence of any underlying muscle movement or human facial features. Consequently, it is not clear whether expression in pareidolia originates from the same mechanism as human faces.
To examine whether expression in illusory faces and human faces are processed via a common mechanism, here we use a rapid adaptation paradigm that has revealed serial dependencies in visual perception [24–26]. When a series of faces constructed from morphing multiple identities together is viewed in quick succession, the perception of a given face morph is biased towards the identity of previously viewed morphs [27]. In addition to identity [27,28], similar serial dependence effects have been demonstrated in faces for traits such as attractiveness [29–33], gender [34], eye gaze [35] and expression [34,36]. Serial dependence is thought to reflect an adaptive process that promotes continuity in our perception of the physical world which is largely stable despite fluctuations in viewing conditions [24,25]. Importantly, as with many forms of visual adaptation, serial dependence generally requires a degree of similarity between the current and preceding stimuli. For example, expression judgements show serial dependence only for faces of the same sex [36], and both identity and attractiveness judgements show serial dependence only for faces presented at the same orientation [28,29]. Examples of face pareidolia are much more visually diverse than human faces, with different features of objects defining the illusory facial ‘features’ in each example. Consequently, it is not clear whether face pareidolia will show serial dependence for expression. If it does, this would indicate that illusory face perception engages similar mechanisms of temporal continuity as real human faces. Additionally, if cross-domain serial adaptation occurs between human faces and illusory faces, this would be evidence for a common mechanism. In a series of experiments, we test these ideas using examples of illusory faces and human faces.
2. Methods
(a) . Participants
A total of 17 university students (five male, 12 female) participated in these experiments. Fourteen did both Experiment 1 and Experiment 2. One did Experiment 1 only and two did Experiment 2 only. Therefore, Experiments 1 and 2 had samples sizes of 15 and 16, respectively. All participants were naive to the purpose of the experiments and were paid $AU20 per hour for their participation. All participants signed written consent and all procedures were approved by the Human Research Ethics Committee of the University of Sydney.
(b) . Apparatus and stimuli
The experiment was programmed with MATLAB software which displayed face images on a standard PC monitor (60 Hz refresh rate, resolution 1024 × 768). The stimuli were 40 real face images and 40 inanimate object images which elicited strong pareidolia percepts. All were RGB images with dimensions 400 by 400 pixels, corresponding to 10.6° by 10.6° of visual angle when viewed from 57 cm. Figure 1 displays examples of the faces used in the study and the full list of faces is available for download. Image stimuli were presented for 250 ms and a mouse-controlled rating bar 400 pixels in length located below the images was used to rate the strength of emotion in the image just seen (figure 1). A roller mouse was used to control the slider along the rating bar and the space bar recorded the participant's response.
Some of the illusory and human face images used in the study were used by the authors in previous studies [19,21,37], others were sourced from the Internet. The human faces were naturalistic, with no cropping or standardization of image properties (e.g. luminance, angle of view, gender, etc.). Both human and illusory faces were selected to fall along a positive–negative valence continuum ranging from angry to happy in four categories (i.e. high angry, low angry, low happy, high happy). These categories were validated by participant ratings (figure 2a).
(c) . Design and procedure
Each experiment consisted of a long series of trials in which each face/pareidolia image was briefly displayed (250 ms) and then rated by the participant for emotional expression on the angry/happy dimension while the screen was blank. There was a pause between image offset and presentation of the rating bar (pause time varied randomly between 800 and 1200 ms). The rating bar had a randomly chosen start position on each trial. The participant was able to change the position with a roller-ball mouse; the left end of the rating scale indicated very angry, the right end very happy and the centre indicated a neutral expression. Pressing the space bar recorded the rating and initiated a pause (random within 550–950 ms) before the next trial's stimulus presentation began. Before commencing the experiment, participants completed 16 practice trials (data not recorded) using eight faces and eight pareidolia images that were not used in the experiment.
Experiment 1 tested for serial effects in two separate conditions: sequences of pareidolia images and sequences of faces. Each condition involved a sequence of 320 trials, composed of 40 pareidolia images (or 40 face images) shown eight times each. The trial order was completely randomized and was done in a different order for each participant. Half of the participants did the face condition first and then the pareidolia condition, while the other half did the reverse order. Experiment 2 followed a similar procedure but involved a random interleaving of the 40 real faces and 40 pareidolia images. There were again eight repetitions of each image, meaning there were a total of 640 trials for each participant presented in a random order. The same group of participants took part in both experiments.
(d) . Serial analysis
Each participant's eight ratings for a given image are averaged into a mean estimate of that image's expression. Error scores are then calculated for every trial by calculating the difference between the current trial's rating and the mean of that image's rating. Any serial effect is revealed by plotting these error scores as a function of the relative expression difference between successive images. This inter-trial expression difference is defined as the mean expression rating for the previous trial's minus the mean rating for current trial's image. A positive difference means the previous trial was happier than the current trial; a negative value means the previous trial was angrier. If ratings are serially independent, error scores are distributed around zero over the range of relative expression difference. For face expression, positive serial dependence has been reported [36], meaning that error scores tend to increase with relative expression difference. In other words, current ratings are biased towards the value of the previous image so that current images are rated as happier following a previous happy face, and angrier following an angry face. The analysis of each trial's rating as a function of the preceding mean expression is done for each individual (using their own mean ratings) and then the serial effects are averaged into a group mean.
To model the group mean data, we fitted a difference-of-Gaussian (DoG) model [24,35]. The model describes how the serial dependence bias varies as a function of the relative difference in expression between the previous and current trials and is defined as
2.1 |
where x is the relative expression dimension (previous trial minus current), is the relative expression difference at which the serial effect is maximal and A is the amplitude of the maximal serial effect. Veridical perception (i.e. no influence from the previous trial) would yield zero amplitude.
3. Results
Expression ratings for Experiment 1 are shown in figure 2a, for both face (red) and pareidolia (blue) images. Ratings were made by adjusting a sliding scale bar with a maximum length of 400 pixels and the y-axis, therefore, shows the full-scale range. Confirming piloting work which binned these images into four levels of expression from very angry to very happy, the group mean ratings show a clear increase with each level of expression that is near-linear. There were 10 pareidolia and 10 face images at each level, making 40 of each kind in total. Each observer rated an image eight times and the mean rating was calculated. Each data point in the plot shows the group mean (n = 15) of these mean ratings at a given expression level, with error bars showing ±1 s.e.m. The mean ratings for both pareidolia and face images were very consistent across observers, as seen by the very small standard errors. Although the pareidolia images have a very striking appearance, their rated expressions had a very similar range to the faces. Overall, the range of expression ratings did not differ significantly between image categories (paired samples t-test: t(14) = 1.268, p = 0.225).
The scatter plot in figure 2b shows that variability in expression ratings broadly comparable for face and pareidolia images. The scatter plot contains 60 points, each one representing the standard deviation of all ratings made by a given participant at a given expression level (i.e. the standard deviation of 80 ratings: 10 images at a given level, each rated eight times). Overall, the expression ratings for faces were slightly less variable than those for pareidolia, as summarized by the black arrows on each axis indicating the mean variability in expression ratings. A paired-sample t-test confirmed that variability was significantly less for face than for pareidolia images (t59 = −3.220; p = 0.002).
The analysis of serial dependence is shown in figure 3 and reveals how the expression rating on a given trial depends on the previous trial. The y-axis shows the serial bias, which is the difference between the rating for the current trial's image and the mean rating for that image. Over trials, this value should tend to zero if ratings are sequentially independent as each trial's rating would be an estimate of the mean rating. A significant bias indicates serial dependence. The x-axis shows the difference between the previous trial's rating and the mean rating for the current trial's stimuli. Positive differences mean the previous image was rated more highly for expression than the current image, and vice versa. The serial dependence is positive if there is a bias which increases with the inter-trial difference.
The serial dependence analyses (figure 3) were first done for individual participants and then averaged into the group mean effect shown in the figure. The continuous lines show the best-fitting DoG function for a one-back serial analysis (figure 3a) and a two-back analysis (figure 3b). The DoG functions fitted to the one-back data show clear positive relationships for both face and pareidolia images, with peaks occurring in +/+ and –/– quadrants and the DoG functions describe the data well (faces, r2 = 0.925; pareidolia, r2 = 0.979). Between the peaks, there is a positive relationship between bias and serial difference and beyond the peaks, the serial effect returns to baseline. This is typical for serial dependence effects [24] and the tuning to relatively small stimulus differences rules out a simple response bias explanation. We used a bootstrap sign-test (10 000 iterations) to evaluate significance. The two image categories did not differ on the width (σ) parameter (p = 0.140). The peaks for both image categories were statistically significant when tested against zero (ps < 0.001) and the peak of the serial effect for pareidolia was significantly greater than the peak for faces (p = 0.008). The larger serial effect for pareidolia is consistent with Cicchini et al.'s optimal observer model which predicts that consecutive stimuli of a given variability should exhibit more serial dependence than consecutive stimuli of lesser variability [27, see eqn (3.6)].
As reported in previous studies [24,30,38], serial dependence effects decline in magnitude if the trials are not consecutive. Figure 3b shows the two-back serial analysis with the best-fitting DoG model (faces, r2 = 0.864; pareidolia, r2 = 0.735). Overall, the peak amplitude in the two-back analysis was smaller than the one-back analysis, for both faces (p = 0.047) and pareidolia (p = 0.015), although the peaks were still significantly greater than zero (face: p = 0.002; pareidolia: p = 0.001). The difference between face and pareidolia was not significant (p = 0.526), and the difference in the width parameter was also not significant (p = 0.799). Finally, as a control, we tested if the data contained n + 1 effects. Logically, there can be no n + 1 serial effect (a future trial cannot influence the present) but data patterns resembling serial dependence can sometimes arise from response bias or central tendency. We computed the n + 1 serial effect for face and pareidolia sequences and fitted the DoG model to the data. The fits did not reveal a significant amplitude for faces or pareidolia. Moreover, subtracting the n + 1 effect from the n − 1 data showed that the n − 1 serial effect was still significant for both stimuli. This indicates our serial effects are not driven by response bias or central tendency.
Experiment 1 established that pareidolia images could be rated for expression with a similar precision to face expression ratings and that sequences of pareidolia images produce serial effects that are qualitatively similar to those arising from faces. Experiment 2 interleaved face and pareidolia images in a random alternation. This produces four pairs of consecutive stimuli: two same-category pairs (face/face and pareidolia/pareidolia) and two cross-category pairs (face/pareidolia and pareidolia/face). Serial effects for the same-category pairs are shown in figure 4a shows the cross-category pairs.
The data in figure 4a confirm those of Experiment 1 (figure 3a) by showing clear positive serial dependencies for both faces and pareidolia that are well described by the DoG model (faces, r2 = 0.870; pareidolia, r2 = 0.985). Using the same bootstrap sign-test as in Experiment 1 (10 000 repetitions), the DoG functions show very significantly non-zero peak amplitudes (ps < 0.001) and, again, the amplitude of the pareidolia function is greater than for faces (p = 0.008). The width parameter (σ) of the DoG function did not differ between the two image categories (p = 0.590). These results closely replicate those from Experiment 1 and show the serial effect is robust for both categories regardless of whether the trials are blocked (Experiment 1) or randomly interleaved.
The advantage of randomly interleaving image categories is that a pair of consecutive trials is equally likely be same-category as cross-category. Moreover, there are two orders of cross-category stimuli (face followed by pareidolia, and pareidolia followed by face). The serial effects for both orders of cross-category stimuli and the best-fitting model are shown in figure 4b (red: face followed by pareidolia (r2 = 0.881); blue: pareidolia followed by face (r2 = 0.866)) and closely resemble those obtained with the same-category stimuli (figure 4a). The bootstrap sign-test confirmed significant peaks in the best-fitting DoG function for each cross-category order (ps < 0.001) and also showed that the amplitude of the effect was greater pairs in which pareidolia preceded a face (figure 4b, blue curve: p = 0.004). There was no difference for the width parameter (p = 0.746).
It is also of interest to compare between the conditions shown in figure 4. For example, the conditions shown in red (face_face versus face_pareidolia) are both computed from pairs of trials that have the same first stimulus (face), but they differ in the stimulus that follows. Does the serial effect of expression from the first stimulus carry over to the second equivalently in each case? A bootstrap sign-test comparing the amplitude of the face-first conditions showed that the within-category serial effect was not significantly larger than the cross-category effect (p = 0.526). The same comparison of serial effects was made between the pairs beginning with a pareidolia image (blue curves) and again the amplitude was not significantly larger for the within-category serial effect (p = 0.357).
4. Discussion
We observed positive serial dependence for perceived expression in face pareidolia. The perceived expression of a given illusory face was pulled towards the direction of the expression (happy or angry) of the preceding illusory face. This is consistent with the positive serial dependence previously observed for expression in human face morphs [36]. Illusory faces in objects are much more varied in the visual features that define the perceived ‘facial expression’ than the relatively homogeneous features in human faces, thus it was not clear whether rapid adaptation of expression would be observed in pareidolia. Adaptation typically relies on considerable similarity between stimuli, and serial dependence for human faces disappears if they are rotated [28,29] or of a different social category [36]. Finding positive serial dependence for illusory faces indicates that pareidolia engages similar mechanisms of temporal continuity as human faces despite the visual heterogeneity of illusory faces. This also suggests that serial dependence of facial expression is unlikely to be driven entirely by low-level visual features, even though it is known that serial dependence occurs for low-level visual attributes such as orientation [24,39,40] and motion [38] in addition to faces.
As well as finding serial dependence of expression for both human and illusory faces, we also found cross-domain serial dependence between randomly interleaved human and illusory faces. The perceived expression of an illusory face on a happy–angry continuum was biased towards the expression of a preceding human face, and vice versa. This provides evidence for a common mechanism underlying expression processing in both human faces and illusory faces. Importantly, it suggests that expression is not tightly bound to the visual features that are specific to human faces (e.g. skin colour, round outer contours). Instead, there appears to be a large degree of tolerance in the system to the visual features that define a facial expression, which are effective even in the absence of the underlying muscular structure associated with expression in real faces [13].
Previous work on expression processing has revealed the importance of several local facial features in defining human expressions [14–17]. By contrast, the cross-domain adaptation of expression that we observe with pareidolia suggests that the underlying mechanism does not require fine-scale visual indicators of expression based on biologically plausible muscle movements [2,3,13]. Instead, the coarse-scale form of expression observed in non-human faces such as pareidolia appears to be sufficient to drive the same expression mechanism as that underlying human faces. However, it is not clear to what degree this would generalize to more nuanced human expressions. It may be that our use of an angry–happy continuum of clear positive and negative valence preferentially engages a coarse-scale expression mechanism that only partially accounts for the processing of the much wider range of human expression. Disentangling the effects of valence, affect and attention as well as both visual and social factors is a key challenge in moving towards a complete understanding of how facial expressions relate to emotional states [4,5].
Several lines of converging psychophysical and neuroimaging evidence suggest that face pareidolia and human faces share common mechanisms [21–23]. This is further supported by the finding that rhesus macaque monkeys also experience face pareidolia [19,20], suggesting that misperceiving faces in objects is a universal feature of the primate face detection system. However, although human and illusory faces both speed up visual search for a target [22], engage social attention via eye gaze direction [23] and share common neural mechanisms [21], there are important differences. Human faces are found even faster than illusory faces in visual search [22], and MEG has shown that the initial ‘face-like’ response to pareidolia only occurs for one-quarter of a second before their neural representation reorganizes to be more similar to objects than faces [21]. Here, we also observed an asymmetry, as illusory faces had a larger influence on expression ratings for subsequent human faces than the reverse order (figure 4b). It is not clear why this is the case, although one possibility is that the novelty and striking expression of pareidolia images capture attention more than human faces because of the unexpected nature of their appearance. Since attention is known to modulate both perceptual and neurophysiological responses, it is possible that this accounts for the enhanced serial effects observed when the preceding image was pareidolia [41]. Indeed, attention has been noted as a key element in serial dependence: attended stimuli exhibit a greater serial effect than unattended or actively ignored stimuli [42,43]. Another possibility is that human faces and pareidolia images may engage expression mechanisms differently, which could manifest as an asymmetry in cross-domain adaptation. For example, opposing positive and negative (i.e. adaptation) serial dependences co-occur in perception [44] and if expression in pareidolia images were to elicit less adaptation of expression mechanisms than genuine face images, then the positive serial effect for pareidolia would be relatively stronger.
Together, our results show that illusory faces drive temporal continuity mechanisms in the visual system just as human faces do. Further, we found that illusory faces and human faces share a common mechanism for expression, indicative that expression processing is broadly tuned rather than tightly linked to human facial features and their specific visual appearance. This suggests that just like face detection [37], our ability to detect expressions is tuned to favour rapid responses to facial information signalling emotional valence and that the benefit of fast, sensitive expression detection outweighs the cost of occasional false positives. Such broad tuning is likely adaptive in the context of social communication, as perceiving illusory expressions in inanimate objects does not share the same likelihood of a serious consequence that may follow missing a relevant emotional cue signalled by another social agent.
Supplementary Material
Ethics
This research was conducted in accordance with the Declaration of Helsinki and was approved by the Human Research Ethics Committee of the University of Sydney (project no. 2016/662).
Data accessibility
Data and analysis scripts can be found on FigShare: https://figshare.com/s/bfa112493d16ad3d8666.
Authors' contributions
D.A.: conceptualization, formal analysis, methodology, writing—original draft; Y.X.: investigation, writing—original draft; S.G.W.: conceptualization, methodology; J.T.: conceptualization, methodology, writing—original draft. All authors gave final approval for publication and agreed to be held accountable for the work performed therein.
Competing interests
We declare we have no competing interests.
Funding
Funded by an Australian Research Council grant (DP190101537) to D.A.
References
- 1.Jack RE, Schyns PG. 2017. Toward a social psychophysics of face communication. Annu. Rev. Psychol. 68, 269-297. ( 10.1146/annurev-psych-010416-044242) [DOI] [PubMed] [Google Scholar]
- 2.Waller BM, Julle-Daniere E, Micheletta J. 2020. Measuring the evolution of facial ‘expression’ using multi-species FACS. Neurosci. Biobehav. Rev. 113, 1-11. ( 10.1016/j.neubiorev.2020.02.031) [DOI] [PubMed] [Google Scholar]
- 3.Taubert J, Japee S. 2021. Using FACS to trace the neural specializations underlying the recognition of facial expressions: a commentary on Waller et al. (2020). Neurosci. Biobehav. Rev. 120, 75-77. ( 10.1016/j.neubiorev.2020.10.016) [DOI] [PubMed] [Google Scholar]
- 4.Calvo MG, Nummenmaa L. 2016. Perceptual and affective mechanisms in facial expression recognition: an integrative review. Cogn. Emot. 30, 1081-1106. ( 10.1080/02699931.2015.1049124) [DOI] [PubMed] [Google Scholar]
- 5.Adolphs R. 2002. Recognizing emotion from facial expressions: psychological and neurological mechanisms. Behav. Cogn. Neurosci. Rev. 1, 21-62. ( 10.1177/1534582302001001003) [DOI] [PubMed] [Google Scholar]
- 6.Langton SRH, Law AS, Burton AM, Schweinberger SR. 2008. Attention capture by faces. Cognition 107, 330-342. ( 10.1016/j.cognition.2007.07.012) [DOI] [PubMed] [Google Scholar]
- 7.Calvo M, Esteves F. 2005. Detection of emotional faces: low perceptual threshold and wide attentional span. Vis. Cogn. 12, 13-27. ( 10.1080/13506280444000094) [DOI] [Google Scholar]
- 8.Yang E, Zald DH, Blake R. 2007. Fearful expressions gain preferential access to awareness during continuous flash suppression. Emotion 7, 882-886. ( 10.1037/1528-3542.7.4.882) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.Nomi JS, Rhodes MG, Cleary AM. 2013. Emotional facial expressions differentially influence predictions and performance for face recognition. Cogn. Emot. 27, 141-149. ( 10.1080/02699931.2012.679917) [DOI] [PubMed] [Google Scholar]
- 10.Becker DV, Anderson US, Mortensen CR, Neufeld SL, Neel R. 2011. The face in the crowd effect unconfounded: happy faces, not angry faces, are more efficiently detected in single- and multiple-target visual search tasks. J. Exp. Psychol. Gen. 140, 637-659. ( 10.1037/a0024060) [DOI] [PubMed] [Google Scholar]
- 11.Ambron E, Foroni F. 2015. The attraction of emotions: irrelevant emotional information modulates motor actions. Psychon. Bull. Rev. 22, 1117-1123. ( 10.3758/s13423-014-0779-y) [DOI] [PubMed] [Google Scholar]
- 12.Pitcher D, Pilkington A, Rauth L, Baker C, Kravitz DJ, Ungerleider LG. 2020. The human posterior superior temporal sulcus samples visual space differently from other face-selective regions. Cereb. Cortex 30, 778-785. ( 10.1093/cercor/bhz125) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Rinn WE. 1984. The neuropsychology of facial expression: a review of the neurological and psychological mechanisms for producing facial expressions. Psychol. Bull. 95, 52-77. ( 10.1037/0033-2909.95.1.52) [DOI] [PubMed] [Google Scholar]
- 14.Chen M-Y, Chen C-C. 2010. The contribution of the upper and lower face in happy and sad facial expression classification. Vision Res. 50, 1814-1823. ( 10.1016/j.visres.2010.06.002) [DOI] [PubMed] [Google Scholar]
- 15.Blais C, Fiset D, Roy C, Saumure Régimbald C, Gosselin F. 2017. Eye fixation patterns for categorizing static and dynamic facial expressions. Emotion 17, 1107-1119. ( 10.1037/emo0000283) [DOI] [PubMed] [Google Scholar]
- 16.Beaudry O, Roy-Charland A, Perron M, Cormier I, Tapp R. 2014. Featural processing in recognition of emotional facial expressions. Cogn. Emot. 28, 416-432. ( 10.1080/02699931.2013.833500) [DOI] [PubMed] [Google Scholar]
- 17.Smith ML, Cottrell GW, Gosselin F, Schyns PG. 2005. Transmitting and decoding facial expressions. Psychol. Sci. 16, 184-189. ( 10.1111/j.0956-7976.2005.00801.x) [DOI] [PubMed] [Google Scholar]
- 18.Gosselin F, Schyns PG. 2001. Bubbles: a technique to reveal the use of information in recognition tasks. Vision Res. 41, 2261-2271. ( 10.1016/s0042-6989(01)00097-9) [DOI] [PubMed] [Google Scholar]
- 19.Taubert J, Wardle SG, Flessert M, Leopold DA, Ungerleider LG. 2017. Face pareidolia in the rhesus monkey. Curr. Biol. 27, 2505-2509.e2. ( 10.1016/j.cub.2017.06.075) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Taubert J, Flessert M, Wardle SG, Basile BM, Murphy AP, Murray EA, Ungerleider LG. 2018. Amygdala lesions eliminate viewing preferences for faces in rhesus monkeys. Proc. Natl Acad. Sci. USA 115, 8043-8048. ( 10.1073/pnas.1807245115) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Wardle SG, Taubert J, Teichmann L, Baker CI. 2020. Rapid and dynamic processing of face pareidolia in the human brain. Nat. Commun. 11, 4518. ( 10.1038/s41467-020-18325-8) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22.Keys RT, Taubert J, Wardle SG. 2021. A visual search advantage for illusory faces in objects. Atten. Percept. Psychophys. 83, 1942-1953. ( 10.3758/s13414-021-02267-4) [DOI] [PubMed] [Google Scholar]
- 23.Palmer CJ, Clifford CWG. 2020. Face pareidolia recruits mechanisms for detecting human social attention. Psychol. Sci. 31, 956797620924814. ( 10.1177/0956797620924814) [DOI] [PubMed] [Google Scholar]
- 24.Fischer J, Whitney D. 2014. Serial dependence in visual perception. Nat. Neurosci. 17, 738-743. ( 10.1038/nn.3689) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.Kiyonaga A, Scimeca JM, Bliss DP, Whitney D. 2017. Serial dependence across perception, attention, and memory. Trends Cogn. Sci. (Regul. Ed.) 21, 493-497. ( 10.1016/j.tics.2017.04.011) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Cicchini GM, Mikellidou K, Burr DC. 2018. The functional role of serial dependence. Proc. R. Soc. B 285, 20181722. ( 10.1098/rspb.2018.1722) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Liberman A, Fischer J, Whitney D. 2014. Serial dependence in the perception of faces. Curr. Biol. 24, 2569-2574. ( 10.1016/j.cub.2014.09.025) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28.Turbett K, Palermo R, Bell J, Hanran-Smith DA, Jeffery L. 2021. Serial dependence of facial identity reflects high-level face coding. Vision Res. 182, 9-19. ( 10.1016/j.visres.2021.01.004) [DOI] [PubMed] [Google Scholar]
- 29.Taubert J, Alais D. 2016. Serial dependence in face attractiveness judgements tolerates rotations around the yaw axis but not the roll axis. Vis. Cogn. 24, 1-12. ( 10.1080/13506285.2016.1196803) [DOI] [Google Scholar]
- 30.Xia Y, Leib AY, Whitney D. 2016. Serial dependence in the perception of attractiveness. J. Vis. 16, 28. ( 10.1167/16.15.28) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 31.Van der Burg E, Rhodes G, Alais D. 2019. Positive sequential dependency for face attractiveness perception. J. Vis. 19, 6. ( 10.1167/19.12.6) [DOI] [PubMed] [Google Scholar]
- 32.Kok R, Taubert J, Van der Burg E, Rhodes G, Alais D. 2017. Face familiarity promotes stable identity recognition: exploring face perception using serial dependence. R. Soc. Open Sci. 4, 160685. ( 10.1098/rsos.160685) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33.Taubert J, Van der Burg E, Alais D. 2016. Love at second sight: sequential dependence of facial attractiveness in an on-line dating paradigm. Sci. Rep. 6, 22740. ( 10.1038/srep22740) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34.Taubert J, Alais D, Burr D. 2016. Different coding strategies for the perception of stable and changeable facial attributes. Sci. Rep. 6, 32239. ( 10.1038/srep32239) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35.Alais D, Kong G, Palmer C, Clifford C. 2018. Eye gaze direction shows a positive serial dependency. J. Vis. 18, 11. ( 10.1167/18.4.11) [DOI] [PubMed] [Google Scholar]
- 36.Liberman A, Manassi M, Whitney D. 2018. Serial dependence promotes the stability of perceived emotional expression depending on face similarity. Atten. Percept. Psychophys. 80, 1461-1473. ( 10.3758/s13414-018-1533-8) [DOI] [PubMed] [Google Scholar]
- 37.Taubert J, Wardle SG, Ungerleider LG. 2020. What does a ‘face cell’ want? Prog. Neurobiol. 195, 101880. ( 10.1016/j.pneurobio.2020.101880) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 38.Alais D, Leung J, Van der Burg E. 2017. Linear summation of repulsive and attractive serial dependencies: orientation and motion dependencies sum in motion perception. J. Neurosci. 37, 4381-4390. ( 10.1523/JNEUROSCI.4601-15.2017) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39.Kim S, Burr D, Cicchini GM, Alais D. 2020. Serial dependence in perception requires conscious awareness. Curr. Biol. 30, R257-R258. ( 10.1016/j.cub.2020.02.008) [DOI] [PubMed] [Google Scholar]
- 40.Cicchini GM, Mikellidou K, Burr D. 2017. Serial dependencies act directly on perception. J. Vis. 17, 6. ( 10.1167/17.14.6) [DOI] [PubMed] [Google Scholar]
- 41.Carrasco M, Ling S, Read S. 2004. Attention alters appearance. Nat. Neurosci. 7, 308-313. ( 10.1038/nn1194) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 42.Rafiei M, Hansmann-Roth S, Whitney D, Kristjánsson A, Chetverikov A. 2021. Optimizing perception: attended and ignored stimuli create opposing perceptual biases. Atten. Percept. Psychophys. 83, 1230-1239. ( 10.3758/s13414-020-02030-1) [DOI] [PubMed] [Google Scholar]
- 43.Fritsche M, de Lange FP. 2019. The role of feature-based attention in visual serial dependence. J. Vis. 19, 21. ( 10.1167/19.13.21) [DOI] [PubMed] [Google Scholar]
- 44.Gekas N, McDermott KC, Mamassian P. 2019. Disambiguating serial effects of multiple timescales. J. Vis. 19, 24. ( 10.1167/19.6.24) [DOI] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
Data and analysis scripts can be found on FigShare: https://figshare.com/s/bfa112493d16ad3d8666.