Skip to main content
PLOS One logoLink to PLOS One
. 2025 Mar 18;20(3):e0311714. doi: 10.1371/journal.pone.0311714

Neural correlates of the uncanny valley effect for robots and hyper-realistic masks

Shona Fitzpatrick 1, Ailish K Byrne 2, Alex Headley 1, Jet G Sanders 3, Helen Petrie 4, Rob Jenkins 1, Daniel H Baker 1,*
Editor: Rae Yule Kim5
PMCID: PMC11957730  PMID: 40100857

Abstract

Viewing artificial objects and images that are designed to appear human can elicit a sense of unease, referred to as the ‘uncanny valley’ effect. Here we investigate neural correlates of the uncanny valley, using still images of androids (robots designed to look human), and humans wearing hyper-realistic silicone masks, as well as still images of real humans, in two experiments. In both experiments, human-like stimuli were harder to distinguish from real human faces than stimuli that were clearly not designed to mimic humans but contain facial features (mechanical robots and Halloween masks). Stimulus evoked potentials (electromagnetic brain responses) did not show convincing differences between faces and either androids or realistic masks when using traditional univariate statistical tests. However, a more sensitive multivariate analysis identified two regions of above-chance decoding, indicating neural differences in the response between human faces and androids/realistic masks. The first time window was around 100–200 ms post stimulus onset, and most likely corresponds to low-level image differences between conditions. The second time window was around 600 ms post stimulus onset, and may reflect top-down processing, and may correspond to the subjective sense of unease characteristic of the uncanny valley effect. Objective neural components might be used in future to rapidly train generative artificial intelligence systems to produce more realistic images that are perceived as natural by human observers.

Introduction

Many people report an aversion to entities that are superficially human-like, but on closer inspection turn out to be artificial. Examples include humanoid robots (androids), puppets, hyper-realistic masks [1], and computer-generated images or movies. The term ‘uncanny valley’ [2] (English translation in [3]) describes the idea that clearly human or clearly artificial entities do not evoke unease, whereas artificial entities that are human-like are disconcerting. Understanding these experiences is increasingly important as artificial entities become more integrated into our everyday lives. However at present relatively little is known about the neural underpinnings of the uncanny valley effect (for a recent review, see [4]). In particular, the root of the uncanny valley effect remains debated: does it arise primarily from bottom-up sensory conflicts, or from higher-level cognitive processes? Resolving this question is critical to understanding its fundamental mechanisms.

Neural responses to faces and bodies in general are well-characterised, and there appear to be specialised brain regions devoted to both (reviewed in [5]). For example, areas of the occipital lobe [6] and fusiform gyrus [7] respond more to faces than to non-face stimuli, and sections of extrastriate cortex are responsive to bodies [8]. There are also electromagnetic event-related potential (ERP) signals associated with face and body stimuli, though their precise role is still debated [9,10]. It seems highly likely that ‘uncanny’ images will activate these same processes, yet it is unclear whether the initial cause of the sense of unease they produce occurs at bottom-up sensory stages [1113] or is modulated by more top-down cognitive factors [14,15].

One previous study by [16] measured functional magnetic resonance imaging (fMRI) responses to moving stimuli designed to elicit an uncanny valley effect. They found repetition suppression effects (repetition suppression is a phenomenon in which the neural response to repeated presentations of identical or similar stimuli is reduced relative to the response on the first presentation) in action-specific brain regions responding to movies of androids that had a biological appearance, but mechanical motion. These effects were stronger than for movies of humans or mechanical robots performing the same actions. A more recent electroencephalography (EEG) study [17] identified a difference in the N400 component (the N400 is an electromagnetic brain potential obtained 400 ms after stimulus onset, typically over centro-parietal electrodes; it has been proposed to reflect the extent to which the stimulus presented was surprising or unexpected) between dynamic and static conditions using the same stimuli. Although this difference was strongest over frontal electrodes, source reconstruction of the N400 itself suggested a left-lateralised source in the temporo-parietal cortex, consistent with the fMRI results [16]. The authors interpret both of these findings as being due to the discrepancy between the human-like appearance and the clearly non-biological motion of the robot.

Our aim was to further investigate neural correlates of the uncanny valley effect, with the expectation that increased understanding will aid efforts to generate more convincingly human robots and avatars in the future. We achieved this through two EEG experiments, in which we measured neural responses to static images. Although previous studies focus on dynamic stimuli, static images allow for a more precise investigation of the neural mechanisms underlying the uncanny valley effect, particularly by eliminating motion-related confounds. In the first experiment, the stimuli were still images of humans, machine-like robots, and human-like robots (see Fig 1a). In the second experiment we aimed to generalise the finding by using images of people wearing no masks, wearing obvious masks (e.g. Halloween masks), and wearing hyper-realistic silicone masks [18] (see Fig 1b). Rather than focus on specific ERP components, we use a multivariate pattern classification approach (a machine learning technique in which an algorithm is trained to decode the neural responses) to identify time windows in which information in the EEG signal can be used to distinguish between pairs of conditions. Our rationale is that timepoints where signals evoked by human faces can be distinguished from those evoked by human-like robots, or hyper-realistic masks, are candidates for a neural signature of the uncanny valley effect.

Fig 1. Illustrative stimuli from the same categories as used in Experiments 1 and 2.

Fig 1

Row (a) shows a human face, an android and a robot, all against white backgrounds. Row (b) shows a human face, a hyper-realistic silicone mask, and a Halloween mask, all against natural backgrounds. Images shown here were taken from a variety of sources that permit reuse in academic contexts and in most cases were not part of the stimulus set from the experiments. The silicone mask image was taken by the authors (subject: RJ, image credit: JGS), and was used in Experiment 2. The individual in this manuscript has given written informed consent (as outlined in PLOS consent form) to publish these case details (i.e. this image).

Materials and methods

Participants

A total of 29 participants completed Experiment 1 (12 male, 17 female), and 30 different participants completed Experiment 2 (7 male, 23 female). Participants were young adults with no history of neurological disorder. None of the participants had previously taken part in a study using these stimuli, and all were naïve to the hypotheses and wore their normal optical correction if required. Written informed consent was collected before each experiment began, and all procedures were approved by the Ethics committee of the Department of Psychology at the University of York. Data collection for Experiment 1 ran from 14th July to 15th September 2022, and data collection for Experiment 2 ran from 12th October 2017 to 14th February 2018.

Apparatus and stimuli

In Experiment 1, the stimulus set consisted of a total of 90 images, evenly split between three categories: real faces, human-like robots, and mechanical robots. Images all showed the head and shoulders of the subject, had white backgrounds, and were sourced from the Internet. In Experiment 2, the stimulus set (first described by [19], but here including additional images) consisted of a total of 296 images, comprising real faces (148 images), people wearing silicone masks (74 images), and people wearing obvious masks of the sort typically worn for carnivals and Halloween celebrations (74 images). The backgrounds of these images were more heterogeneous, and showed the natural surroundings of the subject. While the image backgrounds differed across experiments, we hypothesize that the primary task was not affected, as participants focused on the foreground stimuli. In both experiments, images included examples of both genders, and of varied ethnic backgrounds.

All stimuli were displayed on a ViewPixx display running at 120 Hz, controlled by an Apple Macintosh computer. The display was gamma corrected using a photometer to ensure that the luminance output was linear. EEG data were collected using a 64-channel Waveguard cap and an ANT Neuroscan system, sampling at 1 kHz. The ground electrode was located at position AFz, and all signals were referenced to the whole head average. Low latency digital triggers were sent between the display and the EEG amplifier using an 8-bit parallel cable.

Procedure

Experiment 1: Robots.

Each participant completed three blocks of the first experiment. Within each block, all 90 stimulus images were presented twice in a random order. Stimuli subtended 11 × 11 degrees at the viewing distance of 57 cm, and were shown against a mid-grey background, with a black central fixation cross displayed throughout. The presentation duration was 500 ms, and participants were asked to press a mouse button to indicate if they believed each image was of a human or of a robot/android. After each response there was a random duration blank period with a mean duration of 1000 ms and a standard deviation of 200 ms. Durations were chosen to provide sufficient information for judgment while avoiding task fatigue. Randomized blank periods were designed to reduce carryover effects and prevent anticipatory biases. Each block lasted around 6 minutes.

After the EEG experiment, participants also completed a series of questionnaires using the Qualtrics platform. These involved rating their perception of a subset of the stimuli (8 from each category), using items from the Godspeed questionnaire [20]. Items were selected that were expected to be most closely aligned to measuring the sensation of uncanniness. Inspection time was unlimited. Participants also provided demographic information (age, gender) and completed the GAToRS [21] and AQ [22] questionnaires, however the results of these additional questionnaires are not presented here.

Experiment 2: Hyper-realistic masks.

Participants were shown all 296 images in a random order in each of three blocks. In the first block, stimuli subtended 5 . 5 × 7 . 5 degrees of visual angle when viewed at a distance of 57 cm. In the second block, stimuli doubled in size (width and height), and subtended 11 × 15 degrees at the same viewing distance. In the third block, stimuli doubled in size again, and subtended 22 × 30 degrees. The rationale for the size manipulation was to investigate whether increasing levels of detail made the silicone masks more identifiable [18,19]. However as that is not the main focus of the current paper, and our preliminary analyses indicated no differences between size conditions, we collapse results across size conditions. Stimuli were presented for 250 ms, and participants indicated whether they thought each image contained a real face or a mask, using a two-button trackball. The button assignment (whether the left button indicated a face or a mask, and vice versa) was determined randomly for each participant, but remained constant throughout the whole experiment. Text reminding the participant of the button assignment was present continuously in the lower right corner of the screen, far from the area of the screen where the stimuli were presented. A central fixation cross was also present throughout. After each response there was a random duration blank period with a mean duration of 1000 ms and a standard deviation of 200 ms. Each block lasted around 8 minutes.

An independent group of 20 participants also completed an online questionnaire in which they rated the images along various dimensions. The participants repeated the real face vs mask judgement from the main experiment, and were additionally asked to rate emotional expressiveness, realism, and uncanniness for each image using a 7-point Likert scale. Inspection time was unlimited for these judgments.

Data analysis

We analysed response data by calculating d-prime (d) scores for each condition, derived from the hit rate and false alarm rate [23]. For the human conditions, the hit rate was the proportion of human images correctly identified as human, and the false alarm rate was the proportion of robot or mask images that were incorrectly judged as being human. For the robot and mask conditions, the hit rate was the proportion of robot/mask images correctly identified as not being human, and the false alarm rate was the proportion of human images that were incorrectly judged as being non-human (note that this means the false alarm rate was the same for the robot and android conditions, and for the silicone and Halloween mask conditions). We capped infinite d-prime values (which occur e.g. when the hit rate is 1) at an arbitrary ceiling of 5 to prevent outliers from skewing the results, following established conventions in signal detection theory. We log transformed the reaction times (which typically have positive skew) and performed all averaging and statistical analysis on the logarithmic values.

EEG signals were recorded during each block and saved to disc for subsequent offline analysis. We used components of the EEGlab toolbox [24] to import the data into Matlab and collate data across blocks. We then used Brainstorm [25] to filter the data using a bandpass filter (0.5 to 30 Hz), epoch by condition, and subtract a pre-trial baseline (the mean voltage for the 200 ms before stimulus onset). Five participants were excluded from the EEG analysis of each experiment due to excessive noise. Our attempts to clean up the data from these participants using independent components analysis were unsuccessful. Their behavioural data were unaffected, and are therefore still included in the analysis.

We performed univariate analyses by conducting Bayesian t-tests [26] between ERPs from pairs of conditions at each time point using a JZS prior, using signals pooled across electrodes P6, P8, PO6 and PO8, which are typically associated with visual responses to faces. The resulting Bayes factor score is a summary of the evidence in favour of either the null hypothesis (that the waveforms are equal) or the alternative hypothesis (that they differ). We use the heuristics proposed by Jeffreys [27] that Bayes factors >3 (log10BF10 > 0.5) constitute some evidence supporting the alternative hypothesis, factors >10 (log10BF10 > 1) constitute strong evidence, and factors >30 (log10BF10 > 1.5) constitute very strong evidence.

Multivariate pattern analysis (MVPA) was conducted by training a linear support vector machine algorithm (LibSVM, [28]) to discriminate between patterns of activity across electrodes at a specific time point. MVPA is a statistical technique that involves training a machine learning algorithm to identify patterns in data, and then testing its accuracy at classifying unseen data; in EEG analysis above-chance classification is considered evidence of distinct patterns of neural activity between two conditions. Previous work has indicated that EEG data do not typically require more complex nonlinear algorithms [29]. The patterns came from the human face condition and one of the other conditions, for a single participant. Four examples of each pattern were calculated by averaging over random subsets of 20% of the available trials from a given condition, and these were used to train the classifier. The accuracy of the classifier was tested on the remaining trials (that were not used in training) for each condition. This process was repeated 1000 times with different trial permutations to obtain an average accuracy, where chance performance is at 50% correct. The analysis was carried out at all time points, and for each participant separately. We then averaged classifier accuracy across participants, and calculated one sample Bayesian t-tests comparing to chance performance at each time point as described above.

Data and code availability

Raw data, processed data, and analysis scripts are freely available through the project repository at: doi: 10.17605/OSF.IO/5NZ2H

Results

Experiment 1

We first explored the behavioural results for identification of human versus non-human stimuli. We calculated d-prime scores to compare sensitivity across conditions. Sensitivity was highest for identifying robots (d= 4), but still well above chance for both the human (d= 2.72) and android (d= 2.27) conditions. The Bayes factor score for a one-way ANOVA comparing these three conditions indicated very substantial evidence (log10BF10= 6.67) for a difference between conditions, as illustrated in Fig 2a. Pairwise Bayesian t-tests between conditions indicate very convincing differences in sensitivity between robots and androids (log10BF10= 7.15) and robots and humans (log10BF10= 6.27). The difference between androids and humans (log10BF10= 5.58) was also very substantial. The higher d’ values for robots (Fig 2a) could indicate that these stimuli are more visually salient.

Fig 2. Summary of response data and grand mean ERP for Experiment 1.

Fig 2

Panel (a) shows d-prime scores for identifying images of human (blue diamonds), android (green squares) and robot (red circles) faces. Small points show individual participants, and the larger symbols with error bars indicate the group mean and bootstrapped 95% confidence intervals. Panel (b) plots reaction times in the same format (note the logarithmic x-axis). Panel (c) shows the grand mean ERP across all participants and conditions, pooled across electrodes P6, P8, PO6 and PO8 (see inset). The P100 is a positive evoked potential occurring around 100 ms after onset of a visual stimulus, associated with the initial (low level) visual response; the N170 is a negative potential at 170 ms that is often associated with faces; the P200 is a further positive potential linked to attention and stimulus discrimination. The shaded region around the curve illustrates the 95% confidence interval, and the grey rectangle at the foot indicates the stimulus duration.

Reaction times also differed between conditions, though the effects were rather smaller. Reactions were fastest for identifying robots (RT = 688 ms), compared with humans (RT = 803 ms) and androids (RT = 809 ms). The Bayes factor score for a one-way ANOVA comparing these three conditions indicated strong evidence (log10BF10= 1.44) for a difference between conditions, as illustrated in Fig 2b. Pairwise Bayesian t-tests between conditions indicate very convincing differences in sensitivity between robots and androids (log10BF10= 5.01) and robots and humans (log10BF10= 4.61), whereas the reaction time was equivalent between androids and humans (log10BF10= –0.68).

EEG activity showed a clear visually evoked potential over posterior electrode sites (see Fig 2c), with typical components found in response to visual stimuli pooled across all conditions (the P100, N170 and P200 are indicated in the figure). Pairwise comparisons of conditions are shown in Fig 3a and 3b. In general there is a tendency for the ERP response to human faces to diverge slightly from the other two conditions [30], however the evidence for this divergence was not compelling. Bayes factors exceeded 3 for only a small number of time points around 300–400 ms in the comparison between human and robot images (see yellow bars at y = –8 in Fig 3a), but these differences were small considering the variance in the data.

Fig 3. Univariate and multivariate comparisons across image type.

Fig 3

Panel (a) shows the ERPs comparing human (blue) and robot (red) face images, and panel (b) compares human (blue) and android (green) faces. Panels (c) and (d) show multivariate pattern classification accuracy for the same comparisons. Points at y = –8 and y = 25 indicate Bayes factor scores for comparisons between ERPs (a,b) and comparing classification accuracy to chance (50% correct; c,d).

We also conducted multivariate pattern analysis independently at each time-point for the same two comparisons. The evoked responses for human and robot images caused sufficiently distinct patterns of voltages across the scalp that the pattern classifier could distinguish between them from around 100 ms following stimulus onset, with accuracy up to 72% correct (see Fig 3c). Bayes factors exceeded 30 for much of the time window between 100 and 800 ms, indicating that the decoding was meaningfully above chance performance (50% correct). It was also possible to classify between human and android images (see Fig 3d), however performance was much poorer, with a maximum of 62% correct. Classification accuracy had an initial peak around 100 ms that provided compelling evidence for above chance classification (BF>30), and a later region of above-chance classification between 500 and 700 ms. The early time window likely reflects rapid processing of low-level visual features, such as edges or color contrasts, consistent with P100 and N170 components, whereas the later time window may involve higher-level cognitive processes, such as evaluating emotional content or judging authenticity. In the Discussion we speculate that these two time periods might correspond to distinct types of signal associated with the uncanny valley. More generally, the high classification accuracy for human vs. robot stimuli may be attributed to the salient mechanical elements of robots, whereas the lower accuracy for humanoids reflects their ambiguous human-like appearance, leading to confusion.

Finally, we analysed the rating data from a set of 10 questionnaire items for 8 stimuli from each category. The results are summarised in Fig 4, and in general show differences between stimulus categories along most dimensions. Of particular note, the U-shaped function predicted by the uncanny valley effect was apparent for ratings along the Dislike-Like (Fig 4d), Unfriendly-Friendly (Fig 4e), and Anxious-Relaxed (Fig 4i) dimensions. These are all dimensions with emotional valence, indicating support for the ‘uncanniness’ of our android stimuli. However we note that the android and robot categories were typically rated as being more similar to each other than to the real human faces. Following Experiment 1, we sought to generalise our results to a different stimulus set, and next report the results of Experiment 2 which used hyper-realistic silicone masks.

Fig 4. Ratings of stimuli using items from the Godspeed questionnaire.

Fig 4

Each rating was on a Likert scale from 1–5, and was the average of ratings from 8 stimulus examples. Dots show individual participant scores, with the larger symbols indicating the mean and 95% bootstrapped confidence intervals.

Experiment 2

The results of Experiment 2 were similar to those of Experiment 1, despite using a quite different stimulus set involving images of humans wearing masks, rather than robots. Sensitivity was highest for identifying Halloween masks (d= 4.01), but still well above chance for both the human (d= 2.52) and silicone mask (d= 1.29) conditions. The Bayes factor score for a one-way ANOVA comparing these three conditions indicated very substantial evidence (log10BF10= 13.34) for a difference between conditions, as illustrated in Fig 5a. Pairwise Bayesian t-tests between conditions indicate very convincing differences in sensitivity between Halloween and silicone masks (log10BF10= 9.71) between Halloween masks and humans (log10BF10= 11.59), and between silicone masks and humans (log10BF10= 5.19). Unlike in Experiment 1, there were no convincing reaction time differences between conditions (log10BF10= –0.8), as illustrated in Fig 5b.

Fig 5. Summary of response data and grand mean ERP for Experiment 2.

Fig 5

Panel (a) shows d-prime scores for identifying images of human faces (blue diamonds), silicone masks (green squares) and Halloween masks (red circles). Small points show individual participants, and the larger symbols with error bars indicate the group mean and bootstrapped 95% confidence intervals. Panel (b) plots reaction times in the same format (note the logarithmic x-axis). Panel (c) shows the grand mean ERP across all participants and conditions, pooled across electrodes P6, P8, PO6 and PO8 (see inset). The shaded region around the curve illustrates the 95% confidence interval, and the grey rectangle at the foot indicates the stimulus duration.

The grand average ERP waveform for Experiment 2 (see Fig 5c) had similar initial components as for Experiment 1. The latter portion of the waveforms differed somewhat, most likely owing to the difference in presentation duration across experiments (250 ms versus 500 ms). There was a substantial univariate difference in ERP response between human and Halloween mask conditions extending from around 170 to 230 ms following stimulus onset (see Fig 6a), with Bayes factors exceeding 30. Univariate differences between the human and silicone mask conditions were not compelling (see Fig 6b).

Fig 6. Univariate and multivariate comparisons across image type for Experiment 2.

Fig 6

Panel (a) shows the ERPs comparing human faces (blue) and Halloween masks (red), and panel (b) compares human faces (blue) and silicone masks (green). Panels (c) and (d) show multivariate pattern classification accuracy for the same comparisons. Points at y = –8 and y = 25 indicate Bayes factor scores for comparisons between ERPs (a,b) and comparing classification accuracy to chance (50% correct; c,d).

Multivariate pattern analysis revealed extremely high classification accuracy (up to 97% correct) comparing human faces with Halloween masks. This was convincingly above chance, with a Bayes factor score exceeding 30 from around 100 ms following stimulus onset, and extending across the full time window (see Fig 6c). Classification was also convincingly above chance when comparing human faces with silicone masks (Fig 6d). This timecourse had an initial peak of high accuracy (up to 76% correct) between 100 and 200 ms after stimulus onset, followed by a second peak around 600 ms. This replicates the finding from Experiment 1 that uncanny valley responses might involve two distinct components at different moments in time.

We subsequently obtained ratings from an independent sample of 20 participants using the same stimuli as in the EEG experiment. This time, we asked for explicit ratings of emotional expressiveness, realism, and uncanniness, as well as repeating the binary real face vs mask rating. Real faces were rated highest for emotional expressiveness (M = 3.9) and realism (M = 5.9), and lowest for uncanniness (M = 2.7) (Fig 7a7c). The realistic silicone masks were rated highest for uncanniness (M = 4.5), however this was not dramatically higher than for the Halloween masks (M = 4.3). Arguably making judgements of uncanniness is less appropriate for masks that are not intended to be realistic, though our data do qualitatively conform to the U-shaped function expected by the uncanny valley hypothesis. The pattern of d-prime scores (Fig 7d) was similar to those obtained in the main experiment (Fig 5b), with generally higher scores attributable to the unlimited inspection time permitted in this online follow-up experiment.

Fig 7. Additional ratings of mask images, completed by an independent sample (N = 20) with unlimited inspection time.

Fig 7

Images were rated on three dimensions using seven-point scales (panels a–c), and also judged as being either a real face or a mask, from which d-prime scores were calculated (panel d).

Discussion

Across two experiments using diverse stimuli, we identified a potential neurophysiological signature of the ‘uncanny valley’ effect. EEG responses to androids or silicone masks could be distinguished from responses to human faces at around 100 ms after stimulus onset, and also in a later time window around 500–800 ms after stimulus onset. There were no clear differences in the unimodal ERP response at posterior electrodes, but performance of a multivariate pattern classifier was above chance in these time windows. This is a different pattern from that observed for more obviously non-human stimuli (robots and Halloween masks), where there were both univariate and multivariate differences, and the multivariate discrimination accuracy was above chance for an extended time window. Perceptual judgements indicated that identification performance for uncanny valley stimuli was relatively poor, indicating confusion with real human images. We also confirmed that android images were perceived more negatively than either human or robot images, and that silicone masks were perceived as more uncanny than human faces. The similarity in results across our two experiments is striking and constitutes an internal conceptual replication of our main findings, suggesting that the neural characteristics of the uncanny valley effect are generalizable across stimulus categories.

The early time window when pattern classification is above chance corresponds approximately to the P100 and N170 components of the ERP. The P100 is typically associated with low-level visual responses, and is affected by contrast and spatial frequency content of an image. The N170 component is most often associated with faces, though is also observed for other image categories, and there is still debate about its precise function [9,10]. Similar early components have also been investigated in other ERP studies on the uncanny valley effect [31,32], and comparing human and robot faces [30]. This time window is unlikely to be modulated substantially by top-down influences, so we attribute the early component to image-based differences between stimulus categories [33]. ERP components in later time windows have also been studied in previous work [17,34,35], and may reflect cognitive processing stages, such as determining whether a stimulus conforms to categorical expectations. Differences between stimulus categories at these times are more amenable to top-down influences, and most likely involve higher brain areas outside of occipital cortex. We therefore predict that ERP components at later time points should correspond with perceptual judgements and reports of uncanniness - this is a worthwhile hypothesis for future work to investigate.

Our use of hyper-realistic silicone masks is novel in the context of the study of uncanny valley effects. Previous studies using these masks have demonstrated that they are difficult to distinguish from real faces [18,19], including in applied settings such as simulated passport control [36,37], and show large individual differences [38]. Image analysis indicates that good identification performance for silicone masks is typically based on attention to the region below the eyes [38], however it is plausible that many observers are not explicitly aware of the cues they use to perform this judgement. This might contribute to both the early and late components identified in this study, and presumably also to the subjective sensation of ‘uncanniness’ that is characteristic of the phenomenon.

Another increasingly common situation that triggers the ‘uncanny valley’ experience is in the domain of computer-generated images and movies [39,40]. Artificial intelligence algorithms are now able to generate images and movies based on text prompts (for example “a picture of a girl flying a kite in a field”) that often include human subjects. However, at time of writing, images of humans often contain errors, such as the presence of too many limbs, digits, teeth etc. Synthetic movies often contain continuity errors, and have issues reproducing biological motion. Many of these errors are subtle and take time to spot, but it is also the case that human observers can report that images look ‘wrong’ without explicitly knowing why. The neural uncanny valley effect that we report here might prove a useful index of these instinctive reactions, and could even potentially be used to improve artificial intelligence algorithms. For example, images could be penalised for producing neural responses that differed from those for natural images.

More generally, the advantage of measuring neural responses to ‘uncanny valley’ stimuli is that, without requiring conscious awareness or behavioural responses, they can facilitate detection of near-human stimuli. These types of near-human stimuli are becoming increasingly common in impersonation and identity evasion cases [1]. Simultaneously, we observe a growing market for reducing the uncanny valley effect for the benefit of android and robot integration. Exploring the potential of non-invasive brain recordings will benefit various applied fields as a result.

Conclusion

We have identified neural correlates of the uncanny valley effect that are consistent across two experiments, using androids and hyper-realistic silicone masks. In both cases, perceptual discrimination from real human faces was possible, but more challenging than discriminating from mechanical robots or Halloween masks. Univariate differences in the ERP signal were unconvincing, but a more sensitive multivariate classification analysis identified differences at both early (100–200 ms) and later (around 600 ms) time points. These findings suggest the importance of both bottom up and top down influences on the subjective experience of the uncanny valley. Future work might extend these findings to more dynamic stimuli, and explore potential applications for improving android and avatar generation.

Data Availability

Raw data, processed data, and analysis scripts are freely available through the project repository at: doi: 10.17605/OSF.IO/5NZ2H.

References

  • 1.Sanders JG, Jenkins R. Realistic masks in the real world. In: Bindemann M, editor. Forensic face matching: research and practice. Oxford University Press; 2021. doi: 10.1093/oso/9780198837749.003.0010 [DOI] [Google Scholar]
  • 2.Mori M. The uncanny valley. Energy. 1970;7:33–5. [Google Scholar]
  • 3.Mori M, MacDorman K, Kageki N. The uncanny valley [From the Field]. IEEE Robot Automat Mag. 2012;19(2):98–100. doi: 10.1109/mra.2012.2192811 [DOI] [Google Scholar]
  • 4.Vaitonytė J, Alimardani M, Louwerse MM. Scoping review of the neural evidence on the uncanny valley. Comput Hum Behav Rep. 2023;9:100263. doi: 10.1016/j.chbr.2022.100263 [DOI] [Google Scholar]
  • 5.Hu Y, Baragchizadeh A, O’Toole AJ. Integrating faces and bodies: psychological and neural perspectives on whole person perception. Neurosci Biobehav Rev. 2020;112:472–86. doi: 10.1016/j.neubiorev.2020.02.021 [DOI] [PubMed] [Google Scholar]
  • 6.Gauthier I, Tarr MJ, Moylan J, Skudlarski P, Gore JC, Anderson AW. The fusiform “face area” is part of a network that processes faces at the individual level. J Cogn Neurosci. 2000;12(3):495–504. doi: 10.1162/089892900562165 [DOI] [PubMed] [Google Scholar]
  • 7.Kanwisher N, McDermott J, Chun MM. The fusiform face area: a module in human extrastriate cortex specialized for face perception. J Neurosci. 1997;17(11):4302–11. doi: 10.1523/JNEUROSCI.17-11-04302.1997 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Downing PE, Jiang Y, Shuman M, Kanwisher N. A cortical area selective for visual processing of the human body. Science. 2001;293(5539):2470–3. doi: 10.1126/science.1063414 [DOI] [PubMed] [Google Scholar]
  • 9.Thierry G, Martin CD, Downing P, Pegna AJ. Controlling for interstimulus perceptual variance abolishes N170 face selectivity. Nat Neurosci. 2007;10(4):505–11. doi: 10.1038/nn1864 [DOI] [PubMed] [Google Scholar]
  • 10.Hong Y, Mayes MS, Munasinghe AP, Ratner KG. Scrutinizing whether mere group membership influences the N170 response to faces: results from two preregistered event-related potential studies. J Cogn Neurosci. 2022;34(11):1999–2015. doi: 10.1162/jocn_a_01887 [DOI] [PubMed] [Google Scholar]
  • 11.MacDorman KF. Does mind perception explain the uncanny valley? A meta-regression analysis and (de)humanization experiment. Comput Hum Behav: Artif Hum 2024;2(1):100065. doi: 10.1016/j.chbah.2024.100065 [DOI] [Google Scholar]
  • 12.MacDorman KF, Chattopadhyay D. Reducing consistency in human realism increases the uncanny valley effect; increasing category uncertainty does not. Cognition. 2016;146:190–205. doi: 10.1016/j.cognition.2015.09.019 [DOI] [PubMed] [Google Scholar]
  • 13.Moore RK. A Bayesian explanation of the “Uncanny Valley’’ effect and related psychological phenomena. Sci Rep. 2012;2:864. doi: 10.1038/srep00864 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Gray K, Wegner DM. Feeling robots and human zombies: mind perception and the uncanny valley. Cognition. 2012;125(1):125–30. doi: 10.1016/j.cognition.2012.06.007 [DOI] [PubMed] [Google Scholar]
  • 15.Yam KC, Bigman Y, Gray K. Reducing the uncanny valley by dehumanizing humanoid robots. Comput Hum Behav. 2021;125:106945. doi: 10.1016/j.chb.2021.106945 [DOI] [Google Scholar]
  • 16.Saygin AP, Chaminade T, Ishiguro H, Driver J, Frith C. The thing that should not be: predictive coding and the uncanny valley in perceiving human and humanoid robot actions. Soc Cogn Affect Neurosci. 2012;7(4):413–22. doi: 10.1093/scan/nsr025 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Urgen BA, Kutas M, Saygin AP. Uncanny valley as a window into predictive processing in the social brain. Neuropsychologia. 2018;114:181–5. doi: 10.1016/j.neuropsychologia.2018.04.027 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Sanders JG, Ueda Y, Minemoto K, Noyes E, Yoshikawa S, Jenkins R. Hyper-realistic face masks: a new challenge in person identification. Cogn Res. 2017;2(1). doi: 10.1186/s41235-017-0079-y [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Sanders JG, Ueda Y, Yoshikawa S, Jenkins R. More human than human: a turing test for photographed faces. Cogn Res Princ Implic 2019;4(1):43. doi: 10.1186/s41235-019-0197-9 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Bartneck C, Kulić D, Croft E, Zoghbi S. Measurement instruments for the anthropomorphism, animacy, likeability, perceived intelligence, and perceived safety of robots. Int J of Soc Robotics. 2008;1(1):71–81. doi: 10.1007/s12369-008-0001-3 [DOI] [Google Scholar]
  • 21.Koverola M, Kunnari A, Sundvall J, Laakasuo M. General attitudes towards robots scale (gators): a new instrument for social surveys. Int J of Soc Robotics. 2022;14(7):1559–81. doi: 10.1007/s12369-022-00880-3 [DOI] [Google Scholar]
  • 22.Baron-Cohen S, Wheelwright S, Skinner R, Martin J, Clubley E. The autism-spectrum quotient (AQ): evidence from Asperger syndrome/high-functioning autism, males and females, scientists and mathematicians. J Autism Dev Disord. 2001;31(1):5–17. doi: 10.1023/a:1005653411471 [DOI] [PubMed] [Google Scholar]
  • 23.Macmillan NA, Creelman CD. Detection theory: a user’s guide. Psychology Press; 2005.
  • 24.Delorme A, Makeig S. EEGLAB: an open source toolbox for analysis of single-trial EEG dynamics including independent component analysis. J Neurosci Methods. 2004;134(1):9–21. doi: 10.1016/j.jneumeth.2003.10.009 [DOI] [PubMed] [Google Scholar]
  • 25.Tadel F, Baillet S, Mosher JC, Pantazis D, Leahy RM. Brainstorm: a user-friendly application for MEG/EEG analysis. Comput Intell Neurosci. 2011;2011:879716. doi: 10.1155/2011/879716 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Rouder JN, Speckman PL, Sun D, Morey RD, Iverson G. Bayesian t tests for accepting and rejecting the null hypothesis. Psychon Bull Rev. 2009;16(2):225–37. doi: 10.3758/PBR.16.2.225 [DOI] [PubMed] [Google Scholar]
  • 27.Jeffreys H. Theory of probability. 1961.
  • 28.Chang C-C, Lin C-J. LIBSVM. ACM Trans Intell Syst Technol. 2011;2(3):1–27. doi: 10.1145/1961189.1961199 [DOI] [Google Scholar]
  • 29.Grootswagers T, Wardle SG, Carlson TA. Decoding dynamic brain patterns from evoked responses: a tutorial on multivariate pattern analysis applied to time series neuroimaging data. J Cogn Neurosci. 2017;29(4):677–97. doi: 10.1162/jocn_a_01068 [DOI] [PubMed] [Google Scholar]
  • 30.Geiger AR, Balas B. Robot faces elicit responses intermediate to human faces and objects at face-sensitive ERP components. Sci Rep 2021;11(1):17890. doi: 10.1038/s41598-021-97527-6 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31.Schindler S, Zell E, Botsch M, Kissler J. Differential effects of face-realism and emotion on event-related brain potentials and their implications for the uncanny valley theory. Sci Rep. 2017;7:45003. doi: 10.1038/srep45003 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Mustafa M, Magnor M. EEG based analysis of the perception of computer-generated faces. In: Proceedings of the 13th European Conference on Visual Media Production (CVMP 2016). ACM; 2016. doi: 10.1145/2998559.2998563 [DOI]
  • 33.Coggan DD, Baker DH, Andrews TJ. The role of visual and semantic properties in the emergence of category-specific patterns of neural response in the human brain. eNeuro. 2016;3(4):ENEURO.0158-16.2016. doi: 10.1523/ENEURO.0158-16.2016 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Mustafa M, Guthe S, Tauscher J-P, Goesele M, Magnor M. How human am i?: EEG-based evaluation of virtual characters. In: Proceedings of the 2017 CHI Conference on Human Factors in Computing Systems. ACM; 2017. doi: 10.1145/3025453.3026043 [DOI]
  • 35.Cheetham M, Wu L, Pauli P, Jancke L. Arousal, valence, and the uncanny valley: psychophysiological and self-report findings. Front Psychol. 2015;6:981. doi: 10.3389/fpsyg.2015.00981 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36.Robertson DJ, Sanders JG, Towler A, Kramer RSS, Spowage J, Byrne A, et al. Hyper-realistic face masks in a live passport-checking task. Perception. 2020;49(3):298–309. doi: 10.1177/0301006620904614 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37.Robertson DJ, Davis JP, Sanders JG, Towler A. The super-recogniser advantage extends to the detection of hyper-realistic face masks. Appl Cognit Psychol 2024;38(4):e4222. doi: 10.1002/acp.4222 [DOI] [Google Scholar]
  • 38.Sanders JG, Jenkins R. Individual differences in hyper-realistic mask detection. Cogn Res Princ Implic. 2018;3:24. doi: 10.1186/s41235-018-0118-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39.Moshel ML, Robinson AK, Carlson TA, Grootswagers T. Are you for real? decoding realistic AI-generated faces from neural activity. Vision Res. 2022;199:108079. doi: 10.1016/j.visres.2022.108079 [DOI] [PubMed] [Google Scholar]
  • 40.Gu Z, Jamison K, Sabuncu MR, Kuceyeski A. Human brain responses are modulated when exposed to optimized natural images or synthetically generated images. Commun Biol 2023;6(1):1076. doi: 10.1038/s42003-023-05440-7 [DOI] [PMC free article] [PubMed] [Google Scholar]

Decision Letter 0

Rae Yule Kim

9 Dec 2024

PONE-D-24-42223Neural correlates of the uncanny valley effect for robots and hyper-realistic masksPLOS ONE

Dear Dr. Baker,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please submit your revised manuscript by Jan 23 2025 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Rae Yule Kim

Academic Editor

PLOS ONE

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf.

2. Thank you for stating in your Funding Statement:

“SF was funded by a YorRobots Venables internship. JGS was funded by the Economic and Social Research

Council (Studentship ES/J500215/1) and Prins Bernhard Cultuurfonds (grant E/30/30.13.0630/HVH/IE).

Also supported by BBSRC grant BB/V007580/1 awarded to DHB.”

Please provide an amended statement that declares *all* the funding or sources of support (whether external or internal to your organization) received during this study, as detailed online in our guide for authors at http://journals.plos.org/plosone/s/submit-now.  Please also include the statement “There was no additional external funding received for this study.” in your updated Funding Statement.

Please include your amended Funding Statement within your cover letter. We will change the online submission form on your behalf.

3. Thank you for stating the following in the Acknowledgments Section of your manuscript:

“SF was funded by a YorRobots Venables internship. JGS was funded by the Economic and Social Research Council (Studentship ES/J500215/1) and Prins Bernhard Cultuurfonds (grant E/30/30.13.0630/HVH/IE). Also supported by BBSRC grant BB/V007580/1 awarded to DHB.”

We note that you have provided additional information within the Acknowledgements Section that is not currently declared in your Funding Statement. Please note that funding information should not appear in the Acknowledgments section or other areas of your manuscript. We will only publish funding information present in the Funding Statement section of the online submission form.

Please remove any funding-related text from the manuscript and let us know how you would like to update your Funding Statement. Currently, your Funding Statement reads as follows:

“SF was funded by a YorRobots Venables internship. JGS was funded by the Economic and Social Research

Council (Studentship ES/J500215/1) and Prins Bernhard Cultuurfonds (grant E/30/30.13.0630/HVH/IE).

Also supported by BBSRC grant BB/V007580/1 awarded to DHB.”

Please include your amended statements within your cover letter; we will change the online submission form on your behalf.

4. Please update your submission to use the PLOS LaTeX template. The template and more information on our requirements for LaTeX submissions can be found at http://journals.plos.org/plosone/s/latex.

5. We note that Figure 1 includes an image of a [patient / participant / in the study].

As per the PLOS ONE policy (http://journals.plos.org/plosone/s/submission-guidelines#loc-human-subjects-research) on papers that include identifying, or potentially identifying, information, the individual(s) or parent(s)/guardian(s) must be informed of the terms of the PLOS open-access (CC-BY) license and provide specific permission for publication of these details under the terms of this license. Please download the Consent Form for Publication in a PLOS Journal (http://journals.plos.org/plosone/s/file?id=8ce6/plos-consent-form-english.pdf). The signed consent form should not be submitted with the manuscript, but should be securely filed in the individual's case notes. Please amend the methods section and ethics statement of the manuscript to explicitly state that the patient/participant has provided consent for publication: “The individual in this manuscript has given written informed consent (as outlined in PLOS consent form) to publish these case details”.

If you are unable to obtain consent from the subject of the photograph, you will need to remove the figure and any other textual identifying information or case descriptions for this individual.

5. Please review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice.

Additional Editor Comments (if provided):

The reviewers have expressed a positive view of your paper’s contribution to the field of human-robot interaction and are supportive of its acceptance, pending minor revisions.

In particular, the reviewers recommend enhancing the discussion with real-world examples to underscore the relevance of your findings. Additionally, they suggest providing further clarification on certain aspects of your methodology to improve the overall clarity of the paper.

Please address the reviewers’ comments and include a response letter for each reviewer with your revised manuscript.

We look forward to receiving your revised submission.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Partly

Reviewer #3: Yes

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: Dear Authors,

I commend you on your fascinating study of the neural correlates of the uncanny valley effect. Your methodology is solid, and the findings provide valuable insights into human responses to artificial entities, with important implications for AI and robotics design.

Please find more detailed feedback and suggestions in the reviewer attachment. In summary, I recommend the following:

1. Simplify some technical terms in the abstract and introduction for broader accessibility.

2. Provide more context and explanation in the Methods section, especially regarding statistical analysis and EEG terms.

3. Expand figure captions and define EEG terms like P100 and N170 to aid readers’ understanding.

4. Consider exploring further applications of your findings in the Discussion and including alternative explanations.

5. In the Conclusion, suggest concrete directions for future research.

6. Enhance clarity by adding clearer labels in figures and defining technical terms earlier.

I believe your manuscript has great potential and, with these minor revisions, will be ready for publication.

Best regards,

Reviewer #2: The authors presented an interesting work on detecting the uncanny valley effect for robots and human-like subjects. The statistical analysis and results are well presented; however, the following changes can further improve the manuscript.

1. The motivation and application of the study should also be added to the introduction section.

2. The literature survey for similar works is short and needs improvement.

3. An image of the experimentation procedure should be added.

Reviewer #3: Genearl comments:

This study provides a valuable contribution to the understanding of the neural correlates of the uncanny valley effect through the use of EEG analysis on static images of androids and hyper-realistic silicone masks. The findings, such as the identification of distinct neural responses in early (100-200ms) and late (500-800ms) time windows, offer fresh insights into both bottom-up and top-down processes involved in this phenomenon. Additionally, the study demonstrates consistency across two experiments, enhancing the reliability of its conclusions.

However, there are areas that require further clarification and development to strengthen the manuscript. For example, the introduction does not fully articulate the motivation behind using static images and their advantage over dynamic stimuli in uncovering neural mechanisms. Additionally, the experimental design lacks detail regarding stimulus selection criteria and the rationale behind conditions. While EEG and behavioral data were analyzed, their integration remains insufficient, and the neural mechanisms underlying the temporal windows are not explored in depth. Furthermore, subjective ratings (e.g., uncanniness and realism) are underutilized, as their relationship with neural or behavioral data is not analyzed. Lastly, while the discussion highlights potential applications, such as improving AI-generated images, it does not provide sufficient details on how these applications might be implemented.

Major Issues and Recommendations

1. Insufficient Expression of Research Motivation

The introduction does not clearly articulate the scientific significance of the study, such as why investigating static images can complement existing research on dynamic stimuli. This may lead to ambiguity regarding the study’s uniqueness and contributions.

Recommendation: Emphasize the importance of static image research: “Although existing studies focus on dynamic stimuli, static images allow for a more precise investigation of the neural mechanisms underlying the uncanny valley effect, particularly by eliminating motion-related confounds.”

2. Lack of Clarity in Experimental Objectives

The description of the experimental design and objectives is too vague, failing to specify the criteria for selecting static images or the distinctions between experimental conditions.

Provide a clearer explanation of the study’s goals and design, detailing the selection process for images and the specific distinctions between conditions.

3. Insufficient Discussion of EEG Data Analysis

The ERP differences are only briefly discussed, without exploring the potential neural mechanisms of specific time windows (e.g., whether the 300-400ms differences are linked to cognitive conflict).

Deepen the discussion on ERP mechanisms: “The 300-400ms time window may reflect cognitive processing stages, such as determining whether a stimulus conforms to categorical expectations. Differences between human and robot conditions could indicate the unique processing advantage of human faces.”

4. Weak Integration of Results

There is no clear explanation of how d′ values, reaction times, and EEG data integrate to address the research question, making the relationship between methodologies unclear.

Connect different data types to answer the research question: “By integrating d′ values, reaction times, and EEG data, we explore how neural activity in specific time windows correlates with behavioral metrics like accuracy and response time.”

5. Poor Integration of Subjective Ratings with Neural Data

Perception ratings (e.g., uncanniness, realism) are not linked to EEG data. The connection between subjective experiences and neural signals is critical but underexplored.

Analyze the relationship between ratings and EEG data: “Higher uncanniness ratings may correspond to greater ERP differences in the late time window (600ms), indicating a link between subjective perceptions and neural responses.”

Minor Issues and Recommendations

1. Weak Literature Review

The discussion of existing research is underdeveloped, particularly regarding theoretical disagreements on the neural mechanisms (bottom-up vs. top-down processes).

Expand the literature review: “The root of the uncanny valley effect remains debated: does it arise primarily from bottom-up sensory conflicts, or from higher-level cognitive processes? Resolving this question is critical to understanding its fundamental mechanisms.”

2. Lack of Explanation for Stimulus Presentation Time

The rationale behind the stimulus presentation times (500ms for Experiment 1, 250ms for Experiment 2) and random blank periods is not explained.

Justify the choices: “The 500ms and 250ms durations were chosen to provide sufficient information for judgment while avoiding task fatigue. Randomized blank periods were designed to reduce carryover effects and prevent anticipatory biases.”

3. Insufficient Analysis of Image Background Differences

Experiment 1 used gray backgrounds, while Experiment 2 used natural backgrounds, but the potential impact of this difference is not addressed.

Discuss the impact of background differences: “While the backgrounds differed across experiments, we hypothesize that the primary task was not affected, as participants focused on the foreground stimuli.”

4. Lack of Detail in d′ Value Calculation

The calculation method for d′ values lacks detail, particularly for edge cases (e.g., hit rate = 1 or 0) and the rationale for capping infinite values.

Provide detailed explanations: “d′ values were capped at 5 to prevent outliers from skewing the results, following established conventions in signal detection theory.”

5. Insufficient Interpretation of Figures

Figures (e.g., 2 and 3) are not adequately interpreted, leaving the reader unclear about their implications.

Offer more detailed interpretations: “Figure 2a shows that robot stimuli have significantly higher d′ values than humanoids and humans, emphasizing the role of visual salience.”

6. Limited Context for Multivariate Pattern Analysis

High classification accuracy for human vs. robot stimuli (72%) and lower accuracy for humanoids (62%) is presented without explaining the reasons for this disparity, such as the salience of visual features.

Explore potential causes for classification differences: “The high classification accuracy for human vs. robot stimuli may be attributed to the salient mechanical elements of robots, whereas the lower accuracy for humanoids reflects their ambiguous human-like appearance, leading to confusion.”

7. Lack of Discussion on Cross-Experiment Consistency

While the consistency of findings across experiments is mentioned, its implications for the uncanny valley theory are not explored.

Emphasize the theoretical significance of consistent results: “The consistency of results across humanoids and hyper-realistic silicone masks suggests that the neural characteristics of the uncanny valley effect are generalizable across stimulus categories, indicating its broad applicability.”

8. Insufficient Analysis of Temporal Windows

The mechanisms underlying the two temporal windows (100ms and 500-800ms) are oversimplified, with no mention of their potential links to emotion processing or visual memory.

Provide detailed hypotheses for each window:

100ms window: “This window likely reflects rapid processing of low-level visual features, such as edges or color contrasts, consistent with P100 and N170 components.”

500-800ms window: “This window may involve higher-level cognitive processes, such as evaluating emotional content or judging authenticity.”

The study makes a significant contribution by identifying neural correlates of the uncanny valley effect and providing practical implications for AI and identity detection technologies. However, by addressing the above suggestions, the authors can further enhance the clarity, depth, and impact of their work.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: Yes: Osamah AL-Qalisi

Reviewer #2: No

Reviewer #3: No

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

Attachment

Submitted filename: Peer Review 3 OSQ.pdf

pone.0311714.s001.pdf (157.8KB, pdf)
Attachment

Submitted filename: PONE-D-24-42223-Comments.docx

pone.0311714.s002.docx (17.5KB, docx)

Decision Letter 1

Rae Yule Kim

7 Feb 2025

Neural correlates of the uncanny valley effect for robots and hyper-realistic masks

PONE-D-24-42223R1

Dear Dr. Baker,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice will be generated when your article is formally accepted. Please note, if your institution has a publishing partnership with PLOS and your article meets the relevant criteria, all or part of your publication costs will be covered. Please make sure your user information is up-to-date by logging into Editorial Manager at Editorial Manager® and clicking the ‘Update My Information' link at the top of the page. If you have any questions relating to publication charges, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Rae Yule Kim

Academic Editor

PLOS ONE

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #2: All comments have been addressed

Reviewer #3: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #2: Yes

Reviewer #3: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #2: Yes

Reviewer #3: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #2: Yes

Reviewer #3: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #2: Yes

Reviewer #3: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #2: (No Response)

Reviewer #3: The revised manuscript has significantly improved, addressing previous concerns effectively. The clarity, analysis, and overall quality are now strong. I recommend acceptance in its current form. Well done to the authors.

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #2: No

Reviewer #3: No

**********

Acceptance letter

Rae Yule Kim

PONE-D-24-42223R1

PLOS ONE

Dear Dr. Baker,

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now being handed over to our production team.

At this stage, our production department will prepare your paper for publication. This includes ensuring the following:

* All references, tables, and figures are properly cited

* All relevant supporting information is included in the manuscript submission,

* There are no issues that prevent the paper from being properly typeset

If revisions are needed, the production department will contact you directly to resolve them. If no revisions are needed, you will receive an email when the publication date has been set. At this time, we do not offer pre-publication proofs to authors during production of the accepted work. Please keep in mind that we are working through a large volume of accepted articles, so please give us a few weeks to review your paper and let you know the next and final steps.

Lastly, if your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

If we can help with anything else, please email us at customercare@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Rae Yule Kim

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    Attachment

    Submitted filename: Peer Review 3 OSQ.pdf

    pone.0311714.s001.pdf (157.8KB, pdf)
    Attachment

    Submitted filename: PONE-D-24-42223-Comments.docx

    pone.0311714.s002.docx (17.5KB, docx)
    Attachment

    Submitted filename: responsetoreviewsdocx.docx

    pone.0311714.s003.docx (28.2KB, docx)

    Data Availability Statement

    Raw data, processed data, and analysis scripts are freely available through the project repository at: doi: 10.17605/OSF.IO/5NZ2H.

    Raw data, processed data, and analysis scripts are freely available through the project repository at: doi: 10.17605/OSF.IO/5NZ2H


    Articles from PLOS One are provided here courtesy of PLOS

    RESOURCES