Skip to main content
PLOS One logoLink to PLOS One
. 2024 Oct 7;19(10):e0311225. doi: 10.1371/journal.pone.0311225

Horizontal face information is the main gateway to the shape and surface cues to familiar face identity

Helene Dumont 1,*, Alexia Roux-Sibilon 1,2, Valérie Goffaux 1,3
Editor: Patrick Bruns4
PMCID: PMC11458052  PMID: 39374235

Abstract

Humans preferentially rely on horizontal cues when recognizing face identity. The reasons for this preference are largely elusive. Past research has proposed the existence of two main sources of face identity information: shape and surface reflectance. The access to surface and shape is disrupted by picture-plane inversion while contrast negation selectively impedes access to surface cues. Our objective was to characterize the shape versus surface nature of the face information conveyed by the horizontal range. To do this, we tracked the effects of inversion and negation in the orientation domain. Participants performed an identity recognition task using orientation-filtered (0° to 150°, 30° steps) pictures of familiar male actors presented either in a natural upright position and contrast polarity, inverted, or negated. We modelled the inversion and negation effects across orientations with a Gaussian function using a Bayesian nonlinear mixed-effects modelling approach. The effects of inversion and negation showed strikingly similar orientation tuning profiles, both peaking in the horizontal range, with a comparable tuning strength. These results suggest that the horizontal preference of human face recognition is due to this range yielding a privileged access to shape and surface cues, i.e. the two main sources of face identity information.

Introduction

Among the myriad of essential social cues conveyed by the human face (e.g., age, emotion, familiarity…; [1], identity plays a crucial role in human social organization (e.g., cooperation, kin recognition). The ability to recognize identity in a species is supported by the presence of distinctive identity signals [2]. The study of Sheehan & Nachman (2014) [3] suggests that human social evolution maximized the human face phenotypic and genetic diversity. For instance, this is shown by a larger variation of length and distance among facial features than among non-facial features (i.e. someone with big hands is more likely to have big fingers, but someone with a big nose in not more likely to have big eyes or mouth).

The important function of human face identification relies on specialized visual mechanisms and brain regions (e.g., [47]). Humans reach very high identification performance especially in the case of naturally familiar faces, which can be recognized even with bad quality images or videos and drastic changes in appearance [8, 9]. Such tolerance of familiar face recognition is impressive as variability in the appearance of the same face across pictures taken under different conditions (illumination, viewpoint, expression, camera) often exceeds variability between pictures of different identities taken under the same conditions [8, 10, 11]. The visual mechanisms and information allowing the tolerance of familiar face identification are not yet well understood and are extensively researched in vision sciences [1215].

Recent evidence showed that the information that drives human face identity recognition is optimally conveyed by an orientation range centred around the horizontal axis. Dakin & Watt (2009) [16] filtered familiar face images in the Fourier domain to only keep the information contained in a narrow orientation band. They found that identity recognition performance describes a bell-shaped curve peaking in the horizontal range and reaching its minimum in the vertical range. Such horizontal tuning of face identity recognition has been documented in several studies, and replicates across various methodologies, e.g. orientation filtering [1620], oriented noise masking [19, 21] and oriented phase randomization [22]. It is present from three months of age on [23] and develops over lifespan [17, 24, 25]. The perception of face emotional expressions has also been found to be horizontally-tuned [26, 27]. This horizontal tuning is correlated to face recognition performance in participants exhibiting normal recognition skills [21, 28, 29] and is stronger for the recognition of familiar faces [30], suggesting a link between horizontal tuning and genuine and specialized face identity recognition.

This link between horizontal tuning and face-specialized identity processing is further supported by evidence that the former is disrupted by picture-plane inversion [1822]. Inversion, while largely preserving image content, greatly disrupts face recognition. Faces shown upside-down are harder to recognize than faces shown in a canonical upright position, and this decrease in recognition performance affects faces disproportionately more than other mono-oriented objects or scenes [3136]. Similarly to inversion, another image manipulation known to affect the recognition of faces more than other objects is contrast-polarity reversal (referred to as negation in this paper; [37, 38]).

The disproportionate effects of inversion and negation indicate the existence of visual mechanisms tuned to the natural statistics of the human face. Indeed, over evolution and lifespan, the human visual system presumably builds an internal representation of how faces appear in everyday life, i.e., usually in upright position, lit from above [39] and depicting characteristic contrast relationships (i.e., an alternation of light and dark, with facial features relatively darker than surrounding skin; [4042]). The typical vertical contrast alternation of the human face is thought to be conveyed by the horizontal range of the face image [16, 43]. Such natural statistics are reversed in space and polarity in inverted and negated faces, respectively, which likely explains the drastic influence of these manipulations on face recognition [16]. Past evidence showed that inversion strongly weakens the horizontal tuning of human face identity recognition, with a reduction in the amplitude of the horizontal peak and a doubling in bandwidth [20]. However, whether and how the the polarity reversal of the face contrast alternance caused by negation influences the orientation tuning profile of human face identification remains elusive.

While inversion and negation both alter horizontal face structure, they do so in very distinct ways. Indeed, while inversion flips the edge and shape properties of the face upside-down, negation reverses luminance gradient polarity but preserves edge and shape information. Past behavioural [4446] and neural [47, 48] evidence further indicates that they disrupt distinct aspects of face processing. Inversion is commonly thought to disrupt the so-called holistic processing (i.e. perceiving the face as a whole; [5, 34, 49]) of face shape properties [4951]. On the other hand, negation has been proposed to disrupt the access to face surface cues such as pigmentation [44, 52, 53], and shape-from-shading cues [44, 54, 55]. Although some recent studies found little to no effect of negation on holistic processing [45, 56], others reported that disruption of surface cues with negation could also result in a disruption of holistic processing [57]. Overall, while some studies indicate some functional overlap between these effects, inversion and negation are generally thought to disrupt relatively distinct aspects of face information: shape and surface, respectively.

Our study characterizes the orientation profile of inversion and negation effects on human face recognition in order to infer the shape versus surface nature of the information contained in the horizontal range of the face stimulus. We used familiar face stimuli to tap into genuine and specific face recognition mechanisms [8, 30]. We measured the difference between the orientation tuning profile for natural faces (faces in natural upright orientation and contrast polarity) and that for (i) inverted faces and (ii) negated faces in the same participants. A difference in the horizontal tuning profile of the inversion and negation effects would indirectly point to a differential contribution of this orientation range to shape and surface processing. Alternatively, finding that inversion and negation similarly modulate recognition orientation profile would indicate that the horizontal range conveys either the information driving shape and surface cue processing of the face image, or a more fundamental aspect of the natural statistics of the upright face coming in play prior to the presumed functional dissociation of shape and surface cue processing.

Material and methods

Subjects

40 healthy young adults were recruited (from 15/06/22 to 12/12/22) for the experiment. Only 21 of them, eight males, aged 23.6 (± 4.5) years old, recognized the requested number of celebrities and completed the experiment in exchange for (i) course credits (1 credit per hour of testing) for participants recruited via a psychology course, or (ii) monetary compensation (10 euros per hour of testing) for participants recruited through a Facebook ad. All participants gave their written, informed consent, based on a short explanation of the protocol before starting the experiment. At the end of the experiment (usually after three sessions) participants received detailed information regarding the experiment and had the opportunity to ask any question they might have had during a debriefing. Visual acuity was tested with the Landolt C test of the FrACT10 v1.0 [58] and all logMAR scores indicated normal or corrected to normal vision (logMAR<0). Face perception abilities were measured with a computerized Benton Facial Recognition Task [59] and all scores indicated normal face perception (score = 46.6 ± 3.6). The experimental protocol was approved by local ethical committee (Psychological Sciences Research Institute, UCLouvain).

Stimuli

In order to compare the influence of inversion and negation on the horizontal tuning of face identity recognition, we measured participants’ identification performance with upright-positive (i.e., ‘natural’), inverted, and contrast-negated stimuli that were either full-spectrum, or orientation-filtered in Fourier domain.

Stimuli consisted of greyscale, 302 x 309 pixels, full front or close to full front, faces images of contemporary white American or British famous male actors (18 actors, four pictures each) selected from the Facescrub dataset [60] and a Google search for free to use images. Images were carefully selected to exclude any distinctive accessory (glasses, hat, …) and to have a neutral or close-to-neutral facial expression. Eighteen celebrities were selected from a larger set of actor images with an online pilot running on Pavlovia (Open Science Tools, Nottingham, UK) to ensure that they were recognized by most participants. We only kept the four best recognized images for each recognized actor (see S1 in S1 File), resulting in a total of 72 images.

Images were converted to greyscale and rotated to align the eyes with the horizontal meridian. We then cropped and resized each one using forehead and chin as points of reference to match face size across all images. The resulting images were padded with 100 pixels that were mirroring the face image, to prevent border artefacts during filtering. Next, images were normalized to obtain a mean luminance and root-mean square (RMS) contrast of 0 and 1 respectively and fast Fourier transformed, with the amplitude spectra multiplied with six wrapped Gaussian filters centered on 0° to 150° in steps of 30°, where 0° is vertical and 90° horizontal [16, 19, 20]. The Gaussian filters’ standard deviation was set at 25° to approximately match the standard deviation of the horizontal tuning of the identity recognition performance in upright unfamiliar faces previously documented by Goffaux & Greenwood (2016) [20] and were close to those used by Dakin and Watt (2009; 23°). This broad bandwidth resulted in some overlap between adjacent orientation ranges but was chosen to ensure that recognition was possible even in the most difficult condition (i.e. inverted and 0° filter). The proportion of clipped pixels in an image (i.e., pixels with an intensity below 0 or above 1, on a [0,1] scale) was below 3%. Mean luminance and RMS contrast of all images were set to the mean luminance and RMS contrast of the image set (0.5 and 0.12 respectively). A grey elliptical aperture was created to match the shape of an average image of all the faces in the set, and blend with the background of the experiment. It was blended with each image (filtered and non-filtered), after RMS equalization, to cover most of the background, hair and the neck (Fig 1). Gaussian white noise masks were also designed with 1/f noise, and global luminance and RMS contrast set to match those of the face images. We used Matlab R2014a for these image processing steps.

Fig 1. Example of the stimulus conditions.

Fig 1

Each column represents the orientation-filter condition starting by non-filtered images and going in steps of 30° from 0° to 150° from left to right. Each row represents the stimulus type (natural, inverted, negated).

In the experiment, images were displayed naturally, inverted or negated. Inverted images were presented using a 180° clockwise rotation on Psychopy. Negated images were made prior to the experiment, using Matlab by reversing contrast values, after filtering, luminance and contrast equalization but before the elliptical aperture was blended with the picture.

Procedure

Participants did the experiment on a VIEWpixx monitor (VPixx Technologies Inc., Saint-Bruno, Canada) with a 1920 × 1080 pixels resolution and a 70Hz refresh rate. They used a head-and-chin-rest to support their head and maintain a viewing distance of 60 cm. Face images presented at this distance subtended a visual angle of 9° by 10.4°, which simulates an observer distance ideal for holistic processing [61]; masks had a diameter of 11°. The experiment lasted for three one-hour sessions. We used Psychopy2021.2.3 to code and run the experiment.

During a prescreening phase, each participant selected 10 actors that they were familiar with (see S1 in S1 File for details). To do so, participants performed a 5-alternative forced choice identity recognition task with upright full spectrum images (four images for each of the 18 actor identities as selected based on the online pilot ran with a different participant sample; see Stimuli). Each image appeared on the center of the screen for 500ms followed by the names of five actors: the name of the actor corresponding to the image and four other names, randomly taken out of the 17 remaining names. Participants were instructed to select the name corresponding to face identity. An actor’s identity was selected only if at least three out of the four images were recognized. Once 10 familiar identities were selected, the four images of each identity were displayed again on the monitor for five seconds. With this procedure, the selection of actors was different for each participant and adapted to their knowledge of the identities, thus ensuring a recognition performance rooted into natural familiarity during the main experiment [8]. In the next training, and main experimental trials, three out of the four images per actor were used as stimuli. The remaining images were presented in-between blocks, to remind participants of unmodified selected identities without exposing them with actual stimulus.

Participants performed an identity face-matching task (Fig 2). A trial started with a white fixation dot on a grey background. The first face image (probe) was presented during 500ms followed by a mask presented for 200ms and a second face image (target) for 400ms. Participants had to decide whether faces in the just-seen-pair depicted the same identity or different identities with a keypress (‘S’ for same identity or ‘L’ for different identity). Participants had up to 4400 ms to respond from target onset. After target offset, the fixation dot turned black. When an answer was recorded, the fixation dot briefly turned green for a correct answer and red for an incorrect one. The following trial started after a random duration interval (ranging from 750 to 1350ms). Identity was matching in 50% of the trials. Pairs of each trial were composed after the prescreening phase for each participant (30 pairs for same and 30 pairs for different trials), and replicated across conditions. Importantly, in the case of same trials, both images were always two different images of the same identity, ensuring participants were recognizing identity and not merely matching images. Both images of a trial were presented under same stimulus type and orientation-filter condition. During the piloting phase of this study, we noticed difficulties for participants to reach above chance accuracy when conditions were fully randomized. We thus decided to block stimulus type to facilitate task performance.

Fig 2. Example same trial sequence of the main experiment.

Fig 2

Each trial starts with a fixation of random duration (between 750ms to 1350ms). A probe face is then presented for 500ms, followed by a 200ms mask and a target face for 400ms. Participants can answer from target onset and up to 4000ms after target offset by pressing either ‘S’ for same identity or ‘L’ for different identity. The fixation dot briefly turns green after a correct answer and red after an error.

As our task involved recognition of face images with orientation filtering, coupled with either inversion or negation, participants first started with a training phase. They were familiarized with the task with increasing levels of difficulty, to ensure they would reach stable and sufficient performance before gathering data. The training phase was divided in three steps of increasing difficulty. In the first training step, participants did the face identity matching task with natural, full spectrum face images. Once participants reached an accuracy > 80%, training went on to the second step. In the second step, face images were either natural, inverted or (contrast-)negated and full spectrum, and presented per Stimulus type in 14-trial blocks. Blocks were grouped by three, one for each Stimulus type, and scores were calculated over these groups of three blocks. Inside each group of blocks, the order of blocks (i.e., of Stimulus type) was set randomly. Once participants achieved a mean score > 75%, training went on to the third training step which was exactly the same as the second, except that within each block (natural, inverted or negated), stimuli were presented both filtered in orientations and full spectrum (i.e., two trials of each of the six orientations, from 0° to 150°, and two full spectrum trials). To pass this last part of training participants needed to reach a mean score of at least 75% (calculated over three-blocks groups with the three Stimulus type conditions), twice in a row. This ensured participants had good comprehension of the task and were able to score well above chance level (50%) before starting the testing phase.

The main experiment consisted of 90 blocks (30 blocks per Stimulus type). Each block was composed of 14 trials: two trials of each orientation filter plus two full spectrum trials. This resulted in a total of 60 trials per condition (i.e., combination of each Stimulus type and Orientation filter), 420 trials per Stimulus type, and 1260 trials overall. With 30 images (three per preselected identities), three Stimulus type (i.e., natural, inverted, negated) and seven Orientation Filter (i.e., 0° to 150° in steps of 30° and full spectrum) conditions, there was a total of 630 images, each repeated four times (twice in same trials twice in different trials) throughout the experiment.

Stimulus type was blocked while Orientation filter varied randomly within blocks. Blocks were presented three by three, one per Stimulus type, in a randomized order, with a short break in-between (of a minimum duration of five seconds) and a light tone indicating the beginning of a new block. Longer breaks (of at least 30 seconds) happened every three blocks. Longer breaks started with a score informing participants on their performance over the last three blocks and the number of remaining blocks; and ended with the (non-filtered) picture of each selected actor that was not used during the experimental trials, along with their name.

Data analysis

Through the comparison of inversion and negation effects on the Orientation filter dependence of identity recognition performance, the purpose of this study was to understand the extent to which the horizontal tuning of face identification reflects its contribution to the processing of shape versus surface cues.

During the experiment accuracy was recorded for each participant. All participants scored above chance level. We calculated the mean response time and standard deviation of each participant and considered outlier every answer with a response time faster than mean − 2.5*sd or slower than mean + 2.5*sd, which resulted in a loss of 3.3% of all answers.

From accuracy data, we computed a sensitivity index (d’) to assess participants’ performance in bias-free manner (accuracy was also analyzed and produced very similar results–this analysis can be found in S6 in S1 File). We computed a d’ for each combination of Stimulus type and Orientation filter (with a same answer in the presence of same identity considered as a hit, a different answer in the presence of different identities considered as a correct rejection, a same answer in the presence of different identities considered as a false alarm, and a different answer in the presence of the same identity considered as an omission). Following the log-linear correction [62], we added 0.5 to the hits and false-alarm counts, and calculated the d’ value for each subject in each condition and filter. The normalization of dependent variables is common practice when using complex hierarchical models, as it facilitates model convergence (Stan Development Team, 2023). Sensitivity was thus normalized with a Z transform using this formula:

normd(x)=d(x)mean(d(x))sd(d(x))

where, for a given combination of Orientation filter and Stimulus type x, normd’ is the normalized d’ output, d’ is the non-normalized sensitivity value, mean(d’) is the group-level mean d’, and sd(d’) is the group-level standard deviation of d’.

Performance at the 0° filter was duplicated to have circular filter values from 0° to 180°.

First, to ensure participants scored higher for natural stimulus type, which we use as baseline to calculate inversion and negation effects, average performance (across Orientation filters, excluding the full spectrum trials) in each stimulus type condition was compared to the others using a Bayesian general linear mixed model (GLMM), using participants as a random effect, to consider inter-subject variability in the performance estimations for each condition, with this equation:

normdStimulustype+(Stimulustype|Subject)

We ran the model with default brms priors with four Markov chain Monte Carlo (MCMC) with 6,000 iterations each (2,000 warm up, resulting in 16,000 iterations after warmup), and evaluated the convergence of the chains with diagnostic values such as number of divergent chains post warmup, Rhat statistics and effective sample size.

As expected, the GLMM indicated that participants performed better in natural condition than in inverted and negated conditions (Fig 3).

Fig 3.

Fig 3

a. Group-level normalized sensitivity (normd’) for identity recognition as a function of Orientation filters. The black diamonds represent the mean performance for each condition and the boxplots represent the 25%, 50% (median) and 75% quantiles. Each point is one participant’s recognition performance for an Orientation filter and Stimulus type. The half violins represent the distribution of the data for each condition. The full lines represent the full spectrum performance for each condition. b. Predictions from the Bayesian general linear mixed model comparison of Stimulus type. The non-overlapping 89% credible intervals (CrI) of model predictions indicate a substantive performance difference between each stimulus type. The triangles and error bars represent the group mean and standard error of the data in each condition. The square is the mean in each condition given by the model prediction, and the doted lines are the 89% CrI. c. Population-level averaged recognition performance for natural images and Bayesian Gaussian mixed model predictions. The dots are the population-level averaged normalized sensitivity with the 95% CrI in error bars. The model mean posterior prediction is drawn in black (line) and the 89% CrI of the predictions are represented as the faded ribbons.

Recognition performance for natural face images described a broad Gaussian centered around 90° and was modelled using a Bayesian Gaussian mixed model (Fig 3). The model is similar as the one used by Goffaux & Greenwood (2016) [20] and described in the next section. Results of this analysis are reported and interpreted in S4 in S1 File.

Next, the magnitude of the inversion and negation effects was calculated in each subject and for each Orientation filter by a subtraction method:

Inversioneffect=NaturalnormdInversionnormd
Negationeffect=NaturalnormdNegatednormd

As can be seen on Fig 4, the magnitude of inversion and negation effects as a function of the Orientation filter described a bell-shaped curve peaking around 90° (horizontal range) and decreasing towards 0° and 180° (vertical range). We used a Bayesian non-linear mixed model to fit a Gaussian function to our data. This enabled the comparison of the effects of inversion and negation along the four parameters of the Gaussian tuning function: (i) Peak Location, in degrees: orientation at which the estimated effect is highest; (ii) Standard Deviation: sharpness of the effect’s tuning around the peak orientation; (iii) Peak Amplitude: difference between the effect at its highest (i.e., at the estimated peak orientation) and at its lowest; and (iv) Base Amplitude: minimum size of the effect (i.e., at the orientation where effect estimate is the lowest). The mixed dimension of the model allowed to consider the inter-subject variability during the parameters estimation by setting it as a random effect. The inversion and negation effects were fitted with the following Gaussian equation [20]:

deffectBaseAmplitude+PeakAmplitude*e(OrientationfilterPeakLocation)22StandardDeviation2

Fig 4.

Fig 4

a. Population-level averaged inversion and negation effects and Bayesian non-linear mixed model predictions as a function of Orientation filter. On the y axis, inversion and negation effects are calculated at the individual level by subtracting performance (normd’) in the Inverted and Negated condition from performance in the Natural stimulus type condition. The dots are population-level mean effects of inversion (blue) and negation (yellow), with the 95% confidence interval in the error bars. The model mean posterior predictions for each Stimulus type are drawn in black (line) and the 89% credible intervals (CrI) of predictions are represented as the faded ribbons. Both Gaussians peaked in the horizontal range and decreased towards the vertical orientation. This shows that inversion and negation effects on identity recognition both culminate in the horizontal range. b. 89% CrI of the four Gaussian parameters posterior predictions distribution. A difference is significant when the CrI of the parameters of the inversion effect (blue) and negation effect (yellow) do not overlap. The CrI of Peak Location, Standard Deviation and Peak Amplitude of inversion and negation effect overlapped, indicating highly similar model estimates for both effects. However, there was a significant difference in the estimates of the base amplitude for inversion and negation effects, with a higher CrI for the inversion effect, indicating that the effect of inversion was overall larger that the effect of negation.

We estimated the four Gaussian parameters (see Fig 4B) with the following linear regression models:

Peak Location ~ 1 + Stimulus type + (1 + Stimulus type | Subject)

Standard Deviation ~ 1 + Stimulus type + (1 + Stimulus type | Subject)

Base Amplitude ~ 1 + Stimulus type + (1 + Stimulus type | Subject)

Peak Amplitude ~ 1 + Stimulus type + (1 + Stimulus type | Subject)

For each of these regression models, prior distributions were defined for intercept (i.e. 1), slope (i.e. the effect of Stimulus type), and random effect of Subjects. Priors for the intercepts of the linear regression models were normal distributions, whose means were chosen based on theoretical expectations. For example, the prior for Peak Location intercept was a normal distribution centered on 90°, as we expected the effects of inversion and negation to peak for horizontally oriented filters. Priors for slopes of the linear regression models were also normal distributions, whose means were 0 (i.e., the most probable a priori was that there would be no difference between inversion and negation effects). Precise values of the standard deviations of prior normal distributions were chosen in a compromise between allowing MCMC chains to converge and keeping priors as broad, thus uninformative, as possible, to not influence the model toward a specific estimation. Finally, priors for random effects were exponential distributions (i.e. positive values only, to specify that there cannot be less than zero variance between subjects; see Table 1 for details about the prior distributions used in the model).

Table 1. Priors of Bayesian non-linear Gaussian models.

Parameter of non-linear Gaussian model Peak Location Standard Deviation Peak Amplitude Base Amplitude
Prior for Intercept normal(90, 20) normal(35, 20) normal(1.5, 1) normal(1.5, 1)
Prior for slope (effect of Stimulus Type) normal(0, 20) normal(0, 30) normal(0, 1.5) normal(0, 1.5)
Prior for random effects exponential(0.1) exponential(0.1) exponential(0.1) exponential(0.05)

Normal prior distributions were set on parameters of linear regressions (intercept and slope) and exponential distributions were set on parameters estimating random effects.

The model was fitted with four MCMC with 6,000 iterations each (2,000 warm up, resulting in 16,000 iterations after warmup). The fit of the model was evaluated with convergence of MCMC chains thanks to diagnostic values such as the number of divergent chains post warmup, Rhat statistics, effective sample size and number of divergent chains. The estimated parameters were compared between inversion and negation effects using their 89% credible intervals (CrI) to test for a “significant” difference.

Since the parameter estimates of the effects of inversion and negation were strikingly similar, we explored whether the two effects were functionally linked at the individual level. We ran Pearson correlations between the individual parameters estimates of the inversion and the negation effects for each of the four Gaussian parameters. These results would give us insight on whether inversion and negation disrupt face identification in a similar way.

In Bayesian mixed-effects analysis, “population-level” includes all participants (fixed effects), and “group-level” is used to describe the sub-categories of the multilevel analysis (random effects), here the group-level represents each participant.

The analysis was done using R 4.2.2. We used the brms package [63, 64] to fit all Bayesian models. All data and analysis codes can be found at https://www.openicpsr.org/openicpsr/project/208541/version/V1/view.

Results

At a descriptive level, normalized d’ was higher across all orientations for natural than negated faces, and higher for negated than inverted faces (see Fig 3A and S1 Fig in S1 File).

Bayesian GLMM comparison of Stimulus type showed significant differences in recognition performance between Stimulus type conditions: the 89% CrI of GLMM predictions of the mean performance of the three conditions do not overlap (Inverted CrI = [-.958; -.516], Negated CrI = [-.4; .084], Natural CrI = [.657; 1.148]; see Fig 3B).

Population-level inversion and negation effects as a function of Orientation filter and their Gaussian model are shown in Fig 4A. Inversion and negation effects both showed a Gaussian tuning profile centered on the horizontal range (see Table 2 and Fig 4). Indeed, while inversion and negation disrupted identity recognition across all orientations, both effects peaked in the horizontal range (mean and CrI Peak Location in degrees; inversion: m = 91.52, CrI = [83.40; 98.56]; negation: m = 85.99, CrI = [78.01; 93.31]). The amplitude of the horizontal peak was comparable across inversion and negation effects (mean and CrI Peak Amplitude in d’effect; inversion: m = 1.11, CrI = [.78; 1.41]; negation: m = 1.18, CrI = [.83; 1.49]). The Standard Deviation of the Gaussian tuning functions was also similar across effects (mean and CrI in degrees; inversion: m = 34.06, CrI = [24.66; 43.92]; negation: m = 39.39, CrI = [24.34; 53.77]). The only difference was found at the level of the Base Amplitude of the Gaussian functions relating inversion and negation effects to Orientation filter (mean and CrI in d’effect; inversion: m = 1.20 CrI = [.99; 1.42], negation m = .53, CrI = [.25; .80]). The fact that the base amplitude was larger for inversion than negation (and that this was the only difference between the two effects) shows that inversion impaired recognition performance overall more than negation.

Table 2. Summary of population-level parameter estimates of the Gaussian mixed model.

Inversion Effect Negation Effect
Parameter Median Mean CrI lower CrI upper Median Mean CrI lower CrI upper
Peak Location 91.53 91.52 83.40 98.56 85.92 85.99 78.01 93.31
Standard Deviation 33.99 34.06 24.66 43.92 38.54 39.39 24.34 53.77
Base Amplitude 1.20 1.20 .99 1.42 .54 .53 .25 .80
Peak Amplitude 1.10 1.11 .78 1.41 1.18 1.18 .83 1.49

The 89% CrI of the Base Amplitude parameter on inversion and negation effects do not overlap indicating a significant difference between the two conditions.

Comparison of population-level model parameter estimates between inversion and negation effects highlights their similar tuning to face orientation content (Fig 5). To examine whether the similarity in the Gaussian parameters reflects a functional relationship between inversion and negation effects at the individual subject level, we ran correlation analyses on the group (i.e., participant)-level Gaussian parameters’ posterior predictions. Correlations between inversion and negation effects were very high for all four Gaussian model parameters’ (all r ≥.72, all p < .001). Stronger correlations were found for Peak location (r = .99, p < .0001, CI = [.99; 1]) and Peak amplitude (r = .99, p < .0001, CI = [.97; .99]) parameter estimates than for Base amplitude (r = .72, p < .001, CI = [.42; .88]) and Standard deviation (r = .78, p < .0001, CI = [.53; .91]; Fig 5) parameter estimates. Supplementary analyses (S3 and S7 in S1 File) indicate that the strong functional dependence of the Gaussian tuning parameters of inversion and negation effects described here is due neither to the use of the upright-positive condition as a common comparison condition (i.e., the correlations were similar when computed while controlling for the variance due to the upright-positive condition; see S3 in S1 File), nor to the use of a smooth Gaussian model to fit the data (i.e., results replicate with a Linear Mixed Model; see S7 in S1 File).

Fig 5. Scatter plot of the four group(i.e., participant)-level Gaussian-model parameter estimates of inversion and negation effects with standard deviation for each group-level parameter estimates as error bars; in blue for inversion and in yellow for negation.

Fig 5

The four Gaussian parameters group-level estimates are highly correlated between inversion and negation effects, the Pearson correlation coefficients are indicated for each parameter. We find a strong correlation between the Gaussian parameters of the inversion and negation orientation tuning profiles, yet several differences can be observed graphically. While peak location and amplitude are very similar across inversion and negation effects (close to diagonal), the base amplitude individual estimate was largest for the inversion effect in all participants, indicating that the decrease in overall recognition performance was consistently more drastic for inversion than negation. Most group-level estimates of Standard deviation were located above the diagonal, in favor of the negation effect section, indicating a narrower horizontal tuning for inversion than negation. Pearson correlation coefficients are written for each of the four parameters with *** for p < .0001, and ** for p < .001. Using partial correlations controlling for the Natural condition we found no evidence that these high correlations were driven by the use of the Natural condition to calculate the effects (see S3 in S1 File).

Discussion

Past evidence showed that the horizontally-oriented cues of the face stimulus enable better identity recognition accuracy than facial cues in other orientation ranges. Yet, the nature of the information contained in the horizontal range remained elusive. Face identity recognition is thought to rely on two main and separable sources of information: shape and surface [50, 51, 65]. The drastic recognition impairments caused by inversion and negation have been attributed to the disrupted processing of shape and surface cues, respectively [44, 53, 55, 66, 67]. Here, we exploited the relative functional selectivity of these effects to characterize the shape versus surface nature of the cues carried by the horizontal range of the face image. Participants identified face images filtered to preserve a narrow orientation band selectively and presented in natural (i.e., upright and in natural contrast polarity), inverted or negated conditions. We estimated the orientation tuning profile of the impairments caused by inversion and negation on face identification performance. As expected, inversion and negation both drastically impaired face identification performance compared to when faces were shown in natural viewing conditions (i.e., upright and in normal positive contrast). Overall, the impairment caused by inversion was stronger than negation. However, the orientation profile of the two effects was strikingly similar: both effects peaked in the horizontal range with similar strength and sharpness. Both effects decreased as the orientation neared vertical. The similar peak location and peak amplitude (which is the difference of effect between its highest, i.e., in the horizontal range, and its lowest, i.e., in the vertical range) indicate that inversion and negation both mostly disrupted access to the horizontally-oriented content of the face stimulus. As inversion and negation prevent the face-specialized processing of shape and surface cues, respectively, these results indicate that the horizontal preference of human face recognition is due to this range yielding a privileged access both to face shape and surface, i.e., two main sources of face identity information.

In line with past evidence, identity recognition in natural viewing condition was tuned to the horizontal range (S4 in S1 File; [16, 17, 19, 20], although compared to the reported results, the horizontal tuning we measured here was broader. Goffaux & Greenwood (2016) [20] found that the bandwidth of the horizontal tuning was influenced by the availability of external facial features such as facial outline cues to perform face recognition. Namely, the horizontal tuning was significantly broader when face outline was available. Our stimulus set contained much larger and more natural outline variations than in this previous study, which could explain the relative broadening of the horizontal tuning. Using familiar faces, as in the present study, Dakin and Watt (2009) [16] also seemed to have found a relatively broadly tuned recognition performance (as they do not report it, this assumption is based on their figure). Finally, to ensure participants performed above chance level in negated and inverted trials, we set the orientation filters to have a broader standard deviation than used in other studies (25° compared to e.g. 15° in Goffaux & Greenwood (2016) [20]). This orientation filter breadth resulted in some overlap in the oriented energy across adjacent orientation ranges. This may to some extent explain the shallower horizontal tuning in the Natural condition and why recognition performance was close to ceiling in the horizontal and adjacent oblique orientations, (See S4 in S1 File).

To investigate their orientation tuning profile, we submitted the effects of inversion and contrast negation on face identification to a Bayesian non-linear Gaussian mixed model. This analysis showed that both effects peaked in the horizontal range and followed a strikingly similar orientation tuning profile. Our finding of the disruption of horizontal tuning of face identification with inversion effect is in line with previous studies comparing the orientation tuning of upright and inverted faces [1921]. The disruption of the horizontal tuning with negation we measured however constitutes new knowledge regarding the orientation tuning of face identification.

The only difference between the inversion and negation effects was at the level of the base amplitude of the orientation tuning profile. Combined with the fact that the peak amplitude was as large for inversion as it was for negation, we conclude that inversion impaired recognition performance more strongly than negation overall (i.e., across orientations, see Fig 4a). Studies about contrast negation showed that the presence of pigmentation cues (e.g., skin texture) in the face stimulus increases the strength of the negation effect [52, 53]. Orientation filtering may have reduced the availability of pigmentation cues, and therefore reduced the impact of negation overall. The weaker effect of negation compared to inversion may also indicate that inversion disrupts face perception at an earlier processing stage than negation [45, 47]. Indeed, compared to negation which flips the polarity while preserving the shape and retinotopic location of the contrast variations making up the face stimulus, inversion drastically changes the latter, and may impact visual processing as soon as V1 (see [18] for supportive evidence); such earlier impact may result in an overall stronger impact on behavioral performance of inversion compared to negation.

Through the demonstration of the additivity and functional selectivity of the behavioral and neural effects, past research showed that inversion and negation affect face processing differently [4448]. This view is suggested by the fact that we have found no evidence for an association between the two effects with full-spectrum faces in our data (see S5 in S1 File). Altogether, our findings indicate that despite inversion and negation showing some functional independence in full-spectrum viewing conditions, they affect access to face-oriented content in a similar way. Previous work has demonstrated that the specialized holistic processing of the face stimulus is affected by inversion and, to a lesser extent, by negation [5, 34, 45, 49, 50, 57], the horizontal advantage over other orientations may therefore arise from the essential nature of the horizontal shape and surface cues to allow holistic face processing.

Another plausible account to the similar orientation tuning properties observed across inversion and negation effects is that the horizontal range of face information carries a more primitive source of information yielding common entry point to the functionally separate shape and surface cue processing. The horizontal structure of a human face has indeed been proposed to convey the natural statistical properties of the face stimulus. These natural statistics are described as an alternance of light and dark horizontal bands along the vertical axis of the face image, from forehead to chin (see Fig 1 of Liu-Shuang et al., 2022 [41], the appearance of which is drastically distorted by inversion and negation [16, 19, 20]. As a matter of fact, the horizontal range of face information was shown to drive the human behavioral and neural responses to faces at more generic levels of categorization than the fine-grained level of identity recognition tested here. A preference for the horizontal was indeed found in infants and children in face categorization [24, 68]. Emotion categorization has also been shown to rely on the horizontal range of face information [26, 27, 69, 70]. At the neural level, Jacques et al., (2014) [22] showed that the face horizontal tuning appears as early as the activation of face specialized-processing (in the N170 time window), in other words as early as face detection. Furthermore, Goffaux et al., (2016) [18] showed a preferential activation of face-preferring brain areas such as the Fusiform Face Area and the Occipital Face area to upright faces containing horizonal information. In both cases the neural preference for the horizontal was disrupted with inversion.

A recent unpublished study of our lab suggests that the horizontal range not only conveys the statistical regularities of the human face category, but that it also carries stable identity cues that are optimal for viewpoint-tolerant recognition of identity (see also [19]. Supporting the crucial contribution of horizontal information to fine-grained identity recognition, horizontal tuning is positively correlated to the face identification performance [21, 28, 29] and the horizontal tuning is stronger for the recognition of familiar than unfamiliar faces [30]. Past and present evidence therefore indicates that the horizontal advantage may actively operate at various processing stages leading from early detection to fine-grained identification. Future studies need to investigate the orientation profile of human face processing at different processing stages to uncover the potential shifts in the orientation dependence of the neural representation of faces over the course of their processing by the human brain (see [71] for such an approach).

To conclude, we have shown that the disruption of human face identity recognition by inversion and negation follows the orientation preference of face identity recognition: it is horizontally tuned. Although inversion and negation presumably disrupt distinct sources of face information, their orientation tuning profiles are highly similar and correlated. These results suggest that the horizontal tuning of face identification is due to this range conveying the two main sources of information driving the specialized processing of face identity: shape and surface.

Supporting information

S1 File. File containing further analysis and explanation.

All figures and analysis cited in the manuscript are in this file.

(DOCX)

pone.0311225.s001.docx (1.3MB, docx)

Acknowledgments

We thank Eunice Tshibambe for her help with data acquisition, and M. Umut Canoluk for his help with statistical analysis. This research has benefitted from the statistical consult with Statistical Methodology and Computing Service, technological platform at UCLouvain–SMCS/LIDAM, UCLouvain, via the help of Vincent Bremhorst.

Data Availability

All data and analysis codes are help by openicpsr, in the project named Inversion_Negation_HorizontalTuning. They can be found at https://www.openicpsr.org/openicpsr/project/208541/version/V1/view.

Funding Statement

This work was supported by the FNRS (Fond National de la Recherche Scientifique https://www.frs-fnrs.be/en/) ASP grants n°: 1.A.681.22F granted to HD and the Excellence of Science grant HUMVISCAT- 30991544. VG is an F.R.S.-F.N.R.S. Research Associate. ARS was supported by a FSR Postdoc Fellowship grant of UCLouvain (Université Catholique de Louvain https://uclouvain.be/en/index.html). The funders did not play any role in the study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1.Jack RE, Schyns PG. Toward a Social Psychophysics of Face Communication. Annu Rev Psychol. 2017;68(1):269‑97. [DOI] [PubMed] [Google Scholar]
  • 2.Tibbetts EA, Dale J. Individual recognition: it is good to be different. Trends Ecol Evol. 1 oct 2007;22(10):529‑37. doi: 10.1016/j.tree.2007.09.001 [DOI] [PubMed] [Google Scholar]
  • 3.Sheehan MJ, Nachman MW. Morphological and population genomic evidence that human faces have evolved to signal individual identity. Nat Commun. 16 sept 2014;5(1):4800. doi: 10.1038/ncomms5800 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Bruce V, Young A. Understanding face recognition. Br J Psychol. 1986;77(3):305‑27. doi: 10.1111/j.2044-8295.1986.tb02199.x [DOI] [PubMed] [Google Scholar]
  • 5.Farah MJ. Is face recognition ‘special’? Evidence from neuropsychology. Behav Brain Res. 1 avr 1996;76(1):181‑9. doi: 10.1016/0166-4328(95)00198-0 [DOI] [PubMed] [Google Scholar]
  • 6.Kanwisher N, Yovel G. The fusiform face area: a cortical region specialized for the perception of faces. Philos Trans R Soc B Biol Sci. 29 déc 2006;361(1476):2109‑28. doi: 10.1098/rstb.2006.1934 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Sergent J, louis Signoret J, Bruce V, Rolls ET, Bruce V, Cowey A et al. Functional and anatomical decomposition of face processing: evidence from prosopagnosia and PET study of normal subjects. Philos Trans R Soc Lond B Biol Sci. 1992;335(1273):55‑62. doi: 10.1098/rstb.1992.0007 [DOI] [PubMed] [Google Scholar]
  • 8.Mike Burton A. Why has research in face recognition progressed so slowly? The importance of variability. Q J Exp Psychol. 1 août 2013;66(8):1467‑85. [DOI] [PubMed] [Google Scholar]
  • 9.Sinha P, Balas B, Ostrovsky Y, Russell R. Face Recognition by Humans: Nineteen Results All Computer Vision Researchers Should Know About. Proc IEEE. nov 2006;94(11):1948‑62. [Google Scholar]
  • 10.Adini Y, Moses Y, Ullman S. Face recognition: the problem of compensating for changes in illumination direction. IEEE Trans Pattern Anal Mach Intell. juill 1997;19(7):721‑32. [Google Scholar]
  • 11.Jenkins R, White D, Van Montfort X, Mike Burton A. Variability in photos of the same face. Cognition. 1 déc 2011;121(3):313‑23. doi: 10.1016/j.cognition.2011.08.001 [DOI] [PubMed] [Google Scholar]
  • 12.DiCarlo JJ, Cox DD. Untangling invariant object recognition. Trends Cogn Sci. août 2007;11(8):333‑41. doi: 10.1016/j.tics.2007.06.010 [DOI] [PubMed] [Google Scholar]
  • 13.Hole GJ, George PA, Eaves K, Rasek A. Effects of Geometric Distortions on Face-Recognition Performance. Perception. 1 oct 2002;31(10):1221‑40. doi: 10.1068/p3252 [DOI] [PubMed] [Google Scholar]
  • 14.Jenkins R, Burton AM. Stable face representations. Philos Trans R Soc B Biol Sci. 12 juin 2011;366(1571):1671‑83. doi: 10.1098/rstb.2010.0379 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Johnston RA, Edmonds AJ. Familiar and unfamiliar face recognition: a review. Mem Hove Engl. juill 2009;17(5):577‑96. doi: 10.1080/09658210902976969 [DOI] [PubMed] [Google Scholar]
  • 16.Dakin SC, Watt RJ. Biological « bar codes » in human faces. J Vis. 3 avr 2009;9(4):2.1–10. [DOI] [PubMed] [Google Scholar]
  • 17.Goffaux V, Poncin A, Schiltz C. Selectivity of Face Perception to Horizontal Information over Lifespan (from 6 to 74 Year Old). PLOS ONE. 23 sept 2015;10(9):e0138812. doi: 10.1371/journal.pone.0138812 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Goffaux V, Duecker F, Hausfeld L, Schiltz C, Goebel R. Horizontal tuning for faces originates in high-level Fusiform Face Area. Neuropsychologia. 29 janv 2016;81:1‑11. doi: 10.1016/j.neuropsychologia.2015.12.004 [DOI] [PubMed] [Google Scholar]
  • 19.Goffaux V, Dakin SC. Horizontal information drives the behavioral signatures of face processing. Front Psychol. 2010;1:143. doi: 10.3389/fpsyg.2010.00143 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Goffaux V, Greenwood JA. The orientation selectivity of face identification. Sci Rep. 28 sept 2016;6:34204. doi: 10.1038/srep34204 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Pachai M, Sekuler A, Bennett P. Sensitivity to Information Conveyed by Horizontal Contours is Correlated with Face Identification Accuracy. Front Psychol [Internet]. 2013. [cité 21 juin 2023];4. Disponible sur: https://www.frontiersin.org/articles/10.3389/fpsyg.2013.00074 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Jacques C, Schiltz C, Goffaux V. Face perception is tuned to horizontal orientation in the N170 time window. J Vis. 7 févr 2014;14(2):5. doi: 10.1167/14.2.5 [DOI] [PubMed] [Google Scholar]
  • 23.de Heering A, Maurer D. The effect of spatial frequency on perceptual learning of inverted faces. Vision Res. 28 juin 2013;86:107‑14. doi: 10.1016/j.visres.2013.04.014 [DOI] [PubMed] [Google Scholar]
  • 24.Balas BJ, Schmidt J, Saville A. A face detection bias for horizontal orientations develops in middle childhood. Front Psychol [Internet]. 2015. [cité 25 juin 2023];6. Disponible sur: https://www.frontiersin.org/articles/10.3389/fpsyg.2015.00772 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Obermeyer S, Kolling T, Schaich A, Knopf M. Differences between Old and Young Adults’ Ability to Recognize Human Faces Underlie Processing of Horizontal Information. Front Aging Neurosci [Internet]. 2012. [cité 25 juin 2023];4. Disponible sur: https://www.frontiersin.org/articles/10.3389/fnagi.2012.00003 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Huynh CM, Balas B. Emotion recognition (sometimes) depends on horizontal orientations. Atten Percept Psychophys. 1 juill 2014;76(5):1381‑92. doi: 10.3758/s13414-014-0669-4 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Yu D, Chai A, Chung STL. Orientation information in encoding facial expressions. Vision Res. 1 sept 2018;150:29‑37. doi: 10.1016/j.visres.2018.07.001 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Duncan J, Royer J, Dugas G, Blais C, Fiset D. Revisiting the link between horizontal tuning and face processing ability with independent measures. J Exp Psychol Hum Percept Perform. nov 2019;45(11):1429‑35. doi: 10.1037/xhp0000684 [DOI] [PubMed] [Google Scholar]
  • 29.Little Z, Susilo T. Preference for horizontal information in faces predicts typical variations in face recognition but is not impaired in developmental prosopagnosia. Psychon Bull Rev. 1 févr 2023;30(1):261‑8. doi: 10.3758/s13423-022-02163-4 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.Pachai MV, Sekuler AB, Bennett PJ, Schyns PG, Ramon M. Personal familiarity enhances sensitivity to horizontal structure during processing of face identity. J Vis. 7 juin 2017;17(6):5. doi: 10.1167/17.6.5 [DOI] [PubMed] [Google Scholar]
  • 31.Albonico A, Furubacke A, Barton JJS, Oruc I. Perceptual efficiency and the inversion effect for faces, words and houses. Vision Res. 1 déc 2018;153:91‑7. doi: 10.1016/j.visres.2018.10.008 [DOI] [PubMed] [Google Scholar]
  • 32.Epstein RA, Higgins JS, Parker W, Aguirre GK, Cooperman S. Cortical correlates of face and scene inversion: A comparison. Neuropsychologia. 1 janv 2006;44(7):1145‑58. doi: 10.1016/j.neuropsychologia.2005.10.009 [DOI] [PubMed] [Google Scholar]
  • 33.Robbins R, McKone E. No face-like processing for objects-of-expertise in three behavioural tasks. Cognition. 1 avr 2007;103(1):34‑79. doi: 10.1016/j.cognition.2006.02.008 [DOI] [PubMed] [Google Scholar]
  • 34.Rossion B. Picture-plane inversion leads to qualitative changes of face perception. Acta Psychol (Amst). 1 juin 2008;128(2):274‑89. doi: 10.1016/j.actpsy.2008.02.003 [DOI] [PubMed] [Google Scholar]
  • 35.Valentine T. Upside-down faces: A review of the effect of inversion upon face recognition. Br J Psychol. 1988;79(4):471‑91. doi: 10.1111/j.2044-8295.1988.tb02747.x [DOI] [PubMed] [Google Scholar]
  • 36.Yin RK. Looking at upside-down faces. J Exp Psychol. 1969;81(1):141‑5. [Google Scholar]
  • 37.Galper RE. Recognition of faces in photographic negative. Psychon Sci. 1 oct 1970;19(4):207‑8. [Google Scholar]
  • 38.Nederhouser M, Yue X, Mangini MC, Biederman I. The deleterious effect of contrast reversal on recognition is unique to faces, not objects. Vision Res. 1 juill 2007;47(16):2134‑42. doi: 10.1016/j.visres.2007.04.007 [DOI] [PubMed] [Google Scholar]
  • 39.Hong Liu C, Collin CA, Burton AM, Chaudhuri A. Lighting direction affects recognition of untextured faces in photographic positive and negative. Vision Res. 1 déc 1999;39(24):4003‑9. doi: 10.1016/s0042-6989(99)00109-1 [DOI] [PubMed] [Google Scholar]
  • 40.Gilad S, Meng M, Sinha P. Role of ordinal contrast relationships in face encoding. Proc Natl Acad Sci U S A. 31 mars 2009;106(13):5353‑8. doi: 10.1073/pnas.0812396106 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41.Liu-Shuang J, Yang YF, Rossion B, Goffaux V. Natural Contrast Statistics Facilitate Human Face Categorization. eNeuro [Internet]. 1 sept 2022. [cité 2 août 2023];9(5). Disponible sur: https://www.eneuro.org/content/9/5/ENEURO.0420-21.2022 doi: 10.1523/ENEURO.0420-21.2022 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42.Ohayon S, Freiwald WA, Tsao DY. What makes a cell face selective? The importance of contrast. Neuron. 10 mai 2012;74(3):567‑81. doi: 10.1016/j.neuron.2012.03.024 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43.Goffaux V, Rossion B. Face inversion disproportionately impairs the perception of vertical but not horizontal relations between features. J Exp Psychol Hum Percept Perform. 2007;33(4):995‑1002. doi: 10.1037/0096-1523.33.4.995 [DOI] [PubMed] [Google Scholar]
  • 44.Bruce V, Langton S. The Use of Pigmentation and Shading Information in Recognising the Sex and Identities of Faces. Perception. 1 juill 1994;23(7):803‑22. doi: 10.1068/p230803 [DOI] [PubMed] [Google Scholar]
  • 45.Hole GJ, George PA, Dunsmore V. Evidence for Holistic Processing of Faces Viewed as Photographic Negatives. Perception. 1 mars 1999;28(3):341‑59. doi: 10.1068/p2622 [DOI] [PubMed] [Google Scholar]
  • 46.Kemp R, McManus C, Pigott T. Sensitivity to the Displacement of Facial Features in Negative and Inverted Images. Perception. 1 août 1990;19(4):531‑43. [DOI] [PubMed] [Google Scholar]
  • 47.Itier RJ, Taylor MJ. Inversion and Contrast Polarity Reversal Affect both Encoding and Recognition Processes of Unfamiliar Faces: A Repetition Study Using ERPs. NeuroImage. 1 févr 2002;15(2):353‑72. doi: 10.1006/nimg.2001.0982 [DOI] [PubMed] [Google Scholar]
  • 48.Liu-Shuang J, Ales J, Rossion B, Norcia AM. Separable effects of inversion and contrast-reversal on face detection thresholds and response functions: A sweep VEP study. J Vis. 10 févr 2015;15(2):11. doi: 10.1167/15.2.11 [DOI] [PubMed] [Google Scholar]
  • 49.Rossion B. Distinguishing the cause and consequence of face inversion: The perceptual field hypothesis. Acta Psychol (Amst). 1 nov 2009;132(3):300‑12. doi: 10.1016/j.actpsy.2009.08.002 [DOI] [PubMed] [Google Scholar]
  • 50.Jiang F, Blanz V, Rossion B. Holistic processing of shape cues in face identification: Evidence from face inversion, composite faces, and acquired prosopagnosia. Vis Cogn [Internet]. 1 sept 2011. [cité 4 août 2024]; Disponible sur: https://www.tandfonline.com/doi/full/10.1080/13506285.2011.604360 [Google Scholar]
  • 51.Bozana Meinhardt-Injac Malte Persike, Meinhardt Günter. Holistic Face Processing is Induced by Shape and Texture [Internet]. 2013. [cité 31 juill 2024]. Disponible sur: https://journals.sagepub.com/doi/abs/10.1068/p7462 [DOI] [PubMed] [Google Scholar]
  • 52.Russell R, Sinha P, Biederman I, Nederhouser M. Is pigmentation important for face recognition? Evidence from contrast negation. Perception. 2006;35(6):749‑59. doi: 10.1068/p5490 [DOI] [PubMed] [Google Scholar]
  • 53.Vuong QC, Peissig JJ, Harrison MC, Tarr MJ. The role of surface pigmentation for recognition revealed by contrast reversal in faces and Greebles. Vision Res. 1 mai 2005;45(10):1213‑23. doi: 10.1016/j.visres.2004.11.015 [DOI] [PubMed] [Google Scholar]
  • 54.Kemp R, Pike G, White P, Musselman A. Perception and Recognition of Normal and Negative Faces: The Role of Shape from Shading and Pigmentation Cues. Perception. 1 janv 1996;25(1):37‑52. doi: 10.1068/p250037 [DOI] [PubMed] [Google Scholar]
  • 55.Liu CH, Collin CA, Chaudhuri A. Does Face Recognition Rely on Encoding of 3-D Surface? Examining the Role of Shape-from-Shading and Shape-from-Stereo. Perception. 1 juin 2000;29(6):729‑43. doi: 10.1068/p3065 [DOI] [PubMed] [Google Scholar]
  • 56.Taubert J, Alais D. Identity Aftereffects, but Not Composite Effects, are Contingent on Contrast Polarity. Perception. 1 avr 2011;40(4):422‑36. doi: 10.1068/p6874 [DOI] [PubMed] [Google Scholar]
  • 57.Sormaz M, Andrews TJ, Young AW. Contrast negation and the importance of the eye region for holistic representations of facial identity. J Exp Psychol Hum Percept Perform. 2013;39(6):1667‑77. doi: 10.1037/a0032449 [DOI] [PubMed] [Google Scholar]
  • 58.Bach M. The Freiburg Visual Acuity test—automatic measurement of visual acuity. Optom Vis Sci Off Publ Am Acad Optom. janv 1996;73(1):49‑53. doi: 10.1097/00006324-199601000-00008 [DOI] [PubMed] [Google Scholar]
  • 59.Rossion B, Michel C. Normative accuracy and response time data for the computerized Benton Facial Recognition Test (BFRT-c). Behav Res Methods. déc 2018;50(6):2442‑60. doi: 10.3758/s13428-018-1023-x [DOI] [PubMed] [Google Scholar]
  • 60.Ng HW, Winkler S. A data-driven approach to cleaning large face datasets. In: 2014 IEEE International Conference on Image Processing (ICIP). 2014. p. 343‑7. [Google Scholar]
  • 61.McKone E. Holistic processing for faces operates over a wide range of sizes but is strongest at identification rather than conversational distances. Vision Res. 1 janv 2009;49(2):268‑83. doi: 10.1016/j.visres.2008.10.020 [DOI] [PubMed] [Google Scholar]
  • 62.Hautus MJ. Corrections for extreme proportions and their biasing effects on estimated values of d′. Behav Res Methods Instrum Comput. 1995;27:46‑51. [Google Scholar]
  • 63.Bürkner PC. brms: An R Package for Bayesian Multilevel Models Using Stan. J Stat Softw. 29 août 2017;80:1‑28. [Google Scholar]
  • 64.Bürkner PC. Advanced Bayesian Multilevel Modeling with the R Package brms. R J. 1 juill 2018;10:395‑411. [Google Scholar]
  • 65.Jiang F, Dricot L, Blanz V, Goebel R, Rossion B. Neural correlates of shape and surface reflectance information in individual faces. Neuroscience. 10 nov 2009;163(4):1078‑91. doi: 10.1016/j.neuroscience.2009.07.062 [DOI] [PubMed] [Google Scholar]
  • 66.Freire A, Lee K, Symons LA. The Face-Inversion Effect as a Deficit in the Encoding of Configural Information: Direct Evidence. Perception. 1 févr 2000;29(2):159‑70. doi: 10.1068/p3012 [DOI] [PubMed] [Google Scholar]
  • 67.Maurer D, Grand RL, Mondloch CJ. The many faces of configural processing. Trends Cogn Sci. 1 juin 2002;6(6):255‑60. doi: 10.1016/s1364-6613(02)01903-4 [DOI] [PubMed] [Google Scholar]
  • 68.de Heering A, Goffaux V, Dollion N, Godard O, Durand K, Baudouin JY. Three-month-old infants’ sensitivity to horizontal information within faces. Dev Psychobiol. 2016;58(4):536‑42. doi: 10.1002/dev.21396 [DOI] [PubMed] [Google Scholar]
  • 69.Balas B, Huynh C, Saville A, Schmidt J. Orientation biases for facial emotion recognition during childhood and adulthood. J Exp Child Psychol. 1 déc 2015;140:171‑83. doi: 10.1016/j.jecp.2015.07.006 [DOI] [PubMed] [Google Scholar]
  • 70.Duncan J, Gosselin F, Cobarro C, Dugas G, Blais C, Fiset D. Orientations for the successful categorization of facial expressions and their link with facial features. J Vis. 11 déc 2017;17(14):7. doi: 10.1167/17.14.7 [DOI] [PubMed] [Google Scholar]
  • 71.Jacobs C, Petras K, Moors P, Goffaux V. Contrast versus identity encoding in the face image follow distinct orientation selectivity profiles. PloS One. 2020;15(3):e0229185. doi: 10.1371/journal.pone.0229185 [DOI] [PMC free article] [PubMed] [Google Scholar]

Decision Letter 0

Patrick Bruns

7 Jun 2024

PONE-D-24-18307

Horizontal face information is the main gateway to the configuration and surface cues to familiar face identity.

PLOS ONE

Dear Dr. Dumont,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Your manuscript has been reviewed by two experts in the topic. As you will see below, both reviewers were positive about your work, but they also made several clear and constructive recommendations for further improving the manuscript which I would encourage you to implement in the revision. One point that came up in both reviews is that the relationships between face inversion/configuration cues and between negation/surface cues might not be as straightforward as it is claimed in the manuscript, and I agree with the reviewers that a more neutral framing of the study (possibly including in the title) might therefore be necessary.

Please submit your revised manuscript by Jul 22 2024 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Patrick Bruns

Academic Editor

PLOS ONE

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at 

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and 

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. Please note that funding information should not appear in any section or other areas of your manuscript. We will only publish funding information present in the Funding Statement section of the online submission form. Please remove any funding-related text from the manuscript.

3. Please include your full ethics statement in the ‘Methods’ section of your manuscript file. In your statement, please include the full name of the IRB or ethics committee who approved or waived your study, as well as whether or not you obtained informed written or verbal consent. If consent was waived for your study, please include this information in your statement as well. 

4. Please include captions for your Supporting Information files at the end of your manuscript, and update any in-text citations to match accordingly. Please see our Supporting Information guidelines for more information: http://journals.plos.org/plosone/s/supporting-information

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Yes

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: Yes

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: No

Reviewer #2: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: This paper examines whether horizontal tuning for faces provides similar access to configural and surface cues. Participants performed a same-different task with familiar faces across various orientation filters on natural, inverted, or negated images. The authors conclude that horizontal tuning is equally important for configural cues and surface cues. This work extends a growing literature on horizontal tuning for faces. It replicates prior inversion data and contribute new negation data. I don’t have major comments but I have some questions and suggestions that I hope can improve the paper.

1. Upright accuracy is near ceiling at 93%. What are the accuracies for inverted and negated? How far from floor are they? While the choice to analyse d-prime makes sense, an additional analysis of accuracy would help readers better appreciate the data.

2. Mike Burton and others have shown that understanding how different images of the same person can vary (telling faces together) is just as important as understanding image differences between people (telling faces apart). I think the authors can contribute to this issue by analysing same vs different trials separately and testing for potential interactions with orientation filter, image manipulation, or both.

3. Neither configural or surface cues were directly manipulated, and the relationships between inversion/configuration and between negation/surface are not straightforward (see Sormaz et al 2013 for example). I suggest using inversion and negation in the title and discussing the configural and surface cues in the intro and discussion.

4. Supplementary analysis 5 is useful but the current interpretation is based on an indirect contrast between a significant correlation (.61) and a non-significant one (.41). These correlations might not differ statistically so the authors need to compare them directly (see http://quantpsy.org/corrtest/corrtest2.htm) and adjust their interpretations if needed.

Reviewer #2: The authors present novel research on the importance of horizontally oriented facial information for familiar face recognition by showing how this information relates to two seminal effects in the face processing literature: face inversion effect and contrast reversal effect. The first occurs when faces are viewed upside-down, and the second occurs when faces are viewed as a film negative. In the literature, face inversion is often understood to alter processing of configuration cues (e.g., how individual features relate to one another), while contrast reversal is often understood to alter surface cues (e.g., how 3D shape is inferred from a 2D image). The authors show that both image manipulations, i.e., face inversion and contrast reversal, entail a substantial and rather selective drop in sensitivity to horizontally oriented features. Thus, the authors argue, horizontal tuning of the visual system might emerge as the system attunes to configuration and surface cues.

I have a mostly favourable opinion of this work, which has generally sound methodology and statistical analyses. However, I do take issue with some interpretations and general framing of the paper (immediately below). I also have a few requests for clarification and recommendations.

The authors begin with a claim that the cause for human reliance on horizontal facial information is not well understood and that the visual system adapting to configuration and surface cues might explain this (e.g., lines 24-25). This framing hinges on 1) face inversion and contrast reversal effects being valid measures of each construct, and 2) on each construct being empirically distinct. However, disruption of configural processing is only one possible explanation—though one often shared—for the face inversion effect, for instance. But there are alternatives. Furthermore, the authors support their independence hypothesis with the results of a correlation between face inversion and contrast reversal effects for full spectrum faces being nonsignificant. While I appreciate that, strictly speaking, a correlation of r = 0.43, p = 0.051, is not statistically significant, this is not exactly evidence of their independence. Furthermore, the fact that orientation profile modeling parameters of the face inversion and contrast reversal effects correlated almost perfectly suggests these processes do share some fundamental aspects—though admittedly this correspondence was likely inflated vs. the true correlation, due to the smooth nature of Gaussian fitting (see also below). Nevertheless, my criticism essentially boils down to this: Configuration and surface cues are theoretically loaded constructs that are assumed to be altered in the face inversion and contrast reversal effects respectively. Prefacing horizontal tuning with these does nothing to enrich the conversation—and potentially even derails it in light of the points I raised above. An alternative and perhaps better framing is that researchers have long been trying to explain face inversion and contrast reversal effects. Though a disruption of configuration and surface cue processing are possible explanations, there are alternatives. With the recent surge in studies suggesting horizontal tuning is a core feature of the face processing architecture, the authors verified whether and how face inversion and contrast reversal disrupt orientation tuning. This more neutral framing benefits from its lack of assumptions regarding face inversion and contrast reversal effects and adds to a growing literature that shows the functional importance of horizontally orientated information for various aspects of face processing.

Writing is generally good, but the authors should watch out for overreliance on the word “the,” which at times makes reading arduous. E.g., in the sentence starting at line 289, there are 8 “the” in a single sentence, several of which are superfluous. “The purpose of this study was to understand the extent to which the horizontal tuning of face identification reflects the contribution of feature configuration or of the surface and shading of the face image through the comparison of inversion and negation effects on the Orientation filter dependence of identity recognition performance.”

Line 182: Were mean luminance and RMS contrast set before or after applying the elliptical aperture? Please make this more explicit in text.

Line 216: Were actor identities selected on a participant basis? Please make this more explicit in text.

Line 261: It might benefit the reader to remind what chance level was here.

Line 305: The authors should explain why they chose to analyze normalized d’, rather than simply d’. Also, they should provide a bit more detail about how this was done, as I found myself a bit confused by their description of the procedure.

Line 333: Base amplitude and peak amplitude should be clearly explained here (perhaps put in relation with the data at hand). Peak amplitude is also poorly explained at line 440.

Line 469: The authors should consider the possibility that these coefficients are somewhat inflated due to the (smooth) Gaussian fit. They could perhaps exploratorily look at the relationship between inverted and reversed d’ as a function of orientation (perhaps as a random effect). They may also want to correlate individual full spectrum d’s with orientation d’s to look at whether one orientation is more likely to account for individual differences in performance vs. others. I am fully aware the study may be underpowered if the correlations are only small/moderate, but it might nevertheless be informative about some trend within the data (perhaps mixed effects modeling would again offset some of that).

Figure 3: plotting face inversion effect and contrast reversal effect as “normalized sensitivity (d’)” is a bit counterintuitive as it might initially suggest that sensitivity is highest with horizontal information. This could be improved by e.g., also specifying on the axis that a positive d’ value = sensitivity loss, or by plotting orientation profiles upside-down (i.e., with negative d’s).

Figures 3 and 4: should their order be inverted? It seems raw results (Fig 4) should be presented before fitted curves (Fig 3).

Full spectrum trials might provide valuable context, e.g., by being plotted alongside orientation d’, or even plotting orientation d’ as a proportion of the full spectrum d’ (or again, to investigate individual differences across orientations, see above point).

I think upright fitted orientation profiles belong in the main text rather than supplementary materials.

Credible interval: value used (89%) should be in text (Data and analyses).

Regarding the broad orientation tuning found (especially for upright faces), the authors report having used a 25deg wide orientation filter. According to my calculations, adjacent orientation samples (e.g., 90deg and 120deg) were not independent with this width parameter, as they intersect at around mu +/- 15deg. Briefly, a horizontal (90deg) filter and oblique-horizontal (120deg) filter both sample 105deg information at approx. 37% of the maximum sampling proportion. The 120deg filter also samples 90deg information at 2% of the maximum sampling proportion (and inversely, the 90deg filter samples 120deg information at 2%). In fact, their respective areas overlap by about 8%. This is not a problem in and of itself, but it should be recognized by the authors, and it might partially account for somewhat wider orientation tuning found.

Supplementary materials, line 77-78 : “A hypothesis about the origin of the asymmetry is that it would result from faces statistics that often appear tilted and not perfectly upright.” Looking at Figure 1, it seems 60deg and 120deg locally enhance eye, eyebrow and mouth contrast more vs. the 90deg filter, at least subjectively. Could this be a factor, seeing as those are very important features for face processing? Also, is it possible this was caused by how luminance and RMS contrast were equated?

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2024 Oct 7;19(10):e0311225. doi: 10.1371/journal.pone.0311225.r002

Author response to Decision Letter 0


16 Aug 2024

Dear Dr Bruns,

We have revised our manuscript entitled “Horizontal face information is the main gateway to the shape and surface cues to familiar face identity” (previous title “Horizontal face information is the main gateway to the configuration and surface cues to familiar face identity”) according to the comments and suggestions of the Reviewers who evaluated our work. The revised manuscript and the detailed response to the Reviewers are enclosed. As we revised the manuscript extensively, keeping track of all the changes to the manuscript automatically results in a very messy document. We have therefore decided to highlight the most significant revisions in the Revised Manuscript with Track Changes.docx document by underlining the relevant sections. We briefly summarize our replies and revisions below (detailed replies can be found in the Response to Reviewers.docx).

Following both reviewers’ advice, we toned down the theoretical framing of our study and, in particular, the assumed link between the inversion effect and face configural processing. The revised version stays closer to what can be asserted with greater confidence about inversion and negation: inversion turns edge-related cues upside-down whereas negation flips the luminance gradient of the face surface and preserves face edges. We revised the Title, Abstract, Introduction and Discussion accordingly.

The Reviewers also pointed us to the fact that our evidence for concluding (in our Supplementary analysis 5) about a

functional independence of the effects of inversion and negation in the full spectrum picture condition was rather weak (i.e., based on a non-significant correlation but with a p-value of .051). We therefore corroborated our initial correlation analysis by adding (1) a Bayesian correlation analysis including a Bayes factor, and (2) the confidence intervals of the frequentist and Bayesian correlation coefficients, both of which included zero. Overall, we now conclude that there is insufficient evidence from our data to conclude a correlation between the two effects and acknowledge that this may be due to a lack of statistical power.

Following Reviewer 1’s comments, we fully re-analyzed our data based on accuracy instead of the sensitivity index d’. This analysis gave highly comparable results than our original d’ analysis. As we agree with the Reviewer that the accuracy data may still be if interest for readers, we added this analysis in the Supplementary Material. Reviewer 1 also suggested to analyze same and different trials independently, which is a very interesting view of our experimental design. We report this analysis in our response to the Reviewer but not in the manuscript, as it is unfortunately too underpowered to be meaningful

Following Reviewer 2’s suggestions, we added some clarifications to our Methods, data analysis, and supplementary sections, as well as to our Figures.

Overall, the comments of the two Reviewers allowed us to make significant improvement to the manuscript by clarifying elements of both the Methods and the theoretical foundation of the study. We thank both Reviewers for their helpful and constructive comments, and we hope that the manuscript is now more suitable for publication in your journal.

Yours sincerely,

Hélène Dumont, And on behalf of Alexia Roux-Sibilon, and Valérie Goffaux

Attachment

Submitted filename: Response to Reviewers.docx

pone.0311225.s002.docx (1.2MB, docx)

Decision Letter 1

Patrick Bruns

17 Sep 2024

Horizontal face information is the main gateway to the shape and surface cues to familiar face identity.

PONE-D-24-18307R1

Dear Dr. Dumont,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice will be generated when your article is formally accepted. Please note, if your institution has a publishing partnership with PLOS and your article meets the relevant criteria, all or part of your publication costs will be covered. Please make sure your user information is up-to-date by logging into Editorial Manager at Editorial Manager® and clicking the ‘Update My Information' link at the top of the page. If you have any questions relating to publication charges, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Patrick Bruns

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

The revised manuscript has been reviewed by original Reviewer #1, and I served as second reviewer because original Reviewer #2 was not available at this time. I agree with Reviewer #1 that all points raised during the review process have been thoroughly addressed.

During production of your article, you might want to integrate the supporting figures into the supporting text document (rather than having only the supporting figure captions in the text document), as this will make the document more accessible. Please also use proper item descriptions in the main text, i.e. refer to items as "Supporting Information S1", "Supporting Figure S2" etc. rather than "Supplementary 1" as in l. 168

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: (No Response)

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

**********

Acceptance letter

Patrick Bruns

27 Sep 2024

PONE-D-24-18307R1

PLOS ONE

Dear Dr. Dumont,

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now being handed over to our production team.

At this stage, our production department will prepare your paper for publication. This includes ensuring the following:

* All references, tables, and figures are properly cited

* All relevant supporting information is included in the manuscript submission,

* There are no issues that prevent the paper from being properly typeset

If revisions are needed, the production department will contact you directly to resolve them. If no revisions are needed, you will receive an email when the publication date has been set. At this time, we do not offer pre-publication proofs to authors during production of the accepted work. Please keep in mind that we are working through a large volume of accepted articles, so please give us a few weeks to review your paper and let you know the next and final steps.

Lastly, if your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

If we can help with anything else, please email us at customercare@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Patrick Bruns

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 File. File containing further analysis and explanation.

    All figures and analysis cited in the manuscript are in this file.

    (DOCX)

    pone.0311225.s001.docx (1.3MB, docx)
    Attachment

    Submitted filename: Response to Reviewers.docx

    pone.0311225.s002.docx (1.2MB, docx)

    Data Availability Statement

    All data and analysis codes are help by openicpsr, in the project named Inversion_Negation_HorizontalTuning. They can be found at https://www.openicpsr.org/openicpsr/project/208541/version/V1/view.


    Articles from PLOS ONE are provided here courtesy of PLOS

    RESOURCES