Skip to main content
PLOS Biology logoLink to PLOS Biology
. 2023 Mar 2;21(3):e3002009. doi: 10.1371/journal.pbio.3002009

Illusory object recognition is either perceptual or cognitive in origin depending on decision confidence

Josipa Alilović 1,2, Eline Lampers 1, Heleen A Slagter 3,4, Simon van Gaal 1,2,*
Editor: Christopher Pack5
PMCID: PMC10013920  PMID: 36862734

Abstract

We occasionally misinterpret ambiguous sensory input or report a stimulus when none is presented. It is unknown whether such errors have a sensory origin and reflect true perceptual illusions, or whether they have a more cognitive origin (e.g., are due to guessing), or both. When participants performed an error-prone and challenging face/house discrimination task, multivariate electroencephalography (EEG) analyses revealed that during decision errors (e.g., mistaking a face for a house), sensory stages of visual information processing initially represent the presented stimulus category. Crucially however, when participants were confident in their erroneous decision, so when the illusion was strongest, this neural representation flipped later in time and reflected the incorrectly reported percept. This flip in neural pattern was absent for decisions that were made with low confidence. This work demonstrates that decision confidence arbitrates between perceptual decision errors, which reflect true illusions of perception, and cognitive decision errors, which do not.


The occurrence of perceptual illusions, the misinterpretation of sensory input, is a common characteristic of several neurological and psychiatric disorders. This study shows that such illusions have both perceptual as well as cognitive neural origins in the healthy human brain.

Introduction

Our perceptual experience can deviate substantially from the actual input that reaches our senses. Misperceiving or misinterpreting sensory input is a common characteristic of several neurological and psychiatric disorders, such as schizophrenia, but also during drug-induced hallucinatory states [1,2]. However, even in healthy people and under common circumstances, illusory perceptual experiences may occur, as illustrated by famous illusions such as the Müller-Lyer and the Kanizsa illusion [3]. Further, context and stimulus frequency can act as top-down priors that bias the interpretation of ambiguous visual information [46]. Yet, even in conditions without strong perceptual priors, perception can spontaneously differ from the veridical sensory input. In particular, when perceptual information is sparse or ambiguous, interpretation of the visual input is challenged, resulting occasionally in incorrect perceptual decisions [7,8]. The experiential aspect of such incorrect decisions is often inferred based on participants’ behavioral report. However, even when one trust such introspective reports, the source of these decision errors remains unclear. Most importantly, with behavioral measures it is hard to disentangle whether decision errors originate from true illusions in perception (misperceptions) or stem from guessing behavior (misreports), although procedures to do so have made rapid progress [914]. Here, we investigated this issue and examined if misreported visual stimuli (e.g., face presented, house reported) were misrepresented at the sensory stage or the decision stage.

To determine the nature of misreports, we dissociate sensory from post-sensory or decision-related processes during human perceptual decision-making using electroencephalography (EEG) in combination with multivariate pattern analyses (MVPA). Forty participants performed a challenging perceptual discrimination task in which faces and houses were briefly presented, preceded, and followed by pattern masks, strongly reducing stimulus visibility. Participants discriminated which of the 2 object categories was presented and indicated how confident they were in this decision (high/low confidence) [15,16]. Discrimination correctness was titrated by individually adjusting stimulus duration to reach approximately 70% correct decisions, thereby inducing a large number of decision errors for our analyses (approximately 30%). Multivariate pattern classifiers were then used to characterize the time course of category-specific neural representations of correct and incorrect decisions, at different levels of confidence. Confidence measures served as a proxy for the vividness of perceptual experience [17], which allowed us to separate strong perceptual illusions (incorrect decision with high confidence) from weak perceptual illusions or mere guessing behavior (incorrect decision with low confidence) [7].

We experimentally dissociated perceptual from decision-related neural processes by employing 2 separate category localizer tasks: one uniquely tuned to sensory features of the images and the other sensitive to both sensory features and decision processes. Using a between-task MVPA generalization approach, we could track perceptual and decisional neural processes [1820], while they evolved across time and how these processes differed for correct and incorrect perceptual decisions as a factor of confidence, and thus strength of perceptual experience. This strategy has a similar goal as recent approaches in the field of consciousness science, in which no-report paradigms have been introduced. In studies employing such no-report paradigms, neural processing is typically compared between experimental conditions in which participants are aware of specific stimuli, but do not have a task to do and thus merely passively perceive these stimuli, versus a condition where active report is required. Contrasting these conditions, usually also including an unaware condition is thought to separate neural signals associated with perceptual experience per se versus neural signals of post-perceptual processes (or task-relevance in the broader sense) [2126]. We found that decision errors made with high confidence were associated with neural representations reflecting the misperceived object (e.g., a house could be decoded while a face was presented). This was the case even when the classifier was trained on a localizer task in which these objects were task-irrelevant and unattended, showing that such decision errors stem from perceptual illusions that are sensory in nature.

Results

There were 2 experimental sessions. In the first, participants performed 2 category localizer tasks, while their brain activity was measured using EEG. The localizer tasks were used to train our pattern classifiers (LDA, linear discriminant analysis) [27]. In the sensory localizer task (S1A Fig), participants reported an infrequent contrast change of the central fixation dot (20% of trials). At the same time, streams of house and face images were shown at the center of the screen, which were fully task-irrelevant. In the decision localizer task (S1B Fig), a masked image of a face or of a house was presented and at the end of each trial participants indicated which stimulus category they had perceived. The images were therefore task-relevant and attended. In the sensory localizer, the classifiers’ sensitivity was thus tuned mainly to sensory features of the 2 stimulus categories due to the absence of attention/task-relevance of the face/house images [26,28,29]. Therefore, the sensory localizer is reminiscent of a so-called no-report paradigm [30,31], but then used to train localizers instead of being used as the main task of interest [2224,32]. The decision localizer was, besides sensitive to sensory features, also sensitive to post-perceptual decision processes. Stimulus-response mappings were counterbalanced across blocks in the localizer tasks to prevent motor response preparation from systematically biasing stimulus category decoding. The orientation of the presented images was either left tilting or right tilting, with 50/50 likelihood. This feature of the stimulus was always task irrelevant and was used to test for differences in decoding between task-relevant (category) and task-irrelevant features (orientation) in the main discrimination task, as a function of decision correctness and confidence (see Fig 1A for an illustration of the main task).

Fig 1. Task and behavior.

Fig 1

(A) Each trial started with a central red fixation dot after which a forward mask was shown, followed by an image of a face or a house and a backward mask. Images were either tilted to the right or to the left (task-irrelevant feature, 5° or 355° angle). Note that in the example trial, only a left-tilted house image is shown. Participants reported whether they perceived a house or a face and indicated their confidence in this decision. (B) Perceptual sensitivity (d’) as a function of decision confidence. The underlying data and scripts supporting this figure can be found on FigShare (https://doi.org/10.21942/uva.c.6265233.v1).

In the second session of the experiment, EEG recordings were obtained during the main perceptual discrimination task, which was similar to the decision localizer task of the first session [33], but now images were presented shorter (20 or 30 ms) and performance was staircased during practice blocks in order to achieve approximately 65% to 70% discrimination accuracy. After each decision, participants also provided a confidence rating regarding their estimated accuracy of their face/house discrimination response. Discrimination performance was kept low to elicit many incorrect perceptual decisions and hence to induce enough misreports for subsequent analyses.

Participants correctly discriminated faces and houses in 69.47% of all images (SD = 7.48) in the perceptual discrimination task. There was a slight tendency to respond house more often than face (58% versus 42%), similarly to our previous study [33]. As expected, signal detection theory (SDT)-based perceptual sensitivity (d’) increased with reported confidence level (main effect of confidence: F3,117 = 116.69, p < 0.001, η2 = 0.54, Fig 1B) and ranged from d’ = 0.11 at confidence level 1 (least confidence) to a d’ of 2.07 at confidence level 4 (most confidence).

The time course of category representations: Decision localizer

To investigate the time course of category-specific neural representations, we trained classifiers on EEG data recorded during the 2 localizer tasks and applied them to the main perceptual discrimination task data. We will first report the classification when classifiers were trained on the decision localizer (Fig 2A) and then the cross-classification to the main task (Fig 2B). Category-specific neural representations could be decoded based on the decision localizer (10-fold validation scheme, i.e., within-task decoding, Fig 2A). The generalization across time (GAT) matrix, time-locked to stimulus onset, exhibited the expected mixture of transient and stable on-diagonal decoding as well as persistent off-diagonal decoding profiles observed previously [3337]. Following the analysis approach by Weaver and colleagues [33], who used a similar face/house discrimination task, we used matching time-windows for statistical analysis as used in that earlier study. For these analyses, based on Weaver and colleagues [33], decoding was examined on several stages, first of which a transient peak in an early time-window between 150 and 200 ms (Fig 2A, inset 2) [11,1315]. The timing (150 to 200 ms; the decoding peak was found at approximately 166 ms) and scalp topography of the early peak is probably related to the N170 ERP component, often related to face processing [36,3840], although also observed for letters, words, and biological motion [41,42]. Peaks of N170-like decoding have been associated with neural processes in the occipital face area, superior temporal sulcus, and/or the fusiform face area (FFA) in ventral-temporal cortex [4349]. Previous work has also shown that the strength of this early face-selective component is proportional to the strength (e.g., phase coherence of the stimulus) of the presented stimulus [18,40].

Fig 2.

Fig 2

(A) GAT matrix for the decision localizer (all electrodes) and ROI marked by inset black boxes (numbers 1–4). On the right, covariance/class separability maps for each diagonal ROI, indicating the underlying data contributing to the classification performance. Highlighted channels significantly contributed to the classification performance after cluster-based permutation testing over channels (corrected for multiple comparisons). Covariance/class separability maps 2 and 4 are the same because decoding scores within insets 2 and 4 were obtained using the same training time-window. (B) Classifiers trained on the decision localizer applied to the main discrimination task (cross-task validation procedure, all electrodes), revealing a high degree of generalizability between the 2 tasks. The underlying data and scripts supporting this figure can be found on FigShare (https://doi.org/10.21942/uva.c.6265233.v1). AUC, area under the curve; GAT, generalization across time; ROI, regions of interest.

We also focused on a more stable square-shaped decoding profile between 350 and 500 ms (on-diagonal), capturing late category-specific processes (Fig 2A, inset 3), with a central-parietal topography (late-stage topography). This decoding profile has previously been suggested to reflect global ignition, large-scale feedback processes allowing information to be broadcasted throughout the entire brain, making information explicit for report and decision-making [8,34,37,50].

Additionally, we examined decoding accuracy in a late time-window from 350 to 500 ms, based on training on early 150 to 200 ms classifiers (i.e., off-diagonal decoding, (Fig 2A, inset 4)). Given that decoding is in this case based on early classifiers tuned to sensory features, this stage is thought to reflect longer lasting sensory processes of categorical information across time [33,34,51]. This late-latency off-diagonal decoding pattern, referred to as the “perceptual maintenance” stage, has been suggested to reflect reactivation of early sensory stages through feedback processes [3335].

Finally, although decoding was pronounced during 3 time-windows observed previously by Weaver and colleagues [33] (insets 2, 3, and 4), it was also observed at an even earlier time-window (80 to 130 ms) with a prominent occipital–parietal topography (Fig 2A, inset 1, early-stage topography). Because this early effect was absent in the same task when images where presented vertically [33], this early peak likely reflects differences in the orientations of the images presented, which we varied systematically across trials (note that training and testing was performed on balanced set of left and right oriented images). Although we follow a confirmatory approach by selecting time-windows of interest (TOI) beforehand [33], we will always in addition use across time permutation tests with multiple comparisons corrections to test for (unexpected) effects observed at other time-windows and to further specify the approximate onset and/or duration of our significant effects (see Materials and methods for details).

Cross validation between the decision localizer and the main discrimination task (Fig 2B for all channels; see S2 Fig for occipital–parietal channels) revealed a highly similar decoding profile for the 2 tasks. Classifiers’ performance in differentiating between face and house images were highest on occipital–parietal electrodes, particularly during early latencies (S2 Fig). Since our aim was to examine whether decision errors have a perceptual origin, we report the decoding results based on data from occipital–parietal channels here to maximize sensitivity to neural signals that distinguish between the 2 categories. In the S1 Text, S4 Fig, and S5 Fig, we also report the critical decoding analyses using all channels, which yielded similar results.

Diagonal decoding based on decision localizer

Next, we examined decoding scores as a function of decision correctness (correct/incorrect) and confidence (low confidence (LC): level 1 and 2 versus high confidence (HC): level 3 and 4, as in [33]). This created 4 conditions: HC correct, LC correct, HC incorrect, and LC incorrect trials. GAT matrices and on/off-diagonal time courses for these 4 conditions are shown in Fig 3A–3D. Note that the “testing labels” for classification were based on the decision of the subject in the main task (deciding face or house), not the presented stimulus. For correct trials, the presented stimulus and the decision are by definition the same. For incorrect trials however, the stimulus and the decision are of different categories. This means that if decoding scores are observed that are higher than chance level, the underlying data pattern correlates with the category of the reported stimulus, the decision. We will refer to this situation as positive decoding (with respect to chance level). In contrast, when below chance level decoding is observed, this would reveal that the underlying data pattern correlates with the presented stimulus, not with the reported stimulus (stimulus and response are different on incorrect trials, see Methods for details). The classifier thus finds negative evidence for the decision, and therefore we will refer to this situation as negative decoding (with respect to chance level).

Fig 3. Face/house classification in the main task while training on the decision localizer.

Fig 3

(A) GAT matrices for correct trials and (B) incorrect trials in the main discrimination task (occipital–parietal channels). Trials are separated by confidence—HC and LC—in the decision. (C) On-diagonal and (D) off-diagonal (perceptual maintenance) time courses of correct and incorrect decisions, separated by confidence. Black boxes indicate the time-windows (early, mid, late) used for averaging the AUC scores. Colored horizontal lines indicate periods of significant decoding with respect to chance. Longer-lasting sensory traces are derived by training a classifier on the time-window 150–200 ms and testing it across the entire time-window (panel D). Classification performance was evaluated at each time point using cluster-based permutation testing (two-tailed cluster-permutation, alpha p < 0.05, cluster alpha p < 0.05, N permutations = 1,000). (E) Bar plots show average AUC values for the TOI highlighted in panels (C) and (D). The underlying data and scripts supporting this figure can be found on FigShare (https://doi.org/10.21942/uva.c.6265233.v1). AUC, area under the curve; GAT, generalization across time; HC, high confidence; LC, low confidence; TOI, time-windows of interest.

Statistical analyses were performed on the diagonal and off-diagonal time courses (late sensory traces) (Fig 3C and 3D). We first report the analyses for the diagonal time courses using a 3-way repeated measures ANOVA with the factors Latency, Correctness, and Confidence. As anticipated, category decoding was better for correct than incorrect trials (F1,39 = 135.05, p < 0.001, ηp2 = 0.78) as well as for HC versus LC trials (F1,39 = 27.75, p < 0.001, ηp2 = 0.42). Category decoding differed across the 3 (early, mid, late) diagonal decoding time-windows (F2,78 = 32.16, p < 0.001, ηp2 = 0.45), with differences being modulated by the level of confidence (Latency × Confidence: F2,78 = 6.88, p = 0.002, ηp2 = 0.15; no evidence for an interaction between confidence and correctness: F2,78 = 0.022, p = 0.88, ηp2 = 0.001). As hypothesized, the characteristic pattern of decoding across 3 diagonal time-windows differed between correct and incorrect trials (Correctness × Latency interaction: F2,78 = 36.57, p < 0.001, ηp2 = 0.48), independent of confidence (Correctness × Latency × Confidence interaction: F2,78 = 1.12, p = 0.33, ηp2 = 0.03). As shown in Fig 3C, diagonal decoding profiles were very different for correct and incorrect decisions, especially for the early and mid-processing stages. Most prominently, it can be observed that for correct trials, all processing stages show positive decoding, whereas in sharp contrast the 2 early stages for the incorrect trials showed negative decoding. This negative decoding pattern on incorrect decisions illustrates that early EEG data patterns look more similar to the presented stimulus category than to the reported stimulus category (note again that the decisions were used as the classifier labels). The fact that the late stage decoding flips to positive decoding illustrates that later EEG data patterns are more similar to the reported category than the presented category (Fig 3C). We will come back to these effects later in the Results section.

For correct trials, decoding strength differed across the 3 TOI’s (F1,39 = 47.01, p < 0.001, η2 = 0.55) and was overall better for HC trials than for LC trials (F1,39 = 31.91, p < 0.001, η2 = 0.45). Confidence interacted with latency (TOI’s) (F2,78 = 10.28, p < 0.001, η2 = 0.21), indicated by better decoding in the mid/late processing stage for HC versus LC trials (mid-stage: t39 = −6.92, p < 0.001, d = −1.09; late stage: t39 = −3.06, p = 0.004, d = −0.48; Fig 3C), while we did not find evidence for this confidence modulation for the early processing stage (t39 = −1.25, p = 0.22, d = −0.2). These results are summarized in the bar plots of Fig 3E.

On incorrect decisions, decoding scores across 3 diagonal time-windows differed (F2,78 = 18.84, p < 0.001, ηp2 = 0.33) and while decoding was also higher for HC than for LC trials (F1,39 = 11.64, p = 0.002, ηp2 = 0.23, Fig 3C), the interaction between these 2 factors was not robust, although numerically, decoding was higher in HC compared to LC trials, in particular at the later versus early stage (F2,78 = 2.81, p = 0.07, ηp2 = 0.07, BFexcl = 1.96). A series of planned simple comparisons showed that the flip from negative to positive decoding was stronger for high confidence decisions (results are summarized in Fig 3E). The below-chance decoding during the 2 early diagonal decoding time-windows on incorrect decisions indicated that the classifiers were picking up the presented stimulus category during sensory stages of visual information processing. At the earliest time-windows (80 to 130 ms), we did not find evidence that decoding was modulated by confidence (HC versus LC; t39 = −0.28, p = 0.78, d = −0.05, BF01 = 5.64) and indeed, decoding was below chance-level for both confidence levels (LC: t39 = −5.83, p < 0.001, d = −0.92; HC: t39 = −3.38, p = 0.002, d = −0.53). During the following mid-stage (150 to 200 ms), classifiers also mainly decoded the veridical stimulus category, but interestingly in particular when participants reported low confidence in their decision (LC: t39 = −5.05, p < 0.001, d = −0.8; HC: t39 = −0.35, p = 0.73, d = −0.06, BF01 = 5.54; LC versus HC: t39 = −2.95, p = 0.005, d = −0.47). Then finally, during the late processing stage (350 to 500 ms), patterns of neural activity represented the incorrectly reported stimulus category, but in particular when participants where highly confident in their decision, so when the illusion was strongest (HC: t39 = 3.53, p < 0.001, d = 0.56; LC: t39 = 0.44, p = 0.67, d = −0.07, BF01 = 5.36, LC versus HC: t39 = −3.28, p = 0.002, d = −0.52). These results show a reversal from the presented stimulus category to the reported category for incorrect decisions, especially for high confidence decisions.

Off-diagonal decoding based on the decision localizer

Next, we examined decoding of the reported stimulus category off-diagonally, training on the N170 time-window (150 to 200 ms) using the decision localizer task data, separately for correct and incorrect decisions (only 1 stage). First, on correct decisions, decoding of the reported stimulus category was robust for both confidence levels (all p’s < 0.001), although decoding was higher for HC than LC decisions (t39 = −3.98, p = <0.001, d = −0.63; Fig 3D and 3E, right panel), suggesting that the veridical/perceived stimulus category was maintained perceptually. On incorrect decisions, however, what was perceptually represented off-diagonally was not the presented stimulus, but the misreported stimulus category, and this effect was stronger for high confident than low confident incorrect decisions (HC: t39 = 3.86, p < 0.001, d = 0.61; LC: t39 = 1.41, p = 0.17, d = 0.22, BF01 = 2.35; LC versus HC: t39 = −3.06, p = 0.004, d = −0.48; Fig 3E). Thus, based on the transient decoding peak observed at early processing stages (150 to 200 ms in the decision localizer), the erroneous and misreported stimulus category was represented in a persistent perceptual format later in time, despite never being presented on the screen and never being neurally represented at earlier time points (Fig 3D). Statistical tests across time revealed that the incorrect off-diagonal decoding trace emerged relatively early, starting at approximately 213 ms and peaking at 236 ms after stimulus presentation (Fig 3D, horizontal blue colored bars reflect significant time points after cluster-based permutation testing). Although we observed a perceptual representation of an incorrectly reported category, concluding that errors in decision-making in this task have a true perceptual origin (in the case confidence is high), based on the current results may be premature, because the decoded signal could reflect a combination of perceptual and decision-related processes [20,52]. Therefore, we further examined this issue when training our classifiers on the sensory tuned localizer.

The time course of category representations: Sensory localizer

To explore to what extent errors in decision-making have a true perceptual origin, we performed the same set of analysis, but now while training the classifier on the sensory localizer task (Fig 4A). In this task, participants performed a simple change detection task on the color of the fixation cross making the face/house images fully task irrelevant. Because the sensory localizer task does not have a decision component to the face/house images, it is uniquely sensitive to sensory features. Therefore, we expected that the late square-shaped on-diagonal decoding pattern, possibly related to global ignition [34,37,50], would be strongly reduced or even disappear when cross classifying between the sensory localizer and the main perceptual decision task (Fig 4B). Previous studies have shown that this late square-shaped decoding pattern is likely related to decision processes arising after perceptual analysis, which disappear when visual input is task irrelevant [1820,53]. On the other hand, because decoding is still performed on the face/house images, the sensory stages should be relatively unaffected by this task relevance manipulation [18,49]. In this Results section, we focus on those aspects of decoding that are informative for addressing the question whether decision errors have a perceptual origin, but note that all other effects were similar to the analyses reported in the “decision localizer” section and the details thereof can be found in the Supporting information (including all the ANOVA’s).

Fig 4. Face/house classification based on the sensory localizer.

Fig 4

(A) GAT matrix for the sensory localizer (all electrodes) and ROI marked by inset black boxes (numbers 1–4). On the right, covariance/class separability maps for each diagonal ROI, indicating the underlying data contributing to the classification performance. Highlighted channels significantly contributed to the classification performance after cluster-based permutation testing over channels (corrected for multiple comparisons). Covariance/class separability maps 2 and 4 are the same because decoding scores within insets 2 and 4 were obtained using the same training time-window. (B) Classifiers trained on the sensory localizer applied to the main discrimination task (cross-task validation procedure, all electrodes). (C) On-diagonal and (D) off-diagonal time courses of correct and incorrect decisions, separated by confidence. Black boxes indicate the time-windows (early, mid, late) used for averaging the AUC scores. Colored horizontal lines indicate periods of significant decoding with respect to chance. Longer-lasting sensory traces are derived by training a classifier on the time-window 150–200 ms and testing it across the entire time-window (panel D). Classification performance was evaluated at each time point using cluster-based permutation testing (two-tailed cluster-permutation, alpha p < 0.05, cluster alpha p < 0.05, N permutations = 1,000). (E) Bar plots show average AUC values for the TOI highlighted in panels (C) and (D). The underlying data and scripts supporting this figure can be found on FigShare (https://doi.org/10.21942/uva.c.6265233.v1). AUC, area under the curve; GAT, generalization across time; HC, high confidence; LC, low confidence; ROI, regions of interest; TOI, time-windows of interest.

For correct decisions, decoding was above chance for both confidence levels (all p’s < 0.001, see Fig 4C) in the early and mid-stage, but we did not find evidence for decoding at the late diagonal stage (LC: t39 = 1.104, p = 0.28, d = 0.18, BF01 = 3.33; HC: t39 = 1.34, p = 0.19, d = 0.21, BF01 = 2.55). Thus, as anticipated, the late on-diagonal 350 to 500 ms stage was not robust when training on the sensory localizer and classifying on the main discrimination task, whereas all other stages of visual processing remained relatively intact (Fig 4C). During the perceptual maintenance stage, the activity patterns reflected the veridical/reported stimulus category (both p’s < 0.001, LC versus HC: t39 = −0.98, p = 0.34, d = −0.15, BF01 = 3.76, Fig 4D).

For incorrect decisions, the early and mid-stage were also similar as compared to when the classifiers were trained on the decision localizer (early stage: LC: t39 = −3.4, p = 0.002, d = −0.54; HC: t39 = −3.91, p < 0.001, d = −0.62; mid-stage: LC: t39 = −3.04, p = 0.004, d = −0.48; HC: t39 = −0.11, p = 0.91, d = −0.02, BF01 = 5.83). Specifically, we observed negative decoding for both early stages (again, except for HC trials in the mid-stage), reflecting the processing of the veridical stimulus input. Crucially, also here, the late-stage decoding disappeared (LC: t39 = 0.16, p = 0.88, d = −0.03, BF01 = 5.8; HC: t39 = 1.27, p = 0.21, d = 0.2, BF01 = 2.8, Fig 4C). However, most importantly, during the off-diagonal perceptual maintenance stage using the classifier trained on the sensory localizer data, we could still decode the incorrectly reported stimulus category, in particular when participants reported high confidence in their decision (LC: t39 = 1.78, p = 0.08, d = 0.28; BF01 = 1.4; HC: t39 = 3.03, p = 0.004, d = 0.48, BF01 = 0.12; LC versus HC: t39 = −1.94, p = 0.06, d = −0.31, BF01 = 1.07, Fig 4D and 4E). Again, here as well, cluster-based testing across the entire time-window revealed that the perceptual representation of the incorrect stimulus spontaneously emerged right after the 150 to 200 ms time-window most strongly representing face/house differentiation (see Fig 4D). This finding shows that misreported decisions may be due to true perceptual illusions.

Stimulus orientation decoding

Our results so far show that initial information processing stages are stimulus related, whereas later off-diagonal decoding captures perceptual aspects of decision errors. To show that these effects are specific for task-relevant features of the images (face/house category), we additionally focused on stimulus orientation decoding as a function of correctness and confidence. Stimulus orientation was a task-irrelevant feature of the stimuli and previous studies have shown that it can be decoded already approximately 100 ms post-stimulus [50]. A single isolated decoding peak within the 30 to 130 ms post-stimulus was observed both for within-task decoding (Fig 5A, 10-fold decoding) as well as cross-task classification (decision classifiers ➔ main discrimination task, Fig 5B and 5C). Decoding strength was similar for correct and incorrect decisions (F1,39 = 0.06, p = 0.82, ηp2 = 0.001, BF01 = 5.7) and more evidence was observed for the absence of an effect. We also did not find evidence that confidence modulated orientation decoding in any way (F1,39 = 0.54, p = 0.47, ηp2 = 0.014, BF01 = 4.77; interaction correctness × confidence: F1,39 = 0.18, p = 0.67, ηp2 = 0.005, BFexcl = 4.22).

Fig 5. Tilt (right/left) classification based on decision localizer.

Fig 5

(A) GAT matrix of the decision localizer data for arbitrating left/right tilted stimuli (irrespective of face/house category, based on occipital–parietal channels). The black square-shaped inset indicates the time-window used for averaging AUC decoding scores (30–130 ms). (B) On-diagonal time courses for correct and incorrect decisions separated by confidence. Colored horizontal lines indicate periods of significant decoding with respect to chance. Classification performance was evaluated at each time point using cluster-based permutation testing (two-tailed cluster-permutation, alpha p < 0.05, cluster alpha p < 0.05, N permutations = 1,000). (C) Bar plots show average AUC values for the time-window of interest highlighted in panels (A) and (B). The underlying data and scripts supporting this figure can be found on FigShare (https://doi.org/10.21942/uva.c.6265233.v1). AUC, area under the curve; GAT, generalization across time; HC, high confidence; LC, low confidence.

Discussion

The goal of this study was to investigate whether errors during perceptual decision-making have a sensory or a decisional origin. We isolated sensory versus decision-related neural processes by using 2 distinct functional localizers, differently tuned to these processes [20]. Replicating a wealth of previous studies, the earliest neural responses reflected the physical properties of the presented stimulus [5155]. The crucial finding reported here is however that when participants made an incorrect perceptual decision with high confidence, early stages of visual information processing (<~200 ms post-stimulus) still represented object features of the presented stimulus (e.g., “faceness” in the case a face stimulus was presented), while later stages represented the illusory perceptual properties of the reported stimulus category (e.g., “houseness” in case a face was presented, but a house response was given). More importantly, at least 3 findings reported here demonstrate the sensory and true illusory nature underlying such high confident erroneous perceptual decisions and thereby go beyond traditional interpretations associating late neural processing with largely stimulus-driven perception.

First, persistent sensory-based decoding across time (perceptual maintenance, off-diagonal decoding) on incorrect trials was observed when training a classifier on the N170-like processing stage, a peak in the EEG signal highly associated with face perception [39,44,4951,5658]. Whereas low-level visual features, such as stimulus orientation and position, drive early-latency decoding (<100 ms post-stimulus), slightly later in time decoding has been shown to reflect the integration of sensory features into coherent objects and object categories (as early as approximately 150 ms post-stimulus, [8,19,50,51,59]). Regarding face perception, a recent MEG study showed that decoding of face representations peaked at approximately 160 ms post-stimulus, and this signal correlated with subjective ratings of “faceness” of inanimate objects [57]. By combining MEG and functional magnetic resonance imaging (fMRI) measures, it was shown that these face responses likely originated from the FFA in ventral temporal cortex (see also [49,56]). This mid latency face-like component has been shown to be relatively unchanged when object categories are task irrelevant [18]. Others have also provided evidence for strong occipital contributions to the decoding of perceptually integrated objects and object categories during similar early latencies [34,60]. These findings highlight that the mid latency (150 to 200 ms) decoding stage indexes perceptually coherent category representations, likely originating from (higher level) visual cortices. A novel line of evidence based on decoding and generalization across time [61] has shown that these category representations are decodable for much longer periods of time than originally anticipated, as reflected in off-diagonal decoding patterns, focused on in this work [3335,53]. Also in the data presented here, sensory representations were decodable for relatively long periods of time (see Figs 3 and 4) [35].

A second reason why erroneous decisions in our task are likely caused by true perceptual illusions is that long-lasting late sensory representations were observed even when classifiers were trained on the sensory-tuned localizer task (perceptual maintenance, Fig 4D). Sensory-tuned classifiers were trained on a task in which face/house images were fully task irrelevant and participants focused their attention on detecting small changes in the color of the fixation cross. Therefore, this classifier was only able to pick up (unattended) sensory features of the images that arbitrate between faces and houses, in the absence of most post-perceptual and decision-related processes [1820,26]. That we were successful in isolating sensory features was evident in the strong decrease of the late on-diagonal “square-shaped” decoding profile while training on the sensory localizer task. Another approach to isolate neural signals associated with perceptual experience from decision- and report-related processes is the development of no-report paradigms. In such paradigms, observers may be aware or unaware of presented stimuli, but in all cases do not make perceptual judgments about them. The resulting contrasts (aware versus unaware) is then supposed to isolate perceptual experience, limiting the influence of post-perceptual processes. Also in this line of work, it has been shown that the majority of the high-level activations in frontal and parietal cortex disappear (similar to our late square-shaped decoding), and hence, that such activation patterns are mostly associated with post-perceptual processes and much less so with perceptual experience itself [2126].

Third and finally, the flip from veridical stimulus-related processing relatively early in time towards illusion-based (or report-based) processing later in time was only observed for high confidence decisions and not for low confidence decisions. In other words, the emergence of these perceptual representations in the EEG signal depends on the strength of the perceptual illusion, assuming that the perceptual illusion was strong for high confidence (inaccurate) decisions and weak for low confidence decisions (see [62] for a way to explore “partial errors” by inspecting the electromyography signal on correct responses). Low confidence decision errors were likely the result of guessing [7]. In a related fMRI study, Summerfield and colleagues have shown that misreporting a house for a face in a challenging discrimination task similar to ours, is accompanied by increased fMRI BOLD activity in the FFA, but not in other face responsive regions such as the occipital face area (OFA) [7]. This suggests that the FFA is associated with illusory perception of faces, whereas earlier face processing regions in the cortical hierarchy are not, and reflect the presented or veridical stimulus category [7,63,64]. Although in this fMRI study, discrimination responses were also given with high or low confidence, unfortunately too few high confidence error trials could be obtained to perform the crucial analyses that test for confidence modulations on decision errors. Furthermore, in this study, image contrast varied across trials, and therefore correctness, confidence, and image contrast co-varied, which complicated the separation between perceptual effects and effects of bottom-up input strength (e.g., decision errors were also often low contrast trials). However, together these results highlight that relatively early stages in object processing, both in time as well as in the cortical hierarchy, are linked to the image input, whereas later stages are more internally driven or perceptual in nature. Intriguingly, our results also show that emerging perceptual representations do not necessarily have to follow from bottom-up sensory processes. When participants misreported the presented stimulus category (on incorrect trials), we found no N170-like decoding pattern for high confidence decisions, but we did observe the emergence of perceptual representations of the reported category later in time (Fig 3B). Thus, illusory category-specific perceptual signatures emerged without any clear bottom-up sensory evidence for this percept.

Although these results reveal that high confidence errors are associated with neural representations indicative of true perceptual illusions, they leave open the question what drives these reversals in representations from the veridical to the misreported stimulus category. There are several likely candidate mechanisms, which future studies should aim to arbitrate between. One intuitive candidate may be decision-related feedback to visual cortex. The frontoparietal network involved in perceptual decision-making has strong feedback connections to sensory regions and can continuously inform sensory regions about the unfolding decision variables through feedback connections [65,66]. In general, activity observed in sensory regions can thus be a mixture of both feedforward and feedback processes, especially at longer latencies (>~100 ms) ([39], but see [40]), making it difficult to disentangle sensory and post-sensory or decision processes even at the level of single sensory neurons [6774]. fMRI studies in humans have for example shown that early visual cortex activity may reflect a combination of stimulus-related feedforward activity and post-sensory decision-related feedback [75] or top-down processes such as task set and expectations [76]. The observation that we only observed the flip form veridical to the reported perceptual representations in high confidence trials runs contrary to the interpretation that decision-related feedback may be the best explanation of the observed data patterns. If these effects were purely decision-driven, we should have observed similar decoding patterns for low confidence decisions, although we cannot fully rule out the possibility that these sensory effects were caused by decision-related feedback.

Another possibility is that, because the bottom-up visual input is impoverished due to strong masking, higher-order areas may incorrectly explain the noisy bottom-up signals. Previous evidence suggests that sensory templates can be implemented in a top-down manner even in anticipation of sensory stimulation [7779] and a body of work shows that the FFA is sensitive to top-down factors, such as task-relevance, expectations, and context [76]. For example, seeing a face in ambiguous Mooney images [80] or face-like inanimate objects, the phenomenon know as pareidolia [57], leads to stronger FFA responses than when no face is perceived, despite very similar bottom-up input. That the FFA is sensitive to context is also illustrated by the observation that the FFA activates to simple oval shapes when the context suggests this shape might be a face [81]. Furthermore, Summerfield and colleagues showed effects of task set on object processing [82]. They administered a task in which in 1 block of trials participants had to detect faces among faces, houses, and cars, and in another block, they had to detect houses, among the same 3 categories of stimuli. Thereby the authors manipulated the top-down “perceptual set” while keeping the bottom-up input the same across blocks. When comparing the overall activity in the FFA on blocks in which faces had to be detected versus houses had to be detected, increased FFA activity was observed. The authors concluded that top-down signals from frontal cortices could sensitize visual regions responsible for collecting evidence about the presence of faces. Although we did not directly manipulate context or expectations in our design (faces and house were equiprobable), participants may have developed expectancies about image likelihood over time, which may have then led to explaining the noisy sensory input, probably at the level of object-selective cortex, in an erroneous way [7]. Another related and interesting finding has been reported by Tu and colleagues [8]. In their task, participants had to discriminate faces from houses or cars while EEG and fMRI were measured simultaneously. As the authors have observed in different task contexts [18,83], there were both early (approximately 200 ms after stimulus) and late (approximately 500 ms after stimulus) EEG components that discriminated between stimulus categories. Interestingly, in behavior, participants had an overall face bias: they often misreported non-faces as faces. By combining EEG with fMRI, the authors showed that individual differences in the strength of this face-bias correlated across participants with the strength of the interaction between “early” (e.g., FFA, parahipocampal place area, parietal regions) and “late” neural networks (primarily frontal regions, e.g., anterior cingulate cortex). Therefore, the strength of a face decision bias may depend on the degree of top-down predictive modulations from frontal to sensory cortices. Because no single trial confidence reports were obtained in that study, the authors could not relate their neural data and categorical responses to introspective confidence in the observers’ decisions. Our study therefore extends this previous work by showing that (at least) high confidence decision errors have a perceptual origin. In general, together these studies suggest that both perceptual biases and decision biases depend on interactions between top-down and bottom-up processes (see also [9,12,14,84] for ways aiming to behaviorally disentangle sensory from decision biases).

Predictions, however, do not have to be strategic or conscious, but may fluctuate naturally over time in neural activity, e.g., in object selective cortex [85]. It has been argued that ongoing fluctuations of neural activity are not random (i.e., stochastic noise), but may contain content-specific information, for example, associated with previously experienced stimuli and (perceptual) learning [8693]. It has, for example, been shown recently that when patterns of pre-stimulus neural activity measured with magnetoencephalography matched the stimulus category that was later presented, which had to be discriminated from other object categories by human participants, perceptual sensitivity was improved [85]. Thus, another possibility is that, in the face of uncertainty, when sensory input is weak or ambiguous, the system may settle on the most likely interpretation of the sensory input, integrating the sensory evidence and the ongoing activity present at the moment the stimulus travels through the cortical hierarchy. In our study, when perceptual decisions were incorrect, but made with high confidence, the perceptual system may have settled in a state coding for the specific stimulus category that was perceived, leading to a true illusion in perception. Yet, in low confidence decision error trials, the perceptual system may have not settled on any of the 2 stimulus categories, as strikingly reflected in the absence of significant decoding for either the presented or the reported object category.

To conclude, we find that object categorization errors are associated with a quick reversal in sensory representation from the veridical, presented stimulus category to the reported stimulus category, but only for decision made with high confidence. This finding shows that high-confidence decision errors are caused by true illusions in perception.

Methods

Ethics statement

Participants provided written informed consent prior to the start of the experiment and were tested following a protocol approved by the ethical committee of the Department of Psychology of the University of Amsterdam (project number: 2019-BC-10724).

Participants

A total of 44 participants (33 female, mean age = 23.34; SD = 5.68) from the University of Amsterdam, all right-handed, with reported normal or corrected-to-normal vision and no history of a psychiatric or neurological disorder, were tested in this study. Participants received research credits or money (10 euros per hour) for compensation. Forty participants, all between the age of 18 and 35, completed 2 experimental sessions of this study and comprised the final sample included in all reported analyses (30 female, mean age = 22.83; SD = 3.31). The remaining 4 participants were identified based on their data from session 1, after which they were excluded from further participation/analyses. One participant was excluded from the final analysis because of surpassing the age criterion (age > 35), while the other 3 participants were excluded due to the around-chance performance in 1 or both localizer tasks in the first session.

Materials

All tasks used in the current study were developed and executed using Matlab 8 and Psychtoolbox-3 software within a Matlab environment (Mathworks, RRID:SCR_001622). Stimuli were presented on 1,920 × 1,080 pixels BenQ XL2420Z LED monitor at a 120-Hz refresh rate on a black (RGB: [0 0 0], ± 3 cd/m2) background and were viewed with a distance of 90 cm from the monitor using a chin rest.

Procedure and stimuli

The experiment consisted of 2 sessions, each approximately 3 h long and scheduled on 2 different days for each participant. In the first session, participants completed a change detection task and a discrimination task while we measured their brain activity using EEG. In the change detection task, participants focused on the central fixation cross and reported its color change whenever it changed from red to light red, while a stream of brief house and face images was shown simultaneously on the screen (S1A Fig). In the discrimination task, in every trial, participants saw a masked image of a face or a house (S1A Fig). At the end of each trial, they were asked to indicate which stimulus category they perceived. The discrimination task administered in the second session was highly similar (see the details of the design below), but in addition to reporting which stimulus category they perceived, participants also provided confidence ratings regarding their decision.

MVPA classifiers were trained on EEG data recorded during the first session, separately for the change detection task (“sensory” localizer) and the discrimination task (“decision” localizer), which were then applied to the discrimination task EEG data recorded in the second session.

Session 1: Change detection task (sensory localizer). In this task, participants focused their attention on a centrally presented fixation cross, which was superimposed on a rapidly changing sequence of house and face images (S1A Fig). The fixation cross remained present on the screen throughout a block of trials. Participants’ task was to monitor the fixation cross and report its brief color change, from red (RBG: [255 0 0]) to a lighter shade of red (RBG: [125 0 0]) for 100 ms on 20% of randomly chosen trials. Participants were instructed to press the spacebar whenever they noticed the color change of the fixation cross. Note that this change detection task adds an orienting response to the fixation cross on some trials, which was however unrelated to, and not predictive of, the occurrence of the task-irrelevant face or house stimuli.

Stimuli set consisted of 180 unique houses and 180 unique faces (90 male and 90 female). Face and house images were gray scale stimuli, obtained from [33]. Face and house stimuli were equated for spatial frequency and luminance (for details, see [33]). All stimuli subtended 16 × 20° visual angle, were presented centrally on a black background, and were tilted to the right or left at a 5° or 355° angle, respectively. Each image was shown for 100 ms with the ISI that was jittered between 1,200 and 1,400 ms. The task-relevant color change of the fixation cross could occur only in the ISI, at a randomly determined moment 15 ms after the start of the ISI and 130 ms before its end. Participants were instructed not to pay attention to images of houses and faces while maintaining their fixation at the center of the screen. Images of houses and faces were thus task irrelevant, although they were concurrently processed visually.

By ensuring that participants’ attention was focused on the centrally presented task, the aim was to minimize the possibility that MVPA classifiers were impacted by systematic eye movements that could be strategically deployed to discriminate between stimuli classes [94]. For example, a consistent eye movement towards the top of the stimulus to detect face-defining features (e.g., eyes) could alone drive multivariate differences between houses and faces stimuli.

In total, 200 images of houses and faces (100 of each category, half tilted to the right and half tilted to the left) were shown in each of 6 experimental blocks. Before the start of the first experimental block, each participant completed 1 practice block of 100 trials in order to get familiar with the task.

Session 1: Discrimination task (decision localizer). The task design was highly similar to the task developed by Weaver and colleagues [33]. An overview of the trial procedure can be found in Fig 1A (and S1B Fig). Each trial started with a fixation dot that remained on the screen throughout the duration of the trial. After a fixation-only interval that was jittered between 600 and 1,000 ms, a scrambled mask stimulus appeared on the screen for 50 ms, followed by a target face or a house stimulus, shown for 80 ms. The target image was followed by another 50 ms-long scrambled mask stimulus. A response screen was presented for 1,000 ms immediately after the offset of the second mask stimulus, during which participants needed to give a speeded response indicating whether they saw a house or a face, using a left-handed (“z”) or right-handed (“m”) keyboard response. Stimulus-response mappings were randomized across blocks of trials to prevent motor response preparation before the response screen was shown. The correct stimulus-response mapping (e.g., a left button press for a face and a right button press for a house stimulus) was presented at the beginning of each block. Furthermore, letters F and H (approximate size 4 × 4°), for faces and houses, respectively, were shown during the entire trial in the right and left upper corner (centered approximately 20 × 24° from fixation), or vice versa, depending on the block, as a reminder of the response mapping to the button press.

Face, house, and mask stimuli were obtained from the study by Weaver and colleagues [33]. Stimuli set consisted of 180 unique houses and 180 unique faces of which 90 faces were male and 90 were female. Face and house stimuli were equated for spatial frequency and luminance (for details see [33]). Visual masks were selected from 900 scrambled face and house images (parsed into 12 × 15 tiles and randomly shuffled) that had been made transparent and superimposed. All stimuli were gray scale, subtended 16 × 20° visual angle, and were presented centrally on a black background. Target stimuli and masks were shown tilted to the right or left, at a 5° or 355° angle, respectively.

Participants completed 14 experimental blocks containing 64 trials each, 896 trials in total. In each block, an equal number of faces and houses were shown, half of which were tilted to the right and half to the left. Before the first experimental block, each participant completed 1 practice block of 64 trials to get familiar with the task.

Session 2: Discrimination task. The task and procedure of the second session discrimination task was highly similar to the discrimination task administered in the first session (cf. [33]). Here, participants again viewed a rapid stream of house and face images and their task was to report at the end of the trial whether they had seen a house or a face. Right (5°) or left (355°) tilted house and face images were preceded and followed by visual masks, which were also tilted in the same direction. The trial procedure was identical to the discrimination task of session 1, except that the target face or house was presented for 20 ms or 30 ms. The duration of target stimuli was determined per participant depending on the discrimination performance during practice blocks, with the aim of achieving approximately 65% discrimination accuracy. Specifically, if a participant scored 60% or higher in correctly discriminating target stimuli when they were shown for 20 ms, this timing was used as the target duration for the remaining experimental blocks. Otherwise, target stimuli were shown for 30 ms.

After a 50 ms-long post-mask, a response screen was presented for 1,300 ms, during which participants needed to give a speeded response indicating whether they saw a house or a face using a left-handed (“z”) or right-handed (“m”) keyboard response. Additionally, following the discrimination response, participants needed to indicate their confidence in the accuracy of their discrimination response using a 4-point scale (1 –unsure, 4 –sure). The next trial began after the response had been given or after a 3-s timeout if no response was recorded. An equal number of right- and left-tilted faces and houses were shown in each block.

The discrimination task started with 4 practice blocks of 84 trials each, administered to familiarize participants with the task and to get an indication of target presentation time for the remainder of the task. This was done to ensure that the stimulus presentation time was sufficiently long, so that accuracy scores during the practice averaged to around 65%. Following the practice, each participant completed 16 experimental blocks of 84 trials each.

EEG measurements and preprocessing

The electroencephalogram (EEG) and electrooculogram (EOG) were recorded using the Biosemi Functional Two system (Biosemi.com). A total of 64 sintered AG/AgCl electrodes were positioned according to the 64 standard international 10/20 system, 6 external electrodes were placed on the earlobes and around the eye. The vertical EOG (VEOG) was recorded from 2 external electrodes located above and below the right eye. The horizontal EOG (HEOG) was recorded from 2 external electrodes located next to the external canthi of the eyes. The VEOG was used to detect eyeblinks and the HEOG was used to detect horizontal eye movements. Electrophysiological signals were digitized at a sampling rate of 512 Hz.

EEG data was preprocessed and cleaned before further analysis using custom scripts, the EEGLAB toolbox (v2019_1), and the Amsterdam Decoding and Modeling toolbox (ADAM; [95]). EEG data was re-referenced to the average of the earlobes, high-pass filtered at 0.5 Hz and low-pass filtered at 40 Hz, in separate steps as recommended by EEGlab. Note that a high-pass filter may distort the temporal estimates of the EEG signal [96,97], which was not the case in our dataset (see S3 Fig illustrating this). The continuous data was then epoched from −500 ms to 2,000 ms around stimulus onset. Trials containing jump artifacts were removed from the data using an adapted version of ft_artifact_zvalue muscle artifact detection function from the FieldTrip toolbox. Eye blink artifacts were removed from the data using a standard regression-based algorithm. A baseline correction in the 200-ms pre-stimulus onset interval was performed. Data were then downsampled to 128 Hz.

Statistical analyses

All analyses were done using custom scripts, the EEGLAB toolbox (v2019_1), the Amsterdam Decoding and Modeling toolbox (ADAM; [95]), and JASP software. We took a frequentist statistical approach and chose an alpha of p < 0.05 as the threshold for significance as is typical, to guard against false positives. Note, however, that this threshold is relatively arbitrary, and that frequentist statistics cannot provide evidence for the null hypothesis (absence of an effect) [98]. Therefore, we also report effect sizes, and in case of nonsignificant effects as indicated by p > 0.05, followed up with a Bayesian equivalent of the same test in order to quantify the strength of evidence for the null hypothesis (H0). By convention, Bayes factors from 1 to 3 were considered as anecdotal, 3 to 10 as substantial, and those above 10 as strong evidence in favor of H0.

Behavior

Trials in which participants made a discrimination response <200 ms after the stimulus presentation (3.39%) or gave no response at all (1.86%) were discarded from the analysis. To evaluate participants’ behavioral performance, we computed d-prime (d’, Type-I sensitivity), a measure of perceptual sensitivity to presented stimuli, separately for each confidence level and analyzed using a one-way repeated measures ANOVA with reported confidence ratings [14] as within-participant factor.

Decoding analyses

MVPA was applied to EEG data in order to decode patterns of neural activity specific to house and face stimuli. First, to test if we could decode category-specific neural representations in the localizer tasks, we used a 10-fold cross validation scheme. In this procedure, separately per localizer task, after randomizing the order of trials, the dataset was split into 10 equally sized folds. Equal number of house and face stimuli were present in each fold. We then trained a linear discriminant classifier (LDA) [27] to differentiate between house and face images using 9 folds and tested its performance on the remaining fold (using the standard Matlab function fitcdiscr). This was repeated 10 times, until all data were tested exactly once. Classification performance was evaluated for each participant separately by computing the area under the curve (AUC), which indicates the degree of separability between classes of the receiver operating characteristic (ROC) curve. First, we computed the proportion of correct classifications for each stimulus category, after which the scores were averaged across stimulus categories and over the 10 folds. Prior to training and testing procedures, EEG data was downsampled to 128 Hz and epochs were shortened to −100 to 800 ms, centered on the target stimulus, in order to decrease the computational time needed for MVPA.

Decoding analyses were carried out using EEG data recorded at all electrodes and a set of occipital–parietal electrodes [33]. The occipital–parietal set of electrodes, chosen so that it captures early visual “N170-like” response to houses and faces, consisted of Iz, Oz, O1, O2, POz, PO3, PO4, PO7, PO8, Pz, P1, P2, P3, P4, P5, P6, P7, P8, P9, and P10 electrodes.

To test our specific research questions, we used the cross-task validation scheme to evaluate the performance of each classifier (trained on either the sensory or the decision localizer task data) in differentiating between stimulus classes in the main decision task. We applied the same linear discriminant classifier to raw EEG data from the localizer task using voltages at each time sample to train the classifier, which was then applied to the main discrimination task data. This was done for separately for the sensory and decision localizer tasks. Following this procedure, we could examine whether and when house- and face-specific neural representations were associated with participants’ confidence in their perceptual decision, and whether perceptual decisions, in particular when incorrect, rely on the category-specific pattern of activation that is specific for the reported percept, or alternatively, representations are specific for the presented category. Moreover, critically, we could isolate sensory versus decision-related neural processes as the 2 localizers were differently tuned to these processes.

To examine the representational nature of decision errors, and to uncover how category-specific representations evolve across time, and across distinct processing stages, we used the GAT approach in applying the pattern classifiers [61]. Specifically, a classifier trained on a specific time point was tested on that time point as well as on all other time points. The resulting GAT matrix (training time x testing time) thus reveals dynamic changes of neural representations across time. For example, a classifier trained to decode between house and face images at 170 ms can generalize to a wider time-window, e.g., 170 to 350 ms, in which case it would indicate that the early neural representation was maintained in time. This approach is thus informative of how neural representations change across different stages of visual information processing, also permitting us to examine when in time neural representations differ between the perceived stimulus classes and depending on the correctness and confidence therein.

Decoding analyses were performed separately for correct and incorrect trials as a function of 2 confidence levels. Trials in which participants reported confidence ratings 1 and 2 were aggregated into “low” confidence trials, and those on which ratings 3 and 4 were given into “high” confidence trials. To illustrate, when decoding was performed for incorrect high and low confidence trials, classifiers were trained on the presented image category of the localizer task but were tested using trials in which participants reported seeing incorrect stimulus category, followed by high versus low confidence rating. Besides the stimulus category, we also decoded its orientation, i.e., whether a stimulus was tilted to the right or to the left, separately for correct and incorrect trials, as a function of confidence level. This analysis served 2 purposes. First, we wanted to verify that stimulus processing from the very bottom-up input did not differ between high versus low confidence trials. Low-level visual features such as orientation should be encoded in early brain activity in a bottom-up manner [50], but were not expected to be modulated by confidence due to the early timing of that processing stage. Second, the stimulus orientation was a task-irrelevant feature that participants did not need to do anything with and was therefore useful to tests whether task-irrelevant features are processed differently on high versus low confidence trials. To this end, classifiers were trained to distinguish between right versus left tilt of the stimulus using localizer data and then tested on right versus left images orientations using trials in which the correct versus incorrect stimulus category was reported.

Statistical analyses were performed on average classification scores (AUC) in time-windows that we preselected based on previous empirical work. Typically, studies report 2 processing stages using the GAT approach: an early (<250 to 300 ms) cluster of diagonal decoding reflecting initial sensory processes and a late processing stage (>300 ms), which was found to associate with conscious report [33,34,37]. Additionally, some studies have reported a third stage of visual information processing, starting at early latencies and generalizing off-diagonal, presumably reflecting maintenance of early sensory stimulus representations [33,34,53]. Following this body of work, and the study by Weaver and colleagues [33], who used an almost identical task design as in the present study, we focused our statistical analyses on 2 decoding clusters on-diagonal (classifiers were trained and tested on the same time point) and 1 decoding cluster off-diagonal (training and testing were done on different time points). The 2 on-diagonal clusters were 150 to 200 ms and 350 to 550 ms. The off-diagonal time-window spanned from 150 to 200 ms training time to 350 to 550 ms testing time (see below for more details). Finally, although decoding was pronounced during 3 time-windows observed previously by Weaver and colleagues [33], it was also observed at an even earlier time-window (80 to 130 ms) with a prominent occipital–parietal topography (Fig 2A, inset 1, early-stage topography). Because this early effect was absent in the same task when images where presented vertically [33], this early peak likely reflects differences in the orientations of the images presented, which we varied systematically across trials. This early window was also used as an event of interest in this study.

To inspect the pattern of neural activity that drove classification performance, we computed topographic maps for each classifier set (early, mid, late). Weights resulting from backward decoding models are not interpretable as neural sources [99]. For that reason, we plotted topographic maps resulting from multiplying the data covariance matrix with the classifier weights, yielding activity patterns that are interpretable as neural sources underlying decoding results [95]. These covariance/class separability maps were then normalized across electrodes for each participant (mean activity over electrodes was thus zero).

In order to statistically evaluate classifiers’ performance, we extracted diagonal and off-diagonal traces of GAT matrices. Average AUC values within specified clusters were analyzed using repeated measures ANOVA with factors decoding latency (early, late) and confidence (low, high). Specific hypotheses-driven comparisons between conditions were additionally evaluated using paired-sample t tests on AUC values averaged within our TOI. To evaluate decoding off-diagonal, specifically, when the training was done using 150 to 200 ms classifiers that were then applied to the late 350 to 500 ms time-window, we compared average decoding in this time-window between 2 confidence levels (low versus high) using a paired-sample t tests as well. In cases where a specific hypotheses-driven comparison did not indicate a significant result as indicated by an alpha of p < 0.05, we followed up that null-effect by a Bayesian equivalent of the same test in order to quantify the strength of evidence for the null hypothesis (H0). By convention, Bayes factors from 1 to 3 were considered as anecdotal, 3 to 10 as substantial, and those above 10 as strong evidence in favor of H0. Besides our hypothesis-driven analyses based on our predefined selection of time-windows, we additionally statistically compared each time point to chance (both for on- and off-diagonal decoding) by applying group-level permutation testing with cluster correction for multiple comparisons [100]. We employed two-tailed cluster-permutation testing with an alpha of p < 0.05 and cluster alpha p < 0.05 (N permutations = 1,000, as implemeted in ADAM toolbox). These latter analyses were mainly performed to evaluate whether our hypothesis-driven analyses may have missed potentially relevant effects, and if so, we recommend future studies to replicate these effects to evaluate the robustness of them [98].

Supporting information

S1 Text. Additional results.

BF, Bayes factor; HC, high confidence; LC, low confidence.

(DOCX)

S1 Fig. Trial sequence of the sensory and decision localizer task.

(A) Each trial of the sensory localizer task started with a central red fixation dot for 1,200–1,400 ms, during which an infrequent contrast change of the dot could happen (20% of trials). Participants needed to press the spacebar as soon as they noticed the contrast change. In the same time interval, a house or a face image was briefly shown on the screen, which needed to be ignored. Images were either tilted to the right or to the left (task-irrelevant feature) at a 5° or 355° angle. Note that in the example trial, only a left-tilted house image is shown. (B) Each trial of the decision localizer task started with a central red fixation dot after which a forward mask was shown, followed by an image of a face or a house and a backward mask. Images were either tilted to the right or to the left (task-irrelevant feature) at a 5° or 355° angle. On every trial, participants reported whether they perceived a house or a face and indicated their confidence in this decision.

(TIF)

S2 Fig

(A) GAT matrix for the decision localizer and regions of interest marked by inset black boxes (numbers 1–4). (B) Classifiers trained on the decision localizer applied to the main discrimination task (cross-task validation procedure). (C) GAT matrix for the sensory localizer. (D) Classifiers trained on the sensory localizer applied to the main discrimination task (cross-task validation procedure. (E) Classifiers trained on the sensory localizer applied to the decision localizer (cross-task validation procedure). (F) Classifiers trained on the decision localizer applied to the sensory localizer (cross-task validation procedure). All GAT matrices are based on occipital–parietal electrodes. The underlying data and scripts supporting this figure can be found on FigShare (https://doi.org/10.21942/uva.c.6265233.v1).

(TIF)

S3 Fig. Filtering effects on the timing of ERP components.

ERPs are time-locked to stimulus onset in the passive localizer with 3 different high-pass filtering settings. The region of interest plotted consists of the following electrodes: O1, O2, PO3, PO4, PO7, PO8. The underlying data and scripts supporting this figure can be found on FigShare (https://doi.org/10.21942/uva.c.6265233.v1).

(TIF)

S4 Fig. Face/house classification based on the decision localizer.

(A) On-diagonal time courses of correct and incorrect decisions, separated by confidence (LC, low confidence; HC, high confidence). Black boxes indicate the time-windows (early, mid, late) used for averaging the AUC scores. (B) Perceptual maintenance is derived by training a classifier on the time-window 150–200 ms and testing it across the entire time-window. In (A) and (B), colored horizontal lines indicate periods of significant decoding with respect to chance. Classification performance was evaluated at each time point using cluster-based permutation testing (two-tailed cluster-permutation, alpha p < 0.05, cluster alpha p < 0.05, N permutations = 1,000). (C) Bar plots showing average AUC values for the time-windows of interest highlighted in panels (A) and (B). The underlying data and scripts supporting this figure can be found on FigShare (https://doi.org/10.21942/uva.c.6265233.v1).

(TIF)

S5 Fig. Face/house classification based on the sensory localizer.

(A) On-diagonal time courses of correct and incorrect decisions, separated by confidence (LC, low confidence; HC, high confidence). Black boxes indicate the time-windows (early, mid, late) used for averaging the AUC scores. (B) Perceptual maintenance is derived by training a classifier on the time-window 150–200 ms and testing it across the entire time-window. In (A) and (B), colored horizontal lines indicate periods of significant decoding with respect to chance. Classification performance was evaluated at each time point using cluster-based permutation testing (two-tailed cluster-permutation, alpha p < 0.05, cluster alpha p < 0.05, N permutations = 1,000). (C) Bar plots showing average AUC values for the time-windows of interest highlighted in panels (A) and (B). The underlying data and scripts supporting this figure can be found on Fig Share (https://doi.org/10.21942/uva.c.6265233.v1).

(TIF)

Abbreviations

AUC

area under the curve

EOG

electrooculogram

FFA

fusiform face area

fMRI

functional magnetic resonance imaging

GAT

generalization across time

HC

high confidence

HEOG

horizontal EOG

LC

low confidence

LDA

linear discriminant analysis

MVPA

multivariate pattern analysis

OFA

occipital face area

ROC

receiver operating characteristic

SDT

signal detection theory

TOI

time-windows of interest

VEOG

vertical EOG

Data Availability

The data and analysis scripts used in this article is available on Figshare: https://doi.org/10.21942/uva.c.6265233.v1.

Funding Statement

This research was supported by a grant from the H2020 European Research Council (ERC STG 715605 to SVG). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1.Collerton D, Perry E, McKeith I. Why people see things that are not there: A novel Perception and Attention Deficit model for recurrent complex visual hallucinations. Behav Brain Sci. 2005;28:737–757. doi: 10.1017/S0140525X05000130 [DOI] [PubMed] [Google Scholar]
  • 2.Grossberg S. How hallucinations may arise from brain mechanisms of learning, attention, and volition. J Int Neuropsychol Soc. 2000;6:583–592. doi: 10.1017/s135561770065508x [DOI] [PubMed] [Google Scholar]
  • 3.Parks TE. Illusory figures: A (mostly) atheoretical review. Psychol Bull. 1984;95:282–300. [PubMed] [Google Scholar]
  • 4.de Lange FP, Heilbron M, Kok P. How Do Expectations Shape Perception? Trends Cogn Sci. 2018;22:764–779. doi: 10.1016/j.tics.2018.06.002 [DOI] [PubMed] [Google Scholar]
  • 5.Bar M. Visual objects in context. Nat Rev Neurosci. 2004;5:617–629. doi: 10.1038/nrn1476 [DOI] [PubMed] [Google Scholar]
  • 6.Panichello MF, Cheung OS, Bar M. Predictive feedback and conscious visual experience. Front Psychol. 2013;3:1–8. doi: 10.3389/fpsyg.2012.00620 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Summerfield C, Egner T, Mangels J, Hirsch J. Mistaking a house for a face: Neural correlates of misperception in healthy humans. Cereb Cortex. 2006;16:500–508. doi: 10.1093/cercor/bhi129 [DOI] [PubMed] [Google Scholar]
  • 8.Tu T, Schneck N, Muraskin J, Sajda P. Network configurations in the human brain reflect choice bias during rapid face processing. J Neurosci. 2017;37:12226–12237. doi: 10.1523/JNEUROSCI.1677-17.2017 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Linares D, Aguilar-Lleyda D, López-Moliner J. Decoupling sensory from decisional choice biases in perceptual decision making. Elife. 2019;8:1–16. doi: 10.7554/eLife.43994 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Peters MAK, Ro T, Lau H , Who’s afraid of response bias? Neurosci. Conscious. 2016, niw001 (2016). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Balsdon T, Mamassian P, Wyart V. Separable neural signatures of confidence during perceptual decisions. Elife. 2021;10:1–39. doi: 10.7554/eLife.68491 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Fritsche M, Mostert P, de Lange FP. Opposite Effects of Recent History on Perception and Decision. Curr Biol. 2017;27:590–595. doi: 10.1016/j.cub.2017.01.006 [DOI] [PubMed] [Google Scholar]
  • 13.Akaishi R, Umeda K, Nagase A, Sakai K. Autonomous Mechanism of Internal Choice Estimate Underlies Decision Inertia. Neuron. 2014;81:195–206. doi: 10.1016/j.neuron.2013.10.018 [DOI] [PubMed] [Google Scholar]
  • 14.Fischer J, Whitney D. Serial dependence in visual perception. Nat Neurosci. 2014;17:738–743. doi: 10.1038/nn.3689 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Fleming S, Dolan R. The neural basis of metacognitive ability. Philos Trans R Soc Lond B Biol Sci. 2012;1338–1349. doi: 10.1098/rstb.2011.0417 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Kunimoto C, Miller J, Pashler H. Confidence and accuracy of near-threshold discrimination responses. Conscious Cogn. 2001;10:294–340. doi: 10.1006/ccog.2000.0494 [DOI] [PubMed] [Google Scholar]
  • 17.Salge JH, Pollmann S, Reeder RR. Anomalous visual experience is linked to perceptual uncertainty and visual imagery vividness. Psychol Res. 2020. doi: 10.1007/s00426-020-01364-7 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Philiastides MG, Ratcliff R, Sajda P. Neural representation of task difficulty and decision making during perceptual categorization: a timing diagram. J Neurosci. 2006;26:8965–8975. doi: 10.1523/JNEUROSCI.1655-06.2006 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Ratcliff R, Philiastides MG, Sajda P. Quality of evidence for perceptual decision making is indexed by trial-to-trial variability of the EEG. Proc Natl Acad Sci U S A. 2009;106:6539–6544. doi: 10.1073/pnas.0812589106 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Mostert P, Kok P, De Lange FP. Dissociating sensory from decision processes in human perceptual decision making. Sci Rep. 2015;5:1–13. doi: 10.1038/srep18253 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Hatamimajoumerd E, Ratan Murty NA, Pitts M, Cohen MA. Decoding perceptual awareness across the brain with a no-report fMRI masking paradigm. Curr Biol. 2022;1–11. doi: 10.1016/j.cub.2022.07.068 [DOI] [PubMed] [Google Scholar]
  • 22.Hesse JK, Tsao DY. A new no-report paradigm reveals that face cells encode both consciously perceived and suppressed stimuli. Elife. 2020;9:1–20. doi: 10.7554/eLife.58360 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Canales-Johnson A, Beerendonk L, Chennu S, Davidson MJ, Ince RAA, van Gaal S. Feedback information sharing in the brain distinguishes perception from its consequences during perceptual bistability. Biorxiv. 2021;1–21. [Google Scholar]
  • 24.Frassle S, Sommer J. Binocular Rivalry: Frontal Activity Relates to Introspection and Action But Not to Perception. J Neurosci. 2014;34:1738–1747. doi: 10.1523/JNEUROSCI.4403-13.2014 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Cohen MA, Ortego K, Kyroudis A, Pitts M. Distinguishing the Neural Correlates of Perceptual Awareness and Postperceptual Processing. J Neurosci. 2020. doi: 10.1523/JNEUROSCI.0120-20.2020 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Pitts MA, Padwal J, Fennelly D, Martínez A, Hillyard SA. Gamma band activity and the P3 reflect post-perceptual processes, not visual awareness. Neuroimage. 2014;101:337–350. doi: 10.1016/j.neuroimage.2014.07.024 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Parra LC, Spence CD, Gerson AD, Sajda P. Recipes for the linear analysis of EEG. Neuroimage. 2005;28:326–341. doi: 10.1016/j.neuroimage.2005.05.032 [DOI] [PubMed] [Google Scholar]
  • 28.Dehaene S, Changeux JP, Naccache L, Sackur J, Sergent C. Conscious, preconscious, and subliminal processing: a testable taxonomy. Trends Cogn Sci. 2006;10:204–211. doi: 10.1016/j.tics.2006.03.007 [DOI] [PubMed] [Google Scholar]
  • 29.Lamme VAF. Why visual attention and awareness are different. Trends Cogn Sci. 2003;7:12–18. doi: 10.1016/s1364-6613(02)00013-x [DOI] [PubMed] [Google Scholar]
  • 30.Tsuchiya N, Wilke M, Frässle S, Lamme VAF. No-Report Paradigms: Extracting the True Neural Correlates of Consciousness. Trends Cogn Sci. 2015;19:757–770. doi: 10.1016/j.tics.2015.10.002 [DOI] [PubMed] [Google Scholar]
  • 31.Block N. What Is Wrong with the No-Report Paradigm and How to Fix It. Trends Cogn Sci. 2019;23:1003–1013. doi: 10.1016/j.tics.2019.10.001 [DOI] [PubMed] [Google Scholar]
  • 32.Kapoor V, Dwarakanath A, Safavi S, Werner J, Besserve M, Panagiotaropoulos TI, et al. Decoding internally generated transitions of conscious contents in the prefrontal cortex without subjective reports. Nat Commun. 2022:13. doi: 10.1038/s41467-022-28897-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Weaver MD, Fahrenfort JJ, Belopolsky A, Van Gaal S. Independent neural activity patterns for sensory-and confidence-based information maintenance during category-selective visual processing. eNeuro. 2019;6. doi: 10.1523/ENEURO.0268-18.2018 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Marti S, Dehaene S. Discrete and continuous mechanisms of temporal selection in rapid visual streams. Nat Commun. 2017;8. doi: 10.1038/s41467-017-02079-x [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35.Cichy RM, Pantazis D, Oliva A. Resolving human object recognition in space and time. Nat Neurosci. 2014;17:455–462. doi: 10.1038/nn.3635 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36.Carlson T, Tovar DA, Alink A, Kriegeskorte N. Representational dynamics of object vision: The first 1000 ms. J Vis. 2013;13:1–19. doi: 10.1167/13.10.1 [DOI] [PubMed] [Google Scholar]
  • 37.Kaiser D, Oosterhof NN, Peelen MV. The neural dynamics of attentional selection in natural scenes. J Neurosci. 2016;36:10522–10528. doi: 10.1523/JNEUROSCI.1385-16.2016 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 38.Bentin S, Allison T, Puce A, Perez E, McCarthy G. Electrophysiological studies of face perception in humans. J Cogn Neurosci. 1996;8:551–565. doi: 10.1162/jocn.1996.8.6.551 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39.Rossion B, Jacques C. The N170: understanding the time-course of face perception in the human brain. In The Oxford Handbook of ERP Components. 2011. p. 115–142. [Google Scholar]
  • 40.Harris JA, Te Wu C, Woldorff MG. Sandwich masking eliminates both visual awareness of faces and face-specific brain activity through a feedforward mechanism. J Vis. 2011;11:1–12. doi: 10.1167/11.7.3 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41.Nan W, Liu Y, Zeng X, Yang W, Linag J, Lan Y, et al. The spatiotemporal characteristics of N170s for faces and words: A meta-analysis study. Psych J. 2021. [DOI] [PubMed] [Google Scholar]
  • 42.Eimer M. The Face-Sensitive N170 Component of the Event-Related Brain Potential. Oxford Handb Face Percept. 2012;329–344. [Google Scholar]
  • 43.Linkenkaer-Hansen K, Palva JM, Sams M, Hietanen JK, Aronen HJ, Ilmoniemi RJ. Face-selective processing in human extrastriate cortex around 120 ms after stimulus onset revealed by magneto- and electroencephalography. Neurosci Lett. 1998;253:147–150. doi: 10.1016/s0304-3940(98)00586-2 [DOI] [PubMed] [Google Scholar]
  • 44.Halgren E, Raij T, Marinkovic K, Jousmäki V, Hari R. Cognitive response profile of the human fusiform face area as determined by MEG. Cereb Cortex. 2000;10:69–81. doi: 10.1093/cercor/10.1.69 [DOI] [PubMed] [Google Scholar]
  • 45.Haxby JV, Hoffman EA, Gobbini MI. The distributed human neural system for face perception. Trends Cogn Sci. 2000;4:223–233. doi: 10.1016/s1364-6613(00)01482-0 [DOI] [PubMed] [Google Scholar]
  • 46.Itier RJ, Taylor MJ. N170 or N1? Spatiotemporal Differences between Object and Face Processing Using ERPs. Cereb Cortex. 2004;14:132–142. doi: 10.1093/cercor/bhg111 [DOI] [PubMed] [Google Scholar]
  • 47.Deffke I, Sander T, Heidenreich J, Sommer W, Curio G, Trahms L, et al. MEG/EEG sources of the 170-ms response to faces are co-localized in the fusiform gyrus. Neuroimage. 2007;35:1495–1501. doi: 10.1016/j.neuroimage.2007.01.034 [DOI] [PubMed] [Google Scholar]
  • 48.Rossion B, Jacques C. The N170: Understanding the Time Course of Face Perception in the Human Brain. Oxford Handb Event-Related Potential Components. 2011;115–142. [Google Scholar]
  • 49.Philiastides MG, Sajda P. EEG-informed fMRI reveals spatiotemporal characteristics of perceptual decision making. J Neurosci. 2007;27:13082–13091. doi: 10.1523/JNEUROSCI.3540-07.2007 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 50.King J-R, Pescetelli N, Dehaene S. Brain Mechanisms Underlying the Brief Maintenance of Seen and Unseen Sensory Information. Neuron. 2016;1122–1134. doi: 10.1016/j.neuron.2016.10.051 [DOI] [PubMed] [Google Scholar]
  • 51.Gwilliams L, King JR. Recurrent processes support a cascade of hierarchical decisions. Elife. 2020;9:1–20. doi: 10.7554/eLife.56603 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 52.Philiastides MG, Heekeren HR, Sajda P. Human scalp potentials reflect a mixture of decision- related signals during perceptual choices. J Neurosci. 2014;34:16877–16889. doi: 10.1523/JNEUROSCI.3012-14.2014 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 53.Meijs EL, Mostert P, Slagter HA, De Lange FP, Van Gaal S. Exploring the role of expectations and stimulus relevance on stimulus-specific neural representations and conscious report. Neurosci Conscious. 2019. doi: 10.1093/nc/niz011 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 54.Logothetis NK, Schall JD. Neuronal Correlates of Subjective Visual Perception. Science. 1989; 245:761–763. [DOI] [PubMed] [Google Scholar]
  • 55.Supèr H, Spekreijse H. Two distinct modes of sensory processing observed in monkey primary visual cortex (V1). Nat Neurosci. 2001;4:304–310. doi: 10.1038/85170 [DOI] [PubMed] [Google Scholar]
  • 56.Sadeh B, Podlipsky I, Zhdanov A, Yovel G. Event-related potential and functional MRI measures of face-selectivity are highly correlated: A simultaneous ERP-fMRI investigation. Hum Brain Mapp. 2010;31:1490–1501. doi: 10.1002/hbm.20952 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 57.Wardle SG, Taubert J, Teichmann L, Baker CI. Rapid and dynamic processing of face pareidolia in the human brain. Nat Commun. 2020;11:1–14. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 58.Jacques C, Jonas J, Maillard L, Colnat-Coulbois S, Koessler L, Rossion B. The inferior occipital gyrus is a major cortical source of the face-evoked N170: Evidence from simultaneous scalp and intracerebral human recordings. Hum Brain Mapp. 2019;40:1403–1418. doi: 10.1002/hbm.24455 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 59.Carlson TA, Hogendoorn H, Kanai R, Mesik J, Turret J. High temporal resolution decoding of object position and category. J Vis. 2011;11:1–17. doi: 10.1167/11.10.9 [DOI] [PubMed] [Google Scholar]
  • 60.Fahrenfort JJ, Van Leeuwen J, Olivers CNL, Hogendoorn H. Perceptual integration without conscious access. Proc Natl Acad Sci U S A. 2017;114:3744–3749. doi: 10.1073/pnas.1617268114 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 61.King JR, Dehaene S. Characterizing the dynamics of mental representations: The temporal generalization method. Trends Cogn Sci. 2014;18:203–210. doi: 10.1016/j.tics.2014.01.002 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 62.Rochet N, Spieser L, Casini L, Hasbroucq T, Burle B. Detecting and correcting partial errors: Evidence for efficient control without conscious access. Cogn Affect Behav Neurosci. 2013. doi: 10.3758/s13415-013-0232-0 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 63.Rotshtein P, Henson RNA, Treves A, Driver J, Dolan RJ. Morphing Marilyn into Maggie dissociates physical and identity face representations in the brain. Nat Neurosci. 2005;8:107–113. doi: 10.1038/nn1370 [DOI] [PubMed] [Google Scholar]
  • 64.Stein T, Kaiser D, Fahrenfort JJ, Van Gaal S. The human visual system differentially represents subjectively and objectively invisible stimuli. PLoS Biol. 2021;19:1–24. doi: 10.1371/journal.pbio.3001241 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 65.Wimmer K, Compte A, Roxin A, Peixoto D, Renart A, De La Rocha J. Sensory integration dynamics in a hierarchical network explains choice probabilities in cortical area MT. Nat Commun. 2015:6. doi: 10.1038/ncomms7177 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 66.Haefner RM, Berkes P, Fiser J. Perceptual Decision-Making as Probabilistic Inference by Neural Sampling. Neuron. 2016;90:649–660. doi: 10.1016/j.neuron.2016.03.020 [DOI] [PubMed] [Google Scholar]
  • 67.Lamme VAF, Roelfsema PR. The distinct modes of vision offered by feedforward and recurrent processing. Trends Neurosci. 2000;23:571–579. doi: 10.1016/s0166-2236(00)01657-x [DOI] [PubMed] [Google Scholar]
  • 68.Hupé JM, James AC, Girard P, Lomber SG, Payne BR, Bullier J. Feedback connections act on the early part of the responses monkey visual cortex. J Neurophysiol. 2001;85:134–145. doi: 10.1152/jn.2001.85.1.134 [DOI] [PubMed] [Google Scholar]
  • 69.Bondy AG, Haefner RM, Cumming BG. Feedback determines the structure of correlated variability in primary visual cortex. Nat Neurosci. 2018;21:598–606. doi: 10.1038/s41593-018-0089-1 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 70.Nienborg H, Cumming BG. Decision-related activity in sensory neurons may depend on the columnar architecture of cerebral cortex. J Neurosci. 2014;34:3579–3585. doi: 10.1523/JNEUROSCI.2340-13.2014 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 71.Supèr H, Spekreijse H, Lamme VAF. Two distinct modes of sensory processing observed in monkey primary visual cortex (V1). Nat Neurosci. 2001;4:1–7. [DOI] [PubMed] [Google Scholar]
  • 72.Wilming N, Murphy PR, Meyniel F, Donner TH. Large-scale dynamics of perceptual decision information across human cortex. Nat Commun. 2020;11:1–14. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 73.Nienborg H, Cohen MR, Cumming BG. Decision-related activity in sensory neurons: Correlations among neurons and with behavior. Annu Rev Neurosci. 2012;35:463–483. doi: 10.1146/annurev-neuro-062111-150403 [DOI] [PubMed] [Google Scholar]
  • 74.Foxe JJ, Simpson GV. Flow of activation from V1 to frontal cortex in humans: A framework for defining “early” visual processing. Exp Brain Res. 2002;142:139–150. [DOI] [PubMed] [Google Scholar]
  • 75.Ress D, Heeger DJ. Neuronal correlates of perception in early visual cortex. Nat Neurosci. 2003;6:414–420. doi: 10.1038/nn1024 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 76.Summerfield C, de Lange FP. Expectation in perceptual decision making: neural and computational mechanisms. Nat Rev Neurosci. 2014;1–12. [DOI] [PubMed] [Google Scholar]
  • 77.Blom T, Feuerriegel D, Johnson P, Bode S, Hogendoorn H. Predictions drive neural representations of visual events ahead of incoming sensory information. Proc Natl Acad Sci U S A. 2020;117:7510–7515. doi: 10.1073/pnas.1917777117 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 78.Ekman M, Kok P, De Lange FP. Time-compressed preplay of anticipated events in human primary visual cortex. Nat Commun. 2017;8:1–9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 79.Kok P, Mostert P, De Lange FP. Prior expectations induce prestimulus sensory templates. Proc Natl Acad Sci U S A. 2017;114:10473–10478. doi: 10.1073/pnas.1705652114 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 80.Andrews TJ, Schluppeck D. Neural responses to Mooney images reveal a modular representation of faces in human visual cortex. Neuroimage. 2004;21:91–98. doi: 10.1016/j.neuroimage.2003.08.023 [DOI] [PubMed] [Google Scholar]
  • 81.Cox D, Meyers E, Sinha P. Contextually Evoked Object-Specific Responses in Human Visual Cortex. Science (80-.). 2004;304:115–117. doi: 10.1126/science.1093110 [DOI] [PubMed] [Google Scholar]
  • 82.Summerfield C, Egner T, Greene M, Koechlin E, Mangels J, Hirsch J. Predictive codes for forthcoming perception in the frontal cortex. Science. 2006;314:1311–1314. doi: 10.1126/science.1132028 [DOI] [PubMed] [Google Scholar]
  • 83.Diaz JA, Queirazza F, Philiastides MG. Perceptual learning alters post-sensory processing in human decision-making. Nat Hum Behav. 2017;1:1–9. [Google Scholar]
  • 84.Cicchini GM, Mikellidou K, Burr D. Serial dependencies act directly on perception. J Vis. 2017;17:1–9. doi: 10.1167/17.14.6 [DOI] [PubMed] [Google Scholar]
  • 85.Podvalny E, Flounders MW, King LE, Holroyd T, He BJ. A dual role of prestimulus spontaneous neural activity in visual object recognition. Nat Commun. 2019;10:1–13. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 86.Berkes P, Orbán G, Lengyel M, Fiser J. Enhanced actin depolymerization at the mDia1- bound barbed end. This inhibition occurs in the submillimolar range of P. Science (80-.). 2011;331:83–88. [Google Scholar]
  • 87.Han F, Caporale N, Dan Y. Reverberation of Recent Visual Experience in Spontaneous Cortical Waves. Neuron. 2008;60:321–327. doi: 10.1016/j.neuron.2008.08.026 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 88.Tsodyks M, Kenet T, Grinvald A, Arieli A. Linking spontaneous activity of single cortical neurons and the underlying functional architecture. Science (80-.). 1999;286:1943–1946. doi: 10.1126/science.286.5446.1943 [DOI] [PubMed] [Google Scholar]
  • 89.Kenet T, Bibitchkov D, Tsodyks M, Grinvald A, Arieli A. Spontaneously emerging cortical representations of visual attributes. Nature. 2003;425:954–956. doi: 10.1038/nature02078 [DOI] [PubMed] [Google Scholar]
  • 90.Luczak A, Barthó P, Harris KD. Spontaneous Events Outline the Realm of Possible Sensory Responses in Neocortical Populations. Neuron. 2009;62:413–425. doi: 10.1016/j.neuron.2009.03.014 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 91.Ringach DL. Spontaneous and driven cortical activity: implications for computation. Curr Opin Neurobiol. 2009;19:439–444. doi: 10.1016/j.conb.2009.07.005 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 92.Fiser J, Chiu C, Weliky M. Small modulation of ongoing cortical dynamics by sensory input during natural vision. Nature. 2004;431:573–578. doi: 10.1038/nature02907 [DOI] [PubMed] [Google Scholar]
  • 93.Gilbert CD, Sigman M. Brain States: Top-Down Influences in Sensory Processing. Neuron. 2007;54:677–696. doi: 10.1016/j.neuron.2007.05.019 [DOI] [PubMed] [Google Scholar]
  • 94.Mostert P, Albers AM, Brinkman L, Todorova L, Kok P, de Lange FP. Eye Movement-Related Confounds in Neural Decoding of Visual Working Memory Representations. eNeuro. 2018; 10;5(4):ENEURO.0401-17.2018. doi: 10.1523/ENEURO.0401-17.2018 ; PMCID: PMC6179574. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 95.Fahrenfort JJ, van Driel J, van Gaal S, Olivers CNL. From ERPs to MVPA using the Amsterdam Decoding and Modeling toolbox (ADAM). Front Neurosci. 2018;12. doi: 10.3389/fnins.2018.00368 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 96.van Driel J, Olivers CNL, Fahrenfort JJ. High-pass filtering artifacts in multivariate classification of neural time series data. J Neurosci Methods. 2021;352:109080. doi: 10.1016/j.jneumeth.2021.109080 [DOI] [PubMed] [Google Scholar]
  • 97.Acunzo DJ, MacKenzie G, van Rossum MCW. Systematic biases in early ERP and ERF components as a result of high-pass filtering. J Neurosci Methods. 2012;209:212–218. doi: 10.1016/j.jneumeth.2012.06.011 [DOI] [PubMed] [Google Scholar]
  • 98.Lakens D, Adolfi FG, Albers CJ, Anvari F, Apps MAJ, Argamon SE, et al. Justify your alpha. Nat Hum Behav. 2018;2:168–171. doi: 10.1038/s41562-018-0311-x [DOI] [Google Scholar]
  • 99.Haufe S, Meinecke F, Görgen K, Dähne S, Haynes J, Blankertz B, et al. NeuroImage On the interpretation of weight vectors of linear models in multivariate neuroimaging. Neuroimage. 2014;87:96–110. doi: 10.1016/j.neuroimage.2013.10.067 [DOI] [PubMed] [Google Scholar]
  • 100.Maris E, Oostenveld R. Nonparametric statistical testing of EEG- and MEG-data. J Neurosci Methods. 2007;164:177–190. doi: 10.1016/j.jneumeth.2007.03.024 [DOI] [PubMed] [Google Scholar]

Decision Letter 0

Gabriel Gasque

13 Jan 2022

Dear Simon,

Thank you for submitting your manuscript entitled "The perceptual nature of illusory object recognition" for consideration as a Research Article by PLOS Biology.

Your manuscript has now been evaluated by the PLOS Biology editorial staff, as well as by an academic editor with relevant expertise, and I am writing to let you know that we would like to send your submission out for external peer review. Please accept my apologies for the delay in sending this decision to you.

Before we can send your manuscript to reviewers, we need you to complete your submission by providing the metadata that is required for full assessment. To this end, please login to Editorial Manager where you will find the paper in the 'Submissions Needing Revisions' folder on your homepage. Please click 'Revise Submission' from the Action Links and complete all additional questions in the submission questionnaire.

Once your full submission is complete, your paper will undergo a series of checks in preparation for peer review. Once your manuscript has passed the checks it will be sent out for review. To provide the metadata for your submission, please Login to Editorial Manager (https://www.editorialmanager.com/pbiology) within two working days, i.e. by Jan 17 2022 11:59PM.

If your manuscript has been previously reviewed at another journal, PLOS Biology is willing to work with those reviews in order to avoid re-starting the process. Submission of the previous reviews is entirely optional and our ability to use them effectively will depend on the willingness of the previous journal to confirm the content of the reports and share the reviewer identities. Please note that we reserve the right to invite additional reviewers if we consider that additional/independent reviewers are needed, although we aim to avoid this as far as possible. In our experience, working with previous reviews does save time.

If you would like to send previous reviewer reports to us, please email me at ggasque@plos.org to let me know, including the name of the previous journal and the manuscript ID the study was given, as well as attaching a point-by-point response to reviewers that details how you have or plan to address the reviewers' concerns.

During the process of completing your manuscript submission, you will be invited to opt-in to posting your pre-review manuscript as a bioRxiv preprint. Visit http://journals.plos.org/plosbiology/s/preprints for full details. If you consent to posting your current manuscript as a preprint, please upload a single Preprint PDF.

Given the disruptions resulting from the ongoing COVID-19 pandemic, please expect some delays in the editorial process. We apologise in advance for any inconvenience caused and will do our best to minimize impact as far as possible.

Feel free to email us at plosbiology@plos.org if you have any queries relating to your submission.

Kind regards,

Gabriel

Gabriel Gasque

Senior Editor

PLOS Biology

ggasque@plos.org

Decision Letter 1

Gabriel Gasque

28 Feb 2022

Dear Dr van Gaal,

Thank you for submitting your manuscript entitled "The perceptual nature of illusory object recognition" for consideration as a Research Article at PLOS Biology. Thank you also for your patience as we completed our editorial process, and please accept my apologies for the delay in providing you with our decision. Your manuscript has been evaluated by the PLOS Biology editors, an Academic Editor with relevant expertise, and by three independent reviewers.

As you will see, the reviewers find the conclusions of your manuscript interesting, but they also raise several concerns that will need to be addressed in order to consider the manuscript further for publication. Both Reviewers 1 and 3 ask for several clarifications and offer some suggestions to improve the work, including the rewriting of some parts to have a better context in relation with previous literature. Reviewer 2 is more critical and raises some issues regarding the rationale of the study and the novelty of some of the findings, and also thinks that the link between confidence judgments and perceptual decisions should be better explored.

In light of the reviews (attached below), we will not be able to accept the current version of the manuscript, but we would welcome re-submission of a revised version that takes into account the reviewers' comments. We cannot make any decision about publication until we have seen the revised manuscript and your response to the reviewers' comments. Your revised manuscript is also likely to be sent for further evaluation by the reviewers.

We expect to receive your revised manuscript within 3 months.

Please email us (plosbiology@plos.org) if you have any questions or concerns, or would like to request an extension. At this stage, your manuscript remains formally under active consideration at our journal; please notify us by email if you do not intend to submit a revision so that we may end consideration of the manuscript at PLOS Biology.

**IMPORTANT - SUBMITTING YOUR REVISION**

Your revisions should address the specific points made by each reviewer. Please submit the following files along with your revised manuscript:

1. A 'Response to Reviewers' file - this should detail your responses to the editorial requests, present a point-by-point response to all of the reviewers' comments, and indicate the changes made to the manuscript.

*NOTE: In your point by point response to the reviewers, please provide the full context of each review. Do not selectively quote paragraphs or sentences to reply to. The entire set of reviewer comments should be present in full and each specific point should be responded to individually, point by point.

You should also cite any additional relevant literature that has been published since the original submission and mention any additional citations in your response.

2. In addition to a clean copy of the manuscript, please also upload a 'track-changes' version of your manuscript that specifies the edits made. This should be uploaded as a "Related" file type.

3. Resubmission Checklist

When you are ready to resubmit your revised manuscript, please refer to this resubmission checklist: https://plos.io/Biology_Checklist

To submit a revised version of your manuscript, please go to https://www.editorialmanager.com/pbiology/ and log in as an Author. Click the link labelled 'Submissions Needing Revision' where you will find your submission record.

Please make sure to read the following important policies and guidelines while preparing your revision and fulfil the editorial requests:

a) *PLOS Data Policy*

Please note that as a condition of publication PLOS' data policy (http://journals.plos.org/plosbiology/s/data-availability) requires that you make available all data used to draw the conclusions arrived at in your manuscript. If you have not already done so, you must include any data used in your manuscript either in appropriate repositories, within the body of the manuscript, or as supporting information (N.B. this includes any numerical values that were used to generate graphs, histograms etc.). Please also indicate in each figure legend where the data can be found. For an example see here: http://www.plosbiology.org/article/info%3Adoi%2F10.1371%2Fjournal.pbio.1001908#s5

b) *Published Peer Review*

Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out. Please see here for more details:

https://blogs.plos.org/plos/2019/05/plos-journals-now-open-for-published-peer-review/

c) *Blot and Gel Data Policy*

Please provide the original, uncropped and minimally adjusted images supporting all blot and gel results reported in an article's figures or Supporting Information files. We will require these files before a manuscript can be accepted so please prepare them now, if you have not already uploaded them. Please carefully read our guidelines for how to prepare and upload this data: https://journals.plos.org/plosbiology/s/figures#loc-blot-and-gel-reporting-requirements

d) *Blurb*

Please also provide a blurb which (if accepted) will be included in our weekly and monthly Electronic Table of Contents, sent out to readers of PLOS Biology, and may be used to promote your article in social media. The blurb should be about 30-40 words long and is subject to editorial changes. It should, without exaggeration, entice people to read your manuscript. It should not be redundant with the title and should not contain acronyms or abbreviations. For examples, view our author guidelines: https://journals.plos.org/plosbiology/s/revising-your-manuscript#loc-blurb

e) *Protocols deposition*

To enhance the reproducibility of your results, we recommend that if applicable you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosbiology/s/submission-guidelines#loc-materials-and-methods

Thank you again for your submission to our journal. We hope that our editorial process has been constructive thus far, and we welcome your feedback at any time. Please don't hesitate to contact us if you have any questions or comments.

Sincerely,

Ines

--

Ines Alvarez-Garcia, PhD

Senior Editor

PLOS Biology

ialvarez-garcia@plos.org

on behalf of

Gabriel Gasque, PhD

Senior Editor

PLOS Biology

ggasque@plos.org

--------------------------------------------------------------

Reviewers' comments

Rev. 1: Guillaume Rousselet - note that this reviewer has signed his review

This is a very good article, presenting an ambitious and clever experiment, with thoughtful analyses. I have a lot of relatively minor comments and suggestions to try to improve the presentation, including illustrations and reporting.

## Paradigm

To check that the localisers did differ as intended, it would be useful to illustrate, maybe in Figure 1, cross-classification results: classification of localiser 1 data using the classifier trained with localiser 2 data, and vice versa.

## Illustrations

Figures 1-4: please replace the bar graphs with scatterplots and remove the little star system. P values are not measures of effect sizes or the importance of the results. If the goal is to emphasise effect sizes, it would be more informative to plot distributions of pairwise differences. This is essential to let readers assess how many participants show the effects, and the range of effect sizes. See suggestions here:

https://onlinelibrary.wiley.com/doi/full/10.1111/ejn.13400

Figures 2-4: please increase font size. Would be worth splitting the content of figures 2 and 3 across 2 figures to increase the size of all panels. Figure 4 could do with having 2 rows.

"As shown in Fig. 2E and 2G, diagonal decoding profiles were very different for correct and incorrect decisions" - this would be better reported by adding a new plot with correct and incorrect time-courses superimposed, and plotted separately for low and high confidence. Interactions could also be illustrated by plotting time-courses of differences.

I find this bit difficult to understand: "Most prominently, it can be observed that for correct trials, all processing stages show positive decoding, whereas in sharp contrast the two early stages for the incorrect trials showed negative decoding. This negative decoding pattern on incorrect decisions illustrates that early EEG data patterns look more similar to the presented stimulus category than to the reported stimulus category (note again that the decisions were used as the classifier labels)." By negative decoding, you mean below chance, not negative values? More importantly, could you provide a toy example, with simulated data from one participant, that would clearly demonstrate this dissociation and support your interpretation? Or is there a reference about this type of classifier's behaviour?

## Stat reporting.

The results section could be improved by going beyond outdated statements about the statistical significance of effects. Here are resources to more carefully report frequentist results, without falling into the trap of artificial dichotomies:

Language for communicating frequentist results about treatment effects

https://discourse.datamethods.org/t/language-for-communicating-frequentist-results-about-treatment-effects/934

Scientists rise up against statistical significance

https://www.nature.com/articles/d41586-019-00857-9

Inferential Statistics as Descriptive Statistics: There Is No Replication Crisis if We Don't Expect Replication

https://www.tandfonline.com/doi/full/10.1080/00031305.2018.1543137

Abandon Statistical Significance

https://www.tandfonline.com/doi/full/10.1080/00031305.2018.1527253

"Sensitivity (d') did not differ significantly from zero" - this suggests an equivalence test was performed. Otherwise, you can only conclude that you reserve judgement or that you did not get enough evidence. A simple statement would just state the results and point to a detailed graphical representation. There is no need to dichotomise all p values, unless you're trying to localise in space or time using brain imaging data - hard to see how to do that without some threshold.

In this sentence and possibly others, you seem to imply that the lack of evidence for an effect is evidence for its absence: "This was only the case when participants were highly confident in their incorrect decision (HC: t39=3.86, p<.001, d=0.61; LC: t39=1.41, p=.17, d=0.22, BF01=2.35; LC vs HC: t39=-3.06, p=.004, d=-0.48; Fig. 2H)". The statement should be revised by drawing attention to effect sizes and a supporting illustration. The absence of an effect cannot be derived from p values.

Same issue in lines 307-310 as well as other places, including supplementary information:

"Decoding strength was equal for correct and incorrect decisions (F1,39=0.06, p=.82, ηp2=0.001, BF01=5.7) and confidence did not modulate this decoding profile in any way (F1,39=0.54, p=.47, ηp2=0.014, BF01=4.77; interaction correctness x confidence: F1,39=0.18, p=.67, ηp2=0.005, BFexcl=4.22)."

Lines 270-272: can you provide values to qualify "much weaker"? For instance median difference + range?

275: "all other effects were identical" -> "were similar", as they cannot be identical.

808: please add a justification for your alpha.

## Other points

"To determine the nature of misreports, we present a novel way to dissociate sensory from post-sensory or decision-related processes during human perceptual decision-making using electroencephalography (EEG) in combination with multivariate pattern analyses (MVPA)." I'm struggling to see the novelty in the approach, so maybe you need to be more specific. These articles for instance, contain very similar ingredients:

https://jov.arvojournals.org/article.aspx?articleid=2121325

https://elifesciences.org/articles/68491

https://academic.oup.com/cercor/article/16/4/509/382221

https://www.jneurosci.org/content/26/35/8965

https://www.pnas.org/content/106/16/6539.short

Using EMG could also be a way to assess the motor stage:

https://hal.archives-ouvertes.fr/hal-01292369/document

In addition to refs 63-64, this is a good paper about mapping decoding results:

https://www.sciencedirect.com/science/article/pii/S1053811905003381

"N170 ERP component, specific to face processing" -- that component is also observed for letters, words, and biological motion for instance, so not specific to faces.

"In general, activity observed in sensory regions can thus be a mixture of both feedforward and feedback processes, especially at longer latencies (>~100 ms) (39)". Not sure if this has been confirmed and if it applies to non-moving stimuli, but feedback seems to affect the earliest V1 responses:

https://pubmed.ncbi.nlm.nih.gov/11152714/

See also this in humans:

https://pubmed.ncbi.nlm.nih.gov/11797091/

Both reinforcing your point, that indeed activity in sensory regions is a mixture of feedforward and feedback processes.

"were viewed with a distance of 90 cm from the monitor" - please indicate if a chin rest was used. If not, adjust text to something like "unconstrained viewing distance of approximately 90 cm".

663-5: "Furthermore, letters F and H, for faces and houses respectively, were shown on every trial in the right and left upper corner, or vice versa, depending on the block, as a as a reminder of the response mapping to the button press." For replicability, information is missing about the size, position and duration of these stimuli.

711-3: "EEG data was re-referenced to the average of the earlobes, high-pass filtered at 0.5 Hz and low-pass filtered at 40 Hz." Information is missing about the type of filters. Please also provide a figure illustrating filter distortions, ideally using a synthetic signal:

https://www.sciencedirect.com/science/article/pii/S0165027012002361

https://www.sciencedirect.com/science/article/pii/S0165027021000157

A gentle 0.5 Hz HP filter is unlikely to severely distort your temporal estimates, but it is worth checking and reminding readers about the dangers of HP filtering.

719: "All analyses were done using custom scripts..." - this would be a good place to mention that all the code and data are available online.

## Writing

The section starting line 101, "The time-course of category representations: Decision localizer", is difficult to read. It would help to refer to the little squares 1-4 in the figure. The squares are mentioned in a subsequent paragraph, but that description should be integrated with the earlier description. I would make figure 2A much larger and stand alone, to more carefully guide readers through it. Views of the diagonal, like in panels E and G are very useful, and would help support the text description and the understanding of panel A.

"this neural representation "flipped" later in time" - avoid the quotation marks of mystery, be specific instead. Same for "global ignition" and other expressions, unless they are quotes.

This sentence is hard to read:

"Here we investigated this issue and examined at what level of analysis visual stimuli that were misreported (e.g., face presented, house reported) were actually misrepresented, at the neural level, aiming to dissociate sensory from decision-related explanations."

Suggested revision:

"Here we investigated this issue and examined if misreported visual stimuli (e.g., face presented, house reported) were misrepresented at the sensory or the decision stage."

"As can be seen in Fig. 1, the orientation of the presented images was either left tilting or right tilting, with 50/50 likelihood." In Figure 1 all the images are tilted to the left.

Lines 108-109: "In contrast, generalization across time analyses can" - delete "analyses"? That whole sentence needs to be rephrased to remove the repetition "generalization across time [...] observed when a classifier trained on certain time points generalizes to other data points earlier or later in time".

Line 112: "Category-specific neural representations could be robustly decoded" -- robust to what? I would delete "robust".

Figure 2 caption: "On the right covariance/class separability" - is a comma missing after "right"?

Line 245: "despite never been objectively presented and never been" -> "despite never being presented and never being..." Not sure what an "objective presentation" means. Would be clearer to state "presented on the screen".

803: "covariance/class separability map maps were" - extra "map"?

Signed review: Guillaume Rousselet

Rev. 2:

The manuscript is a report of an EEG experiment where participants were asked to discriminate images of faces and houses near threshold. Multivariate analyses are performed and reveal that early sensory processing is consistent with the physical stimulus whereas later processing is consistent with the percept. The experiment is cleanly designed and analyses, and the manuscript is well-written. But I am not convinced that the results are sufficiently ground-breaking to appear in a highly selective journal like PLOS Biology.

My first concern is that the motivation of the study is overlooking decades of research that have addressed the same question as the authors. This is rather surprising given that the senior authors have themselves largely contributed to this research. Just to give a few examples, Linares et al. (2019) have presented a method to decouple sensory from decisional choice biases in perceptual decision making, Kunimoto et al. (2001) have presented behavioral results on the link between near-threshold perceptual responses and confidence judgments, and Fleming & Dolan (2012) have presented the neural bases for these behavioral results.

My second concern is that the novelty of the results is exaggerated. The authors seem surprised that early sensory processing is consistent with the physical stimulus whereas later processing is consistent with the percept. How could it be otherwise? Again, the neural basis of this hierarchical processing of ambiguous information has been established decades ago (see e.g. Logothetis & Schall, 1989).

My third concern is that the link between confidence judgments and perceptual decisions is only superficially explored. For instance, the four-point rating scale is analyzed as just high versus low confidence, disregarding that the lowest confidence level was linked to chance perceptual performance (Figure 1B). Some analyses commonly found in the metacognition literature (for instance using meta-d'; Maniscalco & Lau, 2012) would have been welcomed. There are also interesting EEG-based analyses worth considering on this topic (e.g. Charles et al., 2013; Balsdon et al., 2021).

Minor comments:

- Figures 2E, F, G, & H: would it be possible to use the same scale on the y-axis to better compare classification performance between correct and incorrect decisions?

- line 213: replace "Figure 3E" by "Figure 2E"

References

Balsdon, T., Mamassian, P. & Wyart, V. (2021). Separable neural signatures of confidence during perceptual decisions. ELife, 10, e68491. https://doi.org/10.7554/elife.68491

Charles, L., Van Opstal, F., Marti, S., & Dehaene, S. (2013). Distinct brain mechanisms for conscious versus subliminal error detection. NeuroImage, 73, 80-94. https://doi.org/10.1016/j.neuroimage.2013.01.054

Fleming, S. M. & Dolan, R. J. (2012). The neural basis of metacognitive ability. Philosophical Transactions of the Royal Society of London Series B, Biological Sciences, 367(1594), 1338-1349. https://doi.org/10.1098/rstb.2011.0417

Kunimoto, C., Miller, J., & Pashler, H. (2001). Confidence and accuracy of near-threshold discrimination responses. Consciousness and cognition, 10(3), 294-340. https://doi.org/10.1006/ccog.2000.0494

Linares, D., Aguilar-Lleyda, D. & López-Moliner, J. (2019). Decoupling sensory from decisional choice biases in perceptual decision making. ELife, 8, e43994. https://doi.org/10.7554/elife.43994

Logothetis, N. K., & Schall, J. D. (1989). Neuronal correlates of subjective visual perception. Science, 245(4919), 761-763. https://doi.org/10.1126/science.2772635

Maniscalco, B. & Lau, H. C. (2012). A signal detection theoretic approach for estimating metacognitive sensitivity from confidence ratings. Consciousness and Cognition, 21(1), 422-430. https://doi.org/10.1016/j.concog.2011.09.021

Rev. 3: Paul Sajda - note that this reviewer has signed his review

In this paper, Aliliovic et al investigate how misinterpretations of sensory stimuli are reflected in EEG activity using an MVPA approach that separates sensory and decision related processes in the neural activity. They find that when subjects are confident in a decision which is an "error" (i.e. their choice does not match the stimulus category) sensory stages of stimulus processing represent the category and then later flip to a representation of the choice category. Note the reason this is important is that it is only seen for the high confidence, erroneous decisions. This work adds to the literature on how sensory evidence is transformed into a decision and how confidence may arbitrate bottom up and top down process.

I very much enjoyed the paper and I think it adds to the literature. I have some comments below that I think might help clarify the presentation a bit, particularly within the context of prior work. To be completely transparent most of the prior work I will reference below I am a co-author, and though I am asking questions related to this prior work, by no means do I require/expect the authors to reference all this work unless they believe it relevant. However there are a number of interesting findings they report that map nicely to some of my prior work that I think would be nice to reconcile/consider. I sign my name at the bottom of this review so that the authors know who I am.

1. Interaction between sensory and decision making processes explain differences between stimulus category and choice.

The interaction and "flipping" of neural representation for high confidence error trials is a very interesting finding showing the interaction between confidence, choice and sensory input. in [1] our group used a similar task (discriminating faces from houses or cars) and simultaneously recorded EEG and fMRI and showed that differences in interactions between "early" and "late" networks could explain decision bias. Note that no reports of confidence were used in our study, however bias was the measure that was representative of the mismatch between stimulus and choice category. I think that prior work and this new paper presented by the authors together show that there is clearly a bottom up and top down interaction occurring when "perpeptual illusions" or "decision bias" is in play -- i.e. when the sensory category does not match the choice category and individuals are confident in their choice. It would be interesting to know the authors thoughts with respect to this prior work. Also, as an aside, did you observe any face bias for the decision errors?

2. Separating sensory and decision processes.

This of course is key to the work presented here. I would recommend adding a little more detail about how the MVPA was trained/optimized and also on interpretation (see below). Also some other work that might be worth considering (though once again just a suggestion) which supports these types of approaches to disassociate different processing stages in perceptual decision making.

[2] used MVPA classifiers on EEG for a face vs car task to separate sensory and evidence related signals. Together with [3] this work showed that early discriminating components are likely related to an iconic memory or "perceptual maintenance stage" (see your manuscript, line 133), and that in fact when comparing the single trial variability of these components using a DDM [4] the components clearly dissociate into a sensory evidence variable and decision/choice evidence variable. [5] uses a MVPA technique to show that multiple decision related signals are expressed across a perceptual decision and that they can be tracked via analysis of the MVPA output across time. This prior work just reinforces the approach the authors take and shows that one can use MVPA to dissociate neural correlate of perceptual and cognitive processes.

3. MVPA classifier

I think it would be helpful to know a little more about the classifier. For example, is it a logistic regression classifier, LDA, etc. Was the optimization regularized? In the supplementary material it would be good to show mathematical expressions for the classifier as well as the equation for what is being optimize and if this is a convex optimization or is regularized in some way. In Fig2 and 3, panel A, plotting significant channels is interesting but it might be better to plot forward models which take into account the multivariate interactions between channels and would also be more interpretable.

4. Minor comments

Figure 1B should be clarified. Is significance between confidence levels relative to d' =0 or is this significance relative to one another? It is clear that for confidence = 1, that this is respect to d' = 0, but not clear for the other confidence levels.

The sensory localizer also seems to be an oddball task (in frequent (1/5) contrast change). Perhaps mention how this will add an orienting response to the sensory localizer component. I don't think it is problematic per se, but it adds another process into what the sensory localizer represents.

References cited above

[1] Tu, T., Schneck, N., Muraskin, J., & Sajda, P. (2017). Network configurations in the human brain reflect choice bias during rapid face processing. Journal of Neuroscience, 37(50), 12226-12237.

[2] Philiastides, M. G., Ratcliff, R. & Sajda, P. Neural representation of task difficulty and decision making during perceptual categorization: a timing diagram. The Journal of neuroscience : the official journal of the Society for Neuroscience 26, 8965-8975 (2006).

[3] Philiastides, M. G. & Sajda, P. EEG-informed fMRI reveals spatiotemporal characteristics of perceptual decision making. The Journal of neuroscience : the official journal of the Society for Neuroscience 27, 13082-13091 (2007).

[4] Ratcliff, R., Philiastides, M. G. & Sajda, P. Quality of evidence for perceptual decision making is indexed by trial-to-trial variability of the EEG. Proceedings of the National Academy of Sciences of the United States of America 106, 6539-6544 (2009).

[5] Philiastides, M. G., Heekeren, H. R. & Sajda, P. Human Scalp Potentials Reflect a Mixture of Decision-Related Signals during Perceptual Choices. The Journal of neuroscience : the official journal of the Society for Neuroscience 34, 16877-16889 (2014).

signed Paul Sajda

Decision Letter 2

Kris Dickson, PhD

7 Dec 2022

Dear Simon,

Thank you for your patience while we considered your revised manuscript "The perceptual nature of illusory object recognition" for publication as a Research Article at PLOS Biology. This revised version of your manuscript has been evaluated by the PLOS Biology editors, the Academic Editor, and the original reviewers.

As the reviewers are now satisfied with the revisions, we are likely to accept this manuscript for publication. At this point, we simply ask that you address a few editorial requests and the required data and other policy-related requests detailed at the bottom of this email.

Editorial Requests:

***Article title - Please consider a modified title that will more completely convey the interesting findings in this work. We'd suggest something like:

Illusory object recognition reflects either sensory perceptual or cognitive origins depending on decision confidence

**Blurb –

Please also provide a blurb which, if the paper is accepted, will be included in our weekly and monthly Electronic Table of Contents (eTOCs), sent out to readers of PLOS Biology. This blurb may also be used to promote your article on social media. The blurb should be about 30-40 words long and is subject to editorial changes. It should, without exaggeration, entice people to read your manuscript, should not be redundant with the title and should not contain acronyms or abbreviations. For examples, view our author guidelines: https://journals.plos.org/plosbiology/s/revising-your-manuscript#loc-blurb

As you address these items, please also take this last chance to review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the cover letter that accompanies your revised manuscript.

We expect to receive your revised manuscript within two weeks.

To submit your revision, please go to https://www.editorialmanager.com/pbiology/ and log in as an Author. Click the link labelled 'Submissions Needing Revision' to find your submission record. Your revised submission must include the following:

- a cover letter that should detail your responses to any editorial requests, if applicable, and whether changes have been made to the reference list

- a Response to Reviewers file that provides a detailed response to the reviewers' comments (if applicable)

- a track-changes file indicating any changes that you have made to the manuscript.

NOTE: If Supporting Information files are included with your article, note that these are not copyedited and will be published as they are submitted. Please ensure that these files are legible and of high quality (at least 300 dpi) in an easily accessible file format. For this reason, please be aware that any references listed in an SI file will not be indexed. For more information, see our Supporting Information guidelines:

https://journals.plos.org/plosbiology/s/supporting-information

*Published Peer Review History*

Please note that you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out. Please see here for more details:

https://blogs.plos.org/plos/2019/05/plos-journals-now-open-for-published-peer-review/

*Press*

Should you, your institution's press office or the journal office choose to press release your paper, please ensure you have opted out of Early Article Posting on the submission form. We ask that you notify us as soon as possible if you or your institution is planning to press release the article.

*Protocols deposition*

To enhance the reproducibility of your results, we recommend that if applicable you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols

Please do not hesitate to contact me should you have any questions.

Sincerely,

Kris

Kris Dickson, Ph.D., (she/her)

Neurosciences Senior Editor/Section Manager,

kdickson@plos.org,

PLOS Biology

------------------------------------------------------------------------

DATA POLICY:

As you are aware, the PLOS Data Policy requires that all data be made available without restriction: http://journals.plos.org/plosbiology/s/data-availability. For more information, please also see this editorial: http://dx.doi.org/10.1371/journal.pbio.1001797

***We note that the DOI link provided on Figshare (https://doi.org/10.21942/uva.c.6265233) is not active. This will need to be updated to allow our readership full access to the underlying summary data.

Note that we do not require all raw data. Rather, we ask that all individual quantitative observations that underlie the data summarized in the figures and results of your paper be made available in one of the following forms:

1) Deposition in a publicly available repository (such as Figshare or Zenodo). Please also provide the accession code or a reviewer link so that we may view your data before publication.

2) Supplementary files (e.g., excel). Please ensure that all data files are uploaded as 'Supporting Information'.

In either case, please make sure the files are invariably referred to (in the manuscript, figure legends, and the Description field when uploading your files) using the following format verbatim: S1 Data, S2 Data, etc. Multiple panels of a single or even several figures can be included as multiple sheets in one excel file that is saved using exactly the following convention: S1_Data.xlsx (using an underscore).

***Regardless of the method selected, please ensure that you provide the individual numerical values that underlie the summary data displayed in the following figure panels as they are essential for readers to assess your analysis and to reproduce it:

Fig 1B; Fig 2A,B; Fig 3A-E; Fig 4A-E; Fig 5A-C; SuppFig 2A-F; SuppFig 3; SuppFig 4A-C; SuppFig 5A-C

***Please also ensure that figure legends in your manuscript and in the supplemental file are updated to include information on WHERE THE UNDERLYING DATA CAN BE FOUND (e.g. “The underlying data supporting Fig X, panel Y can be found in file Z.”). Please also ensure that your supplemental data file/s has a legend.

NOTE: the numerical data provided should include all replicates AND the way in which the plotted mean and errors were derived (it should not present only the mean/average values).

Please ensure that your Data Statement in the submission system accurately describes where your data can be found, with the DOI link being updated to an active file link.

------------------------------------------------------------------------

DATA NOT SHOWN?

- Please note that per journal policy, we do not allow the mention of "data not shown", "personal communication", "manuscript in preparation" or other references to data that is not publicly available or contained within this manuscript. Please double check your submission and either remove mention of any such data or provide figures presenting the results and the data underlying such figure(s).

------------------------------------------------------------------------

Reviewer remarks:

Reviewer's Responses to Questions

Do you want your identity to be public for this peer review?

Reviewer #1: Yes: Guillaume Rousselet

Reviewer #2: No

Reviewer #3: Yes: Paul Sajda

Reviewer #1: The authors have tackled all my comments. I've also read the reply to the other reviewers and I think the new version addresses all their comments as well. Emphasis on novelty is not something I consider important, but the authors have now clearly articulated how their work builds upon previous studies on a similar topic. The revision made an already strong article even stronger: the writing is sharper, the reporting and figures more detailed and informative, and the interpretations more nuanced. I would change the wording of a few things, especially in the results, but that's down to personal preferences, so no need to make any more changes at this stage. The authors should be really proud of what they have achieved here.

Reviewer #2: I would like the authors for taking all the reviewers' comments seriously, including mine, and for writing a detailed response letter. The changes made to the manuscript in the text and the figures are very much welcomed. The analogy to the no-report paradigm definitely helped me clarify a few points.

On the use of confidence reports, the authors are stating that "confidence was not used as a dependent measure per se" (response letter), but in the abstract, they admit "This work demonstrates that decision confidence arbitrates between perceptual decision errors, which reflect true illusions of perception, and cognitive decision errors, which do not". So it seems that confidence was after all an important dependent variable in their design. However, I agree with the authors that an in-depth analysis of this variable would have made this manuscript much longer and more cumbersome.

In summary, I think the results are clean, the analyses are trust-worthy, and thus this manuscript deserves to be nicely published.

Reviewer #3: The reviewers have addressed all my comments and concerns and I recommend the manuscript be accepted.

Signed: Paul Sajda

Decision Letter 3

Kris Dickson, PhD

20 Jan 2023

Dear Dr van Gaal,

Thank you for the submission of your revised Research Article "Illusory object recognition is either perceptual or cognitive in origin depending on decision confidence" for publication in PLOS Biology. On behalf of my colleagues and the Academic Editor, Chris Pack, I am pleased to say that we can in principle accept your manuscript for publication, provided you address any remaining formatting and reporting issues. These will be detailed in an email you should receive within 2-3 business days from our colleagues in the journal operations team; no action is required from you until then. Please note that we will not be able to formally accept your manuscript and schedule it for publication until you have completed any requested changes.

Please take a minute to log into Editorial Manager at http://www.editorialmanager.com/pbiology/, click the "Update My Information" link at the top of the page, and update your user information to ensure an efficient production process.

PRESS

We frequently collaborate with press offices. If your institution or institutions have a press office, please notify them about your upcoming paper at this point, to enable them to help maximize its impact. If the press office is planning to promote your findings, we would be grateful if they could coordinate with biologypress@plos.org. If you have previously opted in to the early version process, we ask that you notify us immediately of any press plans so that we may opt out on your behalf.

We also ask that you take this opportunity to read our Embargo Policy regarding the discussion, promotion and media coverage of work that is yet to be published by PLOS. As your manuscript is not yet published, it is bound by the conditions of our Embargo Policy. Please be aware that this policy is in place both to ensure that any press coverage of your article is fully substantiated and to provide a direct link between such coverage and the published work. For full details of our Embargo Policy, please visit http://www.plos.org/about/media-inquiries/embargo-policy/.

Thank you again for choosing PLOS Biology for publication and supporting Open Access publishing. We look forward to publishing your study. 

Sincerely, 

Kris

Kris Dickson, Ph.D., (she/her)

Neurosciences Senior Editor/Section Manager

PLOS Biology

kdickson@plos.org

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 Text. Additional results.

    BF, Bayes factor; HC, high confidence; LC, low confidence.

    (DOCX)

    S1 Fig. Trial sequence of the sensory and decision localizer task.

    (A) Each trial of the sensory localizer task started with a central red fixation dot for 1,200–1,400 ms, during which an infrequent contrast change of the dot could happen (20% of trials). Participants needed to press the spacebar as soon as they noticed the contrast change. In the same time interval, a house or a face image was briefly shown on the screen, which needed to be ignored. Images were either tilted to the right or to the left (task-irrelevant feature) at a 5° or 355° angle. Note that in the example trial, only a left-tilted house image is shown. (B) Each trial of the decision localizer task started with a central red fixation dot after which a forward mask was shown, followed by an image of a face or a house and a backward mask. Images were either tilted to the right or to the left (task-irrelevant feature) at a 5° or 355° angle. On every trial, participants reported whether they perceived a house or a face and indicated their confidence in this decision.

    (TIF)

    S2 Fig

    (A) GAT matrix for the decision localizer and regions of interest marked by inset black boxes (numbers 1–4). (B) Classifiers trained on the decision localizer applied to the main discrimination task (cross-task validation procedure). (C) GAT matrix for the sensory localizer. (D) Classifiers trained on the sensory localizer applied to the main discrimination task (cross-task validation procedure. (E) Classifiers trained on the sensory localizer applied to the decision localizer (cross-task validation procedure). (F) Classifiers trained on the decision localizer applied to the sensory localizer (cross-task validation procedure). All GAT matrices are based on occipital–parietal electrodes. The underlying data and scripts supporting this figure can be found on FigShare (https://doi.org/10.21942/uva.c.6265233.v1).

    (TIF)

    S3 Fig. Filtering effects on the timing of ERP components.

    ERPs are time-locked to stimulus onset in the passive localizer with 3 different high-pass filtering settings. The region of interest plotted consists of the following electrodes: O1, O2, PO3, PO4, PO7, PO8. The underlying data and scripts supporting this figure can be found on FigShare (https://doi.org/10.21942/uva.c.6265233.v1).

    (TIF)

    S4 Fig. Face/house classification based on the decision localizer.

    (A) On-diagonal time courses of correct and incorrect decisions, separated by confidence (LC, low confidence; HC, high confidence). Black boxes indicate the time-windows (early, mid, late) used for averaging the AUC scores. (B) Perceptual maintenance is derived by training a classifier on the time-window 150–200 ms and testing it across the entire time-window. In (A) and (B), colored horizontal lines indicate periods of significant decoding with respect to chance. Classification performance was evaluated at each time point using cluster-based permutation testing (two-tailed cluster-permutation, alpha p < 0.05, cluster alpha p < 0.05, N permutations = 1,000). (C) Bar plots showing average AUC values for the time-windows of interest highlighted in panels (A) and (B). The underlying data and scripts supporting this figure can be found on FigShare (https://doi.org/10.21942/uva.c.6265233.v1).

    (TIF)

    S5 Fig. Face/house classification based on the sensory localizer.

    (A) On-diagonal time courses of correct and incorrect decisions, separated by confidence (LC, low confidence; HC, high confidence). Black boxes indicate the time-windows (early, mid, late) used for averaging the AUC scores. (B) Perceptual maintenance is derived by training a classifier on the time-window 150–200 ms and testing it across the entire time-window. In (A) and (B), colored horizontal lines indicate periods of significant decoding with respect to chance. Classification performance was evaluated at each time point using cluster-based permutation testing (two-tailed cluster-permutation, alpha p < 0.05, cluster alpha p < 0.05, N permutations = 1,000). (C) Bar plots showing average AUC values for the time-windows of interest highlighted in panels (A) and (B). The underlying data and scripts supporting this figure can be found on Fig Share (https://doi.org/10.21942/uva.c.6265233.v1).

    (TIF)

    Attachment

    Submitted filename: Rebuttal_Reviews PLOS Bio.docx

    Attachment

    Submitted filename: Rebuttal_Alilovic_2nd_revision.docx

    Data Availability Statement

    The data and analysis scripts used in this article is available on Figshare: https://doi.org/10.21942/uva.c.6265233.v1.


    Articles from PLOS Biology are provided here courtesy of PLOS

    RESOURCES