Skip to main content
PLOS Biology logoLink to PLOS Biology
. 2024 Sep 10;22(9):e3002790. doi: 10.1371/journal.pbio.3002790

Multisensory perceptual and causal inference is largely preserved in medicated post-acute individuals with schizophrenia

Tim Rohe 1,2,*, Klaus Hesse 1, Ann-Christine Ehlis 1,3, Uta Noppeney 4
Editor: Simon van Gaal5
PMCID: PMC11466413  PMID: 39255328

Abstract

Hallucinations and perceptual abnormalities in psychosis are thought to arise from imbalanced integration of prior information and sensory inputs. We combined psychophysics, Bayesian modeling, and electroencephalography (EEG) to investigate potential changes in perceptual and causal inference in response to audiovisual flash-beep sequences in medicated individuals with schizophrenia who exhibited limited psychotic symptoms. Seventeen participants with schizophrenia and 23 healthy controls reported either the number of flashes or the number of beeps of audiovisual sequences that varied in their audiovisual numeric disparity across trials. Both groups balanced sensory integration and segregation in line with Bayesian causal inference rather than resorting to simpler heuristics. Both also showed comparable weighting of prior information regarding the signals’ causal structure, although the schizophrenia group slightly overweighted prior information about the number of flashes or beeps. At the neural level, both groups computed Bayesian causal inference through dynamic encoding of independent estimates of the flash and beep counts, followed by estimates that flexibly combine audiovisual inputs. Our results demonstrate that the core neurocomputational mechanisms for audiovisual perceptual and causal inference in number estimation tasks are largely preserved in our limited sample of medicated post-acute individuals with schizophrenia. Future research should explore whether these findings generalize to unmedicated patients with acute psychotic symptoms.


Hallucinations and perceptual abnormalities in psychosis are thought to arise from imbalanced integration of prior information and sensory inputs. This study shows that the neurocomputational mechanisms of multisensory perceptual and causal inference remain remarkably intact in schizophrenia.

Introduction

Hallucinations—percepts in the absence of sources in the external world—are a hallmark of psychotic disorders such as schizophrenia. Individuals with psychosis may for instance hear voices or see people that are not present in their environment. The computational and neural mechanisms that give rise to these perceptual abnormalities remain unclear. Increasing research is guided by the notion that perception relies on probabilistic inference based on 2 distinct sources of information, observers’ top-down prior beliefs and bottom-up noisy sensory signals [17]. According to Bayesian probability theory, the brain should combine these 2 sources of information weighted according to their relative precisions (i.e., inverse of noise or variance), with a greater weight given to more reliable information. Hallucinations may thus arise from abnormal weighting of prior beliefs and incoming sensory evidence [8,9].

Consistent with this conjecture, overreliance on priors correlated with psychotic symptoms in patients with schizophrenia [10,11] and hallucination-like experiences in hallucination-prone individuals [1115]. By contrast, direct comparison between patients with schizophrenia and healthy controls did not consistently reveal overreliance on priors but also underweighting of priors [1619]. This pattern of results suggests that overreliance on priors may be associated with psychotic symptoms in patients or hallucination-like experiences in healthy controls rather than act as a trait-marker for schizophrenia [12]. However, these diverse findings raise the possibility that psychosis in schizophrenia may either increase or decrease the precision or weight of different types of priors [16,17]. While the weight of priors about simple features (e.g., motion) is thought to decrease in psychosis [18], the weight of priors about semantic or related information may increase [20].

The weighting of various pieces of information becomes even more complex when the brain is confronted with multiple sensory signals that may come from same or different causes. In the face of this causal uncertainty, the brain needs to infer whether 2 sensory signals—say the sound of a whispering voice and the sight of articulatory movements—come from a common source (e.g., a single speaker) and should hence be integrated or else be processed independently (e.g., in case of different speakers). Recent research has shown that hallucinations in psychosis are not only unisensory (e.g., hearing voices), but hallucinations are more often multisensory than previously assumed [21]. Moreover, multisensory hallucinations are associated with greater conviction and more distress for patients [22], especially when signals from different sensory modalities are semantically related or occur simultaneously. This pattern suggests that patients perform causal inference and integrate signals from different sensory modalities into unified hallucinatory percepts. The multisensory nature of hallucinations points towards an additional intriguing mechanism for hallucinations: Patients may be more prone to integrate even signals from different sources and thereby attribute greater “reality status” to them. Critically, because multisensory hallucinations rely on causal inference computations, they cannot be mediated solely by local neural mechanisms in early sensory cortices. Instead, they must involve complex perceptual and causal inference processes along cortical hierarchies.

Models of hierarchical Bayesian causal inference [2325] account for this causal inference problem in multisensory perception by explicitly modeling the causal structures that could have generated the sensory signals. When signals happen at the same time and space and are semantically (or numerically) congruent, it is likely that they arise from a common source. Hence, in this common-cause case, the brain should fuse the signals weighted in proportion to their relative sensory precisions into 1 single unified perceptual estimate, giving a stronger weight to the more reliable (i.e., less noisy) signal (i.e., fusion estimate). When signals happen at different times or are semantically incongruent, it is likely that they emanate from different sources. In this case of separate sources, the brain should process them independently (i.e., segregation estimates for the unisensory signals). Critically, the brain does not a priori know whether signals come from common or independent sources. Instead, it needs to infer the underlying causal structure from noisy statistical correspondence cues such as signals happening at the same time or space, being numerically or semantically congruent [2629]. To account for observers’ uncertainty about the signals’ causal structure, the brain should read out a final perceptual estimate by combining the fusion (i.e., common source) and segregation (i.e., separate sources) estimates weighted by the posterior probabilities of each causal structure (i.e., common or independent sources). This decision strategy is referred to as model averaging (for other decisional functions, see [30]). Bayesian causal inference thereby enables a graceful transition from integration for (near-) congruent auditory and visual signals to segregation for incongruent signals. Behaviorally, accumulating research has shown that human observers arbitrate between sensory integration and segregation consistent with models of Bayesian causal inference [23,27,28,3136]. At small discrepancies, they combine signals into one coherent percept which leads to prominent crossmodal biases. At larger disparities, these multisensory interactions and crossmodal biases are attenuated. At the neural level, recent research has shown that the brain accomplishes Bayesian causal inference by dynamically encoding the segregation, fusion, and the final perceptual estimate that accounts for the brain’s causal uncertainty along cortical pathways [27,33,3639]. Multisensory perceptual inference is thus governed by 2 sorts of priors, observers’ perceptual priors about environmental properties (e.g., the number of signals) as well as a causal prior about whether signals come from common or independent sources, which represents observers’ a priori tendency to bind sensory signals. While the former priors influence observer’s perceptual estimates directly, the latter does so indirectly by modulating the strength of cross-sensory interactions.

The intricacies of multisensory perception may explain the inconsistent findings regarding multisensory abnormalities in psychosis [4043] (for a review, see [44]). For example, the rate at which participants with schizophrenia experience the McGurk- or sound-induced flash illusions has been shown to be lower [40,43,45], equal [46], or even higher [41] compared to healthy controls. These inconsistencies may arise from the complex interplay of an individual’s auditory and visual precisions, perceptual and causal priors, and decisional strategies, which may all be altered in psychosis. Bayesian modeling and formal model comparison moves beyond previous descriptive approaches by allowing us to dissociate these distinct computational ingredients [4749]. For example, psychosis may alter how the brain weights prior knowledge and different pieces of sensory information: The brain may over-rely on prior information or even assign a greater weight to information from a specific sensory modality. Psychosis may also increase observers’ tendency to bind signals across different senses as quantified by the causal prior. This could facilitate the emergence of percepts that misbind incongruent sensory signals. Finally, psychosis may alter how observers read out their perceptual estimates from complex (e.g., bimodal) posterior distributions that typically arise through Bayesian causal inference [28]. For example, instead of model averaging that is predominantly observed in healthy individuals [27,28,36], patients may apply suboptimal or heuristic strategies that do not optimally take the causal structure of the signals into account [30,31]. This brief overview highlights the powerful insights that may be obtained by combining Bayesian causal inference models with behavioral and neuroimaging data acquired in more complex and challenging multisensory environments.

This psychophysics-EEG study investigated whether schizophrenia alters the computational and/or neural mechanisms of multisensory perception in a sound-induced flash illusion paradigm. In an inter-sensory selective attention task, patients with schizophrenia and age-matched healthy controls were presented with sequences of a varying number of flashes and beeps. We first assessed whether schizophrenia altered the computations of how observers combined auditory and visual signals into number estimates by comparing normative and approximate Bayesian causal inference (BCI) models. Next, we combined BCI modeling with multivariate EEG analyses to unravel the underlying neural mechanisms. Our results suggest that the core computational mechanisms underlying perceptual and caudal inference are largely preserved in our limited medicated schizophrenia cohort compared to healthy participants at the behavioral, computational, and neural level.

Results

Experimental design and analysis

Twenty-three healthy controls (HCs) and 17 observers with schizophrenia (SCZ) were included in the behavioral and EEG analyses. HC were matched to SCZ individuals in sex, age, and education. Neuropsychological tests revealed comparable performances in attentional and executive functions across the 2 groups (Table 1). Yet, HC had better memory recall and premorbid crystallized intelligence.

Table 1. Demographic, psychopathological and neuropsychological data (across-participants mean ± STD) from HC and participants with SCZ.

HC SCZ t38 p BF10
n 23 17 - - -
Age (years) 35.96 ± 2.50 33.12 ± 1.99 0.84 0.414 0.412
Sex (% w) 43.5 17.6 2.97(χ2) 0.085 -
Education (years) 16.65 ± 0.63 15.44 ± 0.89 1.14 0.269 0.520
PANSS Pos. Sympt. (Σ items) - 13.59 ± 0.97 - - -
PANSS Neg. Sympt. (Σ items) - 17.29 ± 1.75 - - -
PANSS Gen - 28.88 ± 1.71 - - -
LSHS-R - 13.82 ± 8.85 - - -
PCL - 38.69 ± 17.94 - - -
CPZ equiv - 572.59 ± 62.95 - - -
VLMT free recall (Σ trial 1–5) 57.83 ± 1.65 49.41 ± 2.99 2.63 0.009 4.230
VLMT delayed recall (trial 7) 12.61 ± 0.56 10.47 ± 0.64 2.49 0.018 3.287
VLMT recognition (Hits-FA) 13.87 ± 0.32 12.91 ± 0.54 1.61 0.104 0.853
TMT A (sec) 28.60 ± 1.98 25.70 ± 2.99 0.84 0.406 0.413
TMT B (sec) 59.89 ± 3.09 67.75 ± 10.72 −0.80 0.462 0.401
MWT-B (raw value) 30.78 ± 0.67 27.57 ± 1.32 2.34 0.019 2.523
Stroop effect (sec) 28.93 ± 2.34 33.61 ± 3.49 −1.16 0.258 0.527
EHI (Laterality index) 58.19 ± 10.65 69.42 ± 8.23 −0.79 0.439 0.398
BDI (Σ items) 2.09 ± 0.62 13.47 ± 1.68 −7.05 <0.001 >100
CDSS (⌀ items) - 0.33±0.08 - - -

Note: PANSS: Positive and Negative Symptom Scale (with 3 subscales: Positive Symptoms/Pos. Sympt., Negative Symptoms/Neg. Sympt. and General Psychopathology/PANSS Gen); LSHS-R: Launey-Slade hallucination scale—revised; PCL: Paranoia check list; CPZ equiv: Chlorpromazine equivalents (mg); VLMT: Verbal learning and memory test; TMT: Trail Making Test; MWT-B: Mehrfachwahl-Wortschatz-Intelligenztest (Multipe word choice intelligent test); Stroop effect: time for naming incongruent color words—time for naming color plates; EHI: Edinburgh Handedness Inventory; BDI: Beck’s Depression Inventory; CDSS: Calgary Depression Scale for Schizophrenia with 9 subscales.

In a sound-induced flash-illusion (SIFI) paradigm, we presented HC and SCZ with flash-beep sequences and their unisensory counterparts. Across trials, the number of beeps and flashes varied independently according to a 4 (1 to 4 flashes) × 4 (1 to 4 beeps) factorial design (Fig 1A and 1B). Thereby, the paradigm yielded numerically congruent or incongruent flash-beep sequences at 4 levels of audiovisual numeric disparity. In an inter-sensory selective attention task, observers reported either the number of beeps or flashes. The manipulation of numeric disparity across several levels enabled us to characterize how observers gracefully transitioned from integration to segregation as a key feature of Bayesian causal inference [27,35,50,51].

Fig 1. Example trial, experimental design, and behavioral data.

Fig 1

(A) Example trial of the flash-beep paradigm (e.g., 2 flashes and 4 beeps are shown) in which participants either report the number of flashes or beeps. (B) The experimental design factorially manipulated the number of beeps (i.e., 1 to 4), number of flashes (i.e., 1 to 4) and the task relevance of the sensory modality (report number of visual flashes vs. auditory beeps). We reorganized these conditions into a 2 (task relevance: auditory vs. visual report) × 2 (numeric disparity: high vs. low) factorial design for the GLM analyses of the audiovisual crossmodal bias. (C) Response accuracy (across-participants mean ± SEM; n = 40) was computed as correlation between experimentally defined task-relevant and reported signal number. Response accuracy is shown as a function of modality (audiovisual congruent conditions vs. unisensory visual and auditory conditions), task relevance (auditory vs. visual report), and group (HC vs. SCZ). (D) The audiovisual CMB (across-participants mean ± SEM; n = 40) is shown as a function of numeric disparity (1, 2, or 3), task relevance (auditory vs. visual report) and group (HC vs. SCZ). CMB was computed from participants’ behavior (upper panel) and from the prediction of the individually fitted BCI model (lower panel; i.e., model averaging with increasing sensory variances). CMB = 1 for purely visual and CMB = 0 for purely auditory influence. Source data is provided in S1 Data. BCI, Bayesian causal inference; CMB, crossmodal bias; GLM, general linear model; HC, healthy control; SCZ, schizophrenia.

Behavior—scalar variability, response accuracy, and crossmodal bias

Fig 2A shows the reported flash and beep counts in SCZ and HC as a function of the true flash and beep numbers. Both SCZ and HC progressively underestimated the increasing numbers of flashes and beeps (S1 Text, S1 Fig, and S1 Table). This logarithmic compression and the increasing variances for greater number of flashes/beeps was observed similarly in both groups. It is in line with the known scalar variability of numerosity estimates [52,53]. Fig 2 also indicated that observers’ reported flash (resp. beep) counts were biased towards the concurrent incongruent beep (resp. flash) number in the ignored sensory modality. We formally compared SCZ and HC in their ability to selectively estimate either the number of flashes or beeps (see Fig 1C and Table 2) using a 2 × 2 × 2 mixed-model ANOVA, with factors group (SCZ vs. HC), stimulus modality (unisensory visual/auditory vs. audiovisual congruent), and task relevance (TR) (auditory vs. visual report) on response accuracies (n.b. no incongruent conditions were included in this ANOVA). No significant group differences or interactions with group were observed. Instead, Bayes factors provided substantial evidence for comparable accuracies of the flash and beep counts in HC and SCZ (i.e., BFincl < 1/3, Table 2; see S2 Fig for comparable response time results).

Fig 2. Distributions of numeric reports (across-participants mean ± SEM, n = 40) for HC and SCZ patients.

Fig 2

(A) Upper panel: The auditory numeric reports plotted as a function of auditory signal number nA, separately for different visual signal numbers nV. Lower panel: The visual numeric reports plotted as a function of visual signal number nV, separately for different auditory signal numbers nA. (B) Auditory reports for a single beep as a function of visual signal number (upper panel) and visual reports for a single flash as a function of auditory signal number (lower panel). Source data is provided in S2 Data. HC, healthy control; SCZ, schizophrenia.

Table 2. Main and interaction effects of TR, modality and group on response accuracy and effects of task relevance, disparity, and group on the crossmodal bias from classical and Bayesian mixed-model ANOVAs.

F df1, df2 p part. η2 BFincl
Response accuracy
TR 49.548 1,38 <0.001 0.566 >100
Mod 0.590 1,38 0.447 0.015 0.602
Group 0.006 1,38 0.939 <0.001 0.233
TR×Mod 6.693 1,38 0.014 0.150 2.452
TR×Group 0.004 1,38 0.950 <0.001 0.352
Mod×Group 0.625 1,38 0.434 0.016 0.186
TR×Mod×Group 1.691 1,38 0.201 0.043 0.145
Crossmodal bias
TR 633.673 1,38 <0.001 0.943 >100
Disp 92.312 1.7,65.7 <0.001 0.708 >100
Group 0.312 1,38 0.579 0.008 0.189
TR×Disp 90.781 1.6,61.4 <0.001 0.705 >100
TR×Group 0.075 1,38 0.786 0.002 0.068
Disp×Group 0.456 1.7,65.7 0.608 0.012 0.146
TR×Disp×Group 0.914 1.6,61.4 0.388 0.023 0.049

Note: TR = task relevance: auditory vs. visual report; Mod = modality: audiovisual congruent vs. unisensory; Group: HC vs. SCZ; Disp = absolute numeric disparity (1 vs. 2 vs. 3). Effects of the classical mixed-model ANOVA were Greenhouse–Geisser corrected if sphericity was violated.

As expected [54], we observed a significant main effect of task-relevance and a task-relevance × modality interaction on response accuracies. Overall, observers estimated the number of beeps more accurately than the number of flashes. They also estimated the number of beeps better when presented alone than together with flashes, while the reverse was true for observers’ flash counts. Consistent with extensive previous research showing a greater temporal precision for the auditory than the visual sense [55], our results confirm that observers obtained more precise estimates for the number of beeps than for the number of flashes.

Based on Bayesian probability theory, observers should therefore assign a stronger weight to the more precise auditory signal when integrating audiovisual signals into number estimates, resulting in the well-known sound-induced flash illusion [27,56,57]. Consistent with this conjecture, both SCZ and HC were more likely to perceive 2 flashes when a single flash was presented together with 2 sounds (i.e., fission illusion) and a single flash when 2 flashes were presented together with 1 sound (i.e., fusion illusion; see S3 Fig). We quantified these audiovisual interactions using the crossmodal bias (CMB) that ranges from pure visual (CMB = 1) to pure auditory (CMB = 0) influence (Fig 1D; n.b. the crossmodal bias can be computed only for numerically disparate flash-beep sequences). Consistent with the principles of precision-weighted integration [58], the HC and SCZ’s flash reports were biased towards the number of auditory beeps (i.e., CMB < 1, t39 = −11.864, p < 0.001, Cohen’s d = −1.876, BF10 > 100), again with no significant differences between the groups (t38 = 0.434, p = 0.667, d = 0.139, BF10 = 0.336). By contrast, we observed only a small but significant crossmodal bias for auditory reports towards the number of flashes in HC and SCZ (Fig 1D; CMB > 0, t39 = 8.550, p < 0.001, d = 1.352, BF10 > 100)—again with no evidence for group differences (t38 = 0.448, p = 0.657, d = 0.143, BF10 = 0.337). Thus, both HC and SCZ assigned a greater weight to the temporally more reliable auditory sense.

An additional 2 × 3 × 2 mixed-model ANOVA with factors group (SCZ versus HC), audiovisual numeric disparity between beeps and flashes (1, 2 versus 3), and task-relevance (auditory versus visual report) revealed that these crossmodal biases were significantly decreased at large numeric disparities, when signals most likely originated from different sources and should hence be segregated (i.e., task relevance × numeric disparity interaction, Table 2). At large numeric disparities, HC and SCZ were thus able to selectively report the number of flashes (or beeps) with minimal interference from task-irrelevant beeps (or flashes). This task relevance × numeric disparity interaction was qualitatively the key response profile predicted by BCI. It basically demonstrated that observers integrated signals at small conflict size, but segregated signals at large numeric conflicts when it was unlikely that auditory beep and visual flash sequences were generated by one common underlying source. Again, Bayes factors indicated substantial to strong evidence for comparable performance in SCZ and HC (i.e., BFincl < 1/3 or even < 1/10, Table 2).

Overall, these GLM-based analyses of behavioral data (i.e., accuracy, crossmodal bias) suggested that both SCZ and HC combine audiovisual signals into number estimates qualitatively consistent with the principles of BCI. Both groups gave a stronger weight to the more reliable auditory signals leading to the well-known sound-induced flash illusions. Moreover, they arbitrated between sensory integration and segregation depending on the numeric disparity.

Behavior—Bayesian modeling

To quantitatively assess whether HC and SCZ combine audiovisual signals according to shared computational principles, we compared 10 Bayesian models in a 5 × 2 factorial model space spanned by the modeling factors of “decision strategy” (5 levels) and “sensory variance” (2 levels) [28,30,31]. All models conformed to the basic architecture of the BCI model (Fig 3).

Fig 3. The BCI model and factorial model comparisons in HC and SCZ.

Fig 3

(A) The BCI model assumes that audiovisual stimuli are generated depending on a causal prior (pCommon): In case of a common cause (C = 1), the “true” number of audiovisual stimuli (NAV) is drawn from a common numeric prior distribution (with mean μP) leading to noisy auditory (xA) and visual (xV) inputs. In case of independent causes (C = 2), the “true” auditory (NA) and visual (NV) numbers of stimuli are drawn independently from the numeric prior distribution. To estimate the number of auditory and visual stimuli given the causal uncertainty, the BCI model estimates the auditory or visual stimulus number (N^A or N^V, depending on the sensory modality that needs to be reported). In the model-averaging decision strategy, the BCI model combines the forced-fusion estimate of the auditory and visual stimuli (N^AV,C=1) with the task-relevant unisensory visual (N^V,C=2) or auditory estimates (N^A,C=2), each weighted by the posterior probability of a common (C = 1) or independent (C = 2) causes, respectively (i.e., p(C=1|xA,xV) or p(C=2|xA,xV)). (B) The factorial Bayesian model comparison (n = 40) of models with different decision strategies (model averaging, MA; model selection, MS; probability matching, PM; fixed criterion, FC; stochastic fusion, SF) with constant or increasing sensory auditory and visual variances, separately for HC and SCZ. The images show the relative model evidence for each model (i.e., participant-specific Bayesian information criterion of a model relative to the worst model summed over all participants). A larger model evidence indicates that a model provides a better explanation of our data. The bar plots show the protected exceedance probability (i.e., the probability that a given model is more likely than any other model, beyond differences due to chance) for each model factor. The BOR estimates the probability that factor frequencies purely arose from chance. Source data is provided in S3 Data. BCI, Bayesian causal inference; BOR, Bayesian omnibus risk; HC, healthy control; SCZ, schizophrenia.

Along the factor “decision strategy,” we manipulated how observers combined the estimates formed under fusion (i.e., common source) and segregation (i.e., separate sources) assumptions into a final perceptual estimate. While growing research has shown that HC combine audiovisual signals according to the decision strategy of model averaging [27,28,36], we hypothesized that SCZ may resort to suboptimal strategies or even simpler heuristics such as applying a fixed threshold on audiovisual numeric disparity [31]. In other words, rather than arbitrating between sensory integration and segregation according to the posterior probability of a common cause in a Bayesian fashion, SCZ may simply do so based on audiovisual spatial disparity.

In total, we compared the following 5 decisional strategies: (i) Model averaging combines the segregation and fusion estimates weighted by the posterior probabilities of each causal structure. For example, it gives a stronger weight to the fusion estimate when it is likely that the flash and beep sequences are generated by one common cause. (ii) Model selection selects the perceptual estimate from the causal structure with the highest posterior probability. So rather than averaging fusion and segregation estimates, it selectively reports either the fusion or segregation estimates depending on the posterior probabilities of common and independent causes. (iii) Probability matching selects either the fusion or segregation estimates in proportion to their posterior probability. (iv) The fixed-criterion threshold model incorporates the simple heuristic of selecting the segregation estimate when the audiovisual numeric disparity estimate exceeded a fixed threshold. This differs from Bayesian model selection strategies in that observers’ do not apply the threshold to the posterior probability of a common cause, but directly on the numeric disparity estimate—thereby ignoring observers’ uncertainty about this estimate. (v) Finally, the stochastic fusion model employs the suboptimal strategy of selecting either the fusion or segregation estimates with a fixed probability that is estimated from observers’ responses (see Materials and methods for details and [31]).

Additionally, we assessed along the modeling factor “sensory variance” whether auditory and visual noise variances were constant or changed with the number of beeps/flashes as predicted by the scalar-variability of numerosity estimation [52,53] and initial inspection of our data (see Figs 2 and S1).

Comparing all 10 models in this 2 × 5 factorial model space (Fig 3 and S2 Table) revealed that the model-averaging model with increasing sensory variances outperformed all other 9 models in both HC and SCZ observers. Furthermore, a between-group Bayesian model-comparison provided strong evidence that HC and SCZ individuals relied similarly on the 5 decision strategies (BF10 = 0.027). In particular, model-averaging turned out as the “winning” strategy equally often (Fig 3B). These Bayesian model-comparison results further supported the notion that SCZ, like HC, performed multisensory perceptual and causal inference according to the same computational and decision strategies. S4 Fig also shows the BICs as approximations to the model evidence for each of the 5 models that allowed for scalar variability for each patient ranked according to their PANSS score. The figure shows that the model averaging was the dominant strategy in almost all of the patients irrespective of their psychosis severity.

Having established that SCZ and HC combined audiovisual signals in line with Bayesian causal inference and read out the final estimate according to model averaging, we examined whether SCZ may overweight their causal prior about the signals’ causal structure (i.e., pcommon). Contrary to this conjecture, two-sample randomization tests on pcommon did not reveal any significant difference between the 2 groups (Fig 4 and Table 3). In the next step, we asked whether SCZ relied more strongly on their priors about the flash/beep number relative to the sensory inputs as may be hypothesized based on a growing number of studies in unisensory perception [10,11,13,15]. An over-reliance on prior information would be reflected in a greater precision (i.e., smaller variance; σP) of numeric priors (i.e., μP) in SCZ relative to HC. As shown in Fig 4 and Table 3, two-sample randomization tests indeed revealed a significantly smaller variance for the numeric prior in SCZ compared to HC. This more precise numeric prior accounts for the fact that the numeric reports are centered more around the numeric prior mean in SCZ compared to HC (Table 3; cf. Fig 2). It is consistent with previous research [10,11,13,15] and consistent with the general notion that SZC rely more on prior knowledge than new incoming audiovisual evidence. However, the effect was small and did not survive correction for multiple comparisons across the model parameters or when including the 6 additional patients with schizoaffective disorder (cf. S1 Text and S7 Table).

Fig 4. The parameters of the BCI model (across participants mean ± SEM; n = 40) separately plotted for HC and SCZ patients.

Fig 4

The BCI model’s decision strategy applies model averaging with increasing sensory variance. Significant different are indicated by * = p < 0.05. Source data is provided in S4 Data. BCI, Bayesian causal inference; HC, healthy control; SCZ, schizophrenia.

Table 3. Comparison of the BCI model’s parameters between HC and SCZ participants.

pcommon μP σP σA σV ΔσA ΔσV L
t38 −0.534 0.454 2.109 0.938 0.599 −0.275 0.958 −1.658
p 0.593 0.653 0.026 0.366 0.553 0.802 0.331 0.062
pcorr 0.731 0.731 0.210 0.717 0.731 0.811 0.717 0.282
Cohen’s d −0.171 0.145 0.675 0.300 0.191 −0.088 0.306 −0.530
BF10 0.349 0.338 1.725 0.441 0.359 0.321 0.448 0.909

Note: The BCI model’s decision strategy applies model averaging with increasing sensory variance; p-values derived from a two-sample two-sided randomization test (n = 5,000 randomizations). pcorr is multiple-comparison corrected across the 8 parameters using the Benjamini–Hochberg correction for false-discovery rate. pcommon, causal prior; μP, mean of the numeric prior; σP, standard deviation of the numeric prior; σA, standard deviation of the auditory likelihood; σV, standard deviation of the visual likelihood; Δσ, increase of standard deviation per auditory or visual signal number; L, lapse parameter.

Further, we observed a significant positive correlation between SCZ’s PANSS positive symptoms with their visual variance (r = 0.625, t16 = 3.101, p = 0.003, pcorr = 0.024, BF10 = 6.518; randomization test of the correlation). This effect remained significant even after controlling for PANSS negative symptoms and general psychopathology (S3 Table). Further, we correlated the BCI model’s parameters with self-reported hallucinatory experiences (i.e., LSHS-R) and paranoid thinking (i.e., PCL) as sensitive measures of psychosis severity that may also capture subclinical psychotic symptoms. LSHS-R and PCL were associated with greater variability in our sample compared to the PANSS positive symptoms (S5 Fig). Only the auditory variance correlated negatively with hallucinatory experiences at marginal significance (without multiple comparison correction, r = −0.498, t16 = −2.223, p = 0.047, pcorr = 0.141, BF10 = 1.439; S3 Table). In addition, we observed a positive correlation between PANSS negative symptoms and lapse rates in SCZ (r = 0.483, t16 = 2.134, p = 0.029, pcorr = 0.232, BF10 = 1.248). No other significant group differences were revealed when comparing the BCI model’s parameters between the 2 groups (Fig 4 and Table 3).

These quantitative Bayesian modeling results corroborated our initial GLM-based conclusions that both HC and SCZ combined audiovisual signals consistent with Bayesian causal inference. The numeric prior was slightly more precise in SCZ compared to HC (according to classical statistics), which was in line with previous results in unisensory perception [10,11,13,15]. Moreover, PANSS positive symptoms correlated positively with visual variance and the LSHS-R correlated negatively with auditory variance. Collectively, these results suggested that patients with psychotic symptoms may have relied more strongly on prior information and possibly auditory information relative to visual inputs for numeric inferences. The positive correlation between PANSS negative symptoms and lapse parameter indicated that at least some SCZ patients may have found it harder to stay attentive throughout the entire experiment.

Behavior—Adjustment of priors

To further explore whether SCZ over-relied on prior information, we exploited the fact that observers dynamically adapt their priors in response to previous stimuli. Some previous studies have suggested that this dynamic updating of priors is altered in psychosis ([10], but see [18]). Thus, we first examined whether and how SCZ and HC increased their binding tendency (“model-free”) or causal prior (i.e., pcommon from Bayesian modeling analysis) after exposure to congruent or incongruent flash-beep sequences. As expected based on previous findings [27,5961], both SCZ and HC similarly decreased their binding tendency after a trial with greater numeric disparity. As shown in S6A Fig, the differences between the CMB for visual and auditory report increased with the greater numerical disparity of the previous trial. In other words, observers were more capable to selectively report the numeric estimate of the task-relevant sensory modality for large numeric disparities on the previous trials. This was consistent with the idea that after a large numeric disparity trial, observers decreased their prior binding tendency which in turn attenuated audiovisual interactions and crossmodal bias on subsequent trials. Likewise, our Bayesian modeling analysis indicated that observers’ causal prior increased after a congruent trial and decreased after a trial with large numeric disparity (i.e., a main effect of previous disparity; Fig 5A and Table 4). Importantly, we did not observe any significant differences between groups (i.e., no significant disparity × group interaction).

Fig 5. Adjustment of the BCI model’s causal and numeric priors (mean ± between-participants’ SEM; n = 40) in the current trial to the previous trial’s numeric disparity or task-relevant stimulus number in HC and SCZ.

Fig 5

(A) The current causal prior (pCommon) as a function of previous audiovisual numeric disparity (i.e., |nA-nV|). (B) The current numeric prior’s mean (μP) as a function of the previous task-relevant stimulus number (i.e., nA for auditory and nV for visual report). (C) The current numeric prior’s STD (σP) as a function of the previous task-relevant stimulus number. Source data is provided in S5 Data. BCI, Bayesian causal inference; HC, healthy control; SCZ, schizophrenia.

Table 4. Main and interaction effects of previous audiovisual disparity (Disp) or audiovisual task-relevant stimulus number (Stim#) and group on the BCI model’s current causal prior (pcommon), the numeric prior’s mean (μP) and STD (σP) from classical and Bayesian mixed-model ANOVAs.

Parameter F df1, df2 p part. η2 BFincl
p common Disp 5.949 3, 114 <0.001 0.135 17.903
Group 0.291 1, 38 0.593 0.008 0.571
Disp×Group 2.195 3, 114 0.103 0.055 0.852
μ P Stim# 4.820 2.6, 97.3 0.006 0.113 7.382
Group 0.568 1, 38 0.456 0.015 0.574
Stim#×Group 0.127 2.6, 97.3 0.923 0.003 0.130
σ P Stim# 6.169 3, 114 <0.001 0.140 53.296
Group 2.706 1, 38 0.108 0.066 0.786
Stim#×Group 0.924 3, 114 0.432 0.024 0.361

Note: Disp = |nA—nV|; Stim# = nA or nV depending on their task relevance; Group: HC vs. SCZ; effects of the classical mixed-model ANOVAs are Greenhouse–Geisser corrected if sphericity is violated.

Next, we investigated the influence of the number of signals on the previous trial on observers’ numeric estimates. As expected, observers’ numeric estimates were biased towards the number of signals on the previous trial (see S6B Fig). Again, this was also reflected in our Bayesian modeling analysis assessing the updating of observers’ numeric prior (i.e., μP) and its variance (i.e., σP). As expected for Bayesian learners, both SCZ and HC increased their numeric prior’s mean and variance after exposure to a high number of signals in the task-relevant modality (e.g., beeps for auditory report; Fig 5B and 5C; significant main effects in Table 4). But again, no significant differences were observed between groups (i.e., no significant interaction effects with group) suggesting that HC and SCZ also adjust their perceptual priors at shorter timescales similarly.

In summary, our analyses of the behavioral data based on the general linear model (GLM) and formal Bayesian modeling suggest that the medicated SCZ dynamically adapted and combined priors about the signals’ causal structure and the number of signals with new audiovisual inputs comparable to the healthy controls and in line with Bayesian principles.

EEG—Decoding of numeric Bayesian causal inference estimates

To investigate whether SCZ and HC achieve Bayesian causal inference via shared or different neural processes, we combined BCI models with EEG decoding. In particular, we temporally resolved how the brain encodes the auditory (N^A,C=2) and visual (N^V,C=2) segregation, the fusion (N^AV,C=1) and the final BCI perceptual estimates (i.e., N^A or N^V) that combine the forced-fusion estimate with the task-relevant unisensory segregation estimates, weighted by the posterior causal probability of common or separate causes [27]. To track the evolution of these different estimates across time, we trained a linear support-vector regression model (SVR decoder) to decode each of the 4 BCI estimates from EEG patterns of 60 ms time windows. SVR decoders were trained on EEG patterns individually in each participant. We computed the decoding accuracy as the Pearson correlation between the decoded and “true” perceptual estimates to quantify how strongly a perceptual estimate was encoded in EEG patterns for each participant. Decoding performance at chance level was defined as a correlation of zero (i.e., r = 0). At the group level, we then assessed whether the decoded perceptual estimates (as indicated by the correlation coefficients) evolved with different time courses in HC and SCZ.

In both HC and SCZ, the decoders predicted the BCI estimates from EEG patterns significantly better than chance over most periods of the post-stimulus time window (i.e., decoding accuracy r ≈ 0.3; Fig 6A). Yet, the different perceptual estimates evolved with different time courses commonly in both groups. Initially, the EEG activity encoded mainly the visual segregation estimates starting at about approximately 60 to 100 ms (i.e., significant clusters in one-sided cluster-based corrected randomization t test; see S4 Table). Slightly later, the auditory segregation and fusion estimates peaked with the fusion estimates showing a slower decline. The final BCI estimate, which accounts for the signals’ causal structure, rose more slowly and showed a more sustained time course. Moreover, in HC only did the decoding accuracy of the final BCI estimate exceed those of the other estimates at 600 ms. The temporal profiles of the decoding accuracies were largely comparable between HC and SCZ. Bayes factors provided mainly weak evidence for no difference between the groups (Fig 6B), except for the visual segregation-estimate that was associated with a significantly lower and more protracted decoding accuracy in SCZ than HC from 120 to 220 ms (Fig 6A and S4 Table; all further clusters p > 0.05). Because we did not track observers’ eye movements, these lower decoding accuracies in SCZ may potentially be explained by reduced fixation stability in SCZ [62]. Potentially, it may also result from increases in visual uncertainty in SCZ patients that were reported in previous studies on the computational mechanisms of causal inference. These previous studies manipulated the asynchrony of audiovisual stimuli in simultaneity judgment tasks [42,63] and may therefore have been more sensitive to detect differences in sensory uncertainty than the current study that categorically manipulated the number of beeps and flashes. In summary, combining Bayesian modeling and EEG decoding largely confirmed that SCZ and HC performed audiovisual number estimation according to similar neurocomputational mechanisms. However, SCZ encoded the visual signal with less precision resulting in a lower decoding accuracy of the visual segregation estimate. These results suggested that the visual uncertainty may be increased in SCZ participants as previously reported [42]. By characterizing sensory uncertainty across time at millisecond resolution, EEG may be able to reveal sensory differences that may not be reflected in observers’ behavioral responses.

Fig 6. Decoding the BCI model’s numeric estimates from EEG patterns using SVR in HC versus SCZ (n = 40).

Fig 6

(A) Decoding accuracy (Fisher’s z-transformed correlation; across-participants mean) of the SVR decoders as a function of time and group (HC vs. SCZ). Decoding accuracy was computed as correlation coefficient between the given BCI model’s internal estimates and BCI estimates that were decoded from EEG activity patterns using SVR models trained separately for each numeric estimate. The BCI model’s internal numeric estimates comprise of: (i) the unisensory visual (N^V,C=2), (ii) the unisensory auditory (N^A,C=2) estimates under the assumption of independent causes (C = 2), (iii) the forced-fusion estimate (N^AV,C=1) under the assumption of a common cause (C = 1), and (iv) the final BCI estimate (N^A or N^V depending on the sensory modality that is task-relevant) that averages the task-relevant unisensory and the precision-weighted estimate by the posterior probability estimate of each causal structure. Color-coded horizontal solid lines (HC) or dashed lines (SCZ) indicate clusters of significant decoding accuracy (p < 0.05; one-sided one-sample cluster-based corrected randomization t test). Color-coded horizontal dotted lines indicate clusters of significant differences of decoding accuracy between both groups (p < 0.05; two-sided two-sample cluster-based corrected randomization t test). Stimulus onsets are shown along the x-axis. (B) Bayes factors for the comparison between the decoding accuracies of HC and SCZ for each of the BCI estimate (i.e., BF10 > 3 substantial evidence for or BF10 < 1/3 against group differences). Source data is provided in S6 Data. BCI, Bayesian causal inference; EEG, electroencephalography; HC, healthy control; SCZ, schizophrenia; SVR, support-vector regression.

EEG–Multisensory interactions in ERPs

Following previous work, we also analyzed and compared the event-related potentials (ERPs) to unisensory and multisensory stimuli between HC and SCZ [46,6468]. ERPs showed the typical components in response to flashes and beeps (Fig 7), i.e., P1 (approximately 50 ms), N1 (100 ms), P2 (200 ms), N2 (280 ms), and P3 (>300 ms) [69]. As previously reported [70], the unisensory visual P3 component was significantly smaller in SCZ than HC (cluster 560 to 705 ms, p = 0.045). To characterize the neural processes of audiovisual integration, we tested for audiovisual interactions (i.e., AVcongr versus (A + V)) over occipital electrodes. In line with previous studies [69,71], we observed early audiovisual interactions from 75 to 130 ms (i.e., measured from the onset of the first flash-beep slot; p = 0.049; two-sided one-sample cluster-based corrected randomization t test) and later negative audiovisual interactions from 260 to 750 ms after stimulus onset (p < 0.001). Crucially, the ERP interactions did not differ between HC and SCZ (p > 0.05; two-sided two-sample cluster-based corrected randomization test).

Fig 7. Occipital ERPs in response to unisensory and audiovisual stimuli in HC and SCZ.

Fig 7

ERPs (across-participants mean grand averages; n = 40) of HC and SCZ participants elicited by unisensory auditory stimuli (A), unisensory visual stimuli (V), audiovisual congruent conditions (AVcongr) and the difference of these ERPs (i.e., AVcongr−(A+V)) indicating multisensory interactions. The ERPs are averaged across occipital electrodes. Color-coded horizontal dotted lines indicate significant clusters (p < 0.05) of ERPs against baseline (i.e., across HC and SCZ, main effect of condition) in one-sample two-sided cluster-based corrected randomization t tests. The horizontal solid line indicates a significant cluster of ERP differences between HC and SCZ in two-sample two-sided cluster-based corrected randomization t tests. The x-axis shows the stimulus onsets. Source data is provided in S7 Data. ERP, event-related potential; HC, healthy control; SCZ, schizophrenia.

To examine whether SCZ patients and HC may differ more subtly in their multivariate EEG response, we trained a support-vector classifier (SVC) to classify participants [72] as SCZ or HC based on their EEG activation patterns to auditory, visual, and audiovisual stimuli across poststimulus time windows (see S1 Text and S7 Fig). The ERP topographies evolved largely similar in HC and SCZ, so that the EEG-based decoder did not classify participants as HC or SCZ better than chance in one-sided cluster-based corrected randomization tests when corrected across multiple comparisons across time. Overall, our ERP and multivariate analyses thus corroborated that the neural mechanisms underlying audiovisual integration were largely comparable in HC and SCZ patients. However, in line with our EEG results from the BCI analysis, we observed a deficit in visual processing in SCZ as indicated by smaller visual P3 responses.

Discussion

This study combined psychophysics, EEG, and Bayesian modeling to investigate whether and how schizophrenia impacts the computational and neural mechanisms of multisensory causal and perceptual inference. A growing number of studies suggests that schizophrenia may alter the brain’s ability to integrate audiovisual signals into coherent percepts (for a review, see [44]). Many of these studies have used qualitative approaches such as evaluating the frequency of perceptual illusions in schizophrenia compared to healthy controls. Yet, the complex dependencies of multisensory illusion rates on sensory precisions and perceptual as well as causal priors may have contributed to the inconsistent findings reported so far [41,43,46]. In this study, we have therefore employed a more sophisticated factorial design with formal Bayesian modeling to disentangle these different computational components.

Our GLM-based and Bayesian modeling analyses offered convergent evidence that SCZ as their HC counterparts performed multisensory perceptual and causal inference in line with normative principles of Bayesian causal inference. When numeric disparities were small and signals were likely to originate from common sources, individuals with SCZ and HC integrated auditory and visual signals weighted by their relative precisions resulting in crossmodal biases. At large numeric disparities, these crossmodal biases and interactions were reduced. Through formal Bayesian model comparison, we further examined whether SCZ computed Bayesian estimates or resort to simpler heuristics such as segregating signals above a numeric disparity threshold. Our quantitative Bayesian modeling analyses confirmed that SCZ, like their healthy counterparts, combined signals consistent with BCI models and read out final numeric estimates based on the decisional strategy of model averaging.

Next, we examined whether SCZ may overweight prior information about flash/beep number in relation to sensory evidence at long and/or short timescales as previously observed in unisensory perception [10,11,13,15]. In support of an overweighting of priors in psychosis, our Bayesian modeling analysis revealed a significantly more precise numeric prior for SCZ compared to HC. Further, the visual variance correlated positively with the positive symptoms on the PANSS scale. These findings are consistent with previous studies showing overreliance of patients with psychosis on prior knowledge relative to sensory evidence [1015]. However, the effect on prior variance was very small. It was no longer significant after correcting for multiple comparisons or including schizoaffective patients (cf. S1 Text and S7 Table). Likewise, the influence of the previous trial’s flash/beep number and numeric disparity on current perceptual choices was comparable in SCZ and HC.

Furthermore, we cannot exclude the possibility that SCZ showed a bias towards an intermediate signal number, which could be misinterpreted as overreliance on their numeric prior (see Figs 4 and 5C), because they were less vigilant or motivated and therefore failed to count some signals. Indeed, in line with this notion, the lapse parameter correlated positively with SCZs’ negative symptoms (S3 Table) and negatively with SCZs’ memory performance in the VLMT as well as their ability to inhibit inappropriate responses in the Stroop test (S5 Table). Overall, our analyses thus suggested that the medicated SCZ group of our study weighted and dynamically adapted numeric priors in a largely comparable fashion as their healthy counterparts with a small trend towards an overreliance on their numeric prior relative to bottom-up audiovisual evidence. The absence of substantial prior overweighting may be explained by our specific sample of medicated patients with chronic rather than acute SCZ and low expression of psychotic symptoms. Thus, prior overweighting may be a state marker associated with psychotic symptoms rather than a trait marker for schizophrenia [12].

Moreover, to our knowledge, this is the first study assessing prior weighting in a multisensory context. The availability of pieces of evidence furnished simultaneously by 2 sensory modalities may decrease the overall reliance on prior information and may thus decrease our paradigm’s sensitivity to detect small modulations in prior variance and weighting. Further, it is conceivable that distinct mechanisms across the cortical hierarchy support integration of prior and sensory evidence depending on whether information is provided only in one or in several sensory modalities. Given the prevalence of multisensory hallucinations in patients with psychosis [21], future work is required to assess perceptual inference in multisensory situations. For example, one may even develop elegant multisensory conditioning paradigms to provoke conditioned audiovisual hallucinations [11].

Crucially, our multisensory paradigm enabled us to assess not only the weighting of prior knowledge regarding environmental properties such as flash/beep number, but also to the world’s causal structure, i.e., whether signals come from common or independent sources as incorporated in the causal prior. A high causal prior or binding tendency increases crossmodal biases and interactions, while a low causal prior enhances an individual’s ability to selectively report the signal number in the task-relevant sensory modality while ignoring the incongruent number of signals in the irrelevant sensory modality. Changes in the causal prior may thus be closely associated with conflict monitoring, cognitive control, and selective attention mechanisms [51,7375]. Yet, despite growing evidence for impaired selective attention and cognitive control mechanisms in SCZ [7678], our Bayesian modeling analyses did not reveal any significant differences in the causal prior between SCZ and HC.

In summary, our GLM-based and Bayesian modeling analyses showed that SCZ and HC combined perceptual and causal priors with sensory evidence in a manner consistent with the principles of normative Bayesian causal inference. Furthermore, Bayesian statistics provided consistent evidence that the causal and perceptual priors’ weighting and updating was largely maintained in the SCZ group. These findings suggest that both groups dynamically adapt to the statistical structure of multisensory stimuli across short time scales (cf. serial dependencies in perception [79]). The absence of notable computational abnormalities in SCZ could be explained by the fact that our patient group was post-acute and medicated and scored low on the PANSS positive symptom scale. Most importantly, our cohort of SCZ patients also showed near-normal performance on the TMT-B and Stroop tests (cf. Table 1). Both tests measure executive and attentional functions which are particularly relevant for optimal performance in our inter-sensory selective-attention paradigm. Our findings add further evidence to the fact that fundamental mechanisms of perceptual inference are preserved at least in subgroups of medicated SCZ individuals [1619]. Future research is required to explore whether schizophrenia patients with attentional and executive deficits and in more severe psychosis states may exhibit abnormal causal priors, and thus, how they arbitrate between sensory integration and segregation.

Using EEG, we investigated whether SCZ and HC support these computations via similar or different neural mechanisms. For example, in SCZ, the computations may involve compensatory neural systems or exhibit a slower dynamic. To explore these questions, we decoded the auditory and visual segregation, the fusion and the final BCI model’s estimates from scalp EEG data. In HC, the decoding accuracy of the visual segregation-estimate peaked earlier and higher compared to the other estimates. By contrast, in the SCZ group, decoding of the visual estimate showed a more protracted time course initially overlapping with the auditory segregation and fusion estimates. Statistical tests confirmed a significant difference in decoding accuracy between HC and SCZ from approximately 100 to 200 ms. However, apart from this difference in decoding of the unisensory visual estimate and similarly in the visual P3 ERP component, the decoding profiles appeared largely similar across the 2 groups.

To ensure that we did not miss any differences between SCZ and HC that were previously reported in the literature [46,6468], we also performed standard ERP analyses to test for audiovisual interactions. However, these analyses revealed only an attenuated visual P3 component in SCZ, but no significant differences in audiovisual interactions between the 2 groups. Similarly, we were unable to predict group membership (i.e., HC versus SCZ) from multivariate audiovisual EEG responses significantly better than chance.

To conclude, our behavioral, computational, and neuroimaging results consistently demonstrate that audiovisual perception in a sound-induced flash illusion paradigm is based on comparable computational and neural mechanisms in SCZ and HC. Both SCZ and HC combined audiovisual signals into number estimates in line with the computational principles of Bayesian causal inference. However, our small sample size may have prevented us from detecting more subtle potential alterations in SCZ (cf. see limitations below). Our computational modeling approach moves significantly beyond previous studies assessing audiovisual illusion rates by characterizing patients according to their perceptual inference, decisional strategy, and specific parameters in the winning model. Additional time-resolved EEG decoding revealed that both HC and SCZ performed Bayesian causal inference by dynamically encoding the visual and auditory segregation- and fusion-estimates followed by later estimates that flexibly integrate auditory and visual information according to their causal structure. Collectively, our results thus showed that at least in our limited sample of post-acute medicated SCZ patients, the computations and neural mechanisms of hierarchical Bayesian causal inference in audiovisual numeric perception were largely preserved.

However, several factors limit the scope of these conclusions: First, the sample size was small and only adequate to reveal large effect sizes. Our study was able to detect significant larger differences between SCZ and HC in the numeric prior (e.g., Cohen’s d = 0.675) or larger correlations between the BCI model parameters and psychopathology such as between visual variance and participant’s psychotic symptoms (e.g., r = 0.625). Yet, the small sample size may have precluded the detection of more subtle effects. Our choice of a small sample size was guided by previous research into behavioral multisensory deficits in SCZ that reported large effect sizes for differences between SCZ and controls in small sample sizes between 12 and 30 patients [40,80,81]. However, these previous small-sample studies may have led to an unrealistic overestimation of the effect size due to the “winner’s curse” phenomenon, i.e., the tendency for published studies to report effect sizes that exceed the true effect size [82]. In fact, effect sizes of interindividual differences can generally be expected to be much smaller [83]. However, increasing the patient sample size by 6 patients with schizoaffective disorder yielded highly similar results for the CMB, BCI modeling and decoding of BCI estimates and even turned the overreliance on the numeric prior nonsignificant (see S1 Text and S8S11 Figs and S6S8 Tables). Moreover, Bayes factors frequently supported the absence of group differences. Nevertheless, studies with higher power are needed to corroborate that our study did not miss out on subtle alterations in multisensory perceptual inference in SCZ. Second, as a result of our recruitment strategy, our sample included mainly chronic patients with high doses of antipsychotics, relatively intact cognitive functions and low psychosis severity (cf. Table 1 and S5 Fig). By contrast, previous studies reported that overweighting of perceptual priors arises specifically in individuals with acute hallucinations [14], independently from a diagnosis of schizophrenia [11]. Thus, prior overweighting could rather be a state-marker for psychotic episodes than a trait-marker for schizophrenia [12]. Our conclusions of relatively intact multisensory perceptual and causal inference may not hold for individuals with prodromal or acute psychotic states of schizophrenia or patients with persistent post-acute hallucinations. Thus, future studies are needed to assess multisensory perceptual inference in a larger sample of SCZ participants that vary substantially in their psychotic symptoms. It has even been suggested that psychotic symptoms arise on a symptomatic continuum across healthy individuals and patients [84], so that altered weighting of priors in multisensory causal and perceptual inference may be found in nonclinical samples with subclinical psychotic symptoms. In line with this conjecture, a recent study in young nonclinical participants reported that individuals with stronger prodromal psychotic symptoms showed reduced causal priors [50]. Third, our small clinical SCZ sample was still heterogeneous with respect to the duration of illness, comorbidities, the dose and type of antipsychotic medication, and the strength of acute positive and negative symptoms as well as cognitive impairments. This substantial heterogeneity might have obfuscated small alterations in multisensory perceptual and causal inference in SCZ. Fourth, our experimental SIFI paradigm manipulated the number of flashes and beeps to investigate causal inference via numeric disparities. Previous studies have shown a widening of the temporal binding window in SCZ ([42]; see [85] for a review), which may be further investigated in SIFI paradigms with variable audiovisual asynchronies. Indeed, one previous study [42] has varied the audiovisual asynchrony of audiovisual stimuli to assess Bayesian causal inference in SCZ, individuals with autism spectrum disorder (ASD) and controls. Interestingly, the widening of the temporal binding window which was common to both ASD and SCZ could be attributed to different computational parameters. While it resulted from a stronger causal prior or binding tendency in ASD, it arose from larger sensory uncertainty in SCZ. Manipulating audiovisual asynchrony enables more precise estimation of observers’ sensory uncertainty than categorically manipulating the number of flashes and beeps as in our paradigm. Thus, it is possible that subtle increases in sensory uncertainty in SCZ may have gone undetected in our paradigm.

Future research is needed to determine whether deviations from normative Bayesian principles in multisensory perception may occur in larger samples of unmedicated psychotic patients in prodromal or acute stages or patient groups with more pronounced impairments on their attentional and executive functions. Critically, our study focused on simple artificial flash-beep stimuli. This makes it an important future research direction to investigate multisensory perceptual and causal inference in psychosis in more naturalistic situations such as face-to-face communication.

Materials and methods

Ethics statement

The study was performed in line with the principles of the Declaration of Helsinki, and it was approved by the human research review committee of the Medical Faculty of the University of Tuebingen and at the University Hospital Tuebingen (approval number 728/2014BO2).

Participants

After giving written informed consent, 24 healthy volunteers and 26 post-acute in- and out-patients participated in the EEG study. This small sample size was approximately in line with power calculations that assumed large effect sizes (Cohen’s d = 1.00) consistent with previous studies that reported large effect sizes for differences between SCZ and controls in multisensory processing [40,8082]. However, it is important to caution against the ‘winner’s curse’ phenomenon [82], i.e., the tendency for published studies to report inflated effect sizes compared to the actual effect in the population, thereby leading to overly optimistic power calculations for subsequent studies.

Participants were recruited via a psychiatry ward that was specialized for treatment of post-acute psychosis, leading to a nonrepresentative sample of chronic patients with high antipsychotic doses and low psychosis severity (Table 1 and S5 Fig). Healthy participants were included when meeting the inclusion criteria (age 20 to 65, adequate German comprehension, normal or corrected-to-normal vision and audition) and not the exclusion criteria (i.e., no psychiatric or psychosomatic disorders except specific phobias and nicotine addiction; no cardio-vascular disorders, diabetes, and neurological disorders). Healthy participants were matched to patients for age, sex, and education (Table 1). Current or former psychiatric disorders were screened using the questions of the structured clinical interview for DSM IV axis I disorders, SCID-I, German version. Patients were included when meeting the same inclusion and exclusion criteria as healthy controls (except psychiatric disorders). To clinically diagnose the patients, they underwent the full SCID-I interview (see below), and 19 of the 26 patients fulfilled the criteria for schizophrenia and were included (12 paranoid, 1 disorganized, 3 undifferentiated, 3 residual). Two of those 19 SCZ patients were excluded because they showed an unusual response profile in the CMB with a very small, outlying task effect (i.e., task effect <2 STD from the mean task effect) suggesting that these 2 patients only used the beeps to provide the auditory and visual numeric reports. Six patients fulfilled the criteria for schizoaffective disorder and 1 patient for substance-induced psychotic disorder and were excluded. Note that we obtained comparable results when we added the 6 schizoaffective patients to the SCZ sample (i.e., yielding n = 23 patients; S1 Text and S8S11 Figs and S6S8 Tables).

Of the 17 included SCZ patients in our main report, 4 exhibited comorbid major depression (lifetime), 1 patient alcohol dependence (lifetime), 3 patients cannabis dependence (lifetime, 1 also acute), 1 patient cocaine dependence (lifetime), 1 patient social phobia (lifetime and acute), 2 patients compulsive-obsessive disorder (lifetime), 1 patient binge-eating disorder (acute; for general demographics and other sample characteristics, see Table 1). The mean duration of the psychotic illness in SCZ patients was 13.37 ± 10.75 years (mean ± STD). Fifteen patients were medicated with atypical antipsychotics (amisulpride, aripriprazole, quetiapine, olanzapine, paliperidone), 1 patient was medicated with typical and atypical antipsychotics (haloperidol and clozapine), and 1 patient was unmedicated. Acute positive symptoms were rather moderate in the current sample (cf. PANSS Table 1).

One healthy participant did not attend the interview session and was excluded. One healthy participant reported an asymptomatic arteriovenous malformation. Because behavioral data and EEG were inconspicuous, the participant was included. Overall, 23 healthy controls and 17 SCZ patients were analyzed. Data from the 23 healthy participants were reported in [27].

Experimental procedures

Participants took part in 2 sessions on 2 separate days. In the first session, participants underwent clinical and neuropsychological assessments (Table 1). For HC, clinical assessments comprised the screening of the SCID-I interview [86]. For SCZ, clinical assessment comprised the full SCID-I interview, assessment of the Positive and Negative Symptom Scale (PANSS) [87], the German versions of the revised Launay Slade Hallucination Scale (LSHS-R) [88], the Paranoia Checklist with frequency items (PCL) [89], the Calgary Depression Scale for Schizophrenia (CDSS) [90], and a recording of clinical data (e.g., medication). PANSS measures the severity of clinical positive and negative symptoms as well as general psychopathology on seven-point Likert scales in the last 7 days using a semi-structured clinical interview. LSHS-R and PCL questionnaires measure self-reported subclinical and prodromal hallucinatory experiences and paranoid delusional thinking, respectively, on five-point Likert scales, also in nonclinical samples. Distributions of PANSS positive symptoms, LSHS-R and PCL are reported in S5 Fig. The antipsychotic medication was converted to chlorpromazine-equivalent doses [91]. Both groups underwent neuropsychological tests comprising the verbal learning and memory test (VLMT) [92], the Trail Making Test (TMT-A and -B) [93] assessing visual attention and executive functions (i.e., task switching), a test for premorbid crystallized intelligence (Mehrfachwahl-Wortschatz-Intelligenztest, MWT-B) [94], a Stroop test [95] assessing executive functions (i.e., cognitive inhibition), the Edinburgh Handedness Inventory (EHI) [96], and Beck’s Depression Inventory (BDI) [97]. All clinical and neuropsychological assessment were made by trained and experienced clinicians. In the second session, participants underwent a flash-beep paradigm including EEG measurements.

Stimuli

The flash-beep paradigm was an adaptation of the “sound-induced flash illusion” paradigm [57,98]. The visual flash was a circle presented in the center of the screen on a black background (i.e., 100% contrast; Fig 1A) briefly for 1 frame (i.e., 16.7 ms, as defined by the monitor refresh rate of 60 Hz). The maximum grayscale value (i.e., white) of the circle was at radius 4.5° with smoothed inner and outer borders by defining the grayscale values of circles of smaller and larger radius by a Gaussian of 0.9° STD visual angle. The auditory beep was a pure tone (2,000 Hz; approximately 70 dBSPL) of 27 ms duration including a 10 ms linear on/off ramp. Multiple visual flashes and auditory beeps were presented sequentially at a fixed SOA of 66.6 ms (see below).

Experimental design

In the flash-beep paradigm, participants were presented with a sequence of (i) 1, 2, 3, or 4 flashes; and (ii) 1, 2, 3, or 4 beeps (Fig 1A). On each trial, the number of flashes and beeps was independently sampled from 1 to 4 leading to 4 levels of numeric audiovisual disparities (i.e., zero = congruent to 4 = maximal level of disparity; Fig 1B). Each flash and/or beep was presented sequentially in fixed temporal slots that started at 0, 66.7, 133, 200 ms. The temporal slots were filled up sequentially. For example, if the number of beeps was 3, they were presented at 0, 66.6, 133, and 200 ms, while the fourth slot was left empty. Hence, if the same number of flashes and beeps was presented on a particular trial, beeps and flashes were presented in synchrony. On numerically disparate trials, the “surplus” beeps (or flashes) were added in the subsequent fixed time slots (e.g., in case of 2 flashes and 3 beeps: we present 2 flash-beeps at 0 and 66.6 ms in synchrony and a single beep at 133 ms).

Across experimental runs, we instructed participants to selectively report either the number of flashes or beeps and to ignore the stimuli in the task-irrelevant modality. Hence, the 4 × 4 × 2 factorial design manipulated (i) the number of visual flashes (i.e., 1, 2, 3, or 4); (ii) the number of auditory beeps (i.e., 1, 2, 3, or 4); and (iii) the task relevance (auditory- versus visual-selective report) yielding 32 conditions in total (Fig 1B). For analyses of the crossmodal bias, we reorganized trials based on their absolute audiovisual numeric disparity (|nA-nV| ∈ {1,2,3}).

The duration of a flash-beep sequence was determined by the number of sequentially presented flash and/or beep stimuli (see above for the definition of temporal slots). Irrespective of the number of flashes and/or beeps, a response screen was presented 750 ms after the onset of the first flash and beep for a maximum duration of 2.5 s instructing participants to report their perceived number of flashes (or beeps) as accurately as possible by pushing one of 4 buttons. The order of buttons was counterbalanced across runs to decorrelate motor responses from numeric reports. On half of the runs, the buttons from left to right corresponded to 1 to 4 stimuli; on the other half, they corresponded to 4 to 1. After a participant’s response, the next trial started after an inter-trial interval of 1 to 1.75 s.

In every experimental run, each of the 16 conditions was presented 10 times. Healthy control participants completed 4 runs of auditory- and 4 runs of visual-selective report in a counterbalanced fashion (except for 1 participant performing 5 runs of auditory and 3 of visual report) and 2 unisensory runs with visual or auditory stimuli only (i.e., 4 unisensory conditions presented 40 times per run). Thus, HC participants completed 10 runs in total. SCZ participants completed 4 to 8 runs of auditory- and visual-selective report, depending on their endurance, and 1 unisensory visual and 1 auditory run. Thus, SCZ participants completed 6 to 10 runs in total (7.47 ± 0.36, mean ± SEM number of runs). To control for time-dependent differences between groups, the trial numbers of behavioral and EEG data for each HC individual were subsampled to match the average trial number of the SCZ group across all analyses (i.e., trials 1–1186 were taken in HC). Before the actual experiment, participants completed 56 practice trials.

Experimental setup

Psychtoolbox 3.09 [99] (www.psychtoolbox.org) running under MATLAB R2016a (MathWorks) presented audiovisual stimuli and sent trigger pulses to the EEG recording system. Auditory stimuli were presented at ≈ 70 dB SPL via 2 loudspeakers (Logitech Z130) positioned on each side of the monitor. Visual stimuli were presented on an LCD screen with a 60 Hz refresh rate (EIZO FlexScan S2202W). Button presses were recorded using a standard keyboard. Participants were seated in front of the monitor and loudspeakers at a distance of 85 cm in an electrically shielded, sound-attenuated room.

Overview of GLM and Bayesian modeling analyses for behavioral data

We assessed similarities and differences in perceptual and causal inference between SCZ and HC by combining GLM-based and Bayesian modeling analysis approaches. The GLM-based analysis computed (i) the correlation between true and reported number of flashes/beep for congruent and unisensory trials; and (ii) the CMB which quantified the relative influence of the auditory and the visual numeric stimuli on observers’ auditory and visual behavioral numeric reports. The Bayesian modeling analysis fitted BCI models with different decisional strategies and additional heuristic models to the behavioral numeric reports. We then used Bayesian model comparison to determine the model(s) that were the best explanation for the behavioral data in SCZ or HC.

GLM-based analysis of behavioral data

Our GLM-based analysis of behavioral data focused on response accuracy indices for unisensory and congruent conditions and on the crossmodal bias for incongruent audiovisual conditions. To compare overall task performance between HC and SCZ, we computed participants’ response accuracy by correlating participants’ numeric report with the true task-relevant number of auditory or visual stimuli (i.e., Pearson correlation coefficient). We then analyzed Fisher z-transformed correlation coefficients in unisensory and audiovisual congruent conditions using a mixed-model ANOVA with within-participants factors modality (unisensory versus audiovisual) and task relevance (auditory versus visual report) and between-participant factor group (HC versus SCZ) (Fig 1C and Table 2).

We characterized how HC and SCZ participants weight audiovisual signals during multisensory perception by computing the CMB from their responses to audiovisual incongruent stimuli [28,29]. The CMB quantifies the relative influence of the auditory (nA) and the visual (nV) numeric stimuli on observers’ auditory or visual behavioral numeric reports (rA/V):

CMB=rA/VnAnVnA (1)

To account for response biases, we adjusted nA and nV with a linear regression approach across all congruent trials in a participant-specific fashion. In other words, we replaced the true nA and nV in the CMB equations with the nA and nV predicted based on participants’ responses during the congruent conditions [28]. The CMB should be one if observers’ reported auditory or visual numbers were solely influenced by the visual number. CMB should be zero if their reported auditory or visual number were solely influenced by the auditory number.

This CMB-based analysis allowed us to compare the audiovisual weight profiles between HC and SCZ using a mixed-model ANOVA with within-participant factors numeric disparity (1, 2, 3) and task relevance (auditory versus visual report) as well as between-participant factor group (HC versus SCZ) (Fig 1D and Table 2). Further, the analysis allowed us to assess whether the weight profiles were qualitatively in line with the principles of Bayesian causal inference. In particular, we would expect a numeric disparity by task-relevance interaction: the difference in crossmodal bias for visual relative auditory report should be greater for large relative to small numeric disparities. Complementary Bayesian ANOVAs were used to provide evidence for no difference between the groups (see section: “Statistical analysis of behavioral data” below).

Bayesian modeling analysis for behavioral data

The Bayesian modeling analysis fitted several competing computational models to the individual behavioral numeric reports. We then used factorial Bayesian model comparison to determine the model that was the best explanation for observers’ behavioral data and compared the results of the model comparison (Fig 3 and S2 Table) and model parameters between HC and SCZ (Fig 4 and Table 3).

In the following, we will briefly describe the competing models which spanned a 5 × 2 factorial model space. The first factor compared 5 model types that differed in their decision function, i.e., how they combined the estimates of the fusion and the segregation components into a final decision. The 3 BCI model combined the fusion and segregation estimates according to the decisional strategies of model averaging, model selection, and probability matching [30] (please see below for further specification). Two additional heuristic models reported the fusion or segregation estimates with a fixed probability (stochastic fusion model) or depending on whether the numeric disparity was greater than a fixed threshold (i.e., fixed-criterion threshold model; [31]). The second factor compared whether auditory and visual variances were equal across all number of beeps/flashes or increased with stimulus number (i.e., scalar-variability model; [52,53]). Details on the BCI model can be found in [23,31]. Specific details on fitting the BCI model to numeric reports in the current experimental paradigm can be found in [27].

Briefly, the generative model of the BCI model (Fig 3A) assumes that common (C = 1) or independent (C = 2) causes are determined by sampling from a binomial distribution with the causal prior p(C = 1) = pcommon (i.e., prior “binding tendency”) [35]. For a common cause, the “true” number of audiovisual stimuli NAV is drawn from the numeric prior distribution N(μP, σP). For 2 independent causes, the “true” auditory (NA) and visual (NV) numbers of stimuli are drawn independently from this numeric prior distribution. Sensory noise is introduced by drawing the sensory inputs xA and xV independently from normal distributions centered on the true auditory (respectively visual) number of stimuli with parameters σA (respectively σV). Thus, the basic generative model included the following free parameters: the causal prior pcommon, the numeric prior’s mean μP and standard deviation σP, the auditory standard deviation σA, and the visual standard deviation σV. Given the sensory inputs xA and xV, the observer infers the posterior probability of the underlying causal structure by combining the causal prior with the sensory evidence according to Bayes rule:

p(C=1|xA,xV)=p(xA,xV|C=1)pcommonp(xA,xV) (2)

The causal prior quantifies observers’ prior belief that flashes and beeps arise from a common cause and should hence be integrated. In the case of a common cause (C = 1), the optimal audiovisual numeric estimate (N^AV,C=1) is obtained by combining the auditory and visual numeric sensory inputs as well as the numeric prior weighted by their relative precisions (i.e., fusion estimate):

N^AV,C=1=xAσA2+xVσV2+μPσP21σA2+1σV2+1σP2 (3)

In the case of independent causes (C = 2), the optimal numeric estimates of the unisensory auditory (N^A,C=2) and visual (N^V,C=2) stimuli are independent (i.e., segregation estimate):

N^A,C=2=xAσA2+μPσP21σA2+1σP2,N^V,C=2=xVσV2+μPσP21σV2+1σP2 (4)

Crucially, the observers do not know whether signals come from common or independent sources, but needs to infer this from the sensory signals. To account for observers’ causal uncertainty, the model computes a final task-relevant numeric estimate (N^A or N^V) by combining the fusion estimate (i.e., N^AV,C=1 formed under C = 1 assumption) and task-relevant segregation estimates (i.e., either N^V,C=2 or N^A,C=2 for C = 2 assumption) depending on the posterior probabilities of the estimates’ underlying causal structures (p(C=1|xA,xV). The observer can combine the numerical estimates according to different decision strategies [30]:

In the “model averaging” strategy, the observers weigh the estimates in proportion to the posterior probabilities of their underlying causal structures:

N^A=p(C=1|xA,xV)N^AV,C=1+(1p(C=1|xA,xV))N^A,C=2N^V=p(C=1|xA,xV)N^AV,C=1+(1p(C=1|xA,xV))N^V,C=2 (5)

In the “model selection” strategy, the observer selects the numeric estimate that has a higher posterior probability:

N^A={N^A,C=1ifp(C=1|xA,xV)>0.5N^A,C=2ifp(C=1|xA,xV)0.5,N^V={N^V,C=1ifp(C=1|xA,xV)>0.5N^V,C=2ifp(C=1|xA,xV)0.5 (6)

In the “probability matching” strategy, the observer selects the numerical estimate stochastically in proportion to the posterior causal probabilities:

N^A={N^AV,C=1ifp(C=1|xA,xV)>αN^A,C=2ifp(C=1|xA,xV)α,N^V={N^AV,C=1ifp(C=1|xA,xV)>αN^V,C=2ifp(C=1|xA,xV)α,αU(0,1) (7)

We compared these 3 decision strategies with 2 additional heuristic strategies [31]. In a stochastic-fusion model, observers stochastically choose the audiovisual numeric (i.e., fusion) or the task-relevant unisensory (i.e., segregation) estimate with a fixed probability parameter η:

N^A={N^AV,C=1ifη>αN^A,C=2ifηα,N^V={N^AV,C=1ifη>αN^V,C=2ifηα,αU(0,1) (8)

The stochastic-fusion model encompasses a potential “forced” fusion (η = 1) or “forced” segregation (η = 0) as well as any intermediate response behavior (0 < η < 1). Importantly, the stochastic-fusion model responds irrespective of the audiovisual signals’ numeric disparity (cf. Fig 1B). Finally, the fixed-criterion model reports the fusion or segregation estimate whenever the signals’ absolute numeric disparity is below or above a fixed criterion k. Thus, the fixed-criterion model implements a response heuristic without formally computing the posterior probability of the causal structure as the BCI model:

N^A={N^AV,C=1if|xAxV|<kN^A,C=2if|xAxV|k,N^V={N^AV,C=1if|xAxV|<kN^V,C=2if|xAxV|k (9)

For the second model factor, we implemented 2 sensory noise models: Sensory standard deviations were either constant across stimulus numbers or increased in proportion to the stimulus controlled by an incremental sensory standard deviation parameter Δσ:

σA=σA+σA(SA1),σV=σV+σV(SV1) (10)

This additional increase in standard deviation with stimulus number incorporates the notion of scalar variability. Additionally, all models included a lapse rate parameter L to account for random uniformly distributed responses independent of the audiovisual inputs.

We fitted each of the 10 models in our 5 decision-strategy × 2 sensory-noise factorial model space to the numeric reports individually for each participant (for details on the fitting procedure, see [27]). For more efficient parameter estimation, we fitted models jointly to unisensory and audiovisual conditions. Thus, the auditory/visual standard deviations and their increments, the numeric prior’s mean and standard deviation as well as the lapse rate were jointly informed by unisensory and audiovisual conditions.

To obtain maximum likelihood estimates for the 8 parameters of the models (pcommon / k / η, μP, σP, σA, σV, ΔσA, △σV, L), we used a Bayesian optimization algorithm as implemented in the BADS toolbox [100]. This optimization algorithm was initialized with 50 different random parameters. We report the results (i.e., model comparisons and parameters) for models with the highest log likelihood across these initializations. Parameter recovery showed that the parameters could be reliably identified with very small biases using this fitting procedure (S12 Fig). Further, model recovery [31,101] demonstrated that the BCI and heuristic models were identifiable and our model fitting procedure was reliable (S13 Fig).

To identify the optimal model for explaining participants’ data, we compared the 10 candidate models using the Bayesian information criterion (BIC) as an approximation to the model evidence [102]. We performed Bayesian model comparison at the random-effects group level [103], separately for HC and SCZ, as implemented in SPM12 [104]. Thus, we obtained the protected exceedance probability, i.e., the probability that a given model is more likely than any other model, beyond differences due to chance [103], for each of the 10 candidate models (S2 Table). Further, we performed factorial Bayesian model selection on the two-factorial model space using robust Bayesian model-selection procedures (L. Acerbi, personal communication) to compute protected exceedance probabilities for each family across the 2 factors in our 2 × 5 factorial model space (Fig 3B). To assess whether the frequencies of the optimal model differ between HC and SCZ, we applied between-group Bayesian model comparison [103] across the 10 candidate models as implemented in the variational Bayesian approach toolbox [105]. Using the posterior probability that the 2 groups have the same model frequencies, we computed the Bayes factor BF10 quantifying the evidence in favor of different (H1) or equivalent (H0) model frequencies (i.e., assuming equal prior probabilities).

To investigate whether HC and SCZ differ in parameters of the winning model (i.e., the BCI model with “model averaging” and increasing sensory variances), we compared the parameters between groups using a two-sample two-sided randomization t test (n = 5,000 randomizations; Table 3). Further, we correlated the parameters with SCZ patients’ positive and negative symptoms as well as general psychopathy as measured with the PANSS using a randomization test of the correlation (n = 5,000 randomizations).

To investigate whether HC and SCZ differentially adjusted the causal priors depending on the stimulus history, we sorted the current trials according to the previous trials’ absolute audiovisual numeric disparity (i.e., |nA-nV| ∈ {0–3}) and selectively refitted the causal prior of the BCI model (Fig 5A). We entered the causal prior pcommon into a 4 (numeric disparity on previous trial: 0, 1, 2, 3) × 2 group (HC versus SCZ) mixed-model ANOVA. Likewise, we investigated how observers adapted the numeric prior by sorting the current trials according to the previous trials’ task-relevant auditory/visual signals (i.e., nA or nV ∈ {1–4}) and selective refitting of the numeric prior (i.e., numeric prior’s mean μP and standard deviation σP; Fig 5A and 5B). We entered the numeric prior’s mean μP and standard deviation σP into separate 4 (number of task-relevant signals on previous trial: 1, 2, 3, 4) × 2 group (HC versus SCZ) mixed-model ANOVAs. To provide evidence for the null-hypothesis, i.e., no difference between groups, these model parameters were also entered into separate Bayesian mixed-model ANOVAs (see statistical analyses below). Parameter recovery showed that the causal and numeric prior could be reliably identified as a function of the previous trials’ conditions with very small biases using this fitting procedure (S14 Fig).

To generate predictions for the behavioral CMB index and EEG analyses (see below) based on the BCI model, we simulated new xA and xV for 10,000 trials for each of the 32 conditions using the fitted BCI model parameters of each participant (i.e., BCI model with model averaging and increasing sensory variances). For each simulated trial, we computed the BCI model’s (i) unisensory visual (N^V,C=2); (ii) unisensory auditory (N^A,C=2) estimates; (iii) forced-fusion (N^AV,C=1); (iv) final BCI audiovisual numeric estimate (N^A or N^V depending on whether the auditory or visual modality was task relevant). Next, we used the mode of the resulting (kernel-density estimated) distributions for each condition and participant to compute the model predictions for the CMB index (Fig 1D) and decoding the BCI model’s estimates from EEG patterns (see multivariate EEG analysis; Fig 6).

Statistical analyses of behavioral data

To refrain from making parametric assumptions, we compared behavioral measures and parameters (i.e., neuropsychological test scores, CMB and BCI model parameters; Tables 1 and 3) between conditions and groups as well as tested correlations between BCI model parameters and symptom scores using randomization tests with t-values as test statistics (n = 5,000 randomizations). To assess the evidence in favor of the null-hypothesis (i.e., no difference between groups or conditions), we also performed complementary Bayesian analyses. Randomization tests were complemented by Bayes factors BF10 quantifying evidence in favor of a condition or group difference or correlation (H1) relative to the null hypothesis of no difference or correlation (H0) using the bayesFactor toolbox [106]. Bayesian t tests and correlations assumed a Jeffrey–Zellner–Siow prior and t tests assumed a scaling factor s = 0.707 of the Cauchy prior on effects [107].

To evaluate whether experimental factors influenced HC and SCZ differentially or equivalently (e.g., for CMB), we complemented classical with Bayesian mixed-model ANOVAs (Tables 2 and 4) with multivariate Cauchy priors on the effects and uniform model priors. Bayesian ANOVAs allowed to compute inclusion Bayes factors (BFincl) for each experimental factor which quantify the average evidence that including the factor in the ANOVA models improves the models’ fit given a higher complexity of these models [108,109]. For Bayesian analyses, the following interpretations of BF apply: BF > 3 or > 10 provide substantial or strong evidence [110,111] for condition/group differences, correlations or inclusion of a factor, whereas BF < 1/3 or < 1/10 provide substantial or strong evidence for condition/group equivalence, no correlation or exclusion the factor [109]. All classical and Bayesian ANOVAs were computed using JASP 0.17.2.1 [108].

EEG—Data acquisition and preprocessing

EEG signals were recorded from 64 active electrodes positioned in an extended 10–20 montage using electrode caps (actiCap, Brain Products, Gilching, Germany) and two 32 channel DC amplifiers (BrainAmp, Brain Products). Electrodes were referenced to FCz using AFz as ground during recording. Signals were digitized at 1,000 Hz with a high-pass filter of 0.1 Hz. Electrode impedances were kept below 25 kOhm. Preprocessing of EEG data was performed using Brainstorm 3.462 running on Matlab R2015b. EEG data were band-pass filtered (0.25 to 45 Hz). Eye blinks were automatically detected using data from the FP1 electrode (i.e., a blink was detected if the band-pass (1.5 to 15 Hz) filtered EEG signal exceeded 2 times the STD; the minimum duration between 2 consecutive blinks was 800 ms). Signal-space projectors (SSPs) were created from band-pass filtered (1.5 to 15 Hz) 400 ms segments centered on detected blinks. The first spatial component of the SSPs was then used to correct blink artifacts in continuous EEG data. Further, all data were visually inspected for artifacts from blinks (i.e., residual blink artifacts after correction using SSPs), saccades, motion, electrode drifts, or jumps and contaminated segments were discarded from further analysis. On average, we discarded 6.171 ± 0.923% SEM of all trials for HC and 7.216 ± 1.444% for SCZ; the difference was not significant, t38 = −0.638, p = 0.528, d = −0.204, BF10 = 0.366. Finally, EEG data were re-referenced to the average of left and right mastoid electrodes and downsampled to 200 Hz. For analysis of ERPs and decoding analyses (see below), all EEG data were baseline corrected with a 200 ms prestimulus baseline and were analyzed from 100 ms before stimulus onset up to 750 ms after stimulus onset, when the response screen was presented.

For multivariate analyses (see below), single-trial EEG data from the 64 electrodes were binned in sliding time windows of 60 ms using an overlap of 40 ms. Hence, given a sampling rate of 200 Hz, each 60 ms time window included 12 temporal sampling points; 64-electrode EEG activity vectors (for each time sample) were concatenated across the 12 sampling points within each bin resulting in a spatiotemporal EEG activity pattern of 768 features. EEG activity patterns were z scored to control for mean differences between conditions. The first sampling point in the 60 ms time window was taken as the window’s time point in all analyses.

EEG—Analysis of audiovisual interactions in ERPs

In univariate analyses, we assessed whether schizophrenia alters basic sensory components and early audiovisual interactions in occipital ERPs. We averaged trial-wise EEG data time-logged to stimulus onset into ERPs for audiovisual congruent conditions (i.e., controlling for effects of disparity and averaging over effects of task-relevance) and unisensory conditions. We then averaged the ERPs across occipital electrodes (i.e., O1, O2, Oz, PO3, POz, PO4; Fig 7). To analyze early audiovisual interactions, we computed the difference between congruent audiovisual conditions and the corresponding unisensory conditions (i.e., AVcongr–(A + V)). Because attentional and cognitive sets between uni- and bisensory runs might have differed and our experimental design did not include null trials to account for anticipatory effects around stimulus onset [112], the audiovisual interactions need to be interpreted with caution. To test whether the difference waves deviated from zero at the group level across HC and SCZ, we used a nonparametric randomization test (5,000 randomizations) in which we flipped the sign of the individual difference waves using a one-sample t test as a test statistic [113]. To compare the difference waves between both groups, we used a randomization test in which we flipped group membership (5,000 randomizations) using two-sample t tests as a test statistic. To correct for multiple comparisons across the EEG sampling points, we used a cluster-based correction [114] with the sum of the t values across a cluster as cluster-level statistic and an auxiliary cluster-defining threshold of t = 2 for each time point. In a supplementary analysis, we used a group classification approach [72], in which we assessed whether the multivariate EEG response patterns to unisensory and multisensory stimuli differed between HC and SCZ (S1 Text and S7 Fig).

EEG—Multivariate analyses of BCI estimates

To characterize whether the neurophysiological processes underlying Bayesian causal inference differ between HC and SCZ, we decoded the 4 numerical estimates of the BCI model from multivariate EEG patterns across poststimulus time points using linear support-vector regression [27,36] as implemented in LibSVM 3.20 [115]. The individually fitted BCI model provided the (i) unisensory visual (N^V,C=2); (ii) unisensory auditory (N^A,C=2) estimates; (iii) forced-fusion (N^AV,C=1); (iv) final BCI audiovisual numeric estimate (N^A or N^V) for each of the 32 audiovisual conditions (see above). For each numeric estimate and each time window, an SVR model was trained to decode the numeric estimate from single-trial EEG activity patterns (see above for a definition) across all 32 conditions in a leave-one-run-out cross validation scheme. The SVRs’ parameters (C and ν) were optimized using a grid search within each cross-validation fold (i.e., nested cross-validation).

To quantify the decoding accuracies, we computed the Pearson correlation coefficients between the “true” model estimates and the decoded model estimates (Fig 6A). The Fisher’s z-transformed correlation coefficients were tested against zero across HC and SCZ using a one-sided randomization test (i.e., sign flip of correlation coefficient in 5,000 randomizations; 1 sample t test as test statistic). The Fisher’s z-transformed correlation coefficients were compared between HC and SCZ using a two-sided randomization test (i.e., 5,000 randomizations; two-sample t tests as a test statistic). Further, we applied Bayesian two-sample t tests (as described above for behavioral data) to compute Bayes factors that quantify evidence for or against group differences in decoding accuracies across poststimulus time points (Fig 6B).

DRYAD DOI

All raw EEG and behavioral data are available from a Dryad repository [116].

Supporting information

S1 Text. Supporting methods and results.

(DOCX)

pbio.3002790.s001.docx (23KB, docx)
S1 Fig. Numeric reports and predictions from a log-linear regression model as well as the BCI model as a function of the visual signal number (nV), auditory signal number (nA), and task relevance, separately for HC (left 2 columns, n = 23) and SCZ (right 2 columns, n = 17).

(A) Mean numeric reports (across participants mean ± SEM). (B) Mean predictions (across-participants mean) from a log-linear regression model that predicted the numeric reports from the logarithmic visual and auditory signal numbers as well as their interaction (i.e., rA/V = bA * log(nA) + bV * log(nV) + bAxV * log(nV) * log(nA) + c). Statistical results for the regression model can be found in S1 Table. (C) Mean predictions from the BCI model (across-participants mean ± SEM; model-averaging with increasing sensory variances) reproduced the logarithmic compression of numeric reports.

(DOCX)

pbio.3002790.s002.docx (457.7KB, docx)
S2 Fig. Overall reaction time (RT) density in HC and SCZ (n = 40).

When fitting ex-Gaussian functions to the RT distributions, the functions’ mean (t38 = −1.325, p = 0.193, d = −0.424, BF10 = 0.620), variance (t38 = −0.511, p = 0.612, d = −0.164, BF10 = 0.346), and lambda parameter (t38 = −0.607, p = 0.548, d = −0.194, BF10 = 0.361) were not significantly different, but rather equivalent, between both groups.

(DOCX)

pbio.3002790.s003.docx (26.3KB, docx)
S3 Fig. The strength of the fission and fusion illusions (across-participants mean ± SEM) is shown as a function of group (HC vs. SCZ, n = 40).

The illusions were computed as difference in sensitivity (d prime, d’) between the unisensory baseline condition (V1A0 vs. V2A0) and the illusion conditions (fission: V1A2 vs. V2A2; fusion: V1A1 vs. V2A2). For the fission or fusion conditions, response “2” or “1,” respectively, were defined as signal. Thus, this illusion measure accounts for a possible shift in the response criterion which could be confounded with the fission or fusion illusions (Vanes et al, 2016). In contrast to Vanes and colleagues, a mixed-model ANOVAs did not reveal a significant difference in illusion strength between the 2 types of illusions (factor audiovisual illusion, F1,38 = 2.598, p = 0.115, part. η2 = 0.064, BFincl = 0.461) or a difference between HC and SCZ participants (factor group: F1,38 = 1.431, p = 0.239, part. η2 = 0.036, BFincl = 0.440; interaction illusion × group: F1,38 = 1.271, p = 0.267, part. η2 = 0.032, BFincl = 0.273). Note that we only included trials in which participants counted the number of flashes as ≤ 2.

(DOCX)

pbio.3002790.s004.docx (25.2KB, docx)
S4 Fig. Bayesian information criterion values for each of the 5 models with increasing sensory variance are shown for each of the 17 patients in the SCZ group ranked according to their PANSS positive score as an index of psychosis severity.

Model comparison was computed as BIC of each model relative to the fixed-criterion model with increasing sensory variance, i.e., higher is better. The model comparison included 5 models with increasing sensory variance and the 5 decision strategies (MA, model averaging; MS, model selection; PM, probability matching; FC, fixed-criterion model; SF, stochastic fusion model). The SCZ patients were rank-ordered according to their PANSS positive score on the x-axis. In cases of ties, patients were arbitrarily rank-ordered within their pair according to participant number.

(DOCX)

pbio.3002790.s005.docx (122.4KB, docx)
S5 Fig. Characterization of psychosis severity.

(A) Data (across-participants mean ± SEM; n = 17) of the PANSS scale for positive symptoms (7 items, range 1–7) and the scale average in the SCZ sample. (B) Sum score from the LSHS-R questionnaire (12 items, range 0–48) for a subset of the HC sample (n = 5) and the SCZ sample (n = 17). The difference between both groups is significant (paired t test, t20 = −2.216, p = 0.039, d = −1.127). (C) Sum score (across participants mean ± SEM) from the PCL questionnaire (18 items, range 18–90) for a subset of the HC sample (n = 5) and the SCZ sample (n = 16). The difference between both groups is marginally significant (paired t test, t19 = −2.0271, p = 0.057, d = −1.039). Importantly, both the LSHS-R and the PCL scores varied substantially across the patient group. This allowed us to assess whether these sensitive measures of psychotic symptoms may correlate with BCI model parameters (S3 Table). PANSS: Positive and Negative Symptom Scale; LSHS-R: Launey–Slade hallucination scale; PCL: Paranoia check list.

(DOCX)

pbio.3002790.s006.docx (163.3KB, docx)
S6 Fig. Modulation of crossmodal bias (CMB) and numeric reports by previous audiovisual numeric disparity and task-relevant signal number.

(A) The audiovisual crossmodal bias (CMB; across-participants mean ± SEM; n = 40) is shown as a function of a current trial’s numeric disparity (1, 2, or 3) × previous trial’s numeric disparity (0, 1, 2, 3) × task relevance (auditory vs. visual report) × group (HC vs. SCZ). A four-factorial mixed-model ANOVA including these factors revealed a marginally significant task-relevance × previous-disparity interaction (F2.1,80.9 = 2.616, p = 0.076, part. η2 = 0.064; effects of task-relevance, current numeric disparity and their interaction were p < 0.05 similarly as reported in Table 2; all other main and interactions effects, especially involving group, p > 0.05). Moreover, we assessed the more constrained hypothesis that the difference between CMB for A and V report would increase linearly with increasing numeric disparity of the previous trial: We applied contrast weights to the A-V report differences depending on the disparity at previous trial, i.e., −1.5 for 0, −0.5 for 1, 0.5 for 2, and 1.5 for 3 audiovisual numeric disparity on the previous trial. This linear contrast showed a significant linear modulation of the task relevance effect on the CMB by previous disparity (i.e., linear contrast on the interaction of previous disparity and task relevance; t114 = 2.712, p = 0.008). In other words, audiovisual interactions increased when the previous numeric disparity was small, but decreased when the previous numeric disparity was large. CMB = 1 for purely visual and CMB = 0 for purely auditory influence. (B) Visual (left panel) and auditory (right panel) numeric reports (across-participants mean ± SEM, n = 40) plotted as a function of current visual or auditory signal number, previous visual or auditory signal number and group (HC vs. SCZ). A three-factorial mixed-model ANOVA including these factors revealed a significant main effect of previous signal number (F1.7,64.7 = 33.192, p < 0.001, part. η2 = 0.466), an interaction of previous signal number and task relevance (F3,114 = 3.986, p = 0.010, part. η2 = 0.095), and a significant interaction of previous signal number and current signal number (F6.6,250.2 = 8.515, p < 0.001, part. η2 = 0.183). The effect of current signal number and the interaction of current signal number and task relevance were also significant (p < 0.05). All other main and interactions effects, especially involving group, were not significant.

(DOCX)

pbio.3002790.s007.docx (192.2KB, docx)
S7 Fig. Multivariate ERP patterns, the spatial correlation of ERP patterns between HC and SCZ and decoding results (n = 40).

(A) ERP patterns (i.e., topographies) are shown as a function of time, averaged in 100 ms time windows and separately for HC and SCZ participants. (B) Spatial correlation of multivariate ERP patterns between HC and SCZ participants as a function of time, separately for the 4 conditions (N.B.: Spatial correlations before stimulus onset might arise from expectation processes). (C) Decoding accuracy (i.e., fraction of correct classifications) of a decoder trained to classify the participant group (i.e., HC vs. SCZ) from ERP patterns of the 4 conditions. No significant clusters (p > 0.05) of decoding accuracies above the chance level of 0.5 were found in one-sided cluster-based corrected randomization tests. As a reference, the gray dashed lines indicate decoding accuracy (fraction ± 68% CI; averaged across the 4 conditions) of the decoder trained on randomized group membership (n = 5,000 randomizations).

(DOCX)

pbio.3002790.s008.docx (331.7KB, docx)
S8 Fig. Crossmodal bias for the clinical sample including schizophrenia (SCZ, n = 17) and schizoaffective (SCA, n = 6) patients.

Participants’ audiovisual crossmodal biases (CMB; across-participants mean ± SEM; n = 46) are shown as a function of numeric disparity (1, 2, or 3), task relevance (auditory vs. visual report), and group (HC vs. SCZ). CMB = 1 for purely visual and CMB = 0 for purely auditory influence. Source data is provided in S8 Data.

(DOCX)

pbio.3002790.s009.docx (68.6KB, docx)
S9 Fig. Factorial Bayesian model comparison including SCZ (n = 17) and SCA patients (n = 6).

Ten models with different decision strategies, model averaging (MA), model selection (MS), probability matching (PM), fixed criterion (FC), stochastic fusion (SF), each with constant or increasing sensory auditory and visual variances. The image shows the relative model evidence for each model (i.e., participant-specific Bayesian information criterion of a model relative to the worst model summed over all participants). A larger model evidence indicates that a model provides a better explanation of our data. Source data is provided in S9 Data.

(DOCX)

pbio.3002790.s010.docx (76.9KB, docx)
S10 Fig. Parameters of the BCI model (across participants mean ± SEM) separately plotted for HC and SCZ (n = 17) & SCA (n = 6) patients.

The BCI model’s decision strategy applies model averaging with increasing sensory variance. Source data is provided in S10 Data.

(DOCX)

pbio.3002790.s011.docx (126.2KB, docx)
S11 Fig. Decoding the BCI model’s numeric estimates from EEG patterns using support-vector regression (SVR) in HC versus SCZ (n = 40) & SCA (n = 6) patients.

(A) Decoding accuracy (Fisher’s z-transformed correlation; across-participants mean) of the SVR decoders as a function of time and group (HC vs. SCZ & SCA). Decoding accuracy was computed as Pearson correlation coefficient between the given BCI model’s internal estimates and BCI estimates that were decoded from EEG activity patterns using SVR models trained separately for each numeric estimate. Color-coded horizontal solid lines (HC) or dashed lines (SCZ & SCA) indicate clusters of significant decoding accuracy (p < 0.05; one-sided one-sample cluster-based corrected randomization t test). (B) Bayes factors for the comparison between the decoding accuracies of HC and SCZ & SCA for each of the BCI estimate (i.e., BF10 > 3 substantial evidence for or BF10 < 1/3 against group differences). Source data is provided in S11 Data.

(DOCX)

pbio.3002790.s012.docx (131.5KB, docx)
S12 Fig. Results of the parameter recovery (n = 40).

For parameter recovery, the winning model (i.e., model averaging with increasing sensory variances) predicted responses which were then again fitted to obtain recovered parameters with the same fitting procedure as for the main analysis (i.e., initialization with 50 different random parameters; predicted distributions were generated from 5,000 simulated trials per condition). The plots show the recovered parameters as a function of the parameters originally fitted to participants’ behavioral data. The red line is a line with slope 1 and intercept 0.

(DOCX)

pbio.3002790.s013.docx (248.4KB, docx)
S13 Fig. Results of model recovery.

To assess model recovery, each of the 5 models with different decision strategies (model averaging, MA; model selection, MS; probability matching, PM; fixed criterion, FC; stochastic fusion, SF) and increasing sensory variance parameters generated responses in 20 randomly selected participants. Each of the 5 models was then fitted to the generated responses of each model with the same fitting procedure as for the main analysis (i.e., initialization with 50 different random parameters; predicted distributions were generated from 5,000 simulated trials per condition). The confusion matrix shows for each generating model (rows) the fraction of participants in which a fitted model (columns) won the model comparison across the 5 fitted models based on Bayesian information criterion within an individual. The bright diagonal indicates that the generating models won the model comparisons, indicating successful model recovery. In particular, the winning model with model averaging as decision strategy was correctly recovered in 85% of the cases, thereby supporting the validity of our model-based analyses.

(DOCX)

pbio.3002790.s014.docx (67.5KB, docx)
S14 Fig. Results of the parameter recovery (HC and SCZ, n = 40) for the modeling analysis of updated causal and numeric priors.

For parameter recovery, the model averaging model with updated priors predicted responses which were then again fitted to obtain recovered parameters with the same fitting procedure as for the main analysis (i.e., initialization with 50 different random parameters; predicted distributions were generated from 5,000 simulated trials per condition). The plots show the recovered parameters as a function of the parameters originally fitted to participants’ behavioral data. Parameters are pooled across different levels of previous trial conditions. The red line is a line with slope 1 and intercept 0.

(DOCX)

pbio.3002790.s015.docx (134.8KB, docx)
S1 Table. Statistical significance of regression parameter estimates across and between HC and SCZ in the log-linear regression model that predicted the numeric reports from the visual and auditory signal numbers as well as their interaction.

(DOCX)

pbio.3002790.s016.docx (15.7KB, docx)
S2 Table. Results of the Bayesian model comparison of the 2 × 5 factorial model space with factor “decision strategy” and “sensory variance” in HC and SCZ participants.

(DOCX)

pbio.3002790.s017.docx (24.3KB, docx)
S3 Table. Pearson correlations of BCI model parameters with SCZ patients’ (n = 17) positive and negative symptoms measured by PANSS, LSHS-R, and PCL.

(DOCX)

pbio.3002790.s018.docx (17.8KB, docx)
S4 Table. Results of testing the decoding accuracies of the BCI estimates from EEG patterns against zero in HC, SCZ, and their group difference.

(DOCX)

pbio.3002790.s019.docx (14.9KB, docx)
S5 Table. Spearman rank correlations of BCI model parameters with SCZ patients’ neurocognitive test scores (n = 17).

(DOCX)

pbio.3002790.s020.docx (15.8KB, docx)
S6 Table. Main and interaction effects of task relevance (TR), disparity (Disp), and group (HC, n = 23 vs. schizophrenia, n = 17, and schizoaffective patients, n = 6) on the crossmodal bias from classical and Bayesian mixed-model ANOVAs.

(DOCX)

pbio.3002790.s021.docx (14.7KB, docx)
S7 Table. Comparison of BCI model parameters between HC (n = 23) versus schizophrenia (n = 17) & schizoaffective (n = 6) patients.

(DOCX)

pbio.3002790.s022.docx (14.6KB, docx)
S8 Table. Correlations of BCI model parameters with patients’ (schizophrenia, n = 17 and schizoaffective, n = 6) positive and negative symptoms measured by PANSS, LSHS-R, and PCL.

(DOCX)

pbio.3002790.s023.docx (17KB, docx)
S1 Data. Source data for Fig 1 which can be opened with Matlab or GNU Octave.

(ZIP)

pbio.3002790.s024.zip (19.3KB, zip)
S2 Data. Source data for Fig 2 which can be opened with Matlab or GNU Octave.

(ZIP)

pbio.3002790.s025.zip (17.3KB, zip)
S3 Data. Source data for Fig 3 which can be opened with Matlab or GNU Octave.

(ZIP)

pbio.3002790.s026.zip (16.1KB, zip)
S4 Data. Source data for Fig 4 which can be opened with Matlab or GNU Octave.

(ZIP)

pbio.3002790.s027.zip (33.9KB, zip)
S5 Data. Source data for Fig 5 which can be opened with Matlab or GNU Octave.

(ZIP)

pbio.3002790.s028.zip (17.1KB, zip)
S6 Data. Source data for Fig 6 which can be opened with Matlab or GNU Octave.

(ZIP)

pbio.3002790.s029.zip (65.8KB, zip)
S7 Data. Source data for Fig 7 which can be opened with Matlab or GNU Octave.

(ZIP)

pbio.3002790.s030.zip (165.1KB, zip)
S8 Data. Source data for S8 Fig which can be opened with Matlab or GNU Octave.

(ZIP)

pbio.3002790.s031.zip (15.4KB, zip)
S9 Data. Source data for S9 Fig which can be opened with Matlab or GNU Octave.

(ZIP)

pbio.3002790.s032.zip (16.5KB, zip)
S10 Data. Source data for S10 Fig which can be opened with Matlab or GNU Octave.

(ZIP)

pbio.3002790.s033.zip (37KB, zip)
S11 Data. Source data for S11 Fig which can be opened with Matlab or GNU Octave.

(ZIP)

pbio.3002790.s034.zip (73.7KB, zip)

Acknowledgments

We thank Luigi Acerbi for sharing code to perform robust Bayesian model-selection procedures which was updated by David Meijer. We thank Ramona Täglich and Larissa Metzler for help with the data collection.

Abbreviations

ASD

autism spectrum disorder

BCI

Bayesian causal inference

BDI

Beck’s Depression Inventory

BIC

Bayesian information criterion

CDSS

Calgary Depression Scale for Schizophrenia

CMB

crossmodal bias

EEG

electroencephalography

EHI

Edinburgh Handedness Inventory

ERP

event-related potential

GLM

general linear model

HC

healthy control

MWT-B

Mehrfachwahl-Wortschatz-Intelligenztest

PANSS

Positive and Negative Symptom Scale

PCL

Paranoia Checklist

SCZ

schizophrenia

SIFI

sound-induced flash-illusion

SLHS-R

Launay Slade Hallucination Scale

SSP

signal-space projector

SVC

support-vector classifier

SVR

support-vector regression

TMT

Trail Making Test

TR

task relevance

VLMT

verbal learning and memory test

Data Availability

All raw EEG and behavioral data as well as code for presenting audiovisual stimuli, fitting computational models and analysis of behavioral data are available from a Dryad repository (https://doi.org/10.5061/dryad.hhmgqnkr1) [116].

Funding Statement

This work was supported by the Deutsche Forschungsgemeinschaft (https://www.dfg.de/; RO 5587/ 1–1 to TR) and the University of Tübingen (https://uni-tuebingen.de/; Fortüne award 2292–0–0 and 2454–0–0 to TR). Publication costs were covered by the Open Access Publishing Fund of the University of Tübingen. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1.Huys QJ, Daw ND, Dayan P. Depression: a decision-theoretic analysis. Annu Rev Neurosci. 2015;38:1–23. Epub 20150211. doi: 10.1146/annurev-neuro-071714-033928 . [DOI] [PubMed] [Google Scholar]
  • 2.Knill DC, Pouget A. The Bayesian brain: the role of uncertainty in neural coding and computation. Trends Neurosci. 2004;27(12):712–9. Epub 2004/11/16. doi: 10.1016/j.tins.2004.10.007 . [DOI] [PubMed] [Google Scholar]
  • 3.Ma WJ. Bayesian Decision Models: A Primer. Neuron. 2019;104(1):164–175. doi: 10.1016/j.neuron.2019.09.037 . [DOI] [PubMed] [Google Scholar]
  • 4.Noppeney U. Perceptual Inference, Learning, and Attention in a Multisensory World. Annu Rev Neurosci. 2021;44. [DOI] [PubMed] [Google Scholar]
  • 5.Parr T, Rees G, Friston KJ. Computational Neuropsychology and Bayesian Inference. Front Hum Neurosci. 2018;12:61. Epub 20180223. doi: 10.3389/fnhum.2018.00061 ; PubMed Central PMCID: PMC5829460. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Pouget A, Beck JM, Ma WJ, Latham PE. Probabilistic brains: knowns and unknowns. Nat Neurosci. 2013;16(9):1170–8. Epub 20130818. doi: 10.1038/nn.3495 ; PubMed Central PMCID: PMC4487650. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Vilares I, Howard JD, Fernandes HL, Gottfried JA, Kording KP. Differential representations of prior and likelihood uncertainty in the human brain. Curr Biol. 2012;22(18):1641–8. Epub 20120726. doi: 10.1016/j.cub.2012.07.010 ; PubMed Central PMCID: PMC3461114. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Adams RA, Stephan KE, Brown HR, Frith CD, Friston KJ. The computational anatomy of psychosis. Front Psych. 2013;4:47. doi: 10.3389/fpsyt.2013.00047 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Fletcher PC, Frith CD. Perceiving is believing: a Bayesian approach to explaining the positive symptoms of schizophrenia. Nat Rev Neurosci. 2009;10(1):48–58. Epub 2008/12/04. doi: 10.1038/nrn2536 [pii]. . [DOI] [PubMed] [Google Scholar]
  • 10.Cassidy CM, Balsam PD, Weinstein JJ, Rosengard RJ, Slifstein M, Daw ND, et al. A perceptual inference mechanism for hallucinations linked to striatal dopamine. Curr Biol. 2018;28(4):503–14.e4. doi: 10.1016/j.cub.2017.12.059 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Powers AR, Mathys C, Corlett P. Pavlovian conditioning–induced hallucinations result from overweighting of perceptual priors. Science. 2017;357(6351):596–600. doi: 10.1126/science.aan3458 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Adams RA, Napier G, Roiser JP, Mathys C, Gilleen J. Attractor-like dynamics in belief updating in schizophrenia. J Neurosci. 2018;38(44):9471–9485. doi: 10.1523/JNEUROSCI.3163-17.2018 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Benrimoh D, Fisher VL, Seabury R, Sibarium E, Mourgues C, Chen D, et al. Evidence for reduced sensory precision and increased reliance on priors in hallucination-prone individuals in a general population sample. Schizophr Bull. 2024;50(2):349–362. doi: 10.1093/schbul/sbad136 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Kafadar E, Fisher VL, Quagan B, Hammer A, Jaeger H, Mourgues C, et al. Conditioned hallucinations and prior overweighting are state-sensitive markers of hallucination susceptibility. Biol Psychiatry. 2022;92(10):772–780. doi: 10.1016/j.biopsych.2022.05.007 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Teufel C, Subramaniam N, Dobler V, Perez J, Finnemann J, Mehta PR, et al. Shift toward prior knowledge confers a perceptual advantage in early psychosis and psychosis-prone healthy individuals. Proc Natl Acad Sci U S A. 2015;112(43):13401–13406. doi: 10.1073/pnas.1503916112 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Jardri R, Duverne S, Litvinova AS, Denève S. Experimental evidence for circular inference in schizophrenia. Nat Commun. 2017;8(1):1–13. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Schmack K, Rothkirch M, Priller J, Sterzer P. Enhanced predictive signalling in schizophrenia. Hum Brain Mapp. 2017;38(4):1767–1779. doi: 10.1002/hbm.23480 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Valton V, Karvelis P, Richards KL, Seitz AR, Lawrie SM, Seriès P. Acquisition of visual priors and induced hallucinations in chronic schizophrenia. Brain. 2019;142(8):2523–2537. doi: 10.1093/brain/awz171 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Weilnhammer V, Röd L, Eckert AL, Stuke H, Heinz A, Sterzer P. Psychotic Experiences in Schizophrenia and Sensitivity to Sensory Evidence. Schizophr Bull. 2020;46(4):927–936. doi: 10.1093/schbul/sbaa003 ; PubMed Central PMCID: PMC7345814. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Fletcher PC, Teufel CR. The Changing Weight of Expectation: How Shifting Priors Underpin Variability in Hallucination Frequency. Biol Psychiatry. 2022;92(10):752–753. doi: 10.1016/j.biopsych.2022.08.027 . [DOI] [PubMed] [Google Scholar]
  • 21.Dudley R, Watson F, O’Grady L, Aynsworth C, Dodgson G, Common S, et al. Prevalence and nature of multi-sensory and multi-modal hallucinations in people with first episode psychosis. Psychiatry Res. 2023;319:114988. doi: 10.1016/j.psychres.2022.114988 [DOI] [PubMed] [Google Scholar]
  • 22.Dudley R, Aynsworth C, Cheetham R, McCarthy-Jones S, Collerton D. Prevalence and characteristics of multi-modal hallucinations in people with psychosis who experience visual hallucinations. Psychiatry Res. 2018;269:25–30. doi: 10.1016/j.psychres.2018.08.032 [DOI] [PubMed] [Google Scholar]
  • 23.Kording KP, Beierholm U, Ma WJ, Quartz S, Tenenbaum JB, Shams L. Causal inference in multisensory perception. PLoS ONE. 2007;2(9):e943. Epub 2007/09/27. doi: 10.1371/journal.pone.0000943 ; PubMed Central PMCID: PMC1978520. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.Shams L, Beierholm UR. Causal inference in perception. Trends Cogn Sci. 2010;14(9):425–32. Epub 2010/08/14. doi: 10.1016/j.tics.2010.07.001 . [DOI] [PubMed] [Google Scholar]
  • 25.Shams L, Beierholm U. Bayesian causal inference: a unifying neuroscience theory. Neurosci Biobehav Rev. 2022;104619. doi: 10.1016/j.neubiorev.2022.104619 [DOI] [PubMed] [Google Scholar]
  • 26.Parise CV, Ernst MO. Correlation detection as a general mechanism for multisensory integration. Nat Commun. 2016;7(11543). doi: 10.1038/ncomms11543 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Rohe T, Ehlis A-C, Noppeney U. The neural dynamics of hierarchical Bayesian causal inference in multisensory perception. Nat Commun. 2019;10(1):1907. doi: 10.1038/s41467-019-09664-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Rohe T, Noppeney U. Sensory reliability shapes perceptual inference via two mechanisms. J Vis. 2015;15(5):1–16. [DOI] [PubMed] [Google Scholar]
  • 29.Wallace MT, Roberson GE, Hairston WD, Stein BE, Vaughan JW, Schirillo JA. Unifying multisensory signals across time and space. Exp Brain Res. 2004;158(2):252–8. Epub 2004/04/28. doi: 10.1007/s00221-004-1899-9 . [DOI] [PubMed] [Google Scholar]
  • 30.Wozny DR, Beierholm UR, Shams L. Probability matching as a computational strategy used in perception. PLoS Comput Biol. 2010;6(8). Epub 2010/08/12. doi: 10.1371/journal.pcbi.1000871 [pii]. ; PubMed Central PMCID: PMC2916852. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31.Acerbi L, Dokka K, Angelaki DE, Ma WJ. Bayesian comparison of explicit and implicit causal inference strategies in multisensory heading perception. PLoS Comput Biol. 2018;14(7):e1006110. doi: 10.1371/journal.pcbi.1006110 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Aller M, Noppeney U. To integrate or not to integrate: Temporal dynamics of hierarchical Bayesian causal inference. PLoS Biol. 2019;17(4):e3000210. doi: 10.1371/journal.pbio.3000210 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Cao Y, Summerfield C, Park H, Giordano BL, Kayser C. Causal inference in the multisensory brain. Neuron. 2019;102(5):1076–1087.e8. doi: 10.1016/j.neuron.2019.03.043 [DOI] [PubMed] [Google Scholar]
  • 34.Dokka K, Park H, Jansen M, DeAngelis GC, Angelaki DE. Causal inference accounts for heading perception in the presence of object motion. Proc Natl Acad Sci U S A. 2019;116(18):9060–9065. doi: 10.1073/pnas.1820373116 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35.Odegaard B, Shams L. The brain’s tendency to bind audiovisual signals is stable but not general. Psychol Sci. 2016;27(4):583–591. doi: 10.1177/0956797616628860 [DOI] [PubMed] [Google Scholar]
  • 36.Rohe T, Noppeney U. Cortical hierarchies perform Bayesian causal inference in multisensory perception. PLoS Biol. 2015;13(2):e1002073. doi: 10.1371/journal.pbio.1002073 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37.Aller M, Noppeney U. To integrate or not to integrate: Temporal dynamics of Bayesian Causal Inference. PLoS Biol. 2019;17(4):e3000210. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 38.Fang W, Li J, Qi G, Li S, Sigman M, Wang L. Statistical inference of body representation in the macaque brain. Proc Natl Acad Sci U S A. 2019;116(40):20151–7. Epub 2019/09/05. doi: 10.1073/pnas.1902334116 ; PubMed Central PMCID: PMC6778219. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39.Mihalik A, Noppeney U. Causal Inference in Audiovisual Perception. J Neurosci. 2020;40(34):6600–12. Epub 2020/07/17. doi: 10.1523/JNEUROSCI.0051-20.2020 ; PubMed Central PMCID: PMC7486655. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40.de Gelder B, Vroomen J, Annen L, Masthof E, Hodiamont P. Audio-visual integration in schizophrenia. Schizophr Res. 2003;59(2–3):211–8. Epub 2002/11/05. doi: 10.1016/s0920-9964(01)00344-9 [pii]. . [DOI] [PubMed] [Google Scholar]
  • 41.Haß K, Sinke C, Reese T, Roy M, Wiswede D, Dillo W, et al. Enlarged temporal integration window in schizophrenia indicated by the double-flash illusion. Cogn Neuropsychiatry. 2017;22(2):145–158. doi: 10.1080/13546805.2017.1287693 [DOI] [PubMed] [Google Scholar]
  • 42.Noel JP, Stevenson RA, Wallace MT. Atypical audiovisual temporal function in autism and schizophrenia: similar phenotype, different cause. Eur J Neurosci. 2018;47(10):1230–1241. doi: 10.1111/ejn.13911 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43.Vanes LD, White TP, Wigton RL, Joyce D, Collier T, Shergill SS. Reduced susceptibility to the sound-induced flash fusion illusion in schizophrenia. Psychiatry Res. 2016;245:58–65. doi: 10.1016/j.psychres.2016.08.016 [DOI] [PubMed] [Google Scholar]
  • 44.Tseng H-H, Bossong MG, Modinos G, Chen K-M, McGuire P, Allen P. A systematic review of multisensory cognitive–affective integration in schizophrenia. Neurosci Biobehav Rev. 2015;55:444–452. doi: 10.1016/j.neubiorev.2015.04.019 [DOI] [PubMed] [Google Scholar]
  • 45.White TP, Wigton RL, Joyce DW, Bobin T, Ferragamo C, Wasim N, et al. Eluding the illusion? Schizophrenia, dopamine and the McGurk effect. Front Hum Neurosci. 2014;8:565. doi: 10.3389/fnhum.2014.00565 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 46.Balz J, Roa Romero Y, Keil J, Krebber M, Niedeggen M, Gallinat J, et al. Beta/gamma oscillations and event-related potentials indicate aberrant multisensory processing in schizophrenia. Front Psychol. 2016;7:1896. doi: 10.3389/fpsyg.2016.01896 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 47.Bennett D, Silverstein SM, Niv Y. The two cultures of computational psychiatry. JAMA Psychiatry. 2019;76(6):563–564. doi: 10.1001/jamapsychiatry.2019.0231 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 48.Huys QJ, Maia TV, Frank MJ. Computational psychiatry as a bridge from neuroscience to clinical applications. Nat Neurosci. 2016;19(3):404–413. doi: 10.1038/nn.4238 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 49.Jones SA, Noppeney U. Ageing and multisensory integration: A review of the evidence, and a computational perspective. Cortex. 2021;138:1–23. Epub 20210211. doi: 10.1016/j.cortex.2021.02.001 . [DOI] [PubMed] [Google Scholar]
  • 50.Odegaard B, Shams L. The relationship between audiovisual binding tendencies and prodromal features of schizophrenia in the general population. Clin Psychol Sci. 2017;5(4):733–741. [Google Scholar]
  • 51.Odegaard B, Wozny DR, Shams L. The effects of selective and divided attention on sensory precision and integration. Neurosci Lett. 2016;614:24–28. doi: 10.1016/j.neulet.2015.12.039 [DOI] [PubMed] [Google Scholar]
  • 52.Dehaene S. Symbols and quantities in parietal cortex: Elements of a mathematical theory of number representation and manipulation. Sensorimotor Foundations of Higher Cognition. 2007;22:527–574. [Google Scholar]
  • 53.Gallistel CR, Gelman II. Non-verbal numerical cognition: from reals to integers. Trends Cogn Sci. 2000;4(2):59–65. doi: 10.1016/s1364-6613(99)01424-2 . [DOI] [PubMed] [Google Scholar]
  • 54.Andersen TS, Tiippana K, Sams M. Factors influencing audiovisual fission and fusion illusions. Cogn Brain Res. 2004;21(3):301–308. doi: 10.1016/j.cogbrainres.2004.06.004 [DOI] [PubMed] [Google Scholar]
  • 55.Burr D, Banks MS, Morrone MC. Auditory dominance over vision in the perception of interval duration. Exp Brain Res. 2009;198(1):49–57. Epub 20090714. doi: 10.1007/s00221-009-1933-z . [DOI] [PubMed] [Google Scholar]
  • 56.Buergers S, Noppeney U. The role of alpha oscillations in temporal binding within and across the senses. Nat Hum Behav. 2022;6(5):732–42. Epub 20220224. doi: 10.1038/s41562-022-01294-x ; PubMed Central PMCID: PMC7612782. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 57.Shams L, Kamitani Y, Shimojo S. What you see is what you hear. Nature. 2000;408(6814):788. Epub 2000/12/29. doi: 10.1038/35048669 . [DOI] [PubMed] [Google Scholar]
  • 58.Ernst MO, Banks MS. Humans integrate visual and haptic information in a statistically optimal fashion. Nature. 2002;415(6870):429–33. Epub 2002/01/25. doi: 10.1038/415429a . [DOI] [PubMed] [Google Scholar]
  • 59.Gau R, Noppeney U. How prior expectations shape multisensory perception. Neuroimage. 2016;124:876–886. doi: 10.1016/j.neuroimage.2015.09.045 [DOI] [PubMed] [Google Scholar]
  • 60.Hong F, Badde S, Landy MS. Repeated exposure to either consistently spatiotemporally congruent or consistently incongruent audiovisual stimuli modulates the audiovisual common-cause prior. Sci Rep. 2022;12(1):15532. Epub 20220915. doi: 10.1038/s41598-022-19041-7 ; PubMed Central PMCID: PMC9478143. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 61.Nahorna O, Berthommier F, Schwartz JL. Binding and unbinding the auditory and visual streams in the McGurk effect. J Acoust Soc Am. 2012;132(2):1061–1077. doi: 10.1121/1.4728187 . [DOI] [PubMed] [Google Scholar]
  • 62.Benson PJ, Beedie SA, Shephard E, Giegling I, Rujescu D, St Clair D. Simple viewing tests can detect eye movement abnormalities that distinguish schizophrenia cases from controls with exceptional accuracy. Biol Psychiatry. 2012;72(9):716–24. Epub 20120522. doi: 10.1016/j.biopsych.2012.04.019 . [DOI] [PubMed] [Google Scholar]
  • 63.Stevenson RA, Park S, Cochran C, McIntosh LG, Noel J-P, Barense MD, et al. The associations between multisensory temporal processing and symptoms of schizophrenia. Schizophr Res. 2017;179:97–103. doi: 10.1016/j.schres.2016.09.035 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 64.Roa Romero Y, Keil J, Balz J, Gallinat J, Senkowski D. Reduced frontal theta oscillations indicate altered crossmodal prediction error processing in schizophrenia. J Neurophysiol. 2016;116(3):1396–1407. doi: 10.1152/jn.00096.2016 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 65.Roa Romero Y, Keil J, Balz J, Niedeggen M, Gallinat J, Senkowski D. Alpha-band oscillations reflect altered multisensory processing of the McGurk illusion in schizophrenia. Front Hum Neurosci. 2016;10:41. doi: 10.3389/fnhum.2016.00041 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 66.Stekelenburg JJ, Maes JP, Van Gool AR, Sitskoorn M, Vroomen J. Deficient multisensory integration in schizophrenia: an event-related potential study. Schizophr Res. 2013;147(2–3):253–61. doi: 10.1016/j.schres.2013.04.038 [DOI] [PubMed] [Google Scholar]
  • 67.Stone DB, Coffman BA, Bustillo JR, Aine CJ. Multisensory stimuli elicit altered oscillatory brain responses at gamma frequencies in patients with schizophrenia. Front Hum Neurosci. 2014;8:788. doi: 10.3389/fnhum.2014.00788 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 68.Stone DB, Urrea LJ, Aine CJ, Bustillo JR, Clark VP, Stephen JM. Unisensory processing and multisensory integration in schizophrenia: a high-density electrical mapping study. Neuropsychologia. 2011;49(12):3178–3187. doi: 10.1016/j.neuropsychologia.2011.07.017 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 69.Mishra J, Martinez A, Sejnowski TJ, Hillyard SA. Early cross-modal interactions in auditory and visual cortex underlie a sound-induced visual illusion. J Neurosci. 2007;27(15):4120–31. Epub 2007/04/13. doi: 10.1523/JNEUROSCI.4912-06.2007 ; PubMed Central PMCID: PMC2905511. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 70.Oribe N, Hirano Y, Kanba S, Del Re E, Seidman L, Mesholam-Gately R, et al. Progressive reduction of visual P300 amplitude in patients with first-episode schizophrenia: an ERP study. Schizophr Bull. 2015;41(2):460–470. doi: 10.1093/schbul/sbu083 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 71.Shams L, Iwaki S, Chawla A, Bhattacharya J. Early modulation of visual cortex by sound: an MEG study. Neurosci Lett. 2005;378(2):76–81. doi: 10.1016/j.neulet.2004.12.035 [DOI] [PubMed] [Google Scholar]
  • 72.Bae G-Y, Leonard CJ, Hahn B, Gold JM, Luck SJ. Assessing the information content of ERP signals in schizophrenia using multivariate decoding methods. NeuroImage Clin. 2020;25:102179. doi: 10.1016/j.nicl.2020.102179 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 73.Badde S, Navarro KT, Landy MS. Modality-specific attention attenuates visual-tactile integration and recalibration effects by reducing prior expectations of a common source for vision and touch. Cognition. 2020;197:104170. doi: 10.1016/j.cognition.2019.104170 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 74.Ferrari A, Noppeney U. Attention controls multisensory perception via two distinct mechanisms at different levels of the cortical hierarchy. PLoS Biol. 2021;19(11):e3001465. doi: 10.1371/journal.pbio.3001465 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 75.Rohe T, Noppeney U. Reliability-Weighted Integration of Audiovisual Signals Can Be Modulated by Top-down Attention. eNeuro. 2018;5(1):ENEURO. 0315–17.2018. doi: 10.1523/ENEURO.0315-17.2018 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 76.Gold JM, Fuller RL, Robinson BM, Braun EL, Luck SJ. Impaired top–down control of visual search in schizophrenia. Schizophr Res. 2007;94(1–3):148–155. doi: 10.1016/j.schres.2007.04.023 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 77.Heinrichs RW, Zakzanis KK. Neurocognitive deficit in schizophrenia: a quantitative review of the evidence. Neuropsychology. 1998;12(3):426. doi: 10.1037//0894-4105.12.3.426 [DOI] [PubMed] [Google Scholar]
  • 78.Lesh TA, Niendam TA, Minzenberg MJ, Carter CS. Cognitive control deficits in schizophrenia: mechanisms and meaning. Neuropsychopharmacology. 2011;36(1):316–38. Epub 20100915. doi: 10.1038/npp.2010.156 ; PubMed Central PMCID: PMC3052853. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 79.Cicchini GM, Mikellidou K, Burr DC. Serial dependence in perception. Annu Rev Psychol. 2024;75:129–154. doi: 10.1146/annurev-psych-021523-104939 [DOI] [PubMed] [Google Scholar]
  • 80.Pearl D, Yodashkin-Porat D, Katz N, Valevski A, Aizenberg D, Sigler M, et al. Differences in audiovisual integration, as measured by McGurk phenomenon, among adult and adolescent patients with schizophrenia and age-matched healthy control groups. Compr Psychiatry. 2009;50(2):186–192. Epub 2009/02/17. doi: 10.1016/j.comppsych.2008.06.004 [pii]. . [DOI] [PubMed] [Google Scholar]
  • 81.Williams LE, Light GA, Braff DL, Ramachandran VS. Reduced multisensory integration in patients with schizophrenia on a target detection task. Neuropsychologia. 2010;48(10):3128–3136. doi: 10.1016/j.neuropsychologia.2010.06.028 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 82.Button KS, Ioannidis JP, Mokrysz C, Nosek BA, Flint J, Robinson ES, et al. Power failure: why small sample size undermines the reliability of neuroscience. Nat Rev Neurosci. 2013;14(5):365–376. doi: 10.1038/nrn3475 [DOI] [PubMed] [Google Scholar]
  • 83.Gignac GE, Szodorai ET. Effect size guidelines for individual differences researchers. Personal Individ Differ. 2016;102:74–78. [Google Scholar]
  • 84.Verdoux H, van Os J. Psychotic symptoms in non-clinical populations and the continuum of psychosis. Schizophr Res. 2002;54(1–2):59–65. doi: 10.1016/s0920-9964(01)00352-8 [DOI] [PubMed] [Google Scholar]
  • 85.Zhou H-y, Cai X-l, Weigl M, Bang P, Cheung EF, Chan RC. Multisensory temporal binding window in autism spectrum disorders and schizophrenia spectrum disorders: A systematic review and meta-analysis. Neurosci Biobehav Rev. 2018;86:66–76. doi: 10.1016/j.neubiorev.2017.12.013 [DOI] [PubMed] [Google Scholar]
  • 86.Wittchen H-U, Wunderlich U, Gruschwitz S, Zaudig M. SKID I. Strukturiertes Klinisches Interview für DSM-IV. Achse I: Psychische Störungen. Interviewheft und Beurteilungsheft. Eine deutschsprachige, erweiterte Bearb. d. amerikanischen Originalversion des SKID I. 1997. [Google Scholar]
  • 87.Kay SR, Fiszbein A, Opler LA. The positive and negative syndrome scale (PANSS) for schizophrenia. Schizophr Bull. 1987;13(2):261. doi: 10.1093/schbul/13.2.261 [DOI] [PubMed] [Google Scholar]
  • 88.Lincoln TM, Keller E, Rief W. Die erfassung von wahn und halluzinationen in der normalbevölkerung: Deutsche adaptationen des Peters et al. delusions inventory (PDI) und der Launay Slade hallucination scale (LSHS-R).? 2009;55(1):29–40. [Google Scholar]
  • 89.Lincoln TM, Ziegler M, Lüllmann E, Müller MJ, Rief W. Can delusions be self-assessed? Concordance between self-and observer-rated delusions in schizophrenia. Psychiatry Res. 2010;178(2):249–254. doi: 10.1016/j.psychres.2009.04.019 [DOI] [PubMed] [Google Scholar]
  • 90.Addington D, Addington J, Schissel B. A depression rating scale for schizophrenics. Schizophr Res. 1990;3(4):247–251. doi: 10.1016/0920-9964(90)90005-r [DOI] [PubMed] [Google Scholar]
  • 91.Leucht S, Samara M, Heres S, Patel MX, Woods SW, Davis JM. Dose equivalents for second-generation antipsychotics: the minimum effective dose method. Schizophr Bull. 2014;40(2):314–326. doi: 10.1093/schbul/sbu001 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 92.Helmstaedter C, Lendt M, Lux S. VLMT: Verbaler Lern-und Merkfähigkeitstest: Beltz Test, Göttingen, Germany; 2001. [Google Scholar]
  • 93.Reitan RM. Trail Making Test: Manual for administration and scoring: Reitan Neuropsychology Laboratory; 1992. [Google Scholar]
  • 94.Lehrl S. Mehrfachwahl-Wortschatz-Intelligenztest MWT-B. Spitta Verlag GmbH, Balingen, Germany; 2005. [Google Scholar]
  • 95.Bäumler G. Farbe-Wort-Interferenztest (FWIT) nach JR Stroop: Hogrefe Verlag GmbH, Göttingen, Germany; 1984. [Google Scholar]
  • 96.Oldfield RC. The assessment and analysis of handedness: the Edinburgh inventory. Neuropsychologia. 1971;9(1):97–113. doi: 10.1016/0028-3932(71)90067-4 [DOI] [PubMed] [Google Scholar]
  • 97.Hautzinger M, Keller F, Kühner C. Das Beck Depressionsinventar II. Deutsche Bearbeitung und Handbuch zum BDI II. Frankfurt a M: Harcourt Test Services. 2006. [Google Scholar]
  • 98.Shams L, Ma WJ, Beierholm U. Sound-induced flash illusion as an optimal percept. Neuroreport. 2005;16(17):1923–1927. doi: 10.1097/01.wnr.0000187634.68504.bb [DOI] [PubMed] [Google Scholar]
  • 99.Brainard DH. The psychophysics toolbox. Spat Vis. 1997;10(4):433–436. [PubMed] [Google Scholar]
  • 100.Acerbi L, Ma WJ. Practical Bayesian optimization for model fitting with Bayesian adaptive direct search. Adv Neural Inf Process Syst. 2017:1836–46. [Google Scholar]
  • 101.Wilson RC, Collins AG. Ten simple rules for the computational modeling of behavioral data. Elife. 2019;8:e49547. doi: 10.7554/eLife.49547 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 102.Raftery AE. Bayesian model selection in social research. Sociological Methodology 1995, Vol 25. 1995;25:111–63. doi: 10.2307/271063 ISI:A1995BE85S00004. [DOI] [Google Scholar]
  • 103.Rigoux L, Stephan KE, Friston KJ, Daunizeau J. Bayesian model selection for group studies—revisited. Neuroimage. 2014;84:971–985. doi: 10.1016/j.neuroimage.2013.08.065 [DOI] [PubMed] [Google Scholar]
  • 104.Friston KJ, Holmes AP, Worsley KJ, Poline JP, Frith CD, Frackowiak RSJ. Statistical parametric maps in functional imaging: a general linear approach. Hum Brain Mapp. 1994;2(4):189–210. [Google Scholar]
  • 105.Daunizeau J, Adam V, Rigoux L. VBA: a probabilistic treatment of nonlinear models for neurobiological and behavioural data. PLoS Comput Biol. 2014;10(1):e1003441. doi: 10.1371/journal.pcbi.1003441 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 106.Krekelberg B. bayesFactor GitHub2021 [cited 2021 Jan 30]. Available from: https://klabhub.github.io/bayesFactor/. [Google Scholar]
  • 107.Wetzels R, Wagenmakers EJ. A default Bayesian hypothesis test for correlations and partial correlations. Psychon Bull Rev. 2012;19(6):1057–1064. doi: 10.3758/s13423-012-0295-x ; PubMed Central PMCID: PMC3505519. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 108.Wagenmakers E-J, Love J, Marsman M, Jamil T, Ly A, Verhagen J, et al. Bayesian inference for psychology. Part II: Example applications with JASP. Psychon Bull Rev. 2018;25(1):58–76. doi: 10.3758/s13423-017-1323-7 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 109.van den Bergh D, Van Doorn J, Marsman M, Draws T, Van Kesteren E-J, Derks K, et al. A tutorial on conducting and interpreting a Bayesian ANOVA in JASP. LAnnee psychologique. 2020;120(1):73–96. [Google Scholar]
  • 110.Kass RE, Raftery AE. Bayes factors. J Am Stat Assoc. 1995;90(430):773–795. [Google Scholar]
  • 111.Wagenmakers E-J, Marsman M, Jamil T, Ly A, Verhagen J, Love J, et al. Bayesian inference for psychology. Part I: Theoretical advantages and practical ramifications. Psychon Bull Rev. 2018;25(1):35–57. doi: 10.3758/s13423-017-1343-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 112.Teder-Sälejärvi W, McDonald J, Di Russo F, Hillyard S. An analysis of audio-visual crossmodal integration by means of event-related potential (ERP) recordings. Cogn Brain Res. 2002;14(1):106–114. doi: 10.1016/s0926-6410(02)00065-4 [DOI] [PubMed] [Google Scholar]
  • 113.Nichols TE, Holmes AP. Nonparametric permutation tests for functional neuroimaging: a primer with examples. Hum Brain Mapp. 2002;15(1):1–25. doi: 10.1002/hbm.1058 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 114.Maris E, Oostenveld R. Nonparametric statistical testing of EEG-and MEG-data. J Neurosci Methods. 2007;164(1):177–190. doi: 10.1016/j.jneumeth.2007.03.024 [DOI] [PubMed] [Google Scholar]
  • 115.Chang CC, Lin CJ. LIBSVM: a library for support vector machines. ACM Trans Intell Syst Technol. 2011;2(3):27. [Google Scholar]
  • 116.Rohe T, Hesse K, Ehlis A-C, Noppeney U. Data from: Multisensory perceptual and causal inference is largely preserved in medicated post-acute individuals with schizophrenia [Dataset]. 2024. Dryad. doi: 10.5061/dryad.hhmgqnkr1 [DOI] [PMC free article] [PubMed] [Google Scholar]

Decision Letter 0

Christian Schnell, PhD

1 Sep 2023

Dear Dr Rohe,

Thank you for submitting your manuscript entitled "Computations and neural dynamics of audiovisual causal and perceptual inference in schizophrenia" for consideration as a Update Article by PLOS Biology.

Your manuscript has now been evaluated by the PLOS Biology editorial staff and I am writing to let you know that we would like to send your submission out for external peer review. However, please note that we have yet to make a firm call about whether the study offers a sufficient conceptual advance for PLOS Biology and that we will be looking for strong support from the reviewers.

However, before we can send your manuscript to reviewers, we need you to complete your submission by providing the metadata that is required for full assessment. To this end, please login to Editorial Manager where you will find the paper in the 'Submissions Needing Revisions' folder on your homepage. Please click 'Revise Submission' from the Action Links and complete all additional questions in the submission questionnaire.

Once your full submission is complete, your paper will undergo a series of checks in preparation for peer review. After your manuscript has passed the checks it will be sent out for review. To provide the metadata for your submission, please Login to Editorial Manager (https://www.editorialmanager.com/pbiology) within two working days, i.e. by Sep 03 2023 11:59PM.

If your manuscript has been previously peer-reviewed at another journal, PLOS Biology is willing to work with those reviews in order to avoid re-starting the process. Submission of the previous reviews is entirely optional and our ability to use them effectively will depend on the willingness of the previous journal to confirm the content of the reports and share the reviewer identities. Please note that we reserve the right to invite additional reviewers if we consider that additional/independent reviewers are needed, although we aim to avoid this as far as possible. In our experience, working with previous reviews does save time.

If you would like us to consider previous reviewer reports, please edit your cover letter to let us know and include the name of the journal where the work was previously considered and the manuscript ID it was given. In addition, please upload a response to the reviews as a 'Prior Peer Review' file type, which should include the reports in full and a point-by-point reply detailing how you have or plan to address the reviewers' concerns.

During the process of completing your manuscript submission, you will be invited to opt-in to posting your pre-review manuscript as a bioRxiv preprint. Visit http://journals.plos.org/plosbiology/s/preprints for full details. If you consent to posting your current manuscript as a preprint, please upload a single Preprint PDF.

Feel free to email us at plosbiology@plos.org if you have any queries relating to your submission.

Kind regards,

Christian

Christian Schnell, PhD

Senior Editor

PLOS Biology

cschnell@plos.org

Decision Letter 1

Christian Schnell, PhD

1 Nov 2023

Dear Dr Rohe,

Thank you for your patience while your manuscript "Computations and neural dynamics of audiovisual causal and perceptual inference in schizophrenia" was peer-reviewed at PLOS Biology. Please allow me to apologize for the long delay in sending our decision, which was caused by difficulties in finding suitable reviewers and academic editors for your study. In any case, your manuscript has now been evaluated by the PLOS Biology editors, an Academic Editor with relevant expertise, and by several independent reviewers.

In light of the reviews, which you will find at the end of this email, we would like to invite you to revise the work to thoroughly address the reviewers' reports.

As you will see below, the reviewers say that the study is overall well executed and interesting. However, all reviewers raise concerns about the small sample size, concerns about the methodological approach, and mention that the high-level claims are too strong considering the limitations of the work. Apart from the limitation of the small sample size (which should be clearly discussed), it seems that most of the concerns can be addressed with additional analyses and a more careful presentation and discussion of the previous literature and results of the current manuscript. We also encourage you to try to make your paper as accessible as possible to readers who are not familiar with the Bayesian modeling procedures.

Given the extent of revision needed, we cannot make a decision about publication until we have seen the revised manuscript and your response to the reviewers' comments. Your revised manuscript is likely to be sent for further evaluation by all or a subset of the reviewers.

We expect to receive your revised manuscript within 3 months. Please email us (plosbiology@plos.org) if you have any questions or concerns, or would like to request an extension.

At this stage, your manuscript remains formally under active consideration at our journal; please notify us by email if you do not intend to submit a revision so that we may withdraw it.

**IMPORTANT - SUBMITTING YOUR REVISION**

Your revisions should address the specific points made by each reviewer. Please submit the following files along with your revised manuscript:

1. A 'Response to Reviewers' file - this should detail your responses to the editorial requests, present a point-by-point response to all of the reviewers' comments, and indicate the changes made to the manuscript.

*NOTE: In your point-by-point response to the reviewers, please provide the full context of each review. Do not selectively quote paragraphs or sentences to reply to. The entire set of reviewer comments should be present in full and each specific point should be responded to individually, point by point.

You should also cite any additional relevant literature that has been published since the original submission and mention any additional citations in your response.

2. In addition to a clean copy of the manuscript, please also upload a 'track-changes' version of your manuscript that specifies the edits made. This should be uploaded as a "Revised Article with Changes Highlighted" file type.

*Re-submission Checklist*

When you are ready to resubmit your revised manuscript, please refer to this re-submission checklist: https://plos.io/Biology_Checklist

To submit a revised version of your manuscript, please go to https://www.editorialmanager.com/pbiology/ and log in as an Author. Click the link labelled 'Submissions Needing Revision' where you will find your submission record.

Please make sure to read the following important policies and guidelines while preparing your revision:

*Published Peer Review*

Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out. Please see here for more details:

https://blogs.plos.org/plos/2019/05/plos-journals-now-open-for-published-peer-review/

*PLOS Data Policy*

Please note that as a condition of publication PLOS' data policy (http://journals.plos.org/plosbiology/s/data-availability) requires that you make available all data used to draw the conclusions arrived at in your manuscript. If you have not already done so, you must include any data used in your manuscript either in appropriate repositories, within the body of the manuscript, or as supporting information (N.B. this includes any numerical values that were used to generate graphs, histograms etc.). For an example see here: http://www.plosbiology.org/article/info%3Adoi%2F10.1371%2Fjournal.pbio.1001908#s5

*Blot and Gel Data Policy*

We require the original, uncropped and minimally adjusted images supporting all blot and gel results reported in an article's figures or Supporting Information files. We will require these files before a manuscript can be accepted so please prepare them now, if you have not already uploaded them. Please carefully read our guidelines for how to prepare and upload this data: https://journals.plos.org/plosbiology/s/figures#loc-blot-and-gel-reporting-requirements

*Protocols deposition*

To enhance the reproducibility of your results, we recommend that if applicable you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols

Thank you again for your submission to our journal. We hope that our editorial process has been constructive thus far, and we welcome your feedback at any time. Please don't hesitate to contact us if you have any questions or comments.

Sincerely,

Christian

Christian Schnell, PhD

Senior Editor

PLOS Biology

cschnell@plos.org

------------------------------------

REVIEWS:

Reviewer's Responses to Questions

Do you want your identity to be public for this peer review?

Reviewer #1: No

Reviewer #2: Yes: Guillermo Horga, MD PhD

Reviewer #3: No

Reviewer #1: This article investigates audiovisual perception in a sound-induced flash illusion paradigm using behavioral, computational (Bayesian modelling) and neuroimaging analysis in 17 SCZ patients as compared to 23 controls. The results are essentially negative: the authors found comparable behavioural, computational and neural mechanisms/ EEG signatures in SCZ and HC an no difference in the model parameters or winning model. Bayesian inference and more particularly causal inference has recently been proposed to be at the origin of differences in SCZ (as well as ASD) so the line of work is important.

Though dense, the article is very clearly written. It is using a 1-4 beeps x 1-4 flash factorial experimental design which is interesting, the computational analysis is thorough and the EEG decoding exercise interesting.

In my understanding (apologies if I have missed relevant information), the paper seems to suffer from a few limitations however, which may limit its impact and significance:

- One obvious issue with the study is the sample size. 17 SCZ is a small sample. I couldn't find a power analysis to motivate the sample size of 17 (some indication for supporting a somewhat larger sample size of the 26 initial participants). As is common for such clinical studies, the SCZ group is also very heterogeneous both in terms of comorbidities and treatment. The SCZ group has been medicated for a long time and stable with low PANSS positive scores. It is also apparent in the modelling that the individual parameters are very widespread. 


- It is also unclear whether the experimental design is well chosen to identify differences in SCZ. The SOA is not varied (and the choice of SOA is not very clear motivated either) which could perhaps explain the negative results, in particular if the differences are about the temporal binding window. I think better motivation with respect to previous literature and experimental design and maybe replication of conditions where differences in schizophrenia were found would have helped.

- For the modelling part, Figure S5 seems to be showing parameter recovery for the winning model rather than model recovery (as indicated in the caption?). It is unclear then what the performance is in terms of model recovery? Similarly, it is not very clear whether the winning is actually a good model for the data (Wilson & Collins, 2019).

- Some aspects are unclear in the text, e.g. are the behavioural effects that are significant (numeric prior and visual variance) corrected for multiple comparisons?

- Page 12: is the SVM trained on both HC and SCZ data, together or separately? The hypotheses being tested here could be clearer.

- "In both HC and SCZ, the decoders predicted the BCI estimates from EEG patterns significantly better than chance over most periods of the post-stimulus time window" "Significantly better than chance" is vague, it is frustrating to have to go to the figure to have an idea how well the decoding is doing and it is unclear what level of accuracy would be necessary for making clear conclusions.

Reviewer #2: Rohe et al. use an EEG-compatible multimodal sensory-integration task to evaluate causal Bayesian inference in medicated patients with schizophrenia and healthy controls. The authors then use computational modeling and EEG decoding to characterize behavioral and neural computations underlying various levels of inference. Overall, they conclude that neither behavioral nor neural data suggest alterations in inference are present in medicated patients with schizophrenia. This article presents a compelling experimental framework and several interesting analyses. Overall, the experimental design and analytical techniques are of excellent quality, particularly for a clinical study.

Despite my general enthusiasm, however, I have concerns about the characterization of prior literature, interpretation of the results, and some analysis decisions that should be addressed before I can make a recommendation for publication.

In general terms, and despite the obvious quality of this work, it is not clear what the main takeaways and contributions to this literature are. Given the sample characteristics (see below), I believe it is fair to say that this study is not well positioned to rule in/out prior over/underweighting as a model of psychotic states, and therefore it cannot resolve mixed findings in this literature. The paper's contributions should be made clearer (novel methods and proof-of-concept application to this area?) and the conclusions should be very explicitly qualified with these limitations in mind (particularly in the abstract and the discussion sections, and with particular attention to the strength of evidence for the null hypothesis and generalizability of these conclusions).

Major concerns

1. Throughout, the authors treat schizophrenia and psychosis interchangeably, from presentation of the literature to analytical justification. This is problematic. Strong/precise prior models (e.g., Powers et al., 2017) are models of psychosis severity and not models of schizophrenia. Importantly these models assume state-dependence of neurocognitive phenotypes of psychosis (Kafadar et al., 2022). So it is critical to separate state-dependent effects from trait-like effects in schizophrenia patients that could be linked to negative or cognitive symptoms, and which have been associated with distinct, even opposite mechanisms from those associated selectively with positive symptoms (e.g., see work from Rick Adams). This confusion arises throughout the manuscript (specific comments below).

a. Description of previous literature: The discussion of the link between hallucinations and prior weighting (citations 8-17) does not distinguish between findings that compare patients vs. controls and studies that consider psychosis (or psychosis-like) severity. Overall the balance of the literature presented gives the wrong impression. The majority of the literature points to a positive relationship between psychosis severity and prior weighting, even if a minority of studies obtained conflicting findings. The literature review should be balanced in terms of the evidence for prior overweighting in patients with more severe psychotic symptoms (hallucinations more concretely) vs. evidence against this proposition. Again, this should separate psychotic states/severity versus trait-like schizophrenia phenotypes (the latter of which, but not the former, are associated with cognitive deficits and amotivation, which can broadly affect cognitive performance). There should also be a distinction between clinical and non-clinical populations.

b. Data analysis: In general, patient vs control comparisons are tested when analyses across psychosis severity (with particular emphasis on hallucination severity) would also be appropriate. There should also be a clearer justification for the scales and items used in the correlations with psychosis and there should be a clearer description of the distribution of psychosis severity in this sample, with emphasis on whether this small sample has enough variability to really test psychosis-related models which require sufficient range of psychosis severity.

c. Results interpretation: This data doesn't appear to argue against strong prior models of psychosis but rather suggests that inference processes are largely spared in a population of medicated schizophrenia patients with low psychosis severity. To the contrary, the authors do report effects of prior variance that relate to psychosis, consistent with many findings in the literature.

i. In the discussion, the authors note that "we cannot exclude the possibility that SCZ may have been less vigilant … which could be misinterpreted as overreliance on their numeric prior." Wouldn't the authors suspect that individual differences in vigilance would be captured by the lapse term? Which would be consistent with the finding that this term correlates with negative symptoms? To clarify the specificity of the identified relationships between prior variance and positive symptoms, negative and general symptoms should be statistically controlled for in these analyses.

2. Sample: The sample is relatively small which should be acknowledged when arguing for the null hypothesis. The sample is also possibly not representative of the population of patients with schizophrenia as a whole, particularly in early or untreated states. The patients have relatively high doses of antipsychotics, low psychosis severity, and are largely cognitively intact. Furthermore, the recruitment strategy/recruitment sources is not discussed, which may be relevant to characterizing this cohort. Some of these points are acknowledged in the discussion but should be clarified during key takeaway statements throughout the manuscript. This is especially important since the authors seem to be arguing that the results are contradictory with the rest of the literature.

a. Schizoaffective patients were excluded in this analysis which is very uncommon in this literature. Given that including them would substantially increase the sample size I would like to see analyses including these patients.

3. The authors evaluate evidence for trial-by-trial learning by splitting the data into subsets with different trial histories and fitting the BCI model to these subsets. This approach is not standard and should be more carefully validated and supported. The authors should first demonstrate evidence for these effects using model-agnostic approaches (e.g., lagged-regression predicting report as a function of previous trial identity), showing that their model-based results capture qualitative features of the raw data. As it stands, it is unclear whether there are even meaningful trial-history effects on this task, or whether they should be expected based on this literature. Second, they should demonstrate that this fitting approach produces valid and interpretable parameter estimates via a parameter recovery. Alternatively, there are models in the literature that dynamically update priors as a function of trial-by-trial evidence. A model with this structure would be a more appropriate test of this question. That said, an alternative would be to eliminate this section altogether, since I am not convinced it adds substantial value if the obtained parameters too noisy to provide meaningful information (particularly meaningful null findings) in this small sample.

Minor concerns

1. Most psychotic symptoms are not multisensory even though they can manifest in different modalities (e.g., even in patients who report visual and auditory hallucinations, integrated audiovisual hallucinations are rare). The focus on multisensory perception/integration should be more carefully justified. One clear strength is the ability to investigate different priors and hierarchical levels of inference. This is linked to point 1 above; cited studies (citations 30-36) focus on diagnosis comparisons rather than psychosis comparisons although the text does not make this distinction.

2. Average fit quality by group is shown. Is there substantial variability in the winning model by individual? If so, it would be important to show the proportions of winning models across groups and by psychosis severity.

3. Comparisons of model fitted variances (visual, auditory, and prior) with specific symptoms of psychosis (hallucinations vs delusions) should be shown.

Reviewer #3: Strengths of this paper are that it generally seems like a good next step in understanding whether and how integration of prior information and sensory inputs may be impacted in individuals with schizophrenia. The methods are sound, including useful details about the paradigm, analyses methods and statistics. It is also fairly clearly written and results interpreted reasonably. The study is novel in its use of psychophysical and EEG data as well as computational modeling to answer some research questions regarding causal inference. One major weakness is the small sample size. The findings supported some of the predictions, and these are reasonably well addressed and seem interesting. They stand to add somewhat to the body of knowledge advancing how we think about multisensory integration mechanisms from a Bayesian standpoint. However, not all predictions panned out, reducing impact of the paper. Further, while at a theoretical level, the explanation for the modeling is clear—how each parameter indexes decision processes needs to be clarified further. My comments/ queries are aimed at improving the clarity of the findings and placing within the broader context of clinical research as well.

1) It would be useful to briefly explain more about the modeling, especially the fused weighting, and setting up/ providing detailed rationale for the computational aspect in the introduction

2) ? It is mentioned that "the visual variance" correlated with positive symptoms--- This should be explained more in terms of which exact measure and what this correlation signifies in the context of overweighting priors. Were there any other symptom correlates in the SZ group. Any effect of MWT-B or VLMT scores?

3) While the paradigm, analyses and results attempt to answer the question of sensory integration in terms of stimulus complexity , it would be useful to place these results in the context of other findings on overly strong priors e.g. Pavlovian conditioning and failures in perceptual updating and discussion on what implication the intricacies of multisensory integration have on specific symptoms—This would strengthen the discussion on specificity and clinical implications of this work

4) Limitations of the sample size and impact of the results/interpretability should be discussed

Decision Letter 2

Christian Schnell, PhD

17 Jul 2024

Dear Dr Rohe,

Thank you for your patience while we considered your revised manuscript "Computations and neural dynamics of audiovisual causal and perceptual inference in schizophrenia" for publication as a Update Article at PLOS Biology. This revised version of your manuscript has been evaluated by the PLOS Biology editors, the Academic Editor and the original reviewers.

Based on the reviews and on our Academic Editor's assessment of your revision, we are likely to accept this manuscript for publication, provided you satisfactorily address the remaining points raised by the reviewers. Please also make sure to address the following data and other policy-related requests.

* We would like to suggest a different title to improve accessibility: "No evidence for impaired multisensory perceptual and causal inference in medicated post-acute individuals with schizophrenia"

* Please provide information in the methods section that the study has been conducted according to the principles expressed in the Declaration of Helsinki.

* Please also attend to the following points:

** Page 9 line 15, missing the word “bias”

** Page 15 line 34 “accuracyin”

** Fig 1C, the y-axis label “correlation” is not explained or described it seems.

* Please include qualifiers where necessary to indicate that due to the small sample size, the evidence for a certain claim may be limited. Reviewer 1 mentions one of these statements in the abstract where this will be needed. Papers such as Gilles E. Gignac ⁎, Eva T. Szodorai. Effect size guidelines for individual differences researchers, Personality and individual differences (2016) and Button, K. S., Ioannidis, J. P., Mokrysz, C., Nosek, B. A., Flint, J., Robinson, E. S., & Munafo, M. R. (2013) might be useful for further discussing this issue, also considering that your power calculation is based on a large effect size (d=1) which may not be realistic given that effect sizes are usually rather small in this field (e.g., d=0.4).

* DATA POLICY:

You may be aware of the PLOS Data Policy, which requires that all data be made available without restriction: http://journals.plos.org/plosbiology/s/data-availability. For more information, please also see this editorial: http://dx.doi.org/10.1371/journal.pbio.1001797

Note that we do not require all raw data. Rather, we ask that all individual quantitative observations that underlie the data summarized in the figures and results of your paper be made available in one of the following forms:

1) Supplementary files (e.g., excel). Please ensure that all data files are uploaded as 'Supporting Information' and are invariably referred to (in the manuscript, figure legends, and the Description field when uploading your files) using the following format verbatim: S1 Data, S2 Data, etc. Multiple panels of a single or even several figures can be included as multiple sheets in one excel file that is saved using exactly the following convention: S1_Data.xlsx (using an underscore).

2) Deposition in a publicly available repository. Please also provide the accession code or a reviewer link so that we may view your data before publication.

Regardless of the method selected, please ensure that you provide the individual numerical values that underlie the summary data displayed in the following figure panels as they are essential for readers to assess your analysis and to reproduce it: 1CD, 4, 5ABC and similar panels in the supplementary information.

NOTE: the numerical data provided should include all replicates AND the way in which the plotted mean and errors were derived (it should not present only the mean/average values).

Please also ensure that figure legends in your manuscript include information on where the underlying data can be found, and ensure your supplemental data file/s has a legend.

Please ensure that your Data Statement in the submission system accurately describes where your data can be found.

* CODE POLICY

Per journal policy, if you have generated any custom code during the course of this investigation, please make it available without restrictions. Please ensure that the code is sufficiently well documented and reusable, and that your Data Statement in the Editorial Manager submission system accurately describes where your code can be found.

Please note that we cannot accept sole deposition of code in GitHub, as this could be changed after publication. However, you can archive this version of your publicly available GitHub code to Zenodo. Once you do this, it will generate a DOI number, which you will need to provide in the Data Accessibility Statement (you are welcome to also provide the GitHub access information). See the process for doing this here: https://docs.github.com/en/repositories/archiving-a-github-repository/referencing-and-citing-content

As you address these items, please take this last chance to review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the cover letter that accompanies your revised manuscript.

We expect to receive your revised manuscript within two weeks.

To submit your revision, please go to https://www.editorialmanager.com/pbiology/ and log in as an Author. Click the link labelled 'Submissions Needing Revision' to find your submission record. Your revised submission must include the following:

- a cover letter that should detail your responses to any editorial requests, if applicable, and whether changes have been made to the reference list

- a Response to Reviewers file that provides a detailed response to the reviewers' comments (if applicable, if not applicable please do not delete your existing 'Response to Reviewers' file.)

- a track-changes file indicating any changes that you have made to the manuscript.

NOTE: If Supporting Information files are included with your article, note that these are not copyedited and will be published as they are submitted. Please ensure that these files are legible and of high quality (at least 300 dpi) in an easily accessible file format. For this reason, please be aware that any references listed in an SI file will not be indexed. For more information, see our Supporting Information guidelines:

https://journals.plos.org/plosbiology/s/supporting-information

*Published Peer Review History*

Please note that you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out. Please see here for more details:

https://blogs.plos.org/plos/2019/05/plos-journals-now-open-for-published-peer-review/

*Press*

Should you, your institution's press office or the journal office choose to press release your paper, please ensure you have opted out of Early Article Posting on the submission form. We ask that you notify us as soon as possible if you or your institution is planning to press release the article.

*Protocols deposition*

To enhance the reproducibility of your results, we recommend that if applicable you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols

Please do not hesitate to contact me should you have any questions.

Sincerely,

Christian

Christian Schnell, PhD

Senior Editor

cschnell@plos.org

PLOS Biology

------------------------------------------------------------------------

Reviewer remarks:

Reviewer #1: The authors have kindly answered our questions, clarified their methods and discussed the limitations of the study, and updated the manuscript accordingly.

Some aspects of the work are interesting and solid.

However, the limitations remain, particularly the fact that the authors mostly report negative findings with a very small sample (which is not well justified by a power analysis, and is poorly representative of a population of patients with schizophrenia as a whole, particularly in early or untreated states), using a design that had not been used before to show differences in SCZ (so not easily comparable to previous literature). The results concerning the numeric prior (the significance of which, unlike that of the p_common prior, was a bit unclear to me to start with) are now muddled/weaker (not significant) after inclusion of the 6 patients with schizo-affective disorder.

All in all, although I recognise how hard it is to get bigger samples of patients with schizohprenia and/or collecting data on different conditions (e.g. also varying SOAs to explore temporal windows), my opinion is that the contribution and significance of the presented results are limited (if the aim is to present them to a large/general audience). I don't think it is fair for example to express in such broad terms that the results " demonstrate that the neurocomputational mechanisms of multisensory perceptual and causal inference remain remarkably intact in medicated post-acute individuals with schizophrenia" (as written in the abstract) as the results may only hold for the chosen design (which doesn't explore differences in binding windows / replicate situations where differences have been observed) and may miss subtle differences that would be detected in a larger sample.

Reviewer #2 (Guillermo Horga): The authors have been very responsive and addressed all my comments. This is a great paper and an important contribution to the literature.

Decision Letter 3

Christian Schnell, PhD

6 Aug 2024

Dear Tim,

Thank you for the submission of your revised Update Article "Multisensory perceptual and causal inference is largely preserved in medicated post-acute individuals with schizophrenia" for publication in PLOS Biology. On behalf of my colleagues and the Academic Editor, Simon van Gaal, I am pleased to say that we can in principle accept your manuscript for publication, provided you address any remaining formatting and reporting issues. These will be detailed in an email you should receive within 2-3 business days from our colleagues in the journal operations team; no action is required from you until then. Please note that we will not be able to formally accept your manuscript and schedule it for publication until you have completed any requested changes.

Please take a minute to log into Editorial Manager at http://www.editorialmanager.com/pbiology/, click the "Update My Information" link at the top of the page, and update your user information to ensure an efficient production process.

PRESS

We frequently collaborate with press offices. If your institution or institutions have a press office, please notify them about your upcoming paper at this point, to enable them to help maximise its impact. If the press office is planning to promote your findings, we would be grateful if they could coordinate with biologypress@plos.org. If you have previously opted in to the early version process, we ask that you notify us immediately of any press plans so that we may opt out on your behalf.

We also ask that you take this opportunity to read our Embargo Policy regarding the discussion, promotion and media coverage of work that is yet to be published by PLOS. As your manuscript is not yet published, it is bound by the conditions of our Embargo Policy. Please be aware that this policy is in place both to ensure that any press coverage of your article is fully substantiated and to provide a direct link between such coverage and the published work. For full details of our Embargo Policy, please visit http://www.plos.org/about/media-inquiries/embargo-policy/.

Thank you again for choosing PLOS Biology for publication and supporting Open Access publishing. We look forward to publishing your study. 

Sincerely, 

Christian

Christian Schnell, PhD

Senior Editor

PLOS Biology

cschnell@plos.org

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 Text. Supporting methods and results.

    (DOCX)

    pbio.3002790.s001.docx (23KB, docx)
    S1 Fig. Numeric reports and predictions from a log-linear regression model as well as the BCI model as a function of the visual signal number (nV), auditory signal number (nA), and task relevance, separately for HC (left 2 columns, n = 23) and SCZ (right 2 columns, n = 17).

    (A) Mean numeric reports (across participants mean ± SEM). (B) Mean predictions (across-participants mean) from a log-linear regression model that predicted the numeric reports from the logarithmic visual and auditory signal numbers as well as their interaction (i.e., rA/V = bA * log(nA) + bV * log(nV) + bAxV * log(nV) * log(nA) + c). Statistical results for the regression model can be found in S1 Table. (C) Mean predictions from the BCI model (across-participants mean ± SEM; model-averaging with increasing sensory variances) reproduced the logarithmic compression of numeric reports.

    (DOCX)

    pbio.3002790.s002.docx (457.7KB, docx)
    S2 Fig. Overall reaction time (RT) density in HC and SCZ (n = 40).

    When fitting ex-Gaussian functions to the RT distributions, the functions’ mean (t38 = −1.325, p = 0.193, d = −0.424, BF10 = 0.620), variance (t38 = −0.511, p = 0.612, d = −0.164, BF10 = 0.346), and lambda parameter (t38 = −0.607, p = 0.548, d = −0.194, BF10 = 0.361) were not significantly different, but rather equivalent, between both groups.

    (DOCX)

    pbio.3002790.s003.docx (26.3KB, docx)
    S3 Fig. The strength of the fission and fusion illusions (across-participants mean ± SEM) is shown as a function of group (HC vs. SCZ, n = 40).

    The illusions were computed as difference in sensitivity (d prime, d’) between the unisensory baseline condition (V1A0 vs. V2A0) and the illusion conditions (fission: V1A2 vs. V2A2; fusion: V1A1 vs. V2A2). For the fission or fusion conditions, response “2” or “1,” respectively, were defined as signal. Thus, this illusion measure accounts for a possible shift in the response criterion which could be confounded with the fission or fusion illusions (Vanes et al, 2016). In contrast to Vanes and colleagues, a mixed-model ANOVAs did not reveal a significant difference in illusion strength between the 2 types of illusions (factor audiovisual illusion, F1,38 = 2.598, p = 0.115, part. η2 = 0.064, BFincl = 0.461) or a difference between HC and SCZ participants (factor group: F1,38 = 1.431, p = 0.239, part. η2 = 0.036, BFincl = 0.440; interaction illusion × group: F1,38 = 1.271, p = 0.267, part. η2 = 0.032, BFincl = 0.273). Note that we only included trials in which participants counted the number of flashes as ≤ 2.

    (DOCX)

    pbio.3002790.s004.docx (25.2KB, docx)
    S4 Fig. Bayesian information criterion values for each of the 5 models with increasing sensory variance are shown for each of the 17 patients in the SCZ group ranked according to their PANSS positive score as an index of psychosis severity.

    Model comparison was computed as BIC of each model relative to the fixed-criterion model with increasing sensory variance, i.e., higher is better. The model comparison included 5 models with increasing sensory variance and the 5 decision strategies (MA, model averaging; MS, model selection; PM, probability matching; FC, fixed-criterion model; SF, stochastic fusion model). The SCZ patients were rank-ordered according to their PANSS positive score on the x-axis. In cases of ties, patients were arbitrarily rank-ordered within their pair according to participant number.

    (DOCX)

    pbio.3002790.s005.docx (122.4KB, docx)
    S5 Fig. Characterization of psychosis severity.

    (A) Data (across-participants mean ± SEM; n = 17) of the PANSS scale for positive symptoms (7 items, range 1–7) and the scale average in the SCZ sample. (B) Sum score from the LSHS-R questionnaire (12 items, range 0–48) for a subset of the HC sample (n = 5) and the SCZ sample (n = 17). The difference between both groups is significant (paired t test, t20 = −2.216, p = 0.039, d = −1.127). (C) Sum score (across participants mean ± SEM) from the PCL questionnaire (18 items, range 18–90) for a subset of the HC sample (n = 5) and the SCZ sample (n = 16). The difference between both groups is marginally significant (paired t test, t19 = −2.0271, p = 0.057, d = −1.039). Importantly, both the LSHS-R and the PCL scores varied substantially across the patient group. This allowed us to assess whether these sensitive measures of psychotic symptoms may correlate with BCI model parameters (S3 Table). PANSS: Positive and Negative Symptom Scale; LSHS-R: Launey–Slade hallucination scale; PCL: Paranoia check list.

    (DOCX)

    pbio.3002790.s006.docx (163.3KB, docx)
    S6 Fig. Modulation of crossmodal bias (CMB) and numeric reports by previous audiovisual numeric disparity and task-relevant signal number.

    (A) The audiovisual crossmodal bias (CMB; across-participants mean ± SEM; n = 40) is shown as a function of a current trial’s numeric disparity (1, 2, or 3) × previous trial’s numeric disparity (0, 1, 2, 3) × task relevance (auditory vs. visual report) × group (HC vs. SCZ). A four-factorial mixed-model ANOVA including these factors revealed a marginally significant task-relevance × previous-disparity interaction (F2.1,80.9 = 2.616, p = 0.076, part. η2 = 0.064; effects of task-relevance, current numeric disparity and their interaction were p < 0.05 similarly as reported in Table 2; all other main and interactions effects, especially involving group, p > 0.05). Moreover, we assessed the more constrained hypothesis that the difference between CMB for A and V report would increase linearly with increasing numeric disparity of the previous trial: We applied contrast weights to the A-V report differences depending on the disparity at previous trial, i.e., −1.5 for 0, −0.5 for 1, 0.5 for 2, and 1.5 for 3 audiovisual numeric disparity on the previous trial. This linear contrast showed a significant linear modulation of the task relevance effect on the CMB by previous disparity (i.e., linear contrast on the interaction of previous disparity and task relevance; t114 = 2.712, p = 0.008). In other words, audiovisual interactions increased when the previous numeric disparity was small, but decreased when the previous numeric disparity was large. CMB = 1 for purely visual and CMB = 0 for purely auditory influence. (B) Visual (left panel) and auditory (right panel) numeric reports (across-participants mean ± SEM, n = 40) plotted as a function of current visual or auditory signal number, previous visual or auditory signal number and group (HC vs. SCZ). A three-factorial mixed-model ANOVA including these factors revealed a significant main effect of previous signal number (F1.7,64.7 = 33.192, p < 0.001, part. η2 = 0.466), an interaction of previous signal number and task relevance (F3,114 = 3.986, p = 0.010, part. η2 = 0.095), and a significant interaction of previous signal number and current signal number (F6.6,250.2 = 8.515, p < 0.001, part. η2 = 0.183). The effect of current signal number and the interaction of current signal number and task relevance were also significant (p < 0.05). All other main and interactions effects, especially involving group, were not significant.

    (DOCX)

    pbio.3002790.s007.docx (192.2KB, docx)
    S7 Fig. Multivariate ERP patterns, the spatial correlation of ERP patterns between HC and SCZ and decoding results (n = 40).

    (A) ERP patterns (i.e., topographies) are shown as a function of time, averaged in 100 ms time windows and separately for HC and SCZ participants. (B) Spatial correlation of multivariate ERP patterns between HC and SCZ participants as a function of time, separately for the 4 conditions (N.B.: Spatial correlations before stimulus onset might arise from expectation processes). (C) Decoding accuracy (i.e., fraction of correct classifications) of a decoder trained to classify the participant group (i.e., HC vs. SCZ) from ERP patterns of the 4 conditions. No significant clusters (p > 0.05) of decoding accuracies above the chance level of 0.5 were found in one-sided cluster-based corrected randomization tests. As a reference, the gray dashed lines indicate decoding accuracy (fraction ± 68% CI; averaged across the 4 conditions) of the decoder trained on randomized group membership (n = 5,000 randomizations).

    (DOCX)

    pbio.3002790.s008.docx (331.7KB, docx)
    S8 Fig. Crossmodal bias for the clinical sample including schizophrenia (SCZ, n = 17) and schizoaffective (SCA, n = 6) patients.

    Participants’ audiovisual crossmodal biases (CMB; across-participants mean ± SEM; n = 46) are shown as a function of numeric disparity (1, 2, or 3), task relevance (auditory vs. visual report), and group (HC vs. SCZ). CMB = 1 for purely visual and CMB = 0 for purely auditory influence. Source data is provided in S8 Data.

    (DOCX)

    pbio.3002790.s009.docx (68.6KB, docx)
    S9 Fig. Factorial Bayesian model comparison including SCZ (n = 17) and SCA patients (n = 6).

    Ten models with different decision strategies, model averaging (MA), model selection (MS), probability matching (PM), fixed criterion (FC), stochastic fusion (SF), each with constant or increasing sensory auditory and visual variances. The image shows the relative model evidence for each model (i.e., participant-specific Bayesian information criterion of a model relative to the worst model summed over all participants). A larger model evidence indicates that a model provides a better explanation of our data. Source data is provided in S9 Data.

    (DOCX)

    pbio.3002790.s010.docx (76.9KB, docx)
    S10 Fig. Parameters of the BCI model (across participants mean ± SEM) separately plotted for HC and SCZ (n = 17) & SCA (n = 6) patients.

    The BCI model’s decision strategy applies model averaging with increasing sensory variance. Source data is provided in S10 Data.

    (DOCX)

    pbio.3002790.s011.docx (126.2KB, docx)
    S11 Fig. Decoding the BCI model’s numeric estimates from EEG patterns using support-vector regression (SVR) in HC versus SCZ (n = 40) & SCA (n = 6) patients.

    (A) Decoding accuracy (Fisher’s z-transformed correlation; across-participants mean) of the SVR decoders as a function of time and group (HC vs. SCZ & SCA). Decoding accuracy was computed as Pearson correlation coefficient between the given BCI model’s internal estimates and BCI estimates that were decoded from EEG activity patterns using SVR models trained separately for each numeric estimate. Color-coded horizontal solid lines (HC) or dashed lines (SCZ & SCA) indicate clusters of significant decoding accuracy (p < 0.05; one-sided one-sample cluster-based corrected randomization t test). (B) Bayes factors for the comparison between the decoding accuracies of HC and SCZ & SCA for each of the BCI estimate (i.e., BF10 > 3 substantial evidence for or BF10 < 1/3 against group differences). Source data is provided in S11 Data.

    (DOCX)

    pbio.3002790.s012.docx (131.5KB, docx)
    S12 Fig. Results of the parameter recovery (n = 40).

    For parameter recovery, the winning model (i.e., model averaging with increasing sensory variances) predicted responses which were then again fitted to obtain recovered parameters with the same fitting procedure as for the main analysis (i.e., initialization with 50 different random parameters; predicted distributions were generated from 5,000 simulated trials per condition). The plots show the recovered parameters as a function of the parameters originally fitted to participants’ behavioral data. The red line is a line with slope 1 and intercept 0.

    (DOCX)

    pbio.3002790.s013.docx (248.4KB, docx)
    S13 Fig. Results of model recovery.

    To assess model recovery, each of the 5 models with different decision strategies (model averaging, MA; model selection, MS; probability matching, PM; fixed criterion, FC; stochastic fusion, SF) and increasing sensory variance parameters generated responses in 20 randomly selected participants. Each of the 5 models was then fitted to the generated responses of each model with the same fitting procedure as for the main analysis (i.e., initialization with 50 different random parameters; predicted distributions were generated from 5,000 simulated trials per condition). The confusion matrix shows for each generating model (rows) the fraction of participants in which a fitted model (columns) won the model comparison across the 5 fitted models based on Bayesian information criterion within an individual. The bright diagonal indicates that the generating models won the model comparisons, indicating successful model recovery. In particular, the winning model with model averaging as decision strategy was correctly recovered in 85% of the cases, thereby supporting the validity of our model-based analyses.

    (DOCX)

    pbio.3002790.s014.docx (67.5KB, docx)
    S14 Fig. Results of the parameter recovery (HC and SCZ, n = 40) for the modeling analysis of updated causal and numeric priors.

    For parameter recovery, the model averaging model with updated priors predicted responses which were then again fitted to obtain recovered parameters with the same fitting procedure as for the main analysis (i.e., initialization with 50 different random parameters; predicted distributions were generated from 5,000 simulated trials per condition). The plots show the recovered parameters as a function of the parameters originally fitted to participants’ behavioral data. Parameters are pooled across different levels of previous trial conditions. The red line is a line with slope 1 and intercept 0.

    (DOCX)

    pbio.3002790.s015.docx (134.8KB, docx)
    S1 Table. Statistical significance of regression parameter estimates across and between HC and SCZ in the log-linear regression model that predicted the numeric reports from the visual and auditory signal numbers as well as their interaction.

    (DOCX)

    pbio.3002790.s016.docx (15.7KB, docx)
    S2 Table. Results of the Bayesian model comparison of the 2 × 5 factorial model space with factor “decision strategy” and “sensory variance” in HC and SCZ participants.

    (DOCX)

    pbio.3002790.s017.docx (24.3KB, docx)
    S3 Table. Pearson correlations of BCI model parameters with SCZ patients’ (n = 17) positive and negative symptoms measured by PANSS, LSHS-R, and PCL.

    (DOCX)

    pbio.3002790.s018.docx (17.8KB, docx)
    S4 Table. Results of testing the decoding accuracies of the BCI estimates from EEG patterns against zero in HC, SCZ, and their group difference.

    (DOCX)

    pbio.3002790.s019.docx (14.9KB, docx)
    S5 Table. Spearman rank correlations of BCI model parameters with SCZ patients’ neurocognitive test scores (n = 17).

    (DOCX)

    pbio.3002790.s020.docx (15.8KB, docx)
    S6 Table. Main and interaction effects of task relevance (TR), disparity (Disp), and group (HC, n = 23 vs. schizophrenia, n = 17, and schizoaffective patients, n = 6) on the crossmodal bias from classical and Bayesian mixed-model ANOVAs.

    (DOCX)

    pbio.3002790.s021.docx (14.7KB, docx)
    S7 Table. Comparison of BCI model parameters between HC (n = 23) versus schizophrenia (n = 17) & schizoaffective (n = 6) patients.

    (DOCX)

    pbio.3002790.s022.docx (14.6KB, docx)
    S8 Table. Correlations of BCI model parameters with patients’ (schizophrenia, n = 17 and schizoaffective, n = 6) positive and negative symptoms measured by PANSS, LSHS-R, and PCL.

    (DOCX)

    pbio.3002790.s023.docx (17KB, docx)
    S1 Data. Source data for Fig 1 which can be opened with Matlab or GNU Octave.

    (ZIP)

    pbio.3002790.s024.zip (19.3KB, zip)
    S2 Data. Source data for Fig 2 which can be opened with Matlab or GNU Octave.

    (ZIP)

    pbio.3002790.s025.zip (17.3KB, zip)
    S3 Data. Source data for Fig 3 which can be opened with Matlab or GNU Octave.

    (ZIP)

    pbio.3002790.s026.zip (16.1KB, zip)
    S4 Data. Source data for Fig 4 which can be opened with Matlab or GNU Octave.

    (ZIP)

    pbio.3002790.s027.zip (33.9KB, zip)
    S5 Data. Source data for Fig 5 which can be opened with Matlab or GNU Octave.

    (ZIP)

    pbio.3002790.s028.zip (17.1KB, zip)
    S6 Data. Source data for Fig 6 which can be opened with Matlab or GNU Octave.

    (ZIP)

    pbio.3002790.s029.zip (65.8KB, zip)
    S7 Data. Source data for Fig 7 which can be opened with Matlab or GNU Octave.

    (ZIP)

    pbio.3002790.s030.zip (165.1KB, zip)
    S8 Data. Source data for S8 Fig which can be opened with Matlab or GNU Octave.

    (ZIP)

    pbio.3002790.s031.zip (15.4KB, zip)
    S9 Data. Source data for S9 Fig which can be opened with Matlab or GNU Octave.

    (ZIP)

    pbio.3002790.s032.zip (16.5KB, zip)
    S10 Data. Source data for S10 Fig which can be opened with Matlab or GNU Octave.

    (ZIP)

    pbio.3002790.s033.zip (37KB, zip)
    S11 Data. Source data for S11 Fig which can be opened with Matlab or GNU Octave.

    (ZIP)

    pbio.3002790.s034.zip (73.7KB, zip)
    Attachment

    Submitted filename: 2024_05_17_ResponseReviewer_R1_final2.docx

    pbio.3002790.s035.docx (58.9KB, docx)
    Attachment

    Submitted filename: 2024_08_01_ResponseLetter_R2.docx

    pbio.3002790.s036.docx (23KB, docx)

    Data Availability Statement

    All raw EEG and behavioral data as well as code for presenting audiovisual stimuli, fitting computational models and analysis of behavioral data are available from a Dryad repository (https://doi.org/10.5061/dryad.hhmgqnkr1) [116].


    Articles from PLOS Biology are provided here courtesy of PLOS

    RESOURCES