Abstract
Background and objectives
A plethora of studies has investigated and compared social cognition in autism and schizophrenia ever since both conditions were first described in conjunction more than a century ago. Recent computational theories have proposed similar mechanistic explanations for various symptoms beyond social cognition. They are grounded in the idea of a general misestimation of uncertainty but so far, almost no studies have directly compared both conditions regarding uncertainty processing. The current study aimed to do so with a particular focus on estimation of volatility, i.e. the probability for the environment to change.
Methods
A probabilistic decision-making task and a visual working (meta-)memory task were administered to a sample of 86 participants (19 with a diagnosis of high-functioning autism, 21 with a diagnosis of schizophrenia, and 46 neurotypically developing individuals).
Results
While persons with schizophrenia showed lower visual working memory accuracy than neurotypical individuals, no significant group differences were found for metamemory or any of the probabilistic decision-making task variables. Nevertheless, exploratory analyses suggest that there may be an overestimation of volatility in subgroups of participants with autism and schizophrenia. Correlations revealed relationships between different variables reflecting (mis)estimation of uncertainty, visual working memory accuracy and metamemory.
Limitations
Limitations include the comparably small sample sizes of the autism and the schizophrenia group as well as the lack of cognitive ability and clinical symptom measures.
Conclusions
The results of the current study provide partial support for the notion of a general uncertainty misestimation account of autism and schizophrenia.
Introduction
More than a century ago, the terms ‘autistic’ and ‘autism’ were coined to describe the social withdrawal observed in individuals with schizophrenia (SCZ) and a childhood form of SCZ, respectively [1]. While SCZ and autism spectrum disorders (ASD) are defined as distinct entities today [2], a substantial amount of research has investigated the shared characteristics of both conditions. Findings suggest an association from both a genetic [3, 4] and a cognitive-behavioral perspective, particularly within the social domain. Comparative and parallel studies have documented similarly impaired social cognitive abilities in SCZ and ASD relative to neurotypically developing (NT) individuals [5]. This concerns various subdomains, including theory of mind, i.e. the ability to infer others’ mental states [6, 7], eye gaze on faces [8], trustworthiness judgements and emotion identification [9]. In fact, a recent systematic review concluded that apart from emotion recognition there seem to be no clear and consistent differences between ASD and SCZ in terms of social cognitive performance [10].
While social cognition has been studied extensively, only few studies compare the two conditions in other cognitive domains [1]. However, results of separately conducted studies suggest similar decision-making impairments in non-social situations [11–13]. One decision-making bias that has extensively been investigated in persons with SCZ is the so-called “Jumping-to-Conclusions” (JTC) bias. It is usually assessed by the beads task in which beads are sampled from one out of two possible containers (e.g. bags) containing unlike amounts of differently colored beads. Based on the sampled beads, participants have to indicate what they believe to be the bag of origin [14]. Different versions of the task exist: in draws-to-decision versions, participants are free to sample as many beads as they want until they decide on the bag of origin. Here, the JTC bias is characterized by premature decisions, i.e. a decision on the bag of origin after very few beads have been sampled. In graded estimates versions of the task participants indicate their certainty about the bag of origin after each bead. Here, reasoning biases include high (initial) certainty and over-adjustment of the reported estimates, meaning radical belief alterations in response to objectively only modest disconfirmatory evidence [15, 16]. Those biases are similar to the ‘classical’ JTC bias in that they all concern drastic decision-making in light of little evidence. While typically studied in SCZ, the JTC bias has also been found in ASD [17]. Conversely, Brosnan and colleagues reported that persons with ASD gathered more beads before making a decision [18].
Of the few studies directly comparing non-social decision-making in ASD and SCZ, Zhang and colleagues [19] found similar impairments in decision-making under different kinds of uncertainty, suggesting that both conditions may be characterized by misestimation of uncertainties. Such misestimations could also explain the aberrant behavior observed in the aforementioned beads task, where performance relies on Bayesian inference [20]. This perspective fits well with computational theories proposing similar mechanistic explanations based on misestimation of uncertainty for various symptoms of ASD and SCZ [21–23]. According to these theories, symptoms might be the result of (implicit) uncertainty misestimation on different levels in Bayesian belief hierarchies of the brain [24]. One ‘level’ concerns beliefs about the environment’s volatility, i.e. the probability for the environment to change. The results of various studies indicate that both persons with ASD and persons with SCZ overestimate volatility, i.e. they seem to perceive the world as less stable. For example, they exhibit more (maladaptive) switching behavior in reversal learning tasks than NT individuals (ASD: [12, 25–27]; SCZ: [28–31]). Surprising events are thus attributed to a change in the overall stochastic structure of the environment [32] rather than to known uncertainties on lower levels, i.e. the expected uncertainty that arises naturally since some events are more likely than others in a stable but stochastic environment. Hence, new events will become more salient as they might signal a relevant change in the environment when subjective volatility is high. Consequentially, beliefs are updated more drastically. This fits well with the over-adjustment of beliefs observed in the beads task, which in turn has been attributed to a “hypersalience” of new evidence [20].
Aberrant representation of uncertainties has also been described as a ‘failure of metacognition’ [24]. Metacognition can refer to both conscious reflective thought processes and automatic monitoring of one’s own thoughts and cognitions [33]. Metacognitive performance is often determined by comparing self-reports and confidence ratings to actual performance [34]. Interestingly, impaired metacognition has been found in both SCZ [35, 36] and ASD [37, 38]. Further, previous studies have revealed a relationship between the JTC bias, corresponding decision confidence and metacognitive deficits in SCZ [36, 39, 40], but to what extent metacognition relates to higher level uncertainty estimation such as volatility remains to be elucidated.
A general misestimation of uncertainties could thus explain various cognitive-behavioral findings in both ASD and SCZ but it remains unclear if and to what extent both groups differ from each other when compared directly. This study aimed to investigate this question with a focus on volatility processing in a modified beads task and its relationship to belief updating, metacognition and working memory, to account for the potential role of general cognitive capacity.
Materials and methods
Persons with SCZ were contacted through a clinician at St. Olavs Hospital, Trondheim University Hospital, Norway, while persons with ASD were recruited through the patient interest group Autismeforeningen and, like NT control participants, through fliers and social media posts. Participants had to meet the following inclusion criteria: (1) 18 to 60 years of age, (2) no current suicide intent, (3) no substance dependence, (4) IQ above 80, (5) a primary diagnosis from the schizophrenia spectrum (SCZ group) or high-functioning autism/Asperger (ASD group) or no psychiatric diagnosis at all (NT group). All participants in the SCZ group were inpatients who had previously been diagnosed according to the ICD-10 research criteria [41] in a consensus meeting assessing clinical reports with at least two senior psychiatrists or psychologists present, of which at least one had personally examined the patient. Diagnoses were confirmed by clinicians upon inclusion in the study. All participants in the ASD group reported prior diagnoses by independent clinicians. Where available, their diagnoses were confirmed through clinical records and their employer (a business exclusively employing persons with a confirmed ASD diagnosis). For three participants with ASD, no such confirmation was available. For all participants, written informed consent was obtained prior to the study. The study was approved by the Central-Norwegian regional committee for medical and health research ethics (REC Central; reference no.: 2014/1648). In total, 92 participants were recruited, whereof six were excluded since they did not complete enough (≥ 80%) trials of the administered tasks. A subset of the participants filled out additional questionnaires but those results are not reported here.
Measures
Beads task
To measure probabilistic decision-making and subjective volatility, a modified version of the beads task was administered (see Fig 1). Two virtual bags were displayed on screen, containing 80 black and 20 white beads and the converse. Five sequences of 20 beads each were presented to the participants. At the beginning of each sequence, one of the two bags was chosen at random (p = 0.5). Each sequence was then generated based on the probabilities for the different colors to be drawn (p = 0.2 and p = 0.8) and a fixed probability for the bag of origin to change (v = 0.04 for each bead, amounting to a ca. 50% chance to observe a bag change in one sequence). This change probability introduced volatility to the task. Participants were informed about this by written instructions stating: “The chance for the bags to change is small enough that in ca. half of the sequences all 20 beads are coming from the same bag and in ca. half of the sequences the bag of origin changes.” During the instruction, the experimenter emphasized the probabilistic nature of this description and explained that more or fewer bag changes are possible. To support understanding, five practice trials (i.e. five sampled beads) were completed before the main task.
After each bead, participants had 10 seconds to indicate their certainty about the bag of origin. They did so by dragging the marker on a visual analogue scale ranging from 0 = “absolutely sure it comes from the bag with more black beads”, to 1 = “absolutely sure it comes from the bag with more white beads” where 101 different steps on the scale were mapped to probabilities. After each sequence, participants received a visual feedback about the beads they had seen, the bags they had chosen, and the true bag of origin for each bead (see Fig 1). This feedback thus provided a demonstration of how the instructed probabilities could manifest in color changes.
During the task, the instructions, the two bags, and the currently drawn sequence remained on screen. Initial certainty was measured as the average of all indicated probabilities for the first bead of each sequence. Higher values indicate a more JTC-like behavior [14]. Disconfirmatory belief updating was measured as the change in probability rating in favor of a given color whenever this color differed from the color of the two or more preceding beads. The total size of changes was first averaged across the number of occurrences of such events per sequence, and subsequently across sequences. Here, higher values reflect the formerly described over-adjustment behavior [20]. Participants’ perception of the probability for the bags to change, i.e. subjective volatility, was derived from the probabilities participants indicated for each trial n out of N = 20 in each sequence k out of K = 5. In an ideal Bayesian model, those probabilities should be based on all observed draws until the current trial n, as well as the assumed volatility v. A participant’s probability rating should consequently be their guess of the theoretical probability P(xk,n|zk,1,…,zk,n, v), where x is the bag of origin (xk,n = 0 if bag A, xk,n = 1 if bag B) and z is the color of the drawn bead (zk,n = 0 if white, zk,n = 1 if black), with n and k denoting current number of trial and sequence, respectively, and v the probability for a bag change to occur. Volatility v was estimated by finding the parameter value that would minimize the difference between the set of theoretical probabilities and the participant’s estimated probabilities (in the least-squares sense). Correlations between observed and predicted probabilities of this ‘volatility model’ indicated model fit and were moderate to high for the majority of participants, but close to zero for three of them (nASD = 1, nSCZ = 2; see Fig 2). Across the sample, model fit correlated negatively with estimated volatility (ρ = -.70, p < .001), indicating that weaker model fit was associated with higher volatility estimates. For details on the calculation of the theoretical probabilities and parameter estimation, see model description in S1 File. Note that due to the probabilistic nature of the task, the sequences displayed differed between participants. This has the benefit that any observed effects on the group level are independent of the particular sequence chosen. In contrast, administering the same fixed sequence to all participants might introduce particular sequence-dependent biases, which hinders the generalization of any potential results.
Visual working memory task
A visual working memory task developed based on previously published paradigms [42, 43] and a variant of the paradigm used by ten Velden Hegelstad and colleagues [44], was administered to measure both visual working memory and implicit metamemory as a proxy for metacognition. An implicit measure was chosen since uncertainty may be encoded without awareness and not accessible to explicit reports [45]. Working memory accuracy was included as a measure to test whether it was related to uncertainty estimation overall and to control for potential differences in cognitive capacity when interpreting group differences on the beads task variables.
A target shape was presented for one second and then had to be selected from an array of similar shapes (see Fig 3). In this array, thirty shapes that varied along continuous quantitative dimensions were displayed in a circular arrangement corresponding to their continuous modification, i.e. shorter angular distance on the circle meant higher resemblance.
The shapes were generated by drawing lines in a polar plot using the following formula:
where "phase" describes the angle relative to the reference direction (upwards) and "amplitude" the length (radius) of the vector. By varying the "shape" parameter in steps of 12 from 0 to 348 degrees, 30 continuously modified shapes were generated (see S1 Fig for a full display of all shapes). The 30 shapes to choose from remained the same across trials and participants, whereas the target shape was selected randomly for each participant on each trial out of the pool of these 30 shapes.
After selection of the target from the array of shapes, participants also set a capture area reflecting their uncertainty about how accurately they had selected the correct shape. They were instructed to set this area big enough to be sure the target shape was included but not bigger than necessary. To demotivate them from capturing the whole circle array at all times, they were rewarded with eight points when their capture area included the target shape, and punished by point subtraction proportional to the size of overshoot when making it too large, though points did not translate to any real reward after task completion. Visual feedback was provided after each trial (see Fig 3). Three practice trials and 30 test trials were administered. Participants had the option to skip trials if they had completely forgotten the sample shape. Trials with extremely large (>350 degrees) or small (<4 degrees) capture areas were excluded from analysis as they might indicate trials where participants accidentally failed to use the option to skip a trial and tried to adjust for that by not setting an appropriate capture area. Visual working memory accuracy was measured as average error, i.e. the average angular distance of the selected shape from the target shape over all trials, with lower values reflecting higher accuracy. Implicit metamemory was assessed by the proportion of all trials where the capture area included the target shape (‘hits’), with lower values indicating overestimation of actual accuracy.
Procedure
On the day of the assessment, participants were briefed regarding the background of the study and signed the consent form. They then first completed the visual working memory task, followed by the beads task. Duration of each task was ca. 15 minutes, depending on participants’ speed of responding. A short break was introduced between both tasks if required. Task order was not counter-balanced and due to the low similarity and short length of both tasks, no carry-over effects were expected. Demographics were collected on a paper sheet.
Analysis
One-way ANOVAs were conducted and their residuals tested for normality. Only disconfirmatory belief updating and estimated volatility violated the normality assumption. While disconfirmatory belief updating was log-transformed, estimated volatility followed a bimodal distribution and could not be transformed. Results were therefore verified with Kruskal-Wallis tests and for volatility, exploratory analyses using Gaussian Mixture models were conducted. Significant ANOVA F-Tests were followed by Tukey's Honest Significant Difference tests and effect sizes are reported as η2. Significant Kruskal-Wallis tests were followed by Bonferroni corrected Dunn’s Tests and effect sizes are reported as ε2. Age and sex did not differ significantly between the groups, and were not controlled for but see analyses in S2 File for group comparisons after propensity matching for both. Level of education differed significantly between the groups but could not be controlled for independently of the diagnosis as the lowest level included more than half of all patients with SCZ but only one participant from the ASD and the NT group. To gauge whether it could be a confounder for any group differences on the task related variables, significant results were followed up by education level comparisons within the NT group. Spearman correlations were chosen to investigate the relationship between the variables of interest across the whole sample. All confirmatory testing was conducted with a significance level of 0.05, one-sided where specified, using the R programming language (R version 3.5.1 [46]).
Results
Demographic variables are summarized in Table 1.
Table 1. Sample demographics per group (total sample size = 86).
ASD (n = 19) | SCZ (n = 21) | NT (n = 46) | ||||||||
---|---|---|---|---|---|---|---|---|---|---|
n | M (SD) | Md (IQR) | n | M (SD) | Md (IQR) | n | M (SD) | Md (IQR) | p | |
Sex (m/f) | 11/8 | 17/4 | 25/21 | .11 | ||||||
Education (“1”/”2”/”3”) | 1/8/10 | 12/3/2a | 1/13/32 | < .001 | ||||||
Antipsychotic medication | ||||||||||
Amisulpride | 1 | |||||||||
Aripiprazol | 5b | |||||||||
Clozapine | 2 | |||||||||
Olanzapine | 6 | |||||||||
Quetiapine | 2 | |||||||||
Risperidone | 1 | |||||||||
None | 4 | |||||||||
Age | 30.32 (8.85) | 26 (12) | 25.67 (4.74) | 26 (7) | 28.41 (7.64) | 25 (9.75) | .14 |
Sample sizes (n), counts, means (M; with standard deviations SD) and medians (Md; with inter-quartile ranges IQR) are displayed. Education was recorded in Norwegian school system categories corresponding to completion of 1 = secondary school (up to age 16), 2 = 6th form college (up to age 19), 3 = higher education (Bachelor, Master, PhD); p-values for group comparisons are provided only for the demographical variables sex and education (Chi-squared tests) as well as Age (ANOVA).
a missing data from 4 patients
b thereof two with additional Quetiapine treatment
In the beads task, sequences of beads were drawn randomly for each participant, but group comparisons indicated that on average, all groups experienced approximately the same amount of color changes per sequence, F(2,83) = 2.53, η2 = 0.06, p = .09, with MASD = 6.38, MSCZ = 6.30, and MNT = 5.79 [nonparametric analysis: χ2(2) = 4.88, ε2 = 0.06, p = .09]. Similarly, the average number of (hidden) bag changes per sequence did not differ by group, F(2,83) = 1.19, η2 = 0.03, p = .31, with MASD = 0.65, MSCZ = 0.75, and MNT = 0.78 [χ2(2) = 2.22, ε2 = 0.03, p = .33]. Behaviorally, there were no significant group differences in any of the beads task variables: initial certainty, F(2,83) = 0.09, η2 < 0.01, p = .91 [χ2(2) = 0.04, ε2 < 0.001, p = .98] (see Fig 4A); estimated volatility, F(2,83) = 1.92, η2 = 0.04, p = .15 [χ2(2) = 3.30, ε2 = 0.04, p = .19] (see Fig 4C); and log transformed disconfirmatory belief updating, F(2,83) = 1.24, η2 = 0.03, p = .30 [not log transformed for the non-parametric test: χ2(2) = 3.16, ε2 = 0.04, p = .21 (see Fig 4B)]. Average volatility estimates were higher than the instructed value of 0.04 in all groups (see Table 2). Three one-sided one-sample Wilcoxon signed-rank tests confirmed that this was significant for the ASD (Md = 0.14, V = 167, p < .01), the SCZ (Md = 0.40, V = 216, p < .001), and the NT (Md = 0.11, V = 918, p < .001) group. Model fit (correlations between predicted and observed probabilities, see Fig 2) differed significantly between groups, χ2(2) = 6.70, ε2 = 0.08, p = .04, with MdASD = 0.79, MdSCZ = 0.68, and MdNT = 0.84. Post-hoc comparisons revealed a significant difference between the NT and the SCZ group, z = 2.59, padj = .03, but not between the ASD and the NT, z = -0.78, padj > .99, or the ASD and the SCZ group, z = 1.48, padj = .42. To control for potential learning effects over the course of the task, volatility was additionally estimated separately for the first two and the last two sequences of beads. This revealed a slight decrease in volatility towards the end of the task, possibly related to learning effects in response to the visually provided feedback. However, this volatility change did not differ between groups (see S3 File for details).
Table 2. Descriptive summary statistics of the two tasks per group, effect size and p-value from the conducted ANOVAs (total sample size = 86).
ASD (n = 19) | SCZ (n = 21) | NT (n = 46) | η2 | p | ||||
---|---|---|---|---|---|---|---|---|
M (SD) | Md (IQR) | M (SD) | Md (IQR) | M (SD) | Md (IQR) | |||
Beads task | ||||||||
initial certainty | 0.73 (0.16) | 0.70 (0.25) | 0.72 (0.13) | 0.71 (0.17) | 0.71 (0.12) | 0.70 (0.18) | <0.01 | .90 |
disconfirmatory belief updatinga | 0.26 (0.20) | 0.17 (0.24) | 0.33 (0.19) | 0.28 (0.18) | 0.28 (0.19) | 0.21 (0.17) | 0.03b | .42 |
estimated volatility | 0.30 (0.30) | 0.14 (0.56) | 0.37 (0.31) | 0.40 (0.56) | 0.23 (0.24) | 0.11 (0.41) | 0.04 | .16 |
VWM task | ||||||||
proportion hits | 0.57 (0.16) | 0.54 (0.23) | 0.52 (0.13) | 0.50 (0.15) | 0.58 (0.12) | 0.57 (0.14) | 0.04 | .18 |
error | 28.23 (11.07) | 26.31 (16.84) | 35.72 (11.58) | 35.67 (12.77) | 24.93 (9.18) | 23.88 (10.65) | 0.15 | < .001 |
M = mean, SD = standard deviation, Md = median, IQR = interquartile range, VWM = visual working memory
a descriptive data not log-transformed but based on original scale
b effect size based on log transformed data
In the visual working memory task, a non-parametric group comparison of number of skipped trials revealed no significant group differences, χ2(2) = 1.99, ε2 = 0.02, p = .37, with MdASD = 0.00, MdSCZ = 0.00, and MdNT = 0.00. A similar comparison for number of trials where the capture area was out of range (i.e. <4 or >350 degrees), also demonstrated no significant differences between groups, χ2(2) = 0.59, ε2 = 0.01, p = .75, with MdASD = 0.00, MdSCZ = 1.00, and MdNT = 0.00. There was a significant effect of group on average error (i.e. memory inaccuracy), F(2,83) = 8.03, η2 = 0.16, p < .001 [χ2(2) = 12.91, ε2 = 0.15, p < .01] (see Table 2 and Fig 4D). Post-hoc comparisons revealed that the average error in the SCZ group (M = 35.72, SD = 11.58) was significantly larger than in the NT group (M = 24.93, SD = 9.18), padj < .001 [nonparametric analysis: z = -3.59, padj < .001], and numerically but not significantly larger compared to the ASD group (M = 28.23, SD = 11.07), padj = .06 [z = -1.97, padj = .15]. ASD and the NT group did not differ, padj = .47 [z = 1.18, padj = .71]. Within the NT group, level of education was unrelated to memory inaccuracy, F(2,43) = 1.39, η2 = 0.06, p = .26 [χ2(2) = 2.11, ε2 = 0.05, p = .35]. For proportion of hits (i.e. metamemory), no significant group differences were found, F(2,83) = 1.73, η2 = 0.04, p = .18 [χ2(2) = 3.29, ε2 = 0.04, p = .19] (see Fig 4E). To control for potential effects of response times in the visual working memory task, additional analyses were conducted. These revealed that the SCZ group responded faster on average, but that independent of group membership longer response times were associated with larger errors (see S4 File for details).
Disconfirmatory belief updating correlated with both initial certainty (ρ = .48, p < .001) and estimated volatility (ρ = .62, p < .001, see Fig 5), but there was no significant relationship between initial certainty and estimated volatility (ρ = .09, p = .39). There was a strong correlation between proportion of hits (metamemory) and average error (memory inaccuracy) as measured by the visual working memory task (ρ = -.59, p < .001). Across tasks, average error was positively correlated with disconfirmatory belief updating (ρ = .33, p < .01) and with estimated volatility (ρ = .40, p < .001), but not initial certainty (ρ = .04, p = .68). Proportion of hits was not related to initial certainty (ρ = .12, p = .28) or disconfirmatory belief updating (ρ = -.09, p = .42) but was negatively associated with estimated volatility (ρ = -.24, p = .03). Thus, participants with better metamemory as measured by the visual working memory task also tended to estimate the volatility within the beads task more appropriately, with lower values approaching the true volatility that was introduced by the task design.
As visible in Fig 4C, estimated volatility v followed a bimodal distribution, suggesting one high- and one low-volatility cluster. This structure may have masked potential between-group effects in traditional and non-parametric tests. In an exploratory approach, the bimodality of this variable was therefore modeled using Gaussian mixture models in conjunction with a Bayesian estimation method. That approach allowed for the extraction of posterior probability distributions to find the most likely values of the estimated coefficients given the data [47]. The model can be written as
where (μ1, σ1) and (μ2, σ2) are the parameters of the first and second cluster, respectively, and θ is the mixing proportion indicating the relative proportion of subjects who belonged to the first vs. the second cluster. Volatility values (formerly v) are labeled as y to emphasize the fact that they are treated as data in this estimation context. Prior distributions were specified to be weakly informative [48] with the standard-deviations
and the means
Hamiltonian Monte-Carlo (HMC) methods were applied and implemented in the Stan software [49] using the RStan interface [50]. All models were fitted using four independent chains with 2000 iterations per chain where the first 1000 steps were discarded as warm-up samples. The Gelman-Rubin diagnostic [51] was used to ensure convergence and all . Results are reported in terms of the posterior mean value and the 95% highest-density intervals (HDI) which cover the area in which the true parameter value is located with probability 95% given the model structure. In order to detect group-level effects, parameters μ1, μ2 and θ were modeled separately per group and the resulting models were compared using leave-one-out cross-validation (LOOCV [52]). Concretely, a sequence of models of increasing complexity was designed and the leave-one-out information criterion (LOOIC) was calculated for each (see Table 3). This criterion can be interpreted similarly as the AIC and BIC criteria (lower values indicate better fit) but is appropriate for Bayesian models.
Table 3. Model comparison.
Rank | Free variables between groups | LOOIC | SE(LOOIC) | ΔLOOIC | SE(ΔLOOIC) |
---|---|---|---|---|---|
1 | μ2 | -70.20 | 17.05 | – | – |
2 | none | -67.81 | 17.10 | 2.39 | 2.91 |
3 | μ1 and μ2 | -66.86 | 16.59 | 3.34 | 1.37 |
4 | μ1, μ2 and θ | -65.98 | 16.49 | 4.22 | 1.72 |
5 | μ2 and θ | -36.46 | 13.35 | 33.74 | 4.48 |
As can be seen in Table 3, the model which allowed the mean of the high-volatility cluster (μ2) to vary between groups performed best in comparison to the baseline-model in which no group-differences were modeled. The model successfully identified two separate clusters, one that was very close to the optimal volatility value of voptimal = 0.04 with a cluster mean of μ1 = 0.05, HDI = [0.04,0.07] and small variance (σ1 = 0.04, HDI = [0.03,0.06]) and one that was centered at μ2 = 0.51, HDI = [0.43,0.59] (σ2 = 0.17, HDI = [0.12,0.22]) reflecting well the bimodal nature of the distribution. Further, the size of the two clusters was very similar with approximately 53% of the subjects belonging to the first (close-to-optimal) cluster, θ = 0.53, HDI = [0.42,0.65]. Model-fit was excellent as determined by the posterior predictive distributions for all groups in Fig 6A–6C.
Both the ASD and the SCZ group had a slightly elevated mean in the high-volatility cluster. For the ASD group, the effect was bASD = 0.07, HDI = [-0.05,0.19] with a probability of a truly higher volatility in this cluster compared to the NT group of 89%. For the SCZ group, the effect was bSCZ = 0.09, HDI = [-0.01,0.20], with a probability of a truly higher volatility than the NT group of 95%. To check for the effect of education, a median split on estimated volatility was conducted. Within the NT group, level of education was unrelated to volatility ratings being above or below and equal to the median, χ2(2) = 0.04, p = .98. Further, volatility estimates within the median-split groups did not differ by education (below and equal to median: F(1,23) = 2.94, η2 = 0.11, p = .10 [χ2(1) = 2.95, ε2 = 0.12, p = .09]; above: F(2,18) = 0.04, η2 < 0.01, p = .96 [χ2(2) = 0.54, ε2 = 0.03, p = .76]).
Discussion
This study investigated probabilistic decision-making and visual (meta-)memory in persons with schizophrenia (SCZ group), persons with high-functioning autism (ASD group) and neurotypically developing individuals without any psychiatric diagnosis (NT group) to explore if and to what extent groups differed in processing of probabilistic information and subsequent estimation of uncertainty. Unexpectedly, none of the groups differed significantly on any of the probabilistic reasoning measures. Relative to NT individuals, neither participants with SCZ nor persons with ASD showed significantly higher or lower certainty when making their first probability rating, when integrating new evidence with previous beliefs or when interpreting the volatility of the task environment. Similarly, none of the groups differed regarding their (un)certainty about their own visual memory performance (metamemory). However, participants with SCZ showed lower visual working memory accuracy than participants of the NT group.
While the absence of a difference between ASD and SCZ group in subjectively perceived volatility is not unexpected in light of the literature that found overestimation of volatility in both groups (e.g. [26, 28]), it is surprising that neither clinical group differed from the NT group. However, additional analyses revealed two clusters of participants: those who estimated volatility in a near-optimal manner and those who strongly overestimated volatility. Within the second cluster, volatility was higher in individuals with ASD and SCZ compared to the NT group, confirming in part the aforementioned findings of volatility overestimation in those clinical groups. The bimodal distribution itself might indicate qualitatively different processing modes [53]. Such processing modes could be related to the different decision-making strategies proposed in the reinforcement learning literature: a model-based mode, which relies on a cognitive representation of state transitions and a complex model of the task overall, and a model-free mode, which is more habitual and driven by trial-and-error feedback [54]. Participants in the low volatility cluster might be more prone to model-based strategies, whereas participants in the high volatility cluster may be more sensitive to trial-wise fluctuations of colors. While both modes are in theory available to all individuals, the choice of one strategy over the other can vary depending on the task at hand and available cognitive resources (e.g. [55]). Notably, volatility was higher for persons with lower working memory accuracy in the current study.
Nevertheless, the absence of overall group differences in the main analyses seems at odds with studies reporting a general overestimation of volatility in individuals with ASD or SCZ. Crucially, most previous studies did not inform their participants about the actual size of the change probability. Instead, it had to be inferred from exposure to the learning environment (e.g. [29, 31]). In contrast, the current study attempted to induce the same prior belief in all groups by providing explicit instructions about the degree of the task environment’s volatility. While one possible explanation of the bimodal volatility distribution is the aforementioned choice of processing mode, another explanation may be individual differences in understanding of the instructions. It is possible that individuals in the high volatility cluster misunderstood the instructions and assumed bags would change with a probability of 0.5 per bead rather than per sequence. Interestingly, miscomprehension of task instructions has been suggested as an explanation for the Jumping-to-Conclusions (JTC) bias in other versions of the beads task [15]. Similarly, misunderstanding of probabilities has been found to explain the JTC bias, possibly caused by reduced general cognitive abilities [56]. In order to clarify the effect of explicit information about volatility on behavior, future studies should contrast conditions where volatility is explicitly announced against conditions where it is not. Further, the role of working memory and other cognitive ability measures in this context should be elaborated, as they may link to the understanding of probabilities and (mis)comprehension of task instructions.
Importantly, while the volatility-estimating model fitted the data of the majority of participants well, model fit was significantly weaker for the SCZ group. Furthermore, weaker model fit was associated with increased volatility estimates across the sample. This may reflect the aforementioned deviation from task instructions or a different choice of processing mode in participants with high volatility values, causing an increased deviation from the behavior the theoretical model would predict. Nevertheless, the estimated volatility values were still those that fitted the observed behavior best, even if not perfectly, and model fit was still reasonable for the majority of participants with high volatility estimates (see Fig 2).
The absence of differences in other, directly observable JTC related variables was surprising. While it was unclear what to expect for participants with ASD given the few and contradictory findings (see [17, 18]), over-adjustment in response to disconfirmatory evidence has been reported for patients with SCZ [20]. This inconsistency with previous findings may in part be related to the choice of method. For SCZ, group differences seem to be less consistent in graded estimates versions of the beads task [16]. Further, the explicit introduction of volatility in the current study may have contributed to the absence of group differences. Similar beliefs about the task’s volatility across groups could cause similar belief updating, as over-adjustment (i.e. increased disconfirmatory belief updating) is likely related to overestimation of volatility: In an environment that is constantly changing, the newest observations seem most reliable and therefore deserve greater attention. This interpretation is supported by the positive correlation between disconfirmatory belief updating and estimated volatility. Hence, introducing volatility explicitly in the current study may have eliminated the difference between persons who typically overestimate volatility (persons with ASD and SCZ) and those that do not (NT group). Importantly, the volatility parameter of the model used in this study is estimated based on all trial-wise deviations of participants’ probabilistic estimates from an ideal Bayesian observer. The model rests on the assumption that these deviations are mainly caused by a misestimation of the true volatility. Yet, other causes for such deviations are conceivable, even if unlikely. As such, estimated volatility might be affected by “noisy” decision-making (see S3 File for additional analyses that address this question). Nonetheless, the positive correlation with disconfirmatory belief updating seems to substantiate the idea that estimated volatility reflects at least in part a belief about the probability for the bags to change, i.e. subjectively perceived volatility of the environment.
The lack of group differences in metamemory could be the result of measuring it implicitly as opposed to former studies that used explicit self-reports (e.g. [36, 37]). It has been suggested that implicit metacognition relies on a different cognitive system than explicit metacognition and is only minimally dependent on working memory [57]. These findings are also in line with recent reports of intact implicit metacognition in SCZ [44] and metacognitive efficiency in first episode psychosis [58]. Interestingly, metamemory was negatively related to estimated volatility. This suggests, that both misestimation of subjective cognitive capacity and overestimation of environmental uncertainty (such as volatility) might be affected by similar mechanisms, potentially driven by higher-level uncertainty calculations in the belief hierarchy of the human mind, and is in line with the conceptualization of aberrant representation of uncertainties as a ‘failure of metacognition’ [24]. Notably, average metamemory scores were rather low, with proportion of hits of 50% to ca. 60% for each group. On the one hand, this might indicate an overall tendency of participants to overestimate the accuracy with which they had remembered and correctly identified the target shape. On the other, this may in part be due to difficulties in perceptually differentiating between the shape stimuli overall, suggesting that the task in that regard might have been slightly too demanding.
The finding of lower visual working memory accuracy only for participants with SCZ relative to the NT group was little surprising. Working memory deficits are well established in SCZ (e.g. [59, 60]) but not in high-functioning autism, where findings are less consistent and performance, particularly in the visual domain, is often unimpaired (e.g. [61, 62]). Lower visual working memory accuracy was related to disconfirmatory belief updating across the whole sample. This fits well with findings that linked the JTC bias to memory performance [40, 59].
Limitations of the current study include the rather small sample sizes for the ASD and the SCZ group. The power of this study might have been too low to detect actual group differences in some of the measures. This is particularly the case for estimated volatility, where descriptive statistics and additional modelling suggest higher values in parts of the SCZ and the ASD group. The study would further have profited from the inclusion of additional cognitive ability measures. It remains unclear to what extent differences in cognitive ability may have attributed to differences in probability estimation and task comprehension. This similarly concerns the findings for visual working memory and JTC, both of which have been linked to general cognitive ability [63, 64]. While possibly related, educational degree was not controlled for in the analyses, as differences in educational levels were so large, that their effects could not be assessed independently of clinical diagnoses. However, within the NT group, education was unrelated to the main variables of interest, though it is noteworthy that the lowest educational level was underrepresented in this group. Groups were not matched by education prior to data analysis as this has been criticized for possibly leading to the selection of an atypical, high-achieving SCZ sample [65].
Further, this comparative approach was purely diagnosis-based and there was no differentiation between patients by symptoms. However, recent studies have not found any correlations between severity of psychopathological symptoms and volatility estimation [32] or aberrant switching behavior [28, 30] in SCZ. For ASD, the relationship is less clear with some studies reporting no relationship between ASD-typical symptoms and volatility-related behavior [66], some suggesting a relationship with few of the behavioral variables [25], and some not investigating any correlations along those lines [12, 27]. It is unclear whether linear relationships should even be expected in a cross-sectional design as some of the symptoms (e.g. delusions in SCZ, rigid behavior in ASD) may constitute a secondary coping mechanism in response to prior volatility overestimation [21, 22]. Regarding the often investigated relationship between JTC like behavior and delusions, results are similarly inconsistent [16, 67], but point towards an absence of this relationship for certainty and responses to contradictory evidence [67]. Furthermore, type or dose of medication were not controlled for in the current study. Antipsychotic medication might worsen or improve cognitive capacity. However, some of the study’s main variables were similar to those investigated in the JTC bias literature and previous findings actually indicated that JTC is not influenced by antipsychotic drugs (e.g. [68, 69]). Finally, the SCZ group was recruited amongst the most severely ill patients (inpatient care) and a majority were males. It is therefore unclear how well the results can be generalized.
To summarize, this study demonstrates reduced visual working memory accuracy of SCZ patients compared to NT controls. Further, the findings did not reveal any group differences for metamemory but suggest higher overestimation of volatility among some participants with autism and schizophrenia. This partially supports the conceptualization of uncertainty misestimation based approaches to phenomenology of these conditions. Nevertheless, despite similarities in social and non-social cognitive performance, both conditions’ symptomatology is heterogeneous in nature and while overlap of some clinical symptoms exists, many of them are rather particular for one of the conditions, respectively (e.g. rigid behavior in ASD, delusions or hallucinations in SCZ). It remains unclear how, if present, similar underlying mechanisms can account for that and future studies should investigate this more closely, linking subjective volatility estimation to clinical symptoms and cognitive ability in a longitudinal design.
Supporting information
Acknowledgments
We thank all participants for their time and effort as well as Kristin Sandvik, Mari Arnesen and Jonas Kymre for assistance with data collection.
Data Availability
Anonymized raw and processed data are available in an Open Science Framework repository: DOI 10.17605/OSF.IO/UCA5E.
Funding Statement
This work was supported by a grant from the Research Council of Norway (https://www.forskningsradet.no/en/), awarded to GP, with grant number FRIMEDBIO 262338. The funder had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript. The publication charges for this article have been funded by a grant from the publication fund of UiT – The Arctic University of Norway.
References
- 1.Eack SM, Bahorik AL, McKnight SAF, Hogarty SS, Greenwald DP, Newhill CE, et al. Commonalities in social and non-social cognitive impairments in adults with autism spectrum disorder and schizophrenia. Schizophrenia Research. 2013;148(1):24–8. 10.1016/j.schres.2013.05.013 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.American Psychiatric Association. Diagnostic and Statistical Manual of Mental Disorders: DSM-5. 5 ed. Washington DC: American Psychiatric Association; 2013. [Google Scholar]
- 3.Burbach JP, van der Zwaag B. Contact in the genetics of autism and schizophrenia. Trends Neurosci. 2009;32(2):69–72. Epub 2009/01/13. 10.1016/j.tins.2008.11.002 . [DOI] [PubMed] [Google Scholar]
- 4.Craddock N, Owen MJ. The Kraepelinian dichotomy—going, going… but still not gone. The British journal of psychiatry: The journal of mental science. 2010;196(2):92–5. 10.1192/bjp.bp.109.073429 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.Pinkham AE, Morrison KE, Penn DL, Harvey PD, Kelsven S, Ludwig K, et al. Comprehensive comparison of social cognitive performance in autism spectrum disorder and schizophrenia. Psychological Medicine. 2019:1–9. Epub 2019/10/02. 10.1017/S0033291719002708 [DOI] [PubMed] [Google Scholar]
- 6.Lysaker PH, Olesek KL, Warman DM, Martin JM, Salzman AK, Nicolò G, et al. Metacognition in schizophrenia: Correlates and stability of deficits in theory of mind and self-reflectivity. Psychiatry Research. 2011;190:18–22. 10.1016/j.psychres.2010.07.016 [DOI] [PubMed] [Google Scholar]
- 7.Williams D, Happé F. Representing intentions in self and other: Studies of autism and typical development. Developmental Science. 2010;13(2):307–19. 10.1111/j.1467-7687.2009.00885.x [DOI] [PubMed] [Google Scholar]
- 8.Sasson N, Tsuchiya N, Hurley R, Couture SM, Penn DL, Adolphs R, et al. Orienting to social stimuli differentiates social cognitive impairment in autism and schizophrenia. Neuropsychologia. 2007;45(11):2580–8. 10.1016/j.neuropsychologia.2007.03.009 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.Couture SM, Penn DL, Losh M, Adolphs R, Hurley R, Piven J. Comparison of social cognitive functioning in schizophrenia and high functioning autism: More convergence than divergence. Psychol Med. 2010;40(4):569–79. Epub 2009/08/13. 10.1017/S003329170999078X [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Fernandes JM, Cajão R, Lopes R, Jerónimo R, Barahona-Corrêa JB. Social Cognition in Schizophrenia and Autism Spectrum Disorders: A Systematic Review and Meta-Analysis of Direct Comparisons. Frontiers in Psychiatry. 2018;9(504). 10.3389/fpsyt.2018.00504 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Brown EC, Hack SM, Gold JM, Carpenter WT, Fischer BA, Prentice KP, et al. Integrating frequency and magnitude information in decision-making in schizophrenia: An account of patient performance on the Iowa Gambling Task. Journal of Psychiatric Research. 2015;66–67:16–23. 10.1016/j.jpsychires.2015.04.007 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Mussey JL, Travers BG, Klinger LG, Klinger MR. Decision-Making Skills in ASD: Performance on the Iowa Gambling Task. Autism Research. 2015;8(1):105–14. 10.1002/aur.1429 [DOI] [PubMed] [Google Scholar]
- 13.Raffard S, Capdevielle D, Gely-Nargeot M-C, Attal J, Baillard A, Del-Monte J, et al. Insight is not associated with insensitivity to future consequences in schizophrenia. Psychiatry Research. 2011;187(1):307–9. 10.1016/j.psychres.2010.11.020. [DOI] [PubMed] [Google Scholar]
- 14.Huq SF, Garety PA, Hemsley DR. Probabilistic judgements in deluded and non-deluded subjects. The Quarterly Journal of Experimental Psychology Section A. 1988;40(4):801–12. 10.1080/14640748808402300 [DOI] [PubMed] [Google Scholar]
- 15.Balzan RP, Delfabbro PH, Galletly CA, Woodward TS. Over-adjustment or miscomprehension? A re-examination of the jumping to conclusions bias. Aust N Z J Psychiatry. 2012;46(6):532–40. Epub 2012/06/09. 10.1177/0004867411435291 . [DOI] [PubMed] [Google Scholar]
- 16.Moritz S, Woodward TS. Jumping to conclusions in delusional and non-delusional schizophrenic patients. British Journal of Clinical Psychology. 2005;44(2):193–207. 10.1348/014466505X35678 [DOI] [PubMed] [Google Scholar]
- 17.Jansch C, Hare DJ. An investigation of the "jumping to conclusions" data-gathering bias and paranoid thoughts in Asperger syndrome. J Autism Dev Disord. 2014;44(1):111–9. Epub 2013/08/13. 10.1007/s10803-013-1855-2 . [DOI] [PubMed] [Google Scholar]
- 18.Brosnan M, Chapman E, Ashwin C. Adolescents with autism spectrum disorder show a circumspect reasoning bias rather than 'jumping-to-conclusions'. J Autism Dev Disord. 2014;44(3):513–20. Epub 2013/09/05. 10.1007/s10803-013-1897-5 . [DOI] [PubMed] [Google Scholar]
- 19.Zhang L, Tang J, Dong Y, Ji Y, Tao R, Liang Z, et al. Similarities and Differences in Decision-Making Impairments between Autism Spectrum Disorder and Schizophrenia. Frontiers in Behavioral Neuroscience. 2015;9(259). 10.3389/fnbeh.2015.00259 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Speechley WJ, Whitman JC, Woodward TS. The contribution of hypersalience to the "jumping to conclusions" bias associated with delusions in schizophrenia. Journal of Psychiatry & Neuroscience. 2010;35(1):7–17. 10.1503/jpn.090025 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Fletcher PC, Frith CD. Perceiving is believing: A Bayesian approach to explaining the positive symptoms of schizophrenia. Nat Rev Neurosci. 2009;10(1):48–58. Epub 2008/12/04. 10.1038/nrn2536 . [DOI] [PubMed] [Google Scholar]
- 22.Van de Cruys S, Evers K, Van der Hallen R, Van Eylen L, Boets B, de-Wit L, et al. Precise minds in uncertain worlds: Predictive coding in autism. Psychol Rev. 2014;121(4):649–75. Epub 2014/10/28. 10.1037/a0037665 . [DOI] [PubMed] [Google Scholar]
- 23.van Schalkwyk GI, Volkmar FR, Corlett PR. A predictive coding account of psychotic symptoms in autism spectrum disorder. Journal of Autism and Developmental Disorders. 2017;47(5):1323–40. 10.1007/s10803-017-3065-9 [DOI] [PubMed] [Google Scholar]
- 24.Adams RA, Stephan KE, Brown HR, Frith CD, Friston KJ. The computational anatomy of psychosis. Front Psychiatry. 2013;4:47 Epub 2013/06/12. 10.3389/fpsyt.2013.00047 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.D'Cruz A-M, Ragozzino ME, Mosconi MW, Shrestha S, Cook EH, Sweeney JA. Reduced behavioral flexibility in autism spectrum disorders. Neuropsychology. 2013;27(2):152–60. 10.1037/a0031721 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Lawson RP, Mathys C, Rees G. Adults with autism overestimate the volatility of the sensory environment. Nat Neurosci. 2017;20(9):1293–9. Epub 2017/08/02. 10.1038/nn.4615 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Solomon M, Smith AC, Frank MJ, Ly S, Carter CS. Probabilistic reinforcement learning in adults with autism spectrum disorders. Autism Research. 2011;4(2):109–20. 10.1002/aur.177 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28.Culbreth AJ, Gold JM, Cools R, Barch DM. Impaired Activation in Cognitive Control Regions Predicts Reversal Learning in Schizophrenia. Schizophrenia Bulletin. 2016;42(2):484–93. 10.1093/schbul/sbv075 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Kaplan CM, Saha D, Molina JL, Hockeimer WD, Postell EM, Apud JA, et al. Estimating changing contexts in schizophrenia. Brain. 2016;139(7):2082–95. 10.1093/brain/aww095 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30.Li CT, Lai WS, Liu CM, Hsu YF. Inferring reward prediction errors in patients with schizophrenia: A dynamic reward task for reinforcement learning. Front Psychol. 2014;5:1282 Epub 2014/11/27. 10.3389/fpsyg.2014.01282 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 31.Schlagenhauf F, Huys QJM, Deserno L, Rapp MA, Beck A, Heinze H-J, et al. Striatal dysfunction during reversal learning in unmedicated schizophrenia patients. NeuroImage. 2014;89:171–80. 10.1016/j.neuroimage.2013.11.034 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 32.Deserno L, Boehme R, Mathys C, Katthagen T, Kaminski J, Stephan KE, et al. Volatility Estimates Increase Choice Switching and Relate to Prefrontal Activity in Schizophrenia. Biological Psychiatry: Cognitive Neuroscience and Neuroimaging. 2020;5(2):173–83. 10.1016/j.bpsc.2019.10.007. [DOI] [PubMed] [Google Scholar]
- 33.Flavell JH. Metacognition and cognitive monitoring: A new area of cognitive-developmental inquiry. American Psychologist. 1979;34(10):906–11. 10.1037/0003-066X.34.10.906. 1980-09388-001. [DOI] [Google Scholar]
- 34.Norman E, Pfuhl G, Sæle RG, Svartdal F, Låg T, Dahl TI. Metacognition in Psychology. Review of General Psychology. 2019;23(4):403–24. 10.1177/1089268019883821 [DOI] [Google Scholar]
- 35.Lysaker PH, Dimaggio G. Metacognitive capacities for reflection in schizophrenia: Implications for developing treatments. Schizophrenia bulletin. 2014;40(3):487–91. Epub 2014/03/17. 10.1093/schbul/sbu038 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 36.Moritz S, Woodward TS, Jelinek L, Klinge R. Memory and metamemory in schizophrenia: A liberal acceptance account of psychosis. Psychological Medicine. 2008;38:825–32. 10.1017/S0033291707002553 [DOI] [PubMed] [Google Scholar]
- 37.Grainger C, Williams DM, Lind SE. Metacognition, metamemory, and mindreading in high-functioning adults with autism spectrum disorder. J Abnorm Psychol. 2014;123(3):650–9. Epub 2014/06/24. 10.1037/a0036531 . [DOI] [PubMed] [Google Scholar]
- 38.Wilkinson DA, Best CA, Minshew NJ, Strauss MS. Memory Awareness for Faces in Individuals with Autism. Journal of Autism and Developmental Disorders. 2010;40(11):1371–7. 10.1007/s10803-010-0995-x [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39.Buck KD, Warman DM, Huddy V, Lysaker PH. The Relationship of Metacognition with Jumping to Conclusions among Persons with Schizophrenia Spectrum Disorders. Psychopathology. 2012;45(5):271–5. 10.1159/000330892 [DOI] [PubMed] [Google Scholar]
- 40.Takeda T, Nakataki M, Ohta M, Hamatani S, Matsuura K, Ohmori T. Effect of cognitive function on jumping to conclusion in patients with schizophrenia. Schizophrenia Research: Cognition. 2018;12:50–5. 10.1016/j.scog.2018.04.002 PMC6007052. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 41.World Health Organization. The ICD-10 classification of mental and behavioural disorders: Clinical descriptions and diagnostic guidelines Geneva: World Health Organization; 1992. [Google Scholar]
- 42.Pfuhl G, Barrera LB, Living M, Biegler R. Do adjustments in search behavior depend on the precision of spatial memory? Learn Behav. 2013;41(1):77–93. Epub 2012/08/14. 10.3758/s13420-012-0087-1 . [DOI] [PubMed] [Google Scholar]
- 43.Zhang W, Luck SJ. Discrete fixed-resolution representations in visual working memory. Nature. 2008;453(7192):233–5. Epub 2008/04/04. 10.1038/nature06860 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 44.ten Velden Hegelstad W, Kreis I, Tjelmeland H, Pfuhl G. Psychosis and psychotic-like symptoms affect cognitive abilities but not motivation in a foraging task. Frontiers in Psychology. 2020;11 10.3389/fpsyg.2020.01632 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 45.Knoblich G, Stottmeister F, Kircher T. Self-monitoring in patients with schizophrenia. Psychological Medicine. 2004;34(8):1561–9. Epub 2004/11/04. 10.1017/s0033291704002454 [DOI] [PubMed] [Google Scholar]
- 46.R Core Team. R: A language and environment for statistical computing Vienna, Austria: R Foundation for Statistical Computing; 2018. [Google Scholar]
- 47.Gelman A, Carlin JB, Stern HS, Dunson DB, Vehtari A, Rubin DB. Bayesian data analysis: CRC press; 2013. [Google Scholar]
- 48.Chung Y, Gelman A, Rabe-Hesketh S, Liu J, Dorie V. Weakly informative prior for point estimation of covariance matrices in hierarchical models. Journal of Educational and Behavioral Statistics. 2015;40(2):136–57. [Google Scholar]
- 49.Carpenter B, Gelman A, Hoffman MD, Lee D, Goodrich B, Betancourt M, et al. Stan: A probabilistic programming language. Journal of statistical software. 2017;76(1). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 50.Stan Development Team. RStan: The R interface to Stan. R package version 2019. http://mc-stan.org/.
- 51.Gelman A, Rubin DB. Inference from iterative simulation using multiple sequences. Statistical science. 1992;7(4):457–72. [Google Scholar]
- 52.Vehtari A, Gelman A, Gabry J. Practical Bayesian model evaluation using leave-one-out cross-validation and WAIC. Statistics and computing. 2017;27(5):1413–32. [Google Scholar]
- 53.Freeman JB, Dale R. Assessing bimodality to detect the presence of a dual cognitive process. Behavior research methods. 2013;45(1):83–97. 10.3758/s13428-012-0225-x [DOI] [PubMed] [Google Scholar]
- 54.Daw ND, Niv Y, Dayan P. Uncertainty-based competition between prefrontal and dorsolateral striatal systems for behavioral control. Nature neuroscience. 2005;8(12):1704–11. 10.1038/nn1560 [DOI] [PubMed] [Google Scholar]
- 55.Otto AR, Gershman SJ, Markman AB, Daw ND. The curse of planning: Dissecting multiple reinforcement-learning systems by taxing the central executive. Psychological science. 2013;24(5):751–61. 10.1177/0956797612463080 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 56.Klein HS, Pinkham AE. Examining reasoning biases in schizophrenia using a modified “Jumping to Conclusions” probabilistic reasoning task. Psychiatry research. 2018;270:180–6. 10.1016/j.psychres.2018.09.020 [DOI] [PubMed] [Google Scholar]
- 57.Shea N, Boldt A, Bang D, Yeung N, Heyes C, Frith CD. Supra-personal cognitive control and metacognition. Trends in Cognitive Sciences. 2014;18(4):186–93. 10.1016/j.tics.2014.01.006 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 58.Wright A, Nelson B, Fowler D, Greenwood K. Perceptual biases and metacognition and their association with anomalous self experiences in first episode psychosis. Consciousness and Cognition. 2020;77:102847 10.1016/j.concog.2019.102847 [DOI] [PubMed] [Google Scholar]
- 59.Freeman D, Startup H, Dunn G, Cernis E, Wingham G, Pugh K, et al. Understanding jumping to conclusions in patients with persecutory delusions: Working memory and intolerance of uncertainty. Psychol Med. 2014;44(14):3017–24. Epub 2014/07/30. 10.1017/S0033291714000592 . [DOI] [PubMed] [Google Scholar]
- 60.Horan WP, Braff DL, Nuechterlein KH, Sugar CA, Cadenhead KS, Calkins ME, et al. Verbal working memory impairments in individuals with schizophrenia and their first-degree relatives: Findings from the Consortium on the Genetics of Schizophrenia. Schizophrenia Research. 2008;103(1):218–28. 10.1016/j.schres.2008.02.014. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 61.Jiang YV, Palm BE, DeBolt MC, Goh YS. High-precision visual long-term memory in children with high-functioning autism. Journal of Abnormal Psychology. 2015;124(2):447–56. 10.1037/abn0000022 [DOI] [PubMed] [Google Scholar]
- 62.Koshino H, Carpenter PA, Minshew NJ, Cherkassky VL, Keller TA, Just MA. Functional connectivity in an fMRI working memory task in high-functioning autism. NeuroImage. 2005;24(3):810–21. 10.1016/j.neuroimage.2004.09.028 [DOI] [PubMed] [Google Scholar]
- 63.Bentall RP, Rowse G, Shryane N, Kinderman P, Howard R, Blackwood N, et al. The Cognitive and Affective Structure of Paranoid Delusions: A Transdiagnostic Investigation of Patients With Schizophrenia Spectrum Disorders and Depression. Archives of General Psychiatry. 2009;66(3):236–47. 10.1001/archgenpsychiatry.2009.1 [DOI] [PubMed] [Google Scholar]
- 64.Conway ARA, Kane MJ, Engle RW. Working memory capacity and its relation to general intelligence. Trends in Cognitive Sciences. 2003;7(12):547–52. 10.1016/j.tics.2003.10.005 [DOI] [PubMed] [Google Scholar]
- 65.Resnick SM. Matching for education in studies of schizophrenia. Archives of General Psychiatry. 1992;49(3):246 10.1001/archpsyc.1992.01820030078011 [DOI] [PubMed] [Google Scholar]
- 66.Goris J., Silvetti M., Verguts T., Wiersema J. R., Brass M., & Braem S. (2020). Autistic traits are related to worse performance in a volatile reward learning task despite adaptive learning rates. Autism. 10.1177/1362361320962237. [DOI] [PubMed] [Google Scholar]
- 67.Fine C, Gardner M, Craigie J, Gold I. Hopping, skipping or jumping to conclusions? Clarifying the role of the JTC bias in delusions. Cognitive Neuropsychiatry. 2007;12:46–77. 10.1080/13546800600750597 [DOI] [PubMed] [Google Scholar]
- 68.Andreou C, Moritz S, Veith K, Veckenstedt R, Naber D. Dopaminergic modulation of probabilistic reasoning and overconfidence in errors: A double-blind study. Schizophrenia Bulletin. 2014;40:558–65. 10.1093/schbul/sbt064 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 69.Menon M, Mizrahi R, Kapur S. ‘Jumping to conclusions’ and delusions in psychosis: Relationship and response to treatment. Schizophrenia Research. 2008;98(1):225–31. 10.1016/j.schres.2007.08.021 [DOI] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
Anonymized raw and processed data are available in an Open Science Framework repository: DOI 10.17605/OSF.IO/UCA5E.