Abstract
Confidence is a critical metacognitive signal that guides performance. Biases in confidence, such as excessive doubt, are hallmark features of mental health disorders, especially obsessive-compulsive disorder (OCD). Yet, the underlying neurocognitive mechanisms and how they link to learning and decision-making remain elusive. We asked patients with OCD and matched healthy controls to perform a novel rule-shifting task incorporating trial-by-trial confidence ratings. Using a Bayes-optimal model, we identified two distinct confidence biases: while patients with OCD indicated lower overall confidence, their trial-by-trial confidence ratings more accurately tracked task-relevant information, rendering their confidence reports more Bayes-optimal than those of controls. These findings challenge the idea of a simple, unified metacognitive impairment in OCD. Instead, they suggest that OCD is linked to an enhanced responsiveness to environmental evidence and feedback during decision-making.
Keywords: Obsessive-Compulsive Disorder, Rule-Shifting, Decision Making, Learning, Metacognition, Confidence, Cognitive Flexibility
Introduction
Subjective confidence, our internal estimate of the probability that a choice is correct,1 acts as a control signal for adaptive behavior. Well-calibrated confidence determines when to exploit existing beliefs, abandon them, or seek new information, thereby supporting fast and flexible decisions in volatile environments2–4. In humans, low confidence increases information-seeking and strategy switches, whereas high confidence sustains ongoing courses of action5–7. Failures to regulate confidence can manifest clinically as chronic doubt or unfounded certainty and can hamper cognitive flexibility, the ability to update thoughts and actions as environmental demands change8.
Obsessive-compulsive disorder (OCD), characterized by intrusive thoughts (obsessions) and repetitive behaviors or mental acts aimed at reducing distress (compulsions),9 may be exemplary for how (aberrant) confidence and cognitive flexibility interact. Beyond reported deficits in executive functioning10,11, and cognitive flexibility12–20, individuals with OCD often experience chronic doubt and difficulty managing uncertainty10,11,21–23, alongside lowered confidence24–30. Notably, OCD has been characterized by a discrepancy between low confidence and objectively intact performance27,31–34, prompting the question of whether self-doubt in OCD reflects a difficulty in integrating external feedback or a disruption in internal self-monitoring.
To investigate these distinct possibilities, in this study, we assessed confidence judgements under changing rules in a novel task in individuals with OCD and matched healthy controls.
We find that patients with OCD achieve the same accuracy and rule-shifting performance as matched controls yet show systematically altered confidence. Using a Bayes-optimal model, we show that this mismatch between performance and confidence is driven by two distinct processes in OCD: while patients exhibit a stable, negative metacognitive bias, resulting in persistently lower confidence overall, their trial-by-trial confidence updates track external evidence more optimally than those of controls, adhering more closely to Bayesian predictions. These results indicate that persistently lowered confidence in OCD does not arise from faulty evidence integration; instead, it reflects a stable metacognitive bias that dampens subjective certainty even when environmental cues are processed. By dissociating this stable bias from dynamic evidence tracking, our findings provide a more refined understanding of the computational phenotype of OCD.
Results
Experimental Task and Design
We employed a novel, gamified rule-shifting paradigm, administered via a touchscreen application (cf. Figure 1A), to 29 patients with OCD and 29 healthy controls, matched for age and cognitive ability (serving as a proxy for IQ; cf. Methods). Patients with OCD, recruited from the Yale OCD Research Clinic, were clinically diagnosed, exhibited moderate-to-severe symptoms (Y-BOCS total ≥16), and were either unmedicated or on stable selective serotonin reuptake inhibitor (SSRI) treatment (cf. Methods and Supplemental Information for full participant details). On each trial of the task, participants selected one of two (levels 1–2) or three (level 3; cf. Figure 1B) multi-featured stimuli with the goal of identifying a currently rewarded stimulus feature (e.g., a particular color or shape). After making their choice, but before receiving feedback on its accuracy, they rated their confidence in their selection. Reward contingencies changed covertly and at unpredictable intervals (cf. Methods), requiring participants to continuously monitor outcomes, infer the current rule (i.e., rewarded feature), and flexibly adapt their choices and beliefs in response to environmental volatility. Changing conditions entailed both intra-dimensional [ID] shifts (e.g., from one color to another), in which the target feature shifts within the relevant dimension, and extra-dimensional [ED] shifts (e.g., from a specific color to a specific shape), in which the relevant dimension itself shifts (cf. Figure 1C and Methods). This design allowed for the trial-by-trial assessment of learning, adaptation to different types of rule shifts, and associated confidence dynamics.
Figure 1. Rule-shifting task.
(A) Participants used a touchscreen device to complete a new rule-shifting task (the “Cryptic Creatures Task”, implemented in the citizen science app Brain Explorer, www.brainexplorer.net) involving two or three multi-dimensional creatures. Each dimension (e.g., color, shape, presence of “hair”) has specific features (e.g., orange vs. blue, star vs. round, hair vs. no-hair). On each trial, participants chose one creature and then rated their decision confidence in that choice using a slider. As they moved the slider, anchor labels updated (e.g., “totally guessing,” “pretty unsure,” “pretty certain,” “totally certain”). Feedback was shown as either a coin (correct choice) or a cross (incorrect choice), with coins accumulating in the top-right corner of the screen. (B) Unannounced rule shifts occurred every 6–11 trials and required participants to flexibly update their decision strategy. During intra-dimensional (ID) shifts, the currently relevant dimension remained the same but switched from one feature to another (e.g., from “hair” to “no-hair”). During extra-dimensional (ED) shifts, the relevant dimension itself changed (e.g., switching from “hair vs. no-hair” to “star-shape vs. round-shape”), requiring participants to learn which new dimension was now rewarded. (C) The task progressed through three levels of increasing difficulty. At Level 1, participants were shown two creatures differing on two dimensions. At Level 2, additional dimensions were introduced (still two creatures). Finally, at Level 3, a third creature was added, making it more difficult to identify the correct dimension and feature.
Participants Successfully Navigate the Rule-Shifting Task
To assess how well participants adapted to changing task rules across varying difficulty levels, we first assessed overall choice accuracy. A repeated-measures ANOVA on accuracy, with group (OCD vs. Controls) as a between-subject factor and difficulty level (1, 2, 3) as a within-subject factor, revealed a significant main effect of difficulty . This indicates that accuracy decreased as difficulty increased: from level 1 to level , and further at level .
We did not observe a significant effect of group (F(1,56) = 0.229, p = 0.634) or a group × difficulty interaction (F(2,112) = 1.917, p = 0.152). We also did not find any differences when investigating average accuracy after IDs and EDs separately (cf. Supplementary Information), suggesting that the task was solved equally well by both groups.
Because such average performance metrics might obscure more subtle learning differences, we investigated whether and how quickly participants adjusted to the hidden rule shifts using logistic mixed-effects models predicting accuracy. As expected, accuracy was significantly lower on the trial following a shift (β = −0.912,SE = 0.103, p < 0.001). However, accuracy then recovered quickly. Pooling across groups, we examined recovery by linking accuracy to trial distance from the shift (functional form [i.e. log term] selected by Bayesian Information Criterion [BIC] from linear, quadratic, and log terms; cf. Supplemental Information). We found that accuracy increased with distance, consistent with learning the new rule (ID shifts: β = 1.569, SE = 0.079, p < 0.001; ED shifts: β = 0.867, SE = 0.068, p < 0.001; cf. Figure 2A right panel).
Figure 2. Average accuracy and confidence ratings overall and relative to rule shifts in the task.
Participants played a new intra-/extra-dimensional rule-shifting task that included confidence ratings. (A) Participants’ average accuracy dropped when a shift happened (vertical dashed lines) but recovered afterward (right panel). Contrasting behavior between the two groups, we observed no significant difference in overall accuracy between patients with OCD and controls (left panel). (B) Confidence ratings dropped after the shifts and increased with more trials under the new rule (right panel). Patients were significantly less confident than controls throughout the task (left panel). The box plots on the left of each panel show the means of all trials across the entire task for each measure. The right panels display the mean behavioral measures from five trials before to five trials after each shift. Error bars represent standard errors. Statistical significance was determined using two-sided two-sample t-tests (***p<0.001, n.s. = non-significant).
We next investigated whether the groups differed in these more fine-grained adaptation analyses and added the group variable to the mixed-effects models. Main effects for trial distance remained significant (ID shifts: β = 1.634, SE = 0.112, p < 0.001; ED shifts: β = 0.842, SE = 0.068, p < 0.001), and group main effects were non-significant (ID shifts: β = 0.124, SE = 0.172, p = 0.469; ED shifts: β = 0.111, SE = 0.132, p = 0.398). In addition, no group × distance interactions were significant (ID shifts: β = −0.126, SE = 0.155, p = 0.414; ED shifts: β = 0.024, SE = 0.096, p = 0.805), indicating comparable post-shift learning across groups.
Participants With OCD Show Lower Baseline Confidence and Higher Feedback Sensitivity In Their Confidence Ratings
Having established comparable performance across patients and controls, we moved on to investigate potential group differences in confidence judgments. We examined average confidence ratings using the same ANOVA-structure as for accuracy (cf. Methods for details). We found a significant main effect of group and difficulty level 95.168, , but no significant interaction between group and task difficulty 1.386, . Thus, while confidence declined with increased task difficulty in both groups (level 1: ; level 2: level , the OCD group exhibited consistently lower confidence overall ; cf. Figure 2B left panel). This group difference remained significant when controlling for depression scores and was also not explained by medication status or comorbidities in the patient group (cf. Supplementary Information for additional analyses). Consistent with this pattern of lowered confidence individuals with OCD also exhibited significantly slower overall reaction times (RTs) and both groups markedly slowed as task difficulty increased (cf. Supplementary Information for RT analyses).
Next, we analyzed how confidence changed dynamically around rule shifts and in response to feedback. A significant main effect of group mirrored the lowered confidence in participants with OCD seen in the previous analysis (β = −9.194, SE = 3.032, p < 0.01). Moreover, confidence dropped significantly on trials following a rule shift across both groups (β = −5.440, SE = 1.656, p < 0.01), reflecting their uncertainty about the correct choice. Interestingly, this immediate post-shift drop in confidence was more pronounced in the patient group, as indicated by a significant interaction effect between the group variable and the variable denoting a shift on the preceding trial predicting confidence (β = −5.071, SE = 2.340, p < 0.05; cf. trial 1 in Figure 2B right panel).
Subsequently, confidence increased with trial distance (ID shifts: β = 1.236, SE = 0.153, p < 0.001; ED shifts first-order term: β = 258.375, SE = 43.591, p < 0.001; ED shifts second-order term: β = −174.328, SE = 38.689, p < 0.001; model comparison favored a linear form for ID shifts and a quadratic form for ED shifts using orthogonal polynomials, cf. Supplemental Information; cf. Figure 2B right panel). Adding the group variable yielded no significant group × trial distance interactions (ID shifts: β = 0.206, SE = 0.308, p = 0.506; ED shifts first-order term: β= 114.966, SE=86.367, p=0.188; ED shifts second-order term: β=−122.853, SE=76.505, p=0.114). Distance effects remained significant (ID shifts: β=1.132, SE=0.218, p < 0.001; ED shifts first-order term: β = 204.315, SE=60.750, p < 0.01; ED shifts second-order term: β = −119.424, SE = 52.598, p < 0.05) while significant group main effects again indicated significantly lower overall confidence in individuals with OCD (ID shifts: β = −8.611, SE = 3.459, p < 0.05; ED shifts: β = −11.298, SE = 3.313, p < 0.01). These findings hint towards patients having a heightened sensitivity to environmental changes (i.e. rule shifts) in their confidence ratings but ultimately adapting quickly as the new rule becomes more familiar.
To understand if the confidence ratings of participants with OCD were associated with altered sensitivity to performance feedback in general, we examined the impact of accuracy on confidence on the subsequent trial. We used a mixed-effects model predicting confidence at trial t from accuracy at trial t-1, group, and their interaction. As expected, accuracy on the previous trial positively predicted subsequent confidence (β = 11.779, SE = 2.437, p < 0.001; cf. Figure 3A). Consistent with overall group differences, OCD status also significantly predicted lower confidence within this model (main effect of group: β = −16.241, SE = 4.065, p < 0.001). Crucially, we additionally found a significant interaction between group and preceding accuracy (β = 9.697, SE = 3.445, p < 0.01; cf. Figure 3B), indicating that the effect of preceding accuracy on confidence was more pronounced in participants with OCD.
Figure 3. Confidence ratings in patients with OCD are more sensitive to external feedback.
(A) A mixed-effects model indicated that, overall, patients with OCD were less confident than controls and that confidence was positively associated with accuracy on the preceding trial. (A-B) There was a significant interaction between group and preceding accuracy, showing that the influence of feedback on next-trial confidence differed between groups. Follow-up comparisons examining this interaction revealed that participants with OCD exhibited heightened feedback sensitivity: (C) they showed a significantly greater average increase in confidence following positive feedback and (D) a significantly greater average decrease following negative feedback compared to controls (Bonferroni-corrected). Panels (A) and (B) display model estimates or effects illustrating the main effects and interaction. Box plots in (C) and (D) represent mean confidence levels or changes on relevant trials. Error bars represent standard errors and *p < 0.05, **p < 0.01, ***p < 0.001.
To further examine this interaction and directly compare feedback sensitivity, we assessed the average change in confidence after correct and incorrect trials within each group. Both participants with OCD(t(28) = 7.093, p* < 0.001; where * indicates Bonferroni correction) and controls (Wilcoxon signed-rank test: W = 57,z = 3.67, p* < 0.001) showed a significant average increase in confidence after correct trials. However, only participants with OCD exhibited a significant average decrease in confidence after incorrect trials (t(28) = −5.184, p* < 0.001); the control group’s average change did not reach significance (Wilcoxon signed-rank test: W = 138, z = −1.71, p* = 0.175). Direct group comparisons further revealed that participants with OCD showed a significantly greater increase in confidence after correct decisions (Wilcoxon rank-sum test: U = 598.5, z = 2.768, p * < 0.05; cf. Figure 3C) and a significantly greater decrease after incorrect decisions (U = 239.0, Z = −2.823, p *< 0.01; cf. Figure 3D) than controls.
Taken together, these behavioral analyses show two distinct characteristics of confidence ratings in the OCD group within the rule-shifting task: (1) a lower overall level (metacognitive bias) compared to controls, in the context of comparable accuracy, and (2) heightened reactivity to immediate task events like rule-shifts and performance feedback (dynamic effects). While these analyses characterize observable patterns in confidence reporting, they do not clarify whether the heightened dynamic reactivity is adaptive or maladaptive in terms of optimal evidence integration.
Confidence Ratings of Participants with OCD are More Bayes Optimal
In this task, an optimal learner’s confidence is based not on the simple valence of feedback (positive or negative), but on the precision of the information that feedback provides. For example, an incorrect trial can lead to a subsequent increase in confidence if that feedback helps to definitively identify the correct feature. Consequently, the greater feedback reactivity observed in patients with OCD cannot be simply classified as adaptive or maladaptive. This distinction depends on the context of each trial, and the moment-by-moment accumulated evidence. To formally evaluate these confidence dynamics and disentangle stable metacognitive biases from dynamic evidence tracking, we therefore employed a (Bayes-)optimal observer model (cf. Methods). This model provides trial-by-trial estimates of optimal certainty: Bayes-optimal General Certainty (GC), reflecting overall certainty about the rewarding stimulus feature, and Bayes-optimal Choice Certainty (CC), reflecting certainty about the chosen stimulus containing the correct feature (cf. Figure 4A). While these two metrics, as expected, are positively correlated (mean within-subject: rs = 0.504, SD = 0.125), they capture distinct aspects of optimal certainty. GC reflects the overall certainty about the current rule governing the reward contingencies (i.e., inverse entropy scaled to the confidence scale used by participants), while CC is the certainty specifically associated with the features present in the chosen stimulus, i.e. how certain the observer is that one of these features is the rewarding one given the preceding choice and reward history (cf. Figure 4A).
Figure 4. Bayesian-observer certainty is linked to confidence fluctuations.
(A) Time course of the Bayesian observer’s Global Certainty (GC; dashed blue) and Choice Certainty (CC; purple) around an extra-dimensional (ED) rule shift, with the x-axis ranging from pre-shift trials (−2, −1), to the shift trial (extra-dimensional rule shift; dashed line at trial =0), and post-shift trials (1–4). Pre-shift, GC and CC are high under the learned “round-shape” rule. At the ED shift (0), the latent rule changes to “hair”. A negative outcome (‘X’ over the chosen stimulus) at the shift renders the previous rule incompatible, collapsing GC toward zero while CC 50, reflecting the sum of priors for the chosen features (‘hair’, ‘star-shaped’) under a belief state now divided equally between all present features (‘hair’, ‘no-hair’, ‘star-shaped’, ‘round’). On trial 1, a correct choice (yellow coin) and trial 2, an incorrect choice, allow the model to perform hypothesis elimination by reinforcing the correct features and discarding the irrelevant one. This resolution causes GC to rise, and with a final confirmatory choice on trial 3, GC and CC converge at maximal certainty, signifying that the new correct feature has been identified. (B) Across both groups, confidence ratings were significantly predicted by GC and CC. A significant interaction effect revealed that the relationship between CC and subjective confidence was significantly stronger in participants with OCD compared to controls, indicating more optimal tracking of choice-specific evidence in the patient group. Error bars represent standard errors and **p < 0.01, ***p < 0.001.
To relate these signals to subjective confidence, we fit linear mixed-effects models with GC, CC, group (OCD vs. control), their interactions, and covariates. We found that, in this joint model, confidence significantly tracked both normative benchmarks, with a larger effect for CC (β = 5.953, SE = 0.966, p < 0.001) than GC(β = 1.217, SE = 0.556, p < 0.05; cf. Figure 4B, Supplemental Information).
We examined how the relationship between confidence and the normative predictors differed between the OCD and control group. First, we found a persistent baseline difference: a significant main effect of group (β = −9.429, SE = 2.998, p < 0.01) indicated that individuals with OCD reported lower confidence than controls even when statistically controlling for normative certainty, a consistent metacognitive bias in line with preceding results. Second, we found a significant interaction effect between CC and group (β = 3.152, SE = 1.363, p < 0.05). This indicates a stronger alignment between confidence and CC in participants with OCD. The interaction between GC and group did not reach significance in this joint model (β = 1.328, SE = 0.785, p = 0.097; cf. Figure 4 B). For completeness, in a GC-only model specification (GC, group, covariates; without CC), GC remained significant (β=3.510, SE=0.895, p<0.001) and the GC × group interaction was also significant (β = 2.609, SE = 1.266, p < 0.05). Thus, GC does relate to confidence and shows a group interaction when entered alone; however, when GC and CC were modeled together (variance inflation factor [VIF]<3), the GC × group interaction was attenuated.
Taken together, while individuals with OCD exhibited a general negative metacognitive bias (i.e., pervasively lowered confidence), their dynamic updating of confidence was paradoxically more optimal (i.e., closer to the Bayesian observer’s certainty) than in controls, countering a dominant hypothesis in the field that lowered confidence in OCD stems from globally impaired evidence integration.
Discussion
In this study, we investigated the optimality of confidence judgments in individuals with OCD, a condition characterized by excessive doubt and chronic underconfidence9,27,28,35,36. Employing a deterministic rule-shifting task combined with normative Bayesian modeling, our findings reveal a striking dissociation: despite lower confidence levels (metacognitive bias) relative to healthy controls, individuals with OCD exhibited superior trial-by-trial integration of performance feedback and evidence, demonstrating confidence updates that were more optimal as they were more closely aligned with a Bayes-optimal observer. Importantly, their objective performance was comparable to controls, indicating that reduced baseline confidence does not stem from impaired decision making but rather from a stable calibration bias. Our findings challenge notions of broadly impaired metacognition in OCD and instead propose the coexistence of two distinct metacognitive phenomena.
The first of these entails a chronic bias in baseline confidence calibration, manifesting as reduced subjective confidence irrespective of actual performance, available feedback, or Bayes-optimal certainty. This stable offset represents a classic metacognitive bias, a concept distinct from metacognitive sensitivity (i.e., how well confidence tracks accuracy moment-to-moment)1,37,38. It also resonates with extensive literature documenting pervasive, subjective doubt and lowered confidence in many individuals with OCD24–30, frequently uncoupled from actual task performance or objective outcomes27,31,34,36,39. Such a baseline calibration difference may reflect core alterations in internal evaluations of certainty40, independent of external validation or success.
Conversely, the second mechanism encompasses an enhanced capacity for integrating and dynamically responding to trial-specific evidence. Participants with OCD demonstrated significantly greater sensitivity to immediate feedback, updating their confidence judgments more accurately in relation to Bayesian-derived choice certainty compared to controls. This may be linked to hyperactive error-processing27,41–43,43–45 and increased reward prediction error signaling seen in OCD46,47. Rather than indicative of impaired processing, this heightened responsiveness may reflect an adaptive cognitive strategy, allowing accurate performance despite persistently low confidence. Indeed, this heightened evidence sensitivity echoes findings of superior performance in other contexts, like increased information sampling4,8.
These dual metacognitive profiles, impaired stable confidence calibration coexisting with enhanced dynamic evidence tracking, suggest differential involvement of neural substrates or cognitive systems. Baseline confidence alterations, manifesting as a negative metacognitive bias, might implicate hypoactivity in cortico-striatal circuits associated with internal belief states, goal-directed action, and reward valuation48,49. Conversely, enhanced dynamic evidence tracking, reflecting superior sensitivity to immediate feedback and errors, could be mediated by hyperactive fronto-cingulate systems involved in rapid performance monitoring, conflict detection, and exaggerated error signaling42,43,46. Both hypo-50,51 and hyper-activation52,53 and -connectivity in cortico-striatal and fronto-cingulate circuits in OCD may underpin the distinct metacognitive mechanisms identified here. Further investigation is required to elucidate how these different levels within the metacognitive hierarchy interact and influence one another54–57.
Our findings refine and contextualize previous research on metacognition and cognitive flexibility in OCD, using a deterministic task to dissociate confidence calibration from updating. Prior studies using probabilistic tasks have shown that individuals with OCD can track environmental contingencies and sometimes exhibit increased learning rates in action updates, without consistently showing reduced confidence relative to controls34,58. Others have reported reduced confidence and heightened responsiveness to feedback in behavioral adjustments27. In probabilistic environments, reductions in confidence may reflect adaptive responses to ambiguity and individual strategy profiles, complicating interpretation in the absence of a normative benchmark. By employing a deterministic paradigm with unambiguous feedback, we showed that individuals with OCD retained intact confidence updating relative to controls, despite globally reduced confidence.
This task structure may also explain why, consistent with some studies12–14 but not others15–20, participants with OCD in our study achieved accuracy comparable to controls, albeit with longer deliberation times. Removing time pressure and enforcing progression, unlike classical paradigms that often elicit apparent inflexibility15–17,20 likely permitted deliberate, compensatory strategies. The longer RTs in the OCD group are consistent with such adaptations. These findings reinforce that both cognitive and metacognitive processes in OCD (and beyond) may be sensitive to task structure and contextual demands. Moreover, we retained clinical realism in our cohort, with comorbidities present and a subset stably SSRI-treated. Although such heterogeneity can increase variance and attenuate some group contrasts, it strengthens external validity and ensures that the effects we report reflect OCD as it presents in practice. Future work should test whether the dissociation between baseline calibration and dynamic tracking generalizes across OCD subgroups, symptom dimensions, and learning environments with varying degrees of ambiguity and structure.
In conclusion, we show that metacognitive alterations in OCD do not represent a simple failure of judgment. Our findings suggest separable processes of consistently lowered confidence and superior evidence tracking. This reframes our understanding of the disorder, moving beyond a narrative of generalized deficits to one of specific, and potentially opposing, alterations in metacognition. Clinically, this may mean that pathological doubt arises not because patients cannot see the evidence, but because even when they see it more clearly than others, they cannot translate it into a stable feeling of certainty. This insight shifts the focus for future research and treatment towards belief calibration and enhancing subjective certainty independent of external validation and feedback integration.
Methods
Participants
We recruited 31 patients with OCD and 30 healthy controls (with no medical or psychiatric diagnoses) via the Yale OCD Research Clinic’s participant pool and social media platforms (Instagram, Facebook, and Craigslist). All participants were based in the United States, aged 18 to 55, and underwent group-specific screening procedures. To be eligible for the study, patients with OCD had to be unmedicated or stably treated with selective serotonin reuptake inhibitors (SSRIs). Diagnoses of participants with OCD who participated in this study were established by a licensed psychologist and/or psychiatrist using an unstructured clinical diagnostic interview and the Mini International Neuropsychiatric Interview (MINI)59. Within approximately two weeks of behavioral testing, the severity of their obsessions and compulsions was assessed using the Yale-Brown Obsessive Compulsive Scale (Y-BOCS)60. Patients had to have a minimum total score of 16 (moderate severity of symptoms) to be included in the study. Participants recruited as controls had never been diagnosed with any mental health disorder in the past. Controls underwent an additional clinical online assessment using the MINI and the Padua Inventory-Washington State University Revision (PI-WSUR)61, and were enrolled in the study as controls if they did not meet the criteria for any psychiatric diagnoses and did not indicated any obsessive-compulsive symptoms on a clinical level on the PI-WSUR (cf. Supplemental Information for more details on recruitment and clinical assessment).
The target sample size was determined by referencing comparable studies in the field17,18. The study was approved by the Yale University Human Investigation Committee (HIC #0803003626) and was managed using the Research Electronic Data Capture tools (REDCap)62 hosted at Yale University. Participants provided written informed consent via REDCap. The mobile app used for collecting task data was also approved by the University College London Research Ethics Committee.
Procedure
After study enrollment, participants were sent a link (through REDCap) to complete demographic questions and a cognitive ability assessment (ICAR), the total score of which served as a proxy for IQ63. Once the questionnaires were completed, participants were instructed to download the app the study was hosted on (the Brain Explorer; https://brainexplorer.net/) to their touchscreen device and use a study-specific code to access the appropriate cognitive-behavioral task (cf. below). Participants played a newly developed rule-shifting task based on the principles of the ID/ED shift task, a widely-used cognitive flexibility paradigm (cf. below). Participants were compensated for their time with $20 per hour.
The Rule-Shifting Task
The rule-shifting task presented participants with two or three “amoeba-like” stimuli (depending on the difficulty level; cf. below), each comprising multiple dimensions-such as color or shape-and corresponding features (e.g., if the dimension is shape, features might be round or star-shaped). Importantly, within any differing dimension, each displayed stimulus carried a unique feature value. Participants were instructed to identify a “target stimulus” by determining which dimension (e.g., color or shape) was currently relevant and, within that dimension, which feature (e.g., round vs. star-shaped) was rewarded (cf. Figure 1A).
On each trial, participants selected one stimulus and rated their confidence in that choice on a scale from 1 (“totally guessing”) to 100 (“totally certain”). The confidence slider began in a random position, and moving it prompted an anchor label (e.g., “pretty unsure,” “pretty certain”) that updated as the slider was adjusted. After confirming their confidence rating, participants received feedback: a coin (+1 coin) if they chose correctly or a cross (0 coin) if they erred. Coins accumulated in the top-right corner of the screen.
Unannounced rule changes occurred every 6 to 11 trials (irrespective of participant’s performance), requiring participants to detect and adapt to new contingencies (cf. Figure 1B). Two types of rule shifts were implemented:
Intra-dimensional (ID) shifts - The same dimension remained relevant, but its rewarded feature switched (e.g., shape stayed relevant, but “star” was replaced by “round”).
Extra-dimensional (ED) shifts - The previously rewarded dimension (e.g., shape) became irrelevant, and a new dimension (e.g., color) became rewarding.
The task featured three levels of increasing difficulty (cf. Figure 1C): at level 1, participants saw two stimuli and encountered simple ID/ED shifts across 60 trials (two blocks of 30); at level 2, additional dimensions were introduced while maintaining two stimuli, for a total of 70 trials (two blocks of 35); at level 3, a third stimulus was added, increasing the complexity of identifying the correct dimension and feature. This level also consisted of 70 trials (two blocks of 35), bringing the total to 200 trials across all levels.
Task exclusion criteria and final sample characteristics.
We applied task-specific exclusion criteria that were in line with criteria used in similar decision-making tasks with confidence ratings. Participants were excluded from data analysis if their overall task performance was at or below chance level (at level 1 and 2, the chance level was at 50%, and at level 3 the chance level was at 33%; excluded patients: N=1; excluded controls: N=1). Additionally, one participant with OCD was excluded because they changed their medication between recruitment and task completion. We ensured the use of the full confidence scale by checking that participants’ confidence ratings were not the same for more than 90% of the trials (this was not the case for any participant). We also examined reaction times (RTs) as potential indicators of attention drops. Specifically, we log-transformed RTs for each participant and task difficulty level and then, after excluding the first trial of each block, computed an outlier threshold. Following recommendations by64, any trial whose log-RT exceeded three standard deviations was removed from the final dataset (this accounted for 0.36% of total trials).
This resulted in a gender-balanced final sample of 29 patients with OCD (15 identified as female, 12 as male and two as non-binary; Mage = 31, SDage =11.43) and 29 controls (14 identified as female and 15 as male; Mage = 32, SDage = 10.99; cf. Supplemental Information for clinical characteristics of the final sample). After applying these exclusion criteria, the groups were still matched on average age (Wilcoxon rank-sum test: z = −0.327, p = 0.744) and IQ (total ICAR score; z = 0.715, p = 0.471).
Statistical analyses
We used MATLAB 2023a (MathWorks) for initial data preprocessing. Data analysis was implemented using custom code in Python 3.8.8 with the packages: scikit-learn65, scipy66, numpy67, pandas68, statsmodels69, pymer470, pingouin71, seaborn72, and matplotlib73.
The Bayesian observer.
To establish a normative benchmark for participants’ trial-by-trial confidence, we developed and implemented a Bayesian observer model. This computational model simulates an ideal learner that processes the same experimental information as each participant and computes an optimal, dynamically updated estimate of decision certainty. The model operates under the following key assumptions: (1) The relationship between features and outcomes is deterministic (i.e., a specific feature is consistently rewarded until a rule shift). (2) Shifts in the rewarded feature can occur, and these may be within the same dimension (intra-dimensional) or to a new dimension (extra-dimensional). (3) No explicit signals indicate a rule change; the observer must infer shifts solely from trial-by-trial feedback.
Bayesian updating.
At the beginning of each task level, the observer starts with no preference for any particular feature. Its belief about each of the N unique features across the displayed stimuli, , are initialized to a uniform probability: . On each subsequent trial, the observer updates its belief, the posterior probability , about each feature being the currently rewarding one. This update occurs after observing the participant’s choice and the resulting reward outcome (coded as 1 if rewarded, 0 if not), using Bayes’ theorem:
| (1) |
where is the likelihood of outcome given is the rewarding feature, and the sum in the denominator normalizes the posterior distribution across all features.
Given the observer’s knowledge that sudden rule changes can occur, if contradicting evidence is encountered that would otherwise lead to a state where all current hypotheses are invalidated (mathematically, this corresponds to the denominator in Equation 1 becoming zero) with all feature beliefs reaching 0, the beliefs for all features are reset to . This allows the observer to “re-evaluate” all hypotheses from a neutral standpoint and adapt to the new contingency, assuming it is aware of the occurring rule changes.
Likelihood function.
The likelihood term quantifies the support the observed feedback provides for the hypothesis that is the rewarded feature. On each trial, learning is restricted to dimensions that differ across the displayed stimuli. Within those dimensions, each stimulus carries a unique feature value. Let be the set of unique features present in the chosen stimulus on trial . Then:
| (2) |
Quantifying Uncertainty and Mapping to Confidence.
By applying the Bayesian observer model to each participant’s task run, we generated three trial-by-trial normative estimates. The model’s uncertainty about which feature is rewarded is quantified by the entropy of the posterior distribution :
| (3) |
To translate this entropy into an intuitive 0–100 confidence scale, Bayes-optimal General Certainty (GC), was calculated. GC reflects the proportion of uncertainty that has been reduced from the maximum possible entropy:
| (4) |
In addition, we introduced a measure intended to mirror the conceptual basis of participants’ confidence ratings. Let denote the set of unique features present in the chosen stimulus on trial t. We defined the certainty in the chosen stimulus (Bayes-optimal Choice Certainty, CC) as the sum of the posterior probabilities associated with these features:
| (5) |
We scaled the and values so these optimal confidence estimates would align with the 0–100 confidence scale participants used. We then compared these Bayes-optimal estimates to the actual confidence ratings participants provided. This allowed us to quantify how well participant confidence aligned with an ideal Bayesian strategy and to pinpoint any systematic deviations, such as consistent under-confidence or inefficient evidence integration.
Behavioral analyses
To characterize the link between task variables (e.g., difficulty level, dimensional shift) and participants’ behavior and characteristics (e.g., patient versus non-patient) in a trial-by-trial hierarchical manner, we constructed regressions as mixed-effects models, using the Lmer function from the pymer4 package in Python70. All regression models investigating group differences were covaried for age, gender, and an approximation of IQ.
Accuracy in this rule-shifting task was defined as a binary variable capturing whether participants had chosen the correct (coded as 1) or an incorrect stimulus (coded as 0). We z-scored predictors in our regression models except if they were categorical (gender: 1= female, 2= male, 0= non-binary/Other; group: 1= patient with OCD and 0= control) prior to the analysis to allow comparability of regression coefficients. To account for both population-level and individual-level effects, we included both fixed and random effects in our mixed-effects model, performed standard convergence diagnostics, and used BIC to compare candidate fixed- and random-effects structures, selecting the most parsimonious specification (cf. Supplemental Information). Covariates such as demographics were included as fixed effects.
For the regression models predicting confidence on trial t, we shifted the independent task variables such as accuracy and dimensional by one trial (t – 1) as participants only received feedback after rating their confidence on a given trial. An example model in the pymer4 syntax would be as follows:
We fit models with a binary outcome variable (e.g., accuracy: 0= incorrect, 1= correct) and set the family argument to ‘binomial’, resulting in a logistic regression. We compared group means using independent-samples t-tests. If normality or homogeneity assumption was violated, non-parametric alternatives were used. All tests were two-tailed and p-values were adjusted for multiple comparisons using the Bonferroni method where applicable (denoted as p*). Before these analyses, we identified and removed outliers using the interquartile range (IQR) method, excluding any data points above or below represent the first and third quartiles, respectively, and the IQR is defined as .
Supplementary Material
Additional Information
Supplementary Information is available for this paper.
Supplementary Files
This is a list of supplementary files associated with this preprint. Click to download.
• LoosenEtAISupplementalInformationMetacognitiveParadoxOfOCD.pdf
Acknowledgements
We thank the entire team of the Yale OCD Research Clinic for their support with recruitment and clinical evaluation. We also thank Bastien Blain for helpful discussions. This work was supported in part by the Department of Mental Health and Addiction Services of the State of Connecticut, USA, through its support of the Ribicoff Research Facilities at the Connecticut Mental Health Center in New Haven, Connecticut. The views expressed in this paper are those of the authors and do not reflect the position of the Department of Mental Health and Addiction Services or of the State of Connecticut. A.M.L. was a pre-doctoral fellow of the International Max Planck Research School on Computational Methods in Psychiatry and Ageing Research (IMPRS COMP2PSYCH) and the Yale-UCL Collaborative Student Exchange Programme during the completion of this project. The participating institutions of the IMPRS programme are the Max Planck Institute for Human Development, Berlin, Germany, and University College London, London, UK. For more information, see: https://www.mps-ucl-centre.mpg.de/comp2psych. B.Z. was supported by a Young Investigator Grant from the Brain and Behavior Research Fund. B.Z., H.P., A.M., C.B., and C.P. were supported by National Institute of Mental Health grant R01MH116038 (PIs: Pittenger and Anticevic). H.P. was additionally supported by NIMH grants R03MH128847 and R21MH115394 (PIs: Pushkarskaya and Pittenger). C.P was additionally supported by NIMH grant K24MH121571 (PI: Pittenger). T.U.H. was supported by a Sir Henry Dale Fellowship (211155/Z/18/Z; 211155/Z/18/B; 224051/Z/21) from Wellcome & Royal Society (https://wellcome.ac.uk/funding/sir-henry-dale-fellowships), a grant from the Jacobs Foundation (2017–1261-04; https://jacobsfoundation.org/activity/jacobs-foundation-research-fellowship-program/), the Medical Research Foundation (https://www.medicalresearchfoundation.org.uk/what-we-fund), a 2018 NARSAD Young Investigator grant (27023) from the Brain & Behavior Research Foundation (https://www.bbrfoundation.org/grants-prizes/narsad-young-investigator-grants), and a Philip Leverhulme Prize from the Leverhulme Trust (PLP-2021–040; https://www.leverhulme.ac.uk/philip-leverhulme-prizes). This project has received funding from the European Research Council (ERC) under the European Union’s Horizon 2020 research and innovation programme (grant agreement No 946055; https://erc.europa.eu/funding/starting-grants). TUH is also supported by the Carl-Zeiss-Stiftung. This work was supported by the Alexander von Humboldt foundation, (more) precisely the Alexander-von-Humboldt-Professorship award to Peter Dayan. The Wellcome Centre for Human Neuroimaging was supported by core funding from the Wellcome Trust (203147/Z/16/Z). This research was funded in whole, or in part, by the Wellcome Trust (211155/Z/18/Z). For the purpose of Open Access, the author has applied a CC BY public copyright license to any Author Accepted Manuscript version arising from this submission.
Footnotes
Conflict of Interest Statement
All authors declare no conflicts of interest. TUH consults for limbic ltd and holds shares in the company, which is unrelated to the current project. In the past three years, CP has consulted for Biohaven Pharmaceuticals, Ceruvia Neurosciences, UCB BioPharma, Freedom Biosciences, Transcend Therapeutics, Alco Therapeutics, Lucid/Care, Nobilis Therapeutics, Mind Therapeutics, F-Prime Capital, and Madison Avenue Partners. He holds equity in Alco Therapeutics, Mind Therapeutics, and Lucid/Care. He receives or has received research support from Biohaven Pharmaceuticals, Freedom Biosciences, and Transcend Therapeutics. He receives royalties from Oxford University Press and UpToDate. He has filed patents on pharmacological treatments for OCD and related disorders, psychedelic therapeutics, and autoantibodies in OCD. None of these relationships are of relevance to the work described here. For the past three years, BZ has consulted with Biohaven Pharmaceuticals and received royalties from Oxford University Press.
Code and data availability
The code to the main analyses is accessible in a GitHub repository, at https://github.com/DevComPsy/The-Metacognitive-Paradox-of-OCD. Anonymized data will be freely accessible in a linked OSF repository (https://osf.io/c6tnx/) upon publication.
References
- 1.Fleming S. M. & Lau H. C. How to measure metacognition. Front. Hum. Neurosci. 8, (2014). [Google Scholar]
- 2.Balsdon T., Wyart V. & Mamassian P. Confidence controls perceptual evidence accumulation. Nat. Commun. 11, 1753 (2020). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Balsdon T. & Philiastides M. G. Confidence control for efficient behaviour in dynamic environments. Nat. Commun. 15, 9089 (2024). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.Rollwage M. et al. Confidence drives a neural confirmation bias. Nat. Commun. 11, 2634 (2020). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.Desender K., Boldt A. & Yeung N. Subjective Confidence Predicts Information Seeking in Decision Making. Psychol. Sci. 29, 761–778 (2018). [DOI] [PubMed] [Google Scholar]
- 6.Loosen A. M., Skvortsova V. & Hauser T. U. Obsessive-compulsive symptoms and information seeking during the Covid-19 pandemic. Transl. Psychiatry 11, 1–10 (2021). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Río M. del et al. Indecision and recency-weighted evidence integration in non-clinical and clinical settings. 2024.11.11.622934 Preprint at 10.1101/2024.11.11.622934 (2024). [DOI] [Google Scholar]
- 8.Scott W. A. Cognitive complexity and cognitive flexibility. Sociometry 25, 405–414 (1962). [Google Scholar]
- 9.American Psychiatric Association. Diagnostic and Statistical Manual of Mental Disorders. (American Psychiatric Association, 2013). doi: 10.1176/appi.books.9780890425596. [DOI] [Google Scholar]
- 10.Abramowitz J. S., Taylor S. & McKay D. Obsessive-compulsive disorder. Lancet Lond. Engl. 374, 491–499 (2009). [Google Scholar]
- 11.Jacoby R. J. & Abramowitz J. S. Intolerance of Uncertainty in OCD. in Obsessive-compulsive Disorder: Phenomenology, Pathophysiology, and Treatment (eds. Pittenger C.& Pittenger C.) 0 (Oxford University Press, 2017). doi: 10.1093/med/9780190228163.003.0017. [DOI] [Google Scholar]
- 12.Gottwald J. et al. Impaired cognitive plasticity and goal-directed control in adolescent obsessive-compulsive disorder. Psychol. Med. 48, 1900–1908 (2018). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Hybel K. A., Mortensen E. L., Lambek R., Thastum M. & Thomsen P. H. Cool and Hot Aspects of Executive Function in Childhood Obsessive-Compulsive Disorder. J. Abnorm. Child Psychol. 45, 1195–1205 (2017). [DOI] [PubMed] [Google Scholar]
- 14.Menzies L. et al. Neurocognitive endophenotypes of obsessive-compulsive disorder. Brain 130, 3223–3236 (2007). [DOI] [PubMed] [Google Scholar]
- 15.Chamberlain S. R., Fineberg N. A., Blackwell A. D., Robbins T. W. & Sahakian B. J. Motor Inhibition and Cognitive Flexibility in Obsessive-Compulsive Disorder and Trichotillomania. Am. J. Psychiatry 163, 1282–1284 (2006). [DOI] [PubMed] [Google Scholar]
- 16.Chamberlain S. R. et al. Impaired Cognitive Flexibility and Motor Inhibition in Unaffected First-Degree Relatives of Patients with Obsessive-Compulsive Disorder. Am. J. Psychiatry 164, 335–338 (2007). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Chamberlain S. R., Solly J. E., Hook R. W., Vaghi M. M. & Robbins T. W. Cognitive Inflexibility in OCD and Related Disorders. in The Neurobiology and Treatment of OCD: Accelerating Progress (eds. Fineberg N. A. & Robbins T. W.) 125–145 (Springer, 2021). doi: 10.1007/7854_2020_198. [DOI] [Google Scholar]
- 18.Fradkin I., Strauss A. Y., Pereg M. & Huppert J. D. Rigidly Applied Rules? Revisiting Inflexibility in Obsessive Compulsive Disorder Using Multilevel Meta-Analysis. Clin. Psychol. Sci. 6, 481–505 (2018). [Google Scholar]
- 19.Gruner P. & Pittenger C. Cognitive inflexibility in obsessive-compulsive disorder. Neuroscience 345, 243–255 (2017). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Talwar A., Cormack F., Huys Q. J. M. & Roiser J. P. A hierarchical reinforcement learning model explains individual differences in attentional set shifting. Cogn. Affect. Behav. Neurosci. 24, 1008–1022 (2024). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Gillett C. B., Bilek E. L., Hanna G. L. & Fitzgerald K. D. Intolerance of uncertainty in youth with obsessive-compulsive disorder and generalized anxiety disorder: A transdiagnostic construct with implications for phenomenology and treatment. Clin. Psychol. Rev. 60, 100–108 (2018). [DOI] [PubMed] [Google Scholar]
- 22.Knowles K. A. & Olatunji B. O. Intolerance of Uncertainty as a Cognitive Vulnerability for Obsessive-Compulsive Disorder: A Qualitative Review. Clin. Psychol. Publ. Div. Clin. Psychol. Am. Psychol. Assoc. 30, 317–330 (2023). [Google Scholar]
- 23.Pushkarskaya H. et al. Decision-Making Under Uncertainty in Obsessive-Compulsive Disorder. J. Psychiatr. Res. 69, 166–173 (2015). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24.Boschen M. J. & Vuksanovic D. Deteriorating memory confidence, responsibility perceptions and repeated checking: Comparisons in OCD and control samples. Behav. Res. Ther. 45, 2098–2109 (2007). [DOI] [PubMed] [Google Scholar]
- 25.Dar R. Elucidating the mechanism of uncertainty and doubt in obsessive-compulsive checkers. J. Behav. Ther. Exp. Psychiatry 35, 153–163 (2004). [DOI] [PubMed] [Google Scholar]
- 26.Hoven M. et al. Metacognition and the effect of incentive motivation in two compulsive disorders: Gambling disorder and obsessive-compulsive disorder. Psychiatry Clin. Neurosci. 76, 437–449 (2022). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Hoven M., Mulder T., Denys D., van Holst R. J. & Luigjes J. Lower confidence and increased error sensitivity in OCD patients while learning under volatility. Transl. Psychiatry 14, 370 (2024). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28.Lazarov A., Dar R., Liberman N. & Oded Y. Obsessive-compulsive tendencies and undermined confidence are related to reliance on proxies for internal states in a false feedback paradigm. J. Behav. Ther. Exp. Psychiatry 43, 556–564 (2012). [DOI] [PubMed] [Google Scholar]
- 29.Moritz S. et al. Enhanced perceived responsibility decreases metamemory but not memory accuracy in obsessive-compulsive disorder (OCD). Behav. Res. Ther. 45, 2044–2052 (2007). [DOI] [PubMed] [Google Scholar]
- 30.Sarig S., Dar R. & Liberman N. Obsessive-compulsive tendencies are related to indecisiveness and reliance on feedback in a neutral color judgment task. J. Behav. Ther. Exp. Psychiatry 43, 692–697 (2012). [DOI] [PubMed] [Google Scholar]
- 31.Hauser T. U., Allen M., Rees G. & Dolan R. J. Metacognitive impairments extend perceptual decision making weaknesses in compulsivity. Sci. Rep. 7, 6614 (2017). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 32.Rouault M., Seow T., Gillan C. M. & Fleming S. M. Psychiatric Symptom Dimensions Are Associated With Dissociable Shifts in Metacognition but Not Task Performance. Biol. Psychiatry 84, 443–451 (2018). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33.Seow T. X. F. & Gillan C. M. Transdiagnostic Phenotyping Reveals a Host of Metacognitive Deficits Implicated in Compulsivity. Sci. Rep. 10, 1–11 (2020). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34.Vaghi M. et al. Compulsivity Reveals a Novel Dissociation between Action and Confidence. Neuron 96, 348–354.e4 (2017). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35.Tolin D. F. et al. Memory and memory confidence in obsessive–compulsive disorder. Behav. Res. Ther. 39, 913–927 (2001). [DOI] [PubMed] [Google Scholar]
- 36.Dar R., Sarna N., Yardeni G. & Lazarov A. Are people with obsessive-compulsive disorder under-confident in their memory and perception? A review and meta-analysis. Psychol. Med. 52, 2404–2412 (2022). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37.Maniscalco B. & Lau H. A signal detection theoretic approach for estimating metacognitive sensitivity from confidence ratings. Conscious. Cogn. 21, 422–430 (2012). [DOI] [PubMed] [Google Scholar]
- 38.Fleming S. M. & Dolan R. J. The neural basis of metacognitive ability. Philos. Trans. R. Soc. B Biol. Sci. 367, 1338–1349 (2012). [Google Scholar]
- 39.Hoven M. et al. Abnormalities of confidence in psychiatry: an overview and future perspectives. Transl. Psychiatry 9, 1–18 (2019). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40.Dar R., Lazarov A. & Liberman N. Seeking proxies for internal states (SPIS): Towards a novel model of obsessive-compulsive disorder. Behav. Res. Ther. 147, (2021). [Google Scholar]
- 41.Norman L. J. et al. Error Processing and Inhibitory Control in Obsessive-Compulsive Disorder: A Meta-analysis Using Statistical Parametric Maps. Biol. Psychiatry 85, 713–725 (2019). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 42.Riesel A. The erring brain: Error-related negativity as an endophenotype for OCD—A review and meta-analysis. Psychophysiology 56, e13348 (2019). [DOI] [PubMed] [Google Scholar]
- 43.Endrass T., Klawohn J., Schuster F. & Kathmann N. Overactive performance monitoring in obsessive-compulsive disorder: ERP evidence from correct and erroneous reactions. Neuropsychologia 46, 1877–1887 (2008). [DOI] [PubMed] [Google Scholar]
- 44.Ruchsow M. et al. Error-related brain activity in patients with obsessive-compulsive disorder and in healthy controls. J. Psychophysiol. 19, 298 (2005). [Google Scholar]
- 45.Stern E. R. et al. Hyperactive Error Responses and Altered Connectivity in Ventromedial and Frontoinsular Cortices in Obsessive-Compulsive Disorder. Biol. Psychiatry 69, 583–591 (2011). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 46.Hauser T. U. et al. Increased fronto-striatal reward prediction errors moderate decision making in obsessive-compulsive disorder. Psychol. Med. 47, 1246–1258 (2017). [DOI] [PubMed] [Google Scholar]
- 47.Murray G. K. et al. Dopaminergic drug treatment remediates exaggerated cingulate prediction error responses in obsessive-compulsive disorder. Psychopharmacology (Berl.) 236, 2325–2336 (2019). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 48.Hauser T. U. et al. Increased decision thresholds enhance information gathering performance in juvenile Obsessive-Compulsive Disorder (OCD). PLOS Comput. Biol. 13, e1005440 (2017). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 49.Alves-Pinto A. et al. Altered reward-related effective connectivity in obsessive-compulsive disorder: an fMRI study. J. Psychiatry Neurosci. JPN 44, 395–406 (2019). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 50.Vaghi M. M. et al. Hypoactivation and Dysconnectivity of a Frontostriatal Circuit During Goal-Directed Planning as an Endophenotype for Obsessive-Compulsive Disorder. Biol. Psychiatry Cogn. Neurosci. Neuroimaging 2, 655–663 (2017). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 51.Robbins T. W., Vaghi M. M. & Banca P. Obsessive-Compulsive Disorder: Puzzles and Prospects. Neuron 102, 27–47 (2019). [DOI] [PubMed] [Google Scholar]
- 52.Anticevic A. et al. Global resting-state functional magnetic resonance imaging analysis identifies frontal cortex, striatal, and cerebellar dysconnectivity in obsessive-compulsive disorder. Biol. Psychiatry 75, 595–605 (2014). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 53.Maltby N., Tolin D. F., Worhunsky P., O’Keefe T. M. & Kiehl K. A. Dysfunctional action monitoring hyperactivates frontal-striatal circuits in obsessive-compulsive disorder: an event-related fMRI study. NeuroImage 24, 495–503 (2005). [DOI] [PubMed] [Google Scholar]
- 54.Seow T. X. F., Rouault M., Gillan C. M. & Fleming S. M. How Local and Global Metacognition Shape Mental Health. Biol. Psychiatry 90, 436–446 (2021). [DOI] [PubMed] [Google Scholar]
- 55.Rouault M., Will G.-J., Fleming S. M. & Dolan R. J. Low self-esteem and the formation of global self-performance estimates in emerging adulthood. Transl. Psychiatry 12, 272 (2022). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 56.Rouault M., Dayan P. & Fleming S. M. Forming global estimates of self-performance from local confidence. Nat. Commun. 10, 1141 (2019). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 57.Seow T. X. F., Fleming S. M. & Hauser T. U. Metacognitive biases in anxiety-depression and compulsivity extend across perception and memory. PLOS Ment. Health 2, e0000259 (2025). [Google Scholar]
- 58.Marzuki A. A. et al. Atypical action updating in a dynamic environment associated with adolescent obsessive–compulsive disorder. J. Child Psychol. Psychiatry n/a, (2022). [Google Scholar]
- 59.Sheehan D. V. et al. The Mini-International Neuropsychiatric Interview (M.I.N.I.): the development and validation of a structured diagnostic psychiatric interview for DSM-IV and ICD-10. J. Clin. Psychiatry 59 Suppl 20, 22–33;quiz 34–57 (1998). [Google Scholar]
- 60.Steketee G., Frost R. & Bogart K. The Yale-Brown Obsessive Compulsive Scale: interview versus self-report. Behav. Res. Ther. 34, 675–684 (1996). [DOI] [PubMed] [Google Scholar]
- 61.Rubio-Aparicio M. et al. The Padua Inventory-Washington State University Revision of Obsessions and Compulsions: A Reliability Generalization Meta-Analysis. J. Pers. Assess. 102, 113–123 (2020). [DOI] [PubMed] [Google Scholar]
- 62.Harris P. A. et al. Research electronic data capture (REDCap)—A metadata-driven methodology and workflow process for providing translational research informatics support. J. Biomed. Inform. 42, 377–381 (2009). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 63.Condon D. M. & Revelle W. The international cognitive ability resource: Development and initial validation of a public-domain measure. Intelligence 43, 52–64 (2014). [Google Scholar]
- 64.Berger A. & Kiefer M. Comparison of Different Response Time Outlier Exclusion Methods: A Simulation Study. Front. Psychol. 12, (2021). [Google Scholar]
- 65.Pedregosa F. et al. Scikit-learn: Machine Learning in Python. J. Mach. Learn. Res. 12, 2825–2830 (2011). [Google Scholar]
- 66.Virtanen P. et al. SciPy 1.0: fundamental algorithms for scientific computing in Python. Nat. Methods 17, 261–272 (2020). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 67.Harris C. R. et al. Array programming with NumPy. Nature 585, 357–362 (2020). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 68.the pandas development team. pandas-dev/pandas: Pandas. Zenodo 10.5281/zenodo.13819579 (2020). [DOI] [Google Scholar]
- 69.Seabold S. & Perktold J. Statsmodels: Econometric and Statistical Modeling with Python. scipy (2010) doi: 10.25080/Majora-92bf1922-011. [DOI] [Google Scholar]
- 70.Jolly E. Pymer4: Connecting R and Python for Linear Mixed Modeling. J. Open Source Softw. 3, 862 (2018). [Google Scholar]
- 71.Vallat R. Pingouin: statistics in Python. J. Open Source Softw. 3, 1026 (2018). [Google Scholar]
- 72.Waskom M. L. seaborn: statistical data visualization. J. Open Source Softw. 6, 3021 (2021). [Google Scholar]
- 73.Hunter J. D. Matplotlib: A 2D Graphics Environment. Comput. Sci. Eng. 9, 90–95 (2007). [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
The code to the main analyses is accessible in a GitHub repository, at https://github.com/DevComPsy/The-Metacognitive-Paradox-of-OCD. Anonymized data will be freely accessible in a linked OSF repository (https://osf.io/c6tnx/) upon publication.




