Skip to main content
Philosophical Transactions of the Royal Society B: Biological Sciences logoLink to Philosophical Transactions of the Royal Society B: Biological Sciences
. 2021 Feb 22;376(1822):20200131. doi: 10.1098/rstb.2020.0131

Confirmation bias is adaptive when coupled with efficient metacognition

Max Rollwage 1,2,, Stephen M Fleming 1,2,3
PMCID: PMC7935132  PMID: 33612002

Abstract

Biases in the consideration of evidence can reduce the chances of consensus between people with different viewpoints. While such altered information processing typically leads to detrimental performance in laboratory tasks, the ubiquitous nature of confirmation bias makes it unlikely that selective information processing is universally harmful. Here, we suggest that confirmation bias is adaptive to the extent that agents have good metacognition, allowing them to downweight contradictory information when correct but still able to seek new information when they realize they are wrong. Using simulation-based modelling, we explore how the adaptiveness of holding a confirmation bias depends on such metacognitive insight. We find that the behavioural consequences of selective information processing are systematically affected by agents' introspective abilities. Strikingly, we find that selective information processing can even improve decision-making when compared with unbiased evidence accumulation, as long as it is accompanied by good metacognition. These results further suggest that interventions which boost people's metacognition might be efficient in alleviating the negative effects of selective information processing on issues such as political polarization.

This article is part of the theme issue ‘The political brain: neurocognitive and computational mechanisms’.

Keywords: confirmation bias, metacognition, belief updating, adaptive rationality, selective information processing

1. Introduction

Polarization between opposing viewpoints is increasingly prevalent in discussions surrounding political and societal issues [1]. An important cognitive driver of this polarization is the human tendency to discount evidence against one's current position [25], a phenomenon known as confirmation bias [6]. Confirmation bias has been reported in a variety of settings [7], including the formation of clinical diagnosis [8], inference about people's character [9], investment decisions [10], views about societal issues such as capital punishment [11] and climate change [12]. Perhaps most prominently, confirmation bias has been reported in relation to politically charged beliefs, such that people are generally prone to process information in line with their political convictions [1317]. On a societal level, such skewed information intake might lead to entrenched beliefs, and, in turn, the prevalence of dogmatic groupings and widespread polarization [5,18]. In line with this hypothesis, people who show a resistance to belief updating are also more likely to show extreme political beliefs [19], aggression towards opposing political views [20], and authoritarian [21] or dogmatic traits [4,22].

Recent cognitive neuroscience studies have identified the selective integration of choice-consistent information as a key mechanism underpinning this cognitive bias [2325]. For instance, we recently showed that dogmatic participants were characterized by two cognitive alterations in the context of a perceptual decision-making task [4]. First, dogmatic participants showed a reduction in metacognitive ability, manifesting as a selective overconfidence after making errors. Second, metacognitive ability (i.e. the accuracy of confidence judgements) was predictive of post-decision evidence integration, where people with poorer metacognition showed less sensitivity for corrective information.

These results indicate that confidence may act as an internal control signal that guides future information processing [2630]. In studies using magnetoencephalography, we found evidence for this hypothesis, with confidence strongly modulating the extent of neural post-decision processing. Evidence accumulation was largely unbiased after low confidence decisions but displayed a confirmation bias after high confidence decisions [24]. In other words, people appear especially resistant to corrective information when they are highly confident about a wrong decision. However, when confidence is well aligned with performance—when metacognitive ability is high—weighting post-decisional integration by confidence is likely to be less problematic, as people will tend to be less confident after making errors, and, therefore, also open to corrective information. This line of reasoning suggests that people's metacognitive ability might be a crucial driver of the degree to which selective information processing leads to negative behavioural outcomes.

Here, we test a hypothesis that selective information integration might be adaptive when coupled with high metacognitive ability. Our proposal is in line with a broader hypothesis that the ubiquitous nature of selective evidence integration makes it unlikely that this cognitive characteristic is always maladaptive [31,32]. For instance, others have interpreted confirmation bias as a heuristic that reduces computational complexity [33] or allows for robustness against noise [34,35]. Here, we offer an alternative perspective: that confirmation bias is adaptive to the extent it is accompanied by a metacognitive ability to effectively monitor and recognize when we might be wrong [36]. We use simulation-based modelling to compare different evidence integration strategies and test their respective performances. Specifically, we compared unbiased evidence integration with a simple confirmation bias as well as with a confidence-driven confirmation bias (as observed empirically in [24]). Because we expected the performance of a confidence-weighted confirmation bias to depend on the reliability with which confidence judgements indicate choice accuracy, we also investigated the influence of metacognitive ability on the adaptiveness of confirmation bias.

2. Modelling behavioural impact of selective evidence integration strategies

We model a simple situation in which agents make a binary decision between two choice options based on noisy information drawn from a world state that is unknown to the agent. This situation is easily accommodated by existing frameworks for characterizing belief updating [25,37,38], and closely resembles common perceptual decision-making paradigms. This setting also acts as a minimal framework within which more complex decision problems can be modelled. For instance, to reach an opinion about whether human activity causes global warming (the ground truth), we have to form beliefs based on multiple noisy evidence samples (e.g. scientific publications and newspaper articles). Importantly, this process requires the updating of pre-existing beliefs whenever more evidence becomes available. In such a situation, Bayesian belief updating is often used as benchmark model [39,40]. Thus, we incorporated a Bayesian updater as our ‘unbiased agent’ against which other evidence integration strategies can be compared. A Bayesian agent keeps track of its graded belief that one or other choice option is correct by calculating the probability of the chosen option, given the evidence, compared to the alternative: P(choice|evidence)/P(alternative|evidence)

For simplicity, we simulate a situation in which participants only receive two samples of information: Xpre represents the initial information and Xpost represents the additional (or post-decision) evidence. Both Xpre and Xpost are sampled from normal distributions:

XpreN(μ,σpre2)
XpostN(μ,σpost2)
withμ=[1,1],

where the common mean (µ) of these distributions corresponds to the actual world state that needs to be inferred. While pre- and post-decision evidence distributions have the same mean (i.e. indicate the same underlying world state), they might differ in their variance. The variance represents the reliability of the information, with higher variances indicating less reliable information. We simulate different reliabilities of pre- and post-decision evidence as the consequences of a confirmation bias are likely to depend on this balance.

After receiving initial information, agents make an initial decision (decisioninitial) which depends solely on Xpre. If Xpre has a positive value (Xpre > 0), decisioninitial = 1, whereas if Xpre has a negative value (Xpre < 0), decisioninitial = −1. An estimate of confidence in this initial decision is derived by calculating the log-odds in favour of the chosen world state:

LOpre=2μXpreσpre2.

These log-odds can be transformed into a probability of being correct (between 0 and 1) as follows:

confidenceinitial=eLOpre(1+eLOpre).

After the initial decision, the agent receives additional information in the form of Xpost. In order to reach a final decision, both evidence samples can be integrated in an unbiased Bayesian fashion by simply summing the log-odds:

LOfinal=LOpre+LOpost.

Note that by summing the log-odds of pre- and post-decision evidence, the certainty/reliability of these two evidence samples is implicitly considered, i.e. evidence is combined in line with Bayesian principles. The final decision depends on the sign of the posterior log-odds (LOfinal). If the final decision corresponds to the actual state of the world, the agent can be said to have formed an accurate belief and performs the task correctly. We refer to this unbiased decision-maker as a Bayesian agent.

A confirmation bias can be modelled within this framework as an altered incorporation of post-decision evidence dependent on whether this new information confirms or disconfirms the initial decision:

if sign(Xpost) = sign (decisioninitial):

LOfinal=LOpre+(1 + confirmation bias)LOpost.

else if sign(Xpost) ≠ sign (decisioninitial):

LOfinal=LOpre+(1confirmation bias)LOpost.

This form of confirmation bias can range from 0 (no bias) to 1 (where processing of disconfirmatory information is abolished).

In line with empirical observations, we also modelled a situation in which confirmation bias is modulated by confidence [24], such that participants were relatively unbiased in their use of new evidence when less confident, but showed an enhanced confirmation bias after high confidence decisions. To mimic these signatures, we simulated a ‘metacognitive’ agent which shows a confirmation bias when it is confident in an initial choice (confidence = 1), but remains unbiased when unsure (confidence = 0.5):

if sign(Xpost) = sign (decisioninitial):

LOfinal=LOpre+(1+(confirmation bias((confidence0.5)2))LOpost,

else if sign(Xpost) ≠ sign (decisioninitial):

LOfinal=LOpre+(1(confirmation bias((confidence0.5)2))LOpost.

We note that a metacognitive agent differs from a Bayesian agent in that its initial confidence directly modulates the extent to which post-decision evidence is incorporated. For a Bayesian agent, there is of course also a sense in which confidence ‘weights’ the subsequent incorporation of evidence, in that a highly confident decision will require more disconfirming evidence to be overturned. Such updates, however, are in keeping with the linear accumulation of the log-odds of one or other hypothesis. By contrast, our metacognitive agent downweights the processing of disconfirmatory evidence when it is confident, representing a nonlinear effect of confidence on the incorporation of post-decisional log-odds. Thus, our metacognitive agent shows similarities with circular inference models [41,42], in which prior beliefs directly alter sensitivity to new evidence.

For each agent and each combination of evidence strength (see below), we simulated 200 000 trials, with the average decision accuracy across these simulated trials forming our measure of the agent's performance. In what follows, we describe in detail how these simulations were conducted.

3. Methods

(a). Pre- and post-decision evidence strength

As the performance of the different agents might depend on the reliabilities of pre- (Xpre) and post-decisional (Xpost) information, we simulated different information strengths defined as z-scores (μ/σ). We fixed µ = 1 (or −1, respectively) and changed σpre and σpost to vary information strength. We modelled all combinations of pre-decision (μ/σpre = [0.3–1.2]) and post-decision (μ/σpost = [0.3–1.2]) information strengths. In figure 1a,b and d, analyses of the joint effects of these two evidence strengths are presented, whereas in figures 1c and 2a,c, performance is averaged over all evidence strength conditions.

Figure 1.

Figure 1.

Comparison of agents' performance with different biases in information processing. (a) Performance of an unbiased agent that integrates initial and additional information in a Bayesian manner. Depending on the evidence strength, this agent shows different levels of accuracy, with better performance when both evidence samples are strong/reliable. (b) Difference in performance between an unbiased agent and a confirmation-biased agent as a function of the reliability of the initial and additional evidence. A confirmation bias has especially detrimental effects when initial evidence is relatively weak. (c) Comparison of unbiased, confirmation-biased and metacognitive (confidence-weighted) agents as a function of confirmation bias strength. Performance is averaged over all combinations of initial and additional evidence strengths. The vertical line indicates the strength of confirmation bias used in (b) and (d). (d) Difference in performance between an unbiased agent and a metacognitive agent as a function of the reliability of initial and additional evidence. Overall, the metacognitive agent shows only a relatively small disadvantage in comparison to an unbiased agent. In comparison to a simple confirmation bias, the metacognitive agent suffers less performance detriment in situations with weak initial evidence. (b,d) Dark colours indicate more detrimental performance of confirmation bias strategies when compared with unbiased evidence integration.

Figure 2.

Figure 2.

Performance of a metacognitive agent as a function of metacognitive efficiency. (a) Illustration of different decision and confidence models. (i) A model in which the same evidence (Xpre) informs both the initial decision and the initial confidence, resulting in a ratio of meta-d′/d′ = 1. (ii) A situation in which meta-d′ and d′ can differ as the initial decision and confidence rely on separate, though correlated, evidence samples (Xpre and Xconf). The final decision is determined by a combination of the initial (Xpre) and the additional (Xpost) information. The coloured arrows indicate the way in which either an initial decision (confirmation bias) or an initial decision in combination with confidence (confidence-weighted confirmation bias, as in our metacognitive agent) modulate the incorporation of post-decision evidence. (b) Performance difference for a metacognitive agent compared to an unbiased agent. The performance of a confidence-weighted confirmation bias is sensitive to metacognitive efficiency (i.e. the accuracy of confidence ratings), with the greatest benefits obtained when metacognitive ability is high. When the ratio of meta-d′/d′ is above 1, a metacognitive agent outperforms an unbiased agent. Hotter colours indicate better performance of the metacognitive agent. (c) Performance of an unbiased agent compared to metacognitive agents differing in their metacognitive efficiency. Here, we fix the pre- and post-decision evidence strengths to an intermediate level (see Methods) in order to reveal differences between different decision strategies. We simulated 100 agents for each setting and present group means ± s.e.m. Metacognitive agents with lower metacognitive efficiencies (meta-d′/d′ < 1) show significantly lower performance than an unbiased agent (all p < 0.0001), whereas metacognitive agents with higher metacognitive efficiencies (meta-d′/d′ > 1) show significantly better performance than an unbiased agent (p < 0.0001). An agent with meta-d′/d′ = 1 yields performance that is not significantly different from that of an unbiased agent (p = 0.32).

(b). Generative model of metacognitive abilities

The correspondence between confidence and performance can be formally quantified as the ratio of meta-d′/d′ (known as metacognitive efficiency), where meta-d′ reflects metacognitive sensitivity and d′ reflects primary task performance within a signal detection theoretic framework [43]. Several reasons for a dissociation between meta-d′ and d′ have been suggested. For instance, confidence may reflect a noisy read-out of the decision evidence or a decline of decision evidence in working memory prior to a confidence judgement [43], leading to a meta-d′/d′ ratio of less than 1. On the other hand, confidence might be informed by evidence that was not available at the time of decision [4446], or on correlated evidence that is accumulated in parallel [47], both of which may lead meta-d′/d′ ratios to surpass 1.

To model different degrees of metacognitive ability, we relaxed our assumption that confidence is directly derived from the evidence informing the initial decision (figure 2a). Instead, the evidence informing decisions (Xpre) and confidence estimates (Xconf) were modelled as distinct but correlated, and we allowed the reliability of the confidence (σconf2) and decision (σpre2) samples to differ. Specifically, Xpre and Xconf were sampled from a bivariate normal distribution with mean µ and covariance Σ:

[XpreXconf]N(μ,Σ)

and

Σ=[σpre2ρσpreσconfρσpreσconfσconf2].

Σ parametrizes the relationship between Xpre and Xconf, with ρ representing the correlation between these variables. As described by Fleming & Daw [47], in this situation confidence can be inferred based on a combination of Xconf, the initial decision and the covariance of Xconf and Xpre (see the appendix of [47], for further details on this calculation):

confidence=P(decisioninitial=μ|Xconf,decisioninitial,Σ).

Importantly, modelling separate samples of Xconf and Xpre allows for dissociations between meta-d′ and d′, thus making it possible to simulate varying degrees of metacognitive efficiency. For instance, a decrease in the reliability of the evidence informing the confidence rating (σconf2>σpre2) naturally reduces metacognitive efficiency and leads confidence judgements to be less reliable predictors of choice accuracy. To simulate lower metacognitive abilities (meta-d′/d′ < 1), we fixed ρ = 0.8 and varied σconf2(however, we note that our findings are not dependent on the specific value of ρ or σconf2; see the electronic supplementary material, figure S1). σconf2was defined by multiplying σpre2 by the set of coefficients [2.04, 1.56, 1.23], which were selected to result in ratios of meta-d′/d′ of [0.4 0.6 0.8]. We also modelled a situation in which confidence and decision information have the same reliability (σpre2=σconf2), but the evidence samples have a variable degree of correlation (ρ = [0.1, 0.3, 0.5, 0.65]), resulting in values of meta-d′/d′ of [1.4, 1.3, 1.2, 1.1]. Such decorrelations in evidence samples result in increased metacognitive efficiency because there is additional information on which to base an evaluation of the decider [47].

(c). Assessment of metacognitive efficiency

To assess whether manipulations in our generative model of confidence had the intended influence on agents' metacognitive efficiency, we calculated the meta-d′/d′ ratio for each agent [43] based on their simulated behaviour, using the MLE toolbox of Maniscalco and Lau (http://www.columbia.edu/~bsm2105/type2sdt/). Model fits were conducted separately for each level of evidence strength and then averaged over all evidence strengths for each agent.

(d). Robustness and significance

We simulated many trials per condition (200 000 trials for each combination of pre- and post-decision evidence strength) to ensure robustness to noise perturbations. We also sought to provide a statistical test of the modulation of performance by metacognitive efficiency. We simulated 100 agents per metacognitive efficiency setting (with confirmation bias = 1) and compared their performance to an unbiased agent (also simulated 100 times) using a t-test. We also evaluated the effects of metacognitive efficiency on post-decision performance at intermediate pre-decision evidence μ/σpre = [0.8] and post-decision evidence μ/σpre = [0.5] strengths, which are similar to those commonly used in previous laboratory experiments.

4. Results

Depending on the different reliabilities of Xpre and Xpost, an unbiased Bayesian observer achieves different final performances (figure 1a), with more reliable information yielding better performance. In addition to an unbiased observer, we also modelled agents with gradually increasing levels of confirmation bias (figure 1c). As hypothesized, selectively accumulating confirmatory information results in detrimental performance, with higher levels of confirmation bias leading to a more pronounced detriment. At higher levels of confirmation bias, a detriment in performance from 85% correct to approximately 77% correct was observed, which is significant in a 2 alternative forced choice scenario where performance can only vary between 50 and 100% and evidence strength is adjusted to be of intermediate difficulty. As expected, this effect was most notable when the agent received relatively weak initial information but reliable post-decision evidence (figure 1b), as the bias prevents the incorporation of more reliable corrective information.

We next examined the performance of a metacognitive agent with a confidence-weighted confirmation bias. Notably, a metacognitive agent outperforms a simple confirmation bias in all settings and only shows slight impairments in relation to an unbiased agent (figure 1c). We also found that a metacognitive agent clearly outperforms a simple confirmation bias in situations when the initial evidence is weak (figure 1d). While a simple confirmation bias has the strongest decrement in performance in these situations, a metacognitive agent ‘realizes’ it is dealing with weak initial evidence (by assigning lower confidence to these decisions) and thus shows a more equal sensitivity to confirming and disconfirming information.

Up until now we have assumed that agents calculate confidence in an initial choice by directly evaluating the reliability of evidence that informed the decision, resulting in a fixed metacognitive efficiency (meta-d′/d′ = 1). However, empirical evidence shows that people differ in their metacognitive ability [4,48,49]. Thus, we next assessed the degree to which metacognitive efficiency influences the performance of our metacognitive agent.

We found that reduced metacognitive efficiency (meta-d′/d′ < 1) leads metacognitive agents to show impaired performance compared to unbiased agents (all p-values < 0.0001, see cooler colours in figure 2b), as the weighting of new information by confidence becomes less effective. Conversely, when metacognitive efficiency is high (meta-d′/d′ > 1), we found that a metacognitive agent can even outperform an unbiased agent (all p-values < 0.0001, see hotter colours in figure 2b). This result is striking as it suggests that a selective integration of information is not necessarily a ‘bias’ but represents an advantageous strategy for achieving optimal performance in the context of a realistic cognitive architecture (i.e. one in which metacognition is particularly efficient). Figure 2c displays the effect of metacognitive efficiency on the performance of a metacognitive agent at a fixed level of intermediate evidence strength, relative to an unbiased agent. While the magnitude of these differences might appear relatively small (of the order of 2–3% correct), individual differences between human participants performing a similar task cover a similar range (e.g. in [24] the standard deviation of performance across participants was 3.7% correct decisions).

5. Discussion

Here, we investigated the effects of confirmation bias on the accuracy of belief formation. Our central proposal is that when confirmation bias is a feature of a self-aware (metacognitive) agent, it ceases to be detrimental, and may even become adaptive. We used simulation-based modelling to compare the performance of agents with different forms of confirmation bias against an unbiased agent. A simple (non-metacognitive) confirmation bias showed detrimental effects compared to an unbiased agent in all settings. In comparison, a metacognitive agent which modulates the degree of confirmation bias by confidence (as documented empirically in human observers; [24]) outperformed a simple confirmation bias agent, and was in many cases not substantially worse than an unbiased agent. The benefit of weighting a confirmation bias by confidence is that when confidence is low, and errors are more likely, the system becomes open to new and potentially corrective information.

In turn, by simulating varying degrees of metacognitive efficiency, we found that the performance of our metacognitive agent was sensitive to its level of self-awareness. Strikingly, a metacognitive agent with high self-awareness could in some cases even outperform an unbiased agent, indicating that selective information processing might be particularly adaptive when coupled with good metacognitive abilities. These results are in accordance with a view that cognitive biases may have originally evolved for good evolutionary reasons and are often adaptive when considered in the context of the agent's environment, including its broader mental toolkit [31].

Why should high degrees of metacognitive efficiency be advantageous in this case? The core mechanism appears to be the capacity of a confidence estimate to provide a ‘second look’ on a decision, similarly to how an external adviser might give us a separate view on a topic. The benefit of this mechanism depends on the agent's metacognitive ability—as agents with good metacognition provide the most effective ‘internal’ advisory signals. Interestingly, however, a confidence-weighted confirmation bias outperformed a simple confirmation bias in all settings, even when metacognitive ability was relatively low (see the electronic supplementary material, figure S2), suggesting that the mere presence of confidence weighting, rather than acute metacognition per se, may be sufficient to avoid the most deleterious effects of confirmation bias.

While these advantages for metacognitive agents were relatively small in size, they were robust and similar in magnitude to individual differences in performance on comparable laboratory tasks (e.g. [24]). We note that here we modelled a situation in which only two consecutive samples of evidence had to be integrated. Even in this minimal paradigm, increases in metacognitive efficiency could lead to a 2–3% increase in the number of correct decisions. Such a bias towards higher performance on individual, isolated judgements is likely to be magnified in situations requiring the integration of multiple information samples over time. In such situations, even small alterations in information processing might summate over time and lead to substantial changes in belief accuracy.

By incorporating confirmation bias as part of a broader cognitive architecture in which different mental processes can interact with each other (e.g. decisional and metacognitive processes), selective information processing may become adaptive when compared with the same ‘bias’ considered in isolation. In the same spirit, it has been argued that heuristics that may appear as biases in simple and constrained laboratory tasks become beneficial in more complex environments [50]. More broadly, our study indicates that considering cognitive biases in isolation from other mental processes might lead to the wrong conclusions about the impact of a particular cognitive feature on behaviour.

We note that a behavioural benefit for a confidence-weighted confirmation bias over an unbiased agent was only present when simulating agents with ‘hyper’ metacognitive efficiency (i.e. meta-d′/d > 1), such that metacognition became more acute than first-order task performance. This might seem odd at first glance, as it implies that the system is not using all the information available to it at the time of making an initial choice, and only afterwards becomes more sensitive to whether it was right or wrong. However, this kind of pattern is commonly observed in empirical data [45,47], and is thought to be driven either by additional post-decisional processing, differences in the variance of signal compared to noise [51] or (as simulated here) parallel streams of information processing that allows the system to detect and correct its errors [45]. The capacity for rapid error detection is well established in human studies [30,5255] and thus it is reasonable to assume that hyper metacognitive efficiency may be common in the healthy population.

Importantly, a metacognitive system does not need to be more reliable than the decision-maker to achieve high metacognitive ability: it needs only to incorporate partially independent information (as used here; see the electronic supplementary material, figure S1 for simulations using a wider set of parameters). In this respect, our results also contribute to a debate over why it might be useful for the brain to encode a confidence signal separately from representations of decision evidence [56,57]. A metacognitive agent that can realize its own mistakes (and assign low confidence to these decisions) will tend to become more open to new information owing to the confidence weighting applied to selective information processing (interestingly our model predicts that when confidence falls below 0.5 in a two-choice scenario, metacognitive agents should even show a ‘disconfirmation’ bias, and be more prone to seek out information contradicting their current position). Our results suggest that this benefit is only accrued when metacognition is partly independent of first-order cognition.

Selective information processing has been assumed to lead to skewed, entrenched and potentially inaccurate beliefs about a range of societal and political issues [5,18]. However, the current results suggest that the detrimental effects of selective information processing depend on people's broader self-awareness. In turn, metacognitive deficits might represent core drivers of polarized or radical beliefs, owing to their consequence for maladaptive confirmation bias. Interestingly, this hypothesis is in line with empirical observations [4], showing that more dogmatic participants show reduced metacognitive sensitivity which in turn is predictive of reduced post-decision evidence processing.

Recognizing metacognition as a central driver of belief polarization may make it possible to develop new strategies for debiasing decision-making [5]. The contributors to confirmation bias in any given setting are likely to be multifactorial, with more proximal causes (such as measures of cognitive style) having large effect sizes, but providing more limited mechanistic insight. By contrast, identifying small, reliable effect sizes associated with underlying mechanisms (such as the impact of confidence weighting) may bring us closer to the potential for targeted intervention. Excitingly, there are existing interventions that have been shown to boost people's metacognitive ability [58,59]. Our results indicate that cognitive training which improves domain-general self-awareness and metacognitive efficiency may help to alleviate the negative behavioural outcomes of selective information processing, and foster resilience against misinformation and belief polarization.

Acknowledgements

We thank Prof. Ralph Hertwig for insightful conversations regarding the potential adaptiveness of cognitive biases. M.R. is a predoctoral Fellow of the International Max Planck Research School on Computational Methods in Psychiatry and Ageing Research. The participating institutions are the Max Planck Institute for Human Development and University College London (UCL).

Data accessibility

Code of all simulations for this study is available at a dedicated Github repository (https://github.com/MaxRollwage/ConfirmationBias_simulation).

Authors' contributions

M.R. and S.M.F. conceived the study and wrote the paper. M.R. developed the methodology and ran the simulations with input from S.M.F.

Competing interests

We declare we have no competing interests.

Funding

The Wellcome Centre for Human Neuroimaging is supported by core funding from the Wellcome Trust grant no. (203147/Z/16/Z). S.M.F. is supported by a Sir Henry Dale Fellowship jointly funded by the Wellcome Trust and the Royal Society grant no. (206648/Z/17/Z).

References

  • 1.Kohut A, Doherty C, Dimock M, Keeter S. 2012. Partisan polarization surges in Bush, Obama years. Washington, DC: Pew Research Center for the People and the Press RSS. [Google Scholar]
  • 2.Del Vicario M, Scala A, Caldarelli G, Stanley HE, Quattrociocchi W. 2017. Modeling confirmation bias and polarization. Sci. Rep. 7, 40391. ( 10.1038/srep40391) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Lilienfeld SO, Ammirati R, Landfield K. 2009. Giving debiasing away: can psychological research on correcting cognitive errors promote human welfare? Perspect. Psychol. Sci. 4, 390-398. ( 10.1111/j.1745-6924.2009.01144.x) [DOI] [PubMed] [Google Scholar]
  • 4.Rollwage M, Dolan RJ, Fleming SM. 2018. Metacognitive failure as a feature of those holding radical beliefs. Curr. Biol. 28, 4014-4021. ( 10.1016/j.cub.2018.10.053) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Rollwage M, Zmigrod L, de-Wit L, Dolan RJ, Fleming SM. 2019. What underlies political polarization? A manifesto for computational political psychology. Trends Cogn. Sci. 23, 820-822. ( 10.1016/j.tics.2019.07.006) [DOI] [PubMed] [Google Scholar]
  • 6.Nickerson RS. 1998. Confirmation bias: a ubiquitous phenomenon in many guises. Rev. Gen. Psychol. 2, 175-220. ( 10.1037/1089-2680.2.2.175) [DOI] [Google Scholar]
  • 7.Klayman J. 1995. Varieties of confirmation bias. Psychol. Learn. Motiv. 32, 385-418. ( 10.1016/S0079-7421(08)60315-1) [DOI] [Google Scholar]
  • 8.Groopman J. 2008. How doctors think. Boston, MA: Houghton Mifflin Harcourt. [Google Scholar]
  • 9.Snyder M, Swann WB. 1978. Hypothesis-testing processes in social interaction. J. Pers. Soc. Psychol. 36, 1202. ( 10.1037/0022-3514.36.11.1202) [DOI] [Google Scholar]
  • 10.Park J, Konana P, Gu B, Kumar A, Raghunathan R. 2010. Confirmation bias, overconfidence, and investment performance: evidence from stock message boards. McCombs Research Paper Series No. IROM-07-10. Austin, TX: University of Texas.
  • 11.Lord CG, Ross L, Lepper MR. 1979. Biased assimilation and attitude polarization: the effects of prior theories on subsequently considered evidence. J. Pers. Soc. Psychol. 37, 2098. ( 10.1037/0022-3514.37.11.2098) [DOI] [Google Scholar]
  • 12.Sunstein CR, Bobadilla-Suarez S, Lazzaro SC, Sharot T. 2016. How people update beliefs about climate change: good news and bad news. Cornell L. Rev. 102, 1431. ( 10.31228/osf.io/qn5wp) [DOI] [Google Scholar]
  • 13.Kaplan JT, Gimbel SI, Harris S. 2016. Neural correlates of maintaining one's political beliefs in the face of counterevidence. Sci. Rep. 6, 1-11. ( 10.1038/srep39589) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Nyhan B, Reifler J. 2010. When corrections fail: the persistence of political misperceptions. Polit. Behav. 32, 303-330. ( 10.1007/s11109-010-9112-2) [DOI] [Google Scholar]
  • 15.Redlawsk DP. 2002. Hot cognition or cool consideration? Testing the effects of motivated reasoning on political decision making. J. Polit. 64, 1021-1044. ( 10.1111/1468-2508.00161) [DOI] [Google Scholar]
  • 16.Taber CS, Cann D, Kucsova S. 2009. The motivated processing of political arguments. Polit. Behav. 31, 137-155. ( 10.1007/s11109-008-9075-8) [DOI] [Google Scholar]
  • 17.Taber CS, Lodge M. 2012. Motivated skepticism in the evaluation of political beliefs (2006). Crit. Rev. 24, 157-184. ( 10.1080/08913811.2012.711019) [DOI] [Google Scholar]
  • 18.Zmigrod L. 2020. The role of cognitive rigidity in political ideologies: theory, evidence, and future directions. Curr. Opin. Behav. Sci. 34, 34-39. ( 10.1016/j.cobeha.2019.10.016) [DOI] [Google Scholar]
  • 19.Zmigrod L, Rentfrow PJ, Robbins TW. 2019. The partisan mind: is extreme political partisanship related to cognitive inflexibility? J. Exp. Psychol. Gen. 149, 407. ( 10.1037/xge0000661) [DOI] [PubMed] [Google Scholar]
  • 20.Zmigrod L, Rentfrow PJ, Robbins TW. 2019. Cognitive inflexibility predicts extremist attitudes. Front. Psychol. 10, 989. ( 10.3389/fpsyg.2019.00989) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Sinclair AH, Stanley M, Seli P. 2020. Closed-minded cognition: right-wing authoritarianism is negatively related to belief updating following prediction error. Psychonomic Bull. Rev. 27, 1348-1361. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Zmigrod L, Zmigrod S, Rentfrow PJ, Robbins TW. 2019. The psychological roots of intellectual humility: the role of intelligence and cognitive flexibility. Pers. Individ. Diff. 141, 200-208. ( 10.1016/j.paid.2019.01.016) [DOI] [Google Scholar]
  • 23.Palminteri S, Lefebvre G, Kilford EJ, Blakemore S. 2016. Confirmation bias in human reinforcement learning. bioRXiv, 1–21. ( 10.1101/090654) [DOI]
  • 24.Rollwage M, Loosen A, Hauser TU, Moran R, Dolan RJ, Fleming SM. 2020. Confidence drives a neural confirmation bias. Nat. Commun. 11, 1-11. ( 10.1038/s41467-020-16278-6) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Talluri BC, Urai AE, Tsetsos K, Usher M, Donner TH. 2018. Confirmation bias through selective overweighting of choice-consistent evidence. Curr. Biol. 28, 3128-3135. ( 10.1016/j.cub.2018.07.052) [DOI] [PubMed] [Google Scholar]
  • 26.Atiya NAA, Rañó I, Prasad G, Wong-Lin K. 2019. A neural circuit model of decision uncertainty and change-of-mind. Nat. Commun. 10, 1-12. ( 10.1038/s41467-018-07882-8) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Balsdon T, Wyart V, Mamassian P. 2020. Confidence controls perceptual evidence accumulation. Nat. Commun. 11, 1-11. ( 10.1038/s41467-020-15561-w) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Desender K, Boldt A, Yeung N. 2018. Subjective confidence predicts information seeking in decision making. Psychol. Sci. 29, 761-778. ( 10.1177/0956797617744771) [DOI] [PubMed] [Google Scholar]
  • 29.Meyniel F, Dehaene S. 2017. Brain networks for confidence weighting and hierarchical inference during probabilistic learning. Proc. Natl Acad. Sci. USA 114, E3859-E3868. ( 10.1073/pnas.1615773114) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.Murphy PR, Robertson IH, Harty S, O'Connell RG. 2015. Neural evidence accumulation persists after choice to inform metacognitive judgments. eLife 4, 1-23. ( 10.7554/eLife.11946) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31.Gigerenzer G. 2008. Why heuristics work. Perspect. Psychol. Sci. 3, 20-29. ( 10.1111/j.1745-6916.2008.00058.x) [DOI] [PubMed] [Google Scholar]
  • 32.Klayman J, Ha YW. 1987. Confirmation, disconfirmation, and information in hypothesis testing. Psychol. Rev. 94, 211-228. [Google Scholar]
  • 33.Evans JSBT. 1989. Bias in human reasoning: causes and consequences. Mahwah, NJ: Lawrence Erlbaum Associates, Inc. [Google Scholar]
  • 34.Lefebvre G, Summerfield C, Bogacz R. 2020. A normative account of confirmatory biases during reinforcement learning. bioRxiv. ( 10.1101/2020.05.12.090134) [DOI]
  • 35.Tsetsos K, Moran R, Moreland J, Chater N, Usher M, Summerfield C. 2016. Economic irrationality is optimal during noisy decision making. Proc. Natl Acad. Sci. USA 113, 3102-3107. ( 10.1073/pnas.1519157113) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36.Fleming SM, Dolan RJ. 2012. The neural basis of metacognitive ability. Phil. Trans. R. Soc. B 367, 1338-1349. ( 10.1098/rstb.2011.0417) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37.Bronfman ZZ, Brezis N, Moran R, Tsetsos K, Donner T, Usher M. 2015. Decisions reduce sensitivity to subsequent information. Proc. R. Soc. B 282, 20150228. ( 10.1098/rspb.2015.0228) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 38.Fleming SM, van der Putten EJ, Daw ND. 2018. Neural mediators of changes of mind about perceptual decisions. Nat. Neurosci. 21, 617-624. ( 10.1038/s41593-018-0104-6) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39.Mathys C, Daunizeau J, Friston KJ, Stephan KE. 2011. A Bayesian foundation for individual learning under uncertainty. Fron. Hum. Neurosci. 5, 39. ( 10.3389/fnhum.2011.00039) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40.O'Reilly JX, Schuffelgen U, Cuell SF, Behrens TEJ, Mars RB, Rushworth MFS. 2013. Dissociable effects of surprise and model update in parietal and anterior cingulate cortex. Proc. Natl Acad. Sci. USA 110, E3660-E3669. ( 10.1073/pnas.1305373110) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41.Bouttier V, Deneve S, Jardri R. 2019. Confirmation bias is explained by descending loops in the cortical hierarchy. In Conf. on Cognitive Computational Neuroscience, 13–16 September 2019, Berlin, Germany. ( 10.32470/CCN.2019.1433-0) [DOI] [Google Scholar]
  • 42.Jardri R, Duverne S, Litvinova AS, Denève S. 2017. Experimental evidence for circular inference in schizophrenia. Nat. Commun. 8, 14218. ( 10.1038/ncomms14218) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43.Maniscalco B, Lau H. 2012. A signal detection theoretic approach for estimating metacognitive sensitivity from confidence ratings. Conscious Cogn. 21, 422-430. ( 10.1016/j.concog.2011.09.021) [DOI] [PubMed] [Google Scholar]
  • 44.Moran R, Teodorescu AR, Usher M. 2015. Post choice information integration as a causal determinant of confidence: novel data and a computational account. Cognit. Psychol. 78, 99-147. ( 10.1016/j.cogpsych.2015.01.002) [DOI] [PubMed] [Google Scholar]
  • 45.Moreira CM, Rollwage M, Kaduk K, Wilke M, Kagan I. 2018. Post-decision wagering after perceptual judgments reveals bi-directional certainty readouts. Cognition 176, 40-52. ( 10.1016/j.cognition.2018.02.026) [DOI] [PubMed] [Google Scholar]
  • 46.Pleskac TJ, Busemeyer JR. 2010. Two-stage dynamic signal detection: a theory of choice, decision time, and confidence. Psychol. Rev. 117, 864-901. ( 10.1037/a0019737) [DOI] [PubMed] [Google Scholar]
  • 47.Fleming SM, Daw ND. 2017. Self-evaluation of decision-making: a general bayesian framework for metacognitive computation. Psychol. Rev. 124, 91-114. ( 10.1037/rev0000045) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 48.Ais J, Zylberberg A, Barttfeld P, Sigman M. 2016. Individual consistency in the accuracy and distribution of confidence judgments. Cognition 146, 377-386. ( 10.1016/j.cognition.2015.10.006) [DOI] [PubMed] [Google Scholar]
  • 49.Fleming SM, Weil RS, Nagy Z, Dolan RJ, Rees G. 2010. Relating introspective accuracy to individual differences in brain structure. Science 329, 1541-1543. ( 10.1126/science.1191883) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 50.Pleskac TJ, Hertwig R. 2014. Ecologically rational choice and the structure of the environment. J. Exp. Psychol. Gen. 143, 2000. ( 10.1037/xge0000013) [DOI] [PubMed] [Google Scholar]
  • 51.Miyoshi K, Lau H. In press. A decision-congruent heuristic gives superior metacognitive sensitivity under realistic variance assumptions. Psychol. Rev. [DOI] [PubMed]
  • 52.Rabbitt PM. 1966. Errors and error correction in choice-response tasks. J. Exp. Psychol. 71, 264. ( 10.1037/h0022853) [DOI] [PubMed] [Google Scholar]
  • 53.Resulaj A, Kiani R, Wolpert DM, Shadlen MN. 2009. Changes of mind in decision-making. Nature 461, 263-266. ( 10.1038/nature08275) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 54.Berg RVD, Anandalingam K, Zylberberg A, Kiani R, Shadlen MN, Wolpert DM. 2016. A common mechanism underlies changes of mind about decisions and confidence. eLife 5, 1-21. ( 10.7554/eLife.12192) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 55.Yeung N, Summerfield C. 2012. Metacognition in human decision-making: confidence and error monitoring. Phil. Trans. R. Soc. B 367, 1310-1321. ( 10.1098/rstb.2011.0416) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 56.Maniscalco B, Peters MAK, Lau H. 2016. Heuristic use of perceptual evidence leads to dissociation between performance and metacognitive sensitivity. Atten. Percep. Psychophys. 78, 923-937. ( 10.3758/s13414-016-1059-x) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 57.Meyniel F, Sigman M, Mainen ZF. 2015. Confidence as Bayesian probability: from neural origins to behavior. Neuron 88, 78-92. ( 10.1016/j.neuron.2015.09.039) [DOI] [PubMed] [Google Scholar]
  • 58.Baird B, Mrazek MD, Phillips DT, Schooler JW. 2014. Domain-specific enhancement of metacognitive ability following meditation training. J. Exp. Psychol. Gen. 143, 1972. ( 10.1037/a0036882) [DOI] [PubMed] [Google Scholar]
  • 59.Carpenter J, Sherman MT, Kievit RA, Seth AK, Lau H, Fleming SM. 2019. Domain-general enhancements of metacognitive ability through adaptive training. J. Exp. Psychol. Gen. 148, 51. ( 10.1037/xge0000505) [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

Code of all simulations for this study is available at a dedicated Github repository (https://github.com/MaxRollwage/ConfirmationBias_simulation).


Articles from Philosophical Transactions of the Royal Society B: Biological Sciences are provided here courtesy of The Royal Society

RESOURCES