Skip to main content
Elsevier Sponsored Documents logoLink to Elsevier Sponsored Documents
. 2019 Nov-Dec;67:101504. doi: 10.1016/j.ijlp.2019.101504

Advice-taking as a bridge between decision neuroscience and mental capacity

Elisa van der Plas a,, Anthony S David b, Stephen M Fleming a,c,
PMCID: PMC7661837  PMID: 31785723

Abstract

A person's capacity to process advice is an important aspect of decision making in the real world. For example, in decisions about treatment, the way patients respond to the advice of family, friends and medical professionals may be used (intentionally or otherwise) as a marker of the “use or weigh” requirement of decision-making capacity. Here we explore neuroscientific research on decision-making to identify features of advice-taking that help conceptualize this requirement. We focus on studies of the neural and computational basis of decision-making in laboratory settings. These studies originally investigated simple perceptual decisions about ambiguous stimuli, but have more recently been extended to more complex “value-based” decisions involving the comparison of subjective preferences. Value-based decisions are a useful model system for capacity-related decision-making as they do not have an objectively ‘correct’ answer and are instead based on subjective preferences. In this context, advice-taking can be seen as a process in which new evidence for one or other option is integrated, leading to altered behaviour or choices. We use this framework to distinguish between different types of advice-taking: private compliance consists of updating one's privately held beliefs based on new evidence, whereas in the case of public compliance, people change their behaviour at a surface level without shifting their privately-held beliefs. Importantly, both types of advice-taking may lead to similar outcomes but rely on different decision processes. We suggest that understanding how multiple mechanisms drive advice-taking holds promise for targeting decision-making support and improving our understanding of the use and weigh requirement in cases of contested capacity.

Keywords: Capacity, Decision-making capacity, Advice-taking, Decision neuroscience

1. Introduction

The young patient “Z” has been considered by her clinicians not to have full insight into her risky online behaviour and social interactions. The court questions whether Z’s abnormal susceptibility to others is an indication of her inability to ‘use and weigh’:

“…as a consequence of Z's autism, which impacts on her ability to put herself in other people's shoes and make judgments with regard to their intentions towards her, she struggles to think through the consequences of having contact with another individual who may pose a risk to her […]. As a consequence of this inability to weigh up the positives […] Z lacks capacity to make decisions regarding contact with others.” (Cobb [2016] EWCOP 4)

The legal notion of mental capacity refers to an individual's ability to make autonomous decisions about their own welfare, often referred to as decision-making capacity (DMC). In the UK, the concept of capacity in relation to various types of decisions was initially developed via the common law, before being incorporated into a statutory framework in the Adults with Incapacity (Scotland) Act (2000) and the Mental Capacity Act (MCA; 2005) in England and Wales. These instruments are intended to cover the circumstances in which necessary acts of caring can be administered (and decisions about these acts taken) on behalf of individuals lacking capacity to either consent to care or make their own decisions.

According to the MCA, the criteria for assessing mental capacity are twofold – diagnostic and functional. The diagnostic criteria state that a person's lack of capacity must be due to an “impairment of, or a disturbance in the functioning of, the mind or brain” (s. 2(1)). The functional criterion is a test for the capacity to make decisions, and states (s. 3(1)) that an individual is unable to make a decision for himself if he is unable –

  • a)

    to understand the information relevant to the decision,

  • b)

    to retain that information,

  • c)

    to use or weigh that information as part of the process of making the decision, or

  • d)

    to communicate his decision (whether by talking, using sign language or any other means).

While a majority of capacity cases are uncontroversial (e.g. Ruck Keene, Kane, Kim & Owen, 2019, estimate that 95% of applications to the Court of Protection fall into the category of uncontentious applications determined without a hearing), in those that are contested, the most contentious point is the “using or weighing” of information which does not always have obvious outward signs (Case, 2016; Ruck Keene et al., 2019). The MCA 2005 is also clear that the application of idiosyncratic values should not bias the assessment of use or weigh, stating that P “is not to be treated as unable to make a decision merely because he makes an unwise decision” (s.1(4)). In other words, the law aspires to be value-liberal – supporting the right of patients to act for “any reason, rational or irrational, or for no reason at all” (Re MB [1997]; Re T [1992] ER 649). As an example, we can consider the recent case of “C”, who refused treatment for dialysis despite good prognosis because she believed it would compromise her freewheeling lifestyle. In a statement to the court, her daughter wrote: “Although they are not reasons that are easy to understand, I believe that they are not only fully thought through, but also entirely in keeping with both her (unusual) value system and her (unusual) personality” (Kings College Hospital NHS Foundation Trust v C and V [2015] EWCOP 80).

The use or weigh requirement refers to the process of making a decision, rather than the outcome of a decision. Evaluating how such a decision process unfolds is typically contingent on new information being provided to P in the form of information or advice – either formally (in the case of interactions with health professionals) or informally (from friends and family). For instance, in the case extract with which we began this article, we see that the court's impression of Z's decision-making capacity is informed by her excessive susceptibility to the influence of others, suggesting that she did not truly use or weigh information herself.

In this article we focus on the role that decision neuroscience might play in informing our understanding of the advice-taking process, and thereby how we conceptualize the use or weigh requirement. Because the process of advice-taking can be simulated in a laboratory setting, it forms a natural bridge between the literatures on decision neuroscience (e.g. Campbell-Meiklejohn, Bach, Roepstorff, Dolan, & Frith, 2010; Campbell-Meiklejohn, Simonsen, Frith, & Daw, 2017; De Martino, Bobadilla-Suarez, Nouguchi, Sharot, & Love, 2017; Izuma & Adolphs, 2013) and mental capacity (Gomez-Beldarrain, Harries, Garcia-Monco, Ballus, & Grafman, 2004; Kennedy, Dornan, Rutledge, O'Neill, & Kennedy, 2009). Another advantage of bridging between work on DMC and advice-taking is that decision neuroscience offers the promise of a deeper and more nuanced understanding of why certain decision processes may be impaired due to “an impairment of the mind or brain”.

The outline of this article is as follows. First, we start with briefly reviewing the general information-processing view of the mind in cognitive science that grounds a majority of work in decision neuroscience. Second, we review studies of the cognitive and neural basis of simple decisions, before exploring how these frameworks can be extended to accommodate advice-taking and so-called post-decisional processing. We then consider different types of advice-taking behaviour (such as how public compliance may differ from private belief change) and outline the contribution of metacognition (or insight) and mentalising in advice-taking. Finally, we explore how these lines of work in the neurosciences may inform an understanding of the psychological basis of the use or weigh requirement.

Before we begin, we should point out that the information-processing view of most current neuroscientific studies of decision-making is simplistic compared to the often complex, emotionally-charged decisions about which capacity is typically questioned. Further, it is not known how the core principles of decision-making elucidated in the lab generalise to these situations. For example, decisions that usually give rise to capacity assessments do not have an objectively ‘correct’ answer (e.g. treatment decisions, financial decisions, sexual consent), an issue we will return to towards the end of the article. On balance, however, we believe that there are benefits from adopting a decision neuroscience perspective, not least in providing a value-liberal framework with which to begin to quantify the process, rather than the content, of a decision – i.e. the internal workings of the mind of an individual who is making the decision and not what preferences or values constitute this process.

An example of value-based practise in health care is that of a patient who decides not to engage in treatment because the treatment reduces the patient's ability to engage in activity X (e.g. work as a primary school teacher). Good medical practise is to judge this patient's DMC in a way that is unbiased by the patient's personal preferences, e.g. the content of X (Petrova, Dale, & Bill, 2006); even better medical practise would be to judge the patient's DMC based on the process through which the preference for X was generated, in a way that is impartial to the content of X. Our framework does not say what people should choose, but suggests means of support to encourage people to use and weigh new information in light of their current beliefs, with a mechanistic model of decision-making as a guide.

2. Cognitive and computational approaches to decision-making

The dominant model in the cognitive science of decision-making is that a decider receives some kind of input (from the senses or from memory) and then processes these inputs to arrive at a discrete choice. An agent is thought to reduce uncertainty about which decision to make by sampling information about the potential benefits and disadvantages of the various available options (Gold & Shadlen, 2007). The goal of decision neuroscience is to work out what kind of internal processes govern this transformation of input into output.

Research in this field has focused on two different types of decision. First, research on “perceptual decision-making” (PDM) builds on classical psychophysics from the late 19th century (e.g. Peirce & Jastrow, 1884), and aims to understand how subjects discriminate different types of sensory information (such as deciding whether an object is an apple or an orange from only a brief glimpse; see Hanks & Summerfield, 2017, for a recent review). In contrast, research on “value-based decision-making” (VDM) grew out of the field of behavioural economics, and studies the processes involved in choices involving a comparison of subjective preferences (such as deciding whether to eat an apple or an orange; see Rangel, Camerer, & Montague, 2008). In both cases, the decision process is often studied under conditions of uncertainty, such as when sensory information is noisy or ambiguous, or when value-based choices are made about stimuli that deliver a variable and changing level of reward.

A key benefit of studying PDM is that an explicitly ‘correct’ answer exists, allowing a precise quantification of the speed and accuracy of individual choices. An influential task in studies of PDM is the random-dot motion task (Fig. 1a). Random-dot motion stimuli consist of a rapidly moving cloud of dots presented briefly on a computer screen (typically for less than a second), and a given proportion of the dots are arranged to move coherently in a particular direction (e.g. left or right; Britten, Shadlen, Newsome, & Movshon, 1992; Kim & Shadlen, 1999). On each ‘trial’ of the task subjects are asked to decide whether the dot cloud is mostly moving in one or other direction. It is striking that behaviour in these kinds of tasks can be precisely accommodated by computational models that assume the brain receives noisy samples of evidence about the world (e.g. whether the dots are moving left or right), and compares these samples to an internal decision threshold (Fig. 1c; Vickers, 1979; Luce, 1992; Ratcliff, 1978). When these samples are aggregated over time, these models can predict how long the observer takes to make a decision (Smith & Ratcliff, 2004; Wald, 1947) and how confident they are in their choice (Kiani & Shadlen, 2009).

Fig. 1.

Fig. 1

A schematic example of an evidence accumulation framework for perceptual decision-making (PDM) and value-based decision-making (VDM). (a). A decision about whether a patch of dots is moving to the left or right (PDM), or (b) whether to eat an apple or an orange (VDM), are both studied under cases of uncertainty e.g. when sensory information is noisy or ambiguous or when the value of each object is not perfectly known. (c). Computational models of the decision process assume that the brain receives noisy samples of evidence about the world and compares the accumulated evidence to an internal threshold for each available option (the threshold for when accumulated value results in a selection of an apple or an orange, for instance). When these samples are aggregated over time, such models can predict both choice and how long the observer takes to make a decision (response time).

Similar evidence accumulation models have been successfully applied to model both decision time and choice in VDM (Basten, Biele, Heekeren, & Fiebach, 2010; Hutcherson, Bushong, & Rangel, 2015; Krajbich, Armel, & Rangel, 2010; Krajbich, Hare, Bartling, Morishima, & Fehr, 2015). A classic example is the food choice task, on which hungry subjects are presented with images of food items that can be selected for consumption at the end of the experiment (Fig. 1b). How much time it takes for someone to decide is an indication of the subjective difficulty of the decision (which is usually higher when the two options are more similar in value; Krajbich et al., 2010; Fig. 1c). These findings imply that subjects also have internal uncertainty about their own preferences – for instance how confident they are in preferring apples over oranges (De Martino, Fleming, Garrett, & Dolan, 2013; Manski, 1977). We will return to this important point later in the article.

In the context of value-based decision-making, decision neuroscience has also emphasised a distinction between whether individuals tend to act habitually; recalling the value that has often been attached to a previous action in the past, or in a way that foresees the future consequences of sequences of actions (sometimes known as model-based planning; Valentin, Dickinson, & Doherty, 2007). While this distinction is beyond the scope of the current article, it is plausible that the depth of planning may intersect with advice-taking in interesting ways (e.g. an individual who is myopic in their planning may be over-responsive to advice).

3. Decision-making in the brain

Following the creation of simple paradigms for studying decision-making in the lab, there has been increasing interest in the neural basis of these types of decisions (Gold & Shadlen, 2007). We refer the reader to comprehensive reviews on the neural basis of PDM and VDM (Glimcher & Fehr, 2013; Hanks & Summerfield, 2017), and here focus on a few core principles that have emerged from this work. To a rough approximation, different subregions of the cortex are engaged in different, specialized computations. For instance, the visual cortex, in the posterior temporal and occipital lobes, is specifically involved in processing inputs from the eyes to support visual perception. A series of studies in animals and humans over the past two decades has begun to bridge the gap between perception and action to understand how simple choices are made. In studies of PDM, for instance, neural populations in the association cortex – the parietal and frontal lobes – have been found to gradually increase their firing rate up to a point at which a decision is made (Heekeren, Marrett, Bandettini, & Ungerleider, 2004; O'Connell, Dockree, & Kelly, 2012; Roitman & Shadlen, 2002). The idea is that these activity patterns receive input from areas processing sensory input (for instance area V5 of the visual cortex in the case of the random dot motion task), and represent a neural instantiation of the evidence accumulation process inherent to the computational frameworks introduced above.

Intriguingly, the same circuits are also involved when people make more subjective choices in VDM paradigms (e.g. Platt & Glimcher, 1999). Rather than integrating sensory information such as direction of motion, subjective values (such as how much I prefer apples over oranges) are thought to feed into these choice circuits until a decision is made. It is thought that the medial prefrontal cortex, in conjunction with regions important for memory such as the hippocampus (Barron, Dolan, & Behrens, 2013), retrieves information about past experience to construct these subjective values (Chib, Rangel, Shimojo, & Doherty, 2009; Glimcher, 2011; Rangel et al., 2008). On these tasks, there is no objectively ‘correct’ answer, but difficulty can be manipulated by changing the difference in subjective value between the two choice options. In the next section we turn to how these studies of information processing in the service of simple decisions help shed light on the use or weigh process during advice-taking.

4. Post-decisional evidence processing as a model of advice-taking

A majority of work in decision neuroscience has focused on characterising the processes leading up to a value-based or perceptual choice, but less attention has been paid to how people integrate new information after a choice has been made, or how this post-decisional processing may lead to a change of mind. Initial work in this area characterised how evidence accumulation may continue for a short time after the choice, supporting endogenous error monitoring (Murphy, Robertson, Harty, & O'Connell, 2015; Rabbitt, 1966) and changes of mind (Resulaj, Kiani, Wolpert, & Shadlen, 2009; van den Berg et al., 2016). More recently this line of work has been extended to ask how exogenous evidence presented after an initial choice may lead to later changes of mind.

In these tasks, subjects first make a judgment based on some evidence (e.g. an estimation of the direction of a random-dot motion display), after which they are presented with new evidence (e.g. additional motion) and are asked to make a final judgment. The general finding is that people tend to update their final judgment after seeing the new evidence and that this updating is stronger when the new evidence is more reliable or stronger (Bronfman et al., 2015; Fleming, van der Putten, & Daw, 2018), or when the new evidence confirms the initial judgment (known as ‘confirmation bias’; Talluri, Urai, Tsetsos, Usher, & Donner, 2018).

For example, Fleming et al. (2018) asked subjects to make a first judgment about whether a random-dot motion display with variable evidence strength was mostly moving to the left or right. After this decision, people were shown an additional display of motion in the same direction as the first display (post-decision evidence) which the subjects could use to update their confidence about whether their initial decision was correct. The authors also scanned the brains of participants while they received the new evidence using functional Magnetic Resonance Imaging (fMRI), which tracks fluctuations in blood flow (a proxy of neuronal activity) in different regions of the brain. This allowed the identification of regions in which activity fluctuations tracked whether or not people changed their minds. The posterior medial frontal cortex (pMFC), a region that activates when people detect that they have made a mistake and adjust their behaviour accordingly (Dehaene, Posner, & Tucker, 1994; Ridderinkhof, van den Wildenberg, Segalowitz, & Carter, 2004), tracked the strength of post-decision evidence and signalled the need for behavioural adaptation. In contrast, activity in the anterior prefrontal cortex (aPFC), a region involved in metacognition (the ability to know whether we are right or wrong, see below) mediated the impact of new evidence on people's subjective confidence.

Together, these studies have contributed to an emerging understanding of how new evidence is processed to flexibly change an initial judgment when needed. Interestingly, these studies have also shown that the way in which new evidence is integrated is similar across distinct tasks (e.g. perceptual and numerical tasks; Bronfman et al., 2015; Talluri et al., 2018). Ongoing research is now building upon this finding by examining how new evidence is integrated from both non-social and social information sources, and whether non-social and social post-decision evidence provide similar information content (Olsen, Roepstorff & Bang, 2018; Pescetelli & Yeung, 2018). These paradigms ask subjects to make a first (perceptual or value-based) decision and then present the opinion of an ‘adviser’ (Behrens, Hunt, Woolrich, & Rushworth, 2008; Campbell-Meiklejohn et al., 2010; Campbell-Meiklejohn et al., 2017; De Martino et al., 2017; Gomez-Beldarrain et al., 2004; Sniezek & Van Swol, 2001). Interestingly, and in line with the finding that more reliable evidence elicits more changes of mind (Bronfman et al., 2015; Fleming et al., 2018; Talluri et al., 2018), the reliability of the advice is a crucial determinant of how much it engenders a change in subjects' beliefs. In advice-taking settings, this advice reliability can be communicated in the form of the confidence of the adviser (Campbell-Meiklejohn et al., 2010; Campbell-Meiklejohn et al., 2017; Gomez-Beldarrain et al., 2004; Sniezek & Van Swol, 2001), as judgments made with higher confidence are typically more likely to be correct (Fig. 2). In turn, the effects of the advice are strongest when the initial decision is ambiguous or when an agent is uncertain about their initial judgment (Cialdini & Goldstein, 2004; De Martino et al., 2017). This information-processing view of advice gives rise to useful predictions for when advice is most likely to have an impact on its recipient, in that new (sensory or social) evidence would be expected to have a larger impact on uncertain agents. Confidence is often distorted in psychiatric disorders (e.g. more or less confidence than warranted by performance; Rouault, Seow, Gillan, & Fleming, 2018; Stephan et al., 2016); this may explain why some patients have been reported to be over- or under-susceptible to advice, a topic to which we will return in the next section. Together, these studies provide initial evidence that social advice is processed similarly to post-decisional evidence in mediating changes of mind.

Fig. 2.

Fig. 2

People who are more certain about an initial judgment are less likely to change their minds upon being presented with new sensory evidence or social advice. A decider receives a first sample of sensory evidence on a random-dot motion task (first judgment) and is then presented with either new sensory evidence (a) or social advice (b). Based on this new evidence, the decider makes a second judgment. Confidence in the first judgment affects how likely the decider is to take on board the new evidence (integrating the new sensory evidence or following the advice). More (versus less) confident decisions are less (versus more) likely to be updated when new sensory or social evidence becomes available.

5. The role of metacognition in advice-taking

In the previous section we outlined how changes of mind and improvements in performance are tightly coupled: being sensitive to new evidence is especially beneficial when one's initial judgment is incorrect in order to engender a change of mind. However, even if confidence and accuracy are usually tightly linked, individuals vary in the ability to monitor their performance, known as metacognition. Specifically, some people are less able to notice differences in their performance, or believe themselves to be more or less accurate than they objectively are (Rouault, Seow, et al., 2018). It has recently become possible to precisely quantify a person's ability to discriminate between their own correct and incorrect judgments (known as “metacognitive sensitivity”) using a variant of signal detection theory (Fleming & Lau, 2014). There are various ways of calculating metacognitive sensitivity, with more sophisticated methods being uncontaminated by differences in both accuracy and metacognitive bias, a person's overall propensity to report high confidence (Fleming, Weil, Nagy, Dolan, & Rees, 2010; Rouault, Seow, et al., 2018).

Individual differences in metacognitive sensitivity have been found to correlate with the structure and function of the aPFC, a brain area involved in self-evaluation (Christoff & Gabrieli, 2000; Fleming, Huijgen & Dolan, 2012; Hilgenstock, Weiss, & Witte, 2014; Shimamura & Squire, 1986). This region together with the pMFC (discussed above in the context of advice-taking) is reliably activated during the self-evaluation of performance (Fleming & Dolan, 2012). It is notable that areas such as pMFC and aPFC are implicated in both advice-taking and the metacognitive evaluation of recent decisions, suggesting an important link between these two processes. The general link is unsurprising given that knowing whether we are right or wrong (metacognition) is one of the features that enable us to know when new evidence (advice-taking) is beneficial. For example, if you are completely convinced about the best course of action, you may be less likely to take someone's advice or consider another approach. On the other hand, if you are unsure about your decision, you may be more likely to look at what others are doing and follow their advice (Fig. 2 illustrates the general relationship between confidence and advice-taking). In sum, as advice-taking is most beneficial when one's first judgment was wrong, it follows that metacognition is one of the features that contributes to knowing when to take on board advice (Frith, 2012; Shea et al., 2014).

In clinical settings, the term insight is used to encompass a patient's awareness that they have a disorder, their ability to reflect on specific beliefs including their response to hypothetical contradiction, and to some extent follow the clinician's advice, for example regarding treatment (Amador & David, 2004; David, 1990). There is also an association between the concept of clinical insight and metacognition (David, Bedford, Wiffen, & Gilleen, 2012; Koren et al., 2005, Koren et al., 2004). Patients deemed to have poor insight often show impaired decision-making capacity (Owen et al., 2009). Furthermore, patients who lack mental capacity are more likely to experience coercion (Cairns et al., 2005). To the extent to which insight depends on metacognitive mechanisms, it may be that poor insight leads to difficulty with the advice-taking process (and vice-versa).

This link between metacognition and advice-taking is supported by a recent study from Rollwage, Dolan, and Fleming (2018), who measured both metacognitive sensitivity and sensitivity to post-decision evidence, and asked how these aspects were related to a personality feature known as dogmatism (measured as the extent to which individuals were accepting of conflicting views on political issues). Metacognitive sensitivity predicted the extent to which subjects integrated new evidence on the perceptual task, supporting adaptive changes of mind. In turn, both of these features of decision-making were attenuated among those with higher levels of dogmatism. This finding shows that metacognitive sensitivity may promote adaptive changes of mind when new evidence becomes available. On top of that, this study suggests similar processes may govern the processing of new evidence in both low-level perceptual discrimination tasks and the broader, more subjective decisions about topics such as politics.

6. The role of mentalising in advice-taking

When a decision is difficult to make, it can be advantageous to use the full range of available evidence that is at one's disposal to reduce choice uncertainty, including the advice of others. However, as briefly mentioned at the beginning of this paper, from an information-processing perspective, advice-taking should not only be dependent on the decider's confidence but also on the reliability of the new evidence (De Martino et al., 2017). In other words, efficient advice-taking involves assessing not only the probability that oneself is correct (metacognition) but also the probability that the adviser is correct (Burke, Tobler, Baddeley, & Schultz, 2010; De Martino et al., 2017; Harvey & Fischer, 1997; Pescetelli & Yeung, 2019). Advisors' expressions of certainty are typically a useful source of information about the advisers' reliability, as people who say that they are confident are usually also more likely to be correct (Campbell-Meiklejohn et al., 2010; Campbell-Meiklejohn et al., 2017; Gomez-Beldarrain et al., 2004; Sniezek & Van Swol, 2001). This process is complicated when the fidelity of the advisers' confidence ratings is not representative of their accuracy (the adviser's metacognitive sensitivity). Put differently, it is often sensible to take advisers' certainty estimates with a pinch of salt (Bahrami et al., 2012, Bahrami et al., 2010; Bang et al., 2017), and learn, over the course of repeated interactions, which advisers' confidence estimates are more reliable than others (Hertz et al., 2017; Pescetelli & Yeung, 2019). Pescetelli and Yeung (2019) show that, over the course of repeated interactions with an adviser, the extent to which ‘trust’ relationships are built/broken predict advice-taking. In psychiatric settings, this may be related to trusting treatment providers to give the right advice.

The general set of processes involved in making inferences about others' preferences and mental states is referred to as mentalising or Theory of Mind (ToM; Frith and Frith, 2012, Frith and Frith, 2003). ToM and metacognition may have a shared mechanism (Carruthers, 2009; Fleming & Daw, 2017) that involves understanding that others' world views may differ from our own and using this discrepancy to reason, for example, how reliable others' advice may be (Behrens et al., 2008). This form of ToM develops relatively late in human development (around five years; Wimmer & Perner, 1983) probably through external feedback that shows that others' worldviews can be different from one's own (Gopnik & Wellman, 2010; Heyes & Frith, 2014). In cases of advice-taking about treatment decisions (which are highly dependent on personal subjective preferences) mentalising may play a crucial role in understanding why someone decides one way or another. A well-established link between psychiatric disorders and mentalising is supportive of this idea (Pousa et al., 2008; but see, Frith & Happé, 1994), in addition to more recent neuroimaging studies that explain clinical insight as a dysregulated interplay between mentalising and metacognition (Holt et al., 2011; Modinos, Renken, Ormel, & Aleman, 2011). For the patient, mentalising may be involved in understanding that advice can be biased by the dissimilar worldviews of others or the fidelity of the clinician's metacognitive representations. For the clinician, mentalising may be involved in understanding the patient's motives or idiosyncratic preferences. Thus, ToM reasoning in advice-taking is a two-way process – a sound understanding of another person's views is important both for the patient to assess the reliability of the clinician and for the clinician to understand the decision-process of the patient (Fig. 3).

Fig. 3.

Fig. 3

A schematic illustration of how mentalising and metacognition may play a role in the advice-taking and advice-giving process between a patient and a clinician. Metacognition (indicated in orange) may be used by the patient to understand whether additional evidence is needed (how confident the patient is that their own opinion is correct and that their perceptions are ‘real’). For the clinician, metacognition is involved in expressing the right level of confidence in the advice (how confident the clinician is that their advice is correct). Mentalising (indicated in blue) plays a role in understanding why someone is deciding one way or another, which may be relevant for a clinician to know whether a patient is thinking through the consequences of different courses of action and whether the advice needs to be communicated differently to achieve a given level of influence. For the patient, mentalising may be used to understand the fidelity of the adviser's confidence (the clinician's metacognition). (For interpretation of the references to colour in this figure legend, the reader is referred to the web version of this article.)

7. Private versus public compliance

Up until this point we have considered how models of evidence integration in simple decisions can be extended to characterise the process of revising and updating a belief on the basis of new information. These private changes in belief are thought to affect what the subject truly feels or beliefs about the decision. In contrast, there are plenty scenarios in which subjects publicly comply in order to fit in with a social group (Bobek, Hageman, & Kelliher, 2011; Bond & Smith, 1996) or under the influence of an authority figure (Milgram, 1963), but do not alter their privately held beliefs. As an example, consider a clinician who advises a patient to take a particular medication and a patient who complies with this advice. It may be that the patient blindly follows the advice, or choses to do so tactically (e.g. to hasten their discharge from hospital) irrespective of whether (s)he agrees that the treatment is actually needed (Fig. 4b). Here, the outcome of the decision, i.e. following the advice, appears superficially similar to a case in which a patient complies because (s)he appreciates the need for the treatment – i.e. shows private as well as public compliance (Fig. 4a). Despite an apparently similar decision outcome in both cases, the consequences of advice-taking for behaviour may be radically different. For example, in the case of public compliance (i.e. the patient outwardly expresses their approval but is not genuinely convinced) a patient may be less likely to follow the advice when the adviser is removed from the situation (e.g. at home when medication needs to be taken). Yet in the case of private compliance, the patient's agreement is more likely to persist in time (Edelson, Sharot, Dolan, & Dudai, 2011; Izuma & Adolphs, 2013).

Fig. 4.

Fig. 4

A schematic illustration of private and public compliance. (a). An example of private compliance, in which the use or weigh process consists of engaging with new evidence (e.g. the reliability of the advice) leading to a shift in private beliefs. (b). An example of public compliance, where advice leads to public statements of agreement without a shift in privately held beliefs. The resulting decision outcome is similar in panels (a) and (b), i.e. in agreement with the advice, but has different consequences for behaviour.

A classic laboratory demonstration of supposedly public compliance is the Asch study, in which subjects were asked to judge the length of lines. When asked privately, subjects had an approximately 99% chance of getting the decision correct. In contrast, when subjects made the same decision in the company of six mock participants who were instructed to give the wrong answer, 40% followed the mock participants – a classic case of succumbing to social advice even when this advice-taking results in an incorrect answer (Asch, 1955). Deutsch and Gerard shed new light on the original interpretation of this behaviour as (automatic) public compliance. They experimentally removed the social pressure from the Asch paradigm, identifying two distinct types of compliance: that which was mitigated when the social pressure was removed (normative conformity to social expectations/public compliance), and that which persisted even in the absence of social pressure (accepting others' views because they are informative/private compliance; Deutsch & Gerard, 1955). This suggests that advice-taking may be motivated both by a public willingness to conform to social rules and by private convictions. This view has been supported by psychological experiments (Freedman & Fraser, 1966; Sowden et al., 2018) and, more recently, also by computational simulations (Constant, Ramstead, Veissière, & Friston, 2019; Toelch & Dolan, 2015).

These studies suggest that the decision processes underpinning public and private compliance can be distinguished using experimental techniques, and that advice-taking often consists of combining these two types of belief change. This distinction is supported by studies that have investigated the neural basis of public versus private compliance. For example, Izuma and Adolphs (2013) asked students to update value-based decisions based on the advice of a liked or disliked group, and tested whether the updated preferences still persisted four months later. Strikingly, the researchers found that, when activity in the dorsomedial prefrontal cortex (dmPFC) was low, the subjects showed public compliance that did not lead to a persistent shift in preferences. In contrast, when dmPFC activation was high, the shift in preference was persistent — representing a neural signature of private belief change. A similar dmPFC region has also been shown to take into account the reliability of other's views about everyday products in the form of Amazon star ratings (De Martino et al., 2017) and is involved in observational learning, where it tracks the actions of others (Burke et al., 2010). Together these studies suggest a role of the dmPFC in using or weighing others' advice in a way that is informed by its reliability (Behrens et al., 2008; Campbell-Meiklejohn et al., 2010; Campbell-Meiklejohn et al., 2017), which may be more relevant to informational (private) motivations than to social (public) motivations.

8. Neuro-computational insights into using or weighing

What is meant by the ability to “use or weigh” in the context of DMC? The contentious status of use or weigh in cases in which capacity is in dispute suggests that a perspective from decision neuroscience may shed light on the kinds of processes that support the ability to use or weigh decision evidence as per the MCA requirement. In the current article we have argued that advice-taking provides a “model system” for understanding the use and weigh requirement, given that how individuals respond to (formal, informal or hypothetical) advice is often at the heart of both capacity assessments and capacity disputes.

In this last section we consider how insights from decision neuroscience can be generalised to more complex, real-world decision problems. Most studies of advice-taking have used simple PDM tasks in which there is an objectively correct answer (such as the random-dot motion task). It remains to be seen how post-decision processing and advice-taking operate in tasks such as the food choice task introduced earlier as a canonical example of VDM. More broadly, there is a substantial challenge in translating insights from laboratory decision-making paradigms into the emotionally-charged, often one-off decisions faced by individuals subject to capacity assessments. However, studies of moral decision-making suggest that similar domain-general evidence-accumulation frameworks also hold for even the most serious choice problems (Shenhav & Greene, 2014). For instance, the classic trolley dilemma asks how people trade-off the motivations to save certain lives at the cost to others' (Fischer & Ravizza, 1992; Foot, 1967; Thomson, 1985). People with higher dorsolateral prefrontal cortex (dlPFC) activity are more outcome-based, in that they are more likely to consider killing one to save five morally justifiable (Greene, Nystrom, Engell, Darley, & Cohen, 2004; Greene, Sommerville, Nystrom, Darley, & Cohen, 2001), possibly because the dlPFC is involved in suppressing automatic emotional responses in complex social situations (Sanfey, Rilling, Aronson, Nystrom, & Cohen, 2013). Importantly however, when the dilemmas have actual consequences or when the act of killing is made more salient, people who first approve of the more outcome-based decision often change their preferences (Gold, Pulford, & Colman, 2015; Thomson, 1985), suggesting an important role of emotions in changes of mind about moral issues (for a possible computational basis, see: Crockett, 2013). Indeed, some aspects of the advice-taking process – such as metacognition and mentalising – may provide an enabling context that supports the ability to use or weigh decision evidence about a range of decision problems. We focus here on three aspects of metacognition that exemplify this point:

  • 1)

    Metacognition is a broadly domain-general resource – having good metacognition on one task predicts having good metacognition on another, unrelated task (McCurdy et al., 2013; Song et al., 2011; Mazaniceux, Fleming, Souchay, & Moulin, 2018), and this domain-generality may be supported by abstract representations of self-performance in anterior prefrontal cortex (Morales, Lau, & Fleming, 2018; Rouault, Mcwilliams, Allen, & Fleming, 2018). Understanding how metacognition operates in laboratory situations may therefore inform how metacognition (and its failures) affect advice-taking in real-world contexts.

  • 2)

    Metacognitive sensitivity is not necessarily associated with greater intelligence or intellectual ability. In a recent study, we collected data from almost 1000 individuals in a general population sample, quantifying both their metacognitive awareness of making good or bad perceptual choices, and a proxy for IQ (Rouault, Seow, et al., 2018). While there was a strong link between IQ and the speed/accuracy of the initial decision, there was no relationship between IQ and people's ability to evaluate their decision-making. Metacognitive sensitivity may therefore track an individual's awareness of their decision process independently of whether it is determined by a particular set of values or intellectual style.

  • 3)

    People have awareness of the consistency or accuracy of even value-based, subjective choices. In other words, people have a sense of when they are making a good choice, and this self-knowledge may share a similar neural basis to metacognition about other types of decision (De Martino et al., 2013). In turn, people with greater metacognitive sensitivity about value-based choices tend to make more consistent decisions over time (Folke, Jacobsen, Fleming, & De Martino, 2016). This self-knowledge about our preferences may be akin to a “higher-order desire” that endorses our first-order motives – we “want to want” something. Frankfurt (1971) identifies the presence of such second-order desires with autonomy, and empirical work on metacognition about VDM may inform an emerging understanding of autonomy and agency in DMC. This perspective holds promise for integrating work on metacognition of decision-making (as indexing second-order desires) and notions of free will and autonomy that are important in relation to decision-making capacity (Zürcher, Elger, & Trachsel, 2019). Indeed, it seems that practitioners “on the ground” often inform their assessment of whether a patient is using or weighing decision-evidence according to whether an individual has insight into the relevant evidence supporting a particular choice, and can reflect on the consequences of their actions (Owen, Freyenhagen, Hotopf, & Martin, 2015).

More broadly, the interplay between metacognition, mentalising and advice-taking processes may provide a useful conceptual framework for thinking about apparent “failures” of DMC. For instance, one could usefully ask whether the lack of capacity of Z in the opening example is due to a problem with mentalising, metacognition about her own preferences, or with the processes involved in incorporating social influence with her own beliefs (either privately or publicly). Knowledge of this sort could be used to provide targeted support, e.g. by providing metacognitive training (Carpenter et al., 2019) or by seeking to reduce or remove social pressures. For instance, Owen, Freyenhagen, Richardson, and Hotopf (2009) note that in the case of delusions, patients' metacognition may be intact, in that deluded beliefs and behaviours may be entirely self-consistent. However, mentalising may be less so, in the sense that delusions by definition do not map onto a shared, social reality. A finer-grained conceptualisation of the neuro-computational processes supporting advice-taking will, we hope, provide a framework for discussing such cases.

Author contributions

All authors conceptualized the paper. E.v.d.P. and S.M.F. wrote the paper with substantial comments from A.S.D.

Declarations of Competing Interest

None.

Funding

The authors are all core researchers in the ‘metacognition’ workstream of the Mental Health and Justice Project, which is funded by a grant from the Wellcome Trust (203376/2/16/Z). The Wellcome Centre for Human Neuroimaging is supported by core funding from the Wellcome Trust (203147/Z/16/Z). S.M.F. is supported by a Sir Henry Dale Fellowship jointly funded by the Wellcome Trust and Royal Society (206648/Z/17/Z). A.S.D. is also supported by the National Institute of Health Research Biomedical Research Centre at UCLH.

List of included cases

WBC v Z [2016] EWCOP 4.

Kings College Hospital NHS Foundation Trust v C and V [2015] EWCOP 80.

Re MB [1997] EWCA 3093.

Re T [1992] All ER 649.

Contributor Information

Elisa van der Plas, Email: elisa.plas.18@ucl.ac.uk.

Stephen M. Fleming, Email: stephen.fleming@ucl.ac.uk.

References

  1. Amador X.F., David A.S. 2004. Insight and Psychosis: Awareness of Illness in Schizophrenia and Related Disorders. [DOI] [Google Scholar]
  2. Asch S. Opinions and social pressure. Nature. 1955;193(5):31–35. doi: 10.1038/1761009b0. [DOI] [Google Scholar]
  3. Bahrami B., Olsen K., Bang D., Roepstorff A., Rees G., Frith C. What failure in collective decision-making tells us about metacognition. Philosophical Transactions of the Royal Society of London. Series B, Biological Sciences. 2012;367(1594):1350–1365. doi: 10.1098/rstb.2011.0420. [DOI] [PMC free article] [PubMed] [Google Scholar]
  4. Bahrami B., Olsen K., Latham P.E., Roepstorff A., Rees G., Frith C.D. Optimally interacting minds. Science. 2010;329(5995) doi: 10.1126/science.1185718. [DOI] [PMC free article] [PubMed] [Google Scholar]
  5. Bang D., Aitchison L., Moran R., Herce Castanon S., Rafiee B., Mahmoodi A.…Summerfield C. Confidence matching in group decision-making. Nature Human Behaviour. 2017;1(6):1–7. doi: 10.1038/s41562-017-0117. [DOI] [Google Scholar]
  6. Barron H.C., Dolan R.J., Behrens T.E.J. Online evaluation of novel choices by simultaneous representation of multiple memories. Nature Neuroscience. 2013;16(10):1492–1498. doi: 10.1038/nn.3515. [DOI] [PMC free article] [PubMed] [Google Scholar]
  7. Basten U., Biele G., Heekeren H.R., Fiebach C.J. How the brain integrates costs and benefits during decision making. Proceedings of the National Academy of Sciences. 2010;107(50):21767–21772. doi: 10.1016/0020-711X(72)90050-X. [DOI] [PMC free article] [PubMed] [Google Scholar]
  8. Behrens T.E.J., Hunt L.T., Woolrich M.W., Rushworth M.F.S. Associative learning of social value. Nature. 2008;456:245–249. doi: 10.1038/nature07538. [DOI] [PMC free article] [PubMed] [Google Scholar]
  9. van den Berg R., Anandalingam K., Zylberberg A., Kiani R., Shadlen M.N., Wolpert D.M. A common mechanism underlies changes of mind about decisions and confidence. ELife. 2016 doi: 10.7554/elife.12192. [DOI] [PMC free article] [PubMed] [Google Scholar]
  10. Bobek D.D., Hageman A.M., Kelliher C.F. The social norms of tax compliance: Scale development, social desirability, and presentation effects. Advances in Accounting Behavioral Research. 2011;74(1):49–64. doi: 10.1108/S1475-1488(2011)0000014005. [DOI] [Google Scholar]
  11. Bond R., Smith P.B. Culture and conformity: A meta-analysis of studies using asch's (1952b, 1956) line judgment task. Psychological Bulletin. 1996;119(1):111–137. doi: 10.1037/0033-2909.119.1.111. [DOI] [Google Scholar]
  12. Britten K.H., Shadlen M.N., Newsome W.T., Movshon J.A. The analysis of visual motion: A comparison of neuronal and psychophysical performance. The Journal of Neuroscience : The Official Journal of the Society for Neuroscience. 1992;12(12):4745–4765. doi: 10.1523/JNEUROSCI.12-12-04745.1992. [DOI] [PMC free article] [PubMed] [Google Scholar]
  13. Bronfman Z.Z., Brezis N., Moran R., Tsetsos K., Donner T., Usher M., Usher M. Decisions reduce sensitivity to subsequent information. Proceedings of the Biological Sciences. 2015;282(1810):1–45. doi: 10.1098/rspb.2015.0228. [DOI] [PMC free article] [PubMed] [Google Scholar]
  14. Burke C.J., Tobler P.N., Baddeley M., Schultz W. Neural mechanisms of observational learning. Proceedings of the National Academy of Sciences. 2010;107(32):14431–14436. doi: 10.1073/pnas.1003111107. [DOI] [PMC free article] [PubMed] [Google Scholar]
  15. Cairns R., Buchanan A., David A.S., Hayward P., Richardson G., Szmukler G. Prevalence and predictors of mental incapacity in psychiatric in-patients. British Journal of Psychiatry. 2005;187(4):379–385. doi: 10.1192/bjp.187.4.379. [DOI] [PubMed] [Google Scholar]
  16. Campbell-Meiklejohn D., Bach D.R., Roepstorff A., Dolan R.J., Frith C.D. Report how the opinion of others affects our valuation of objects. Current Biology. 2010;20(13):1165–1170. doi: 10.1016/j.cub.2010.04.055. [DOI] [PMC free article] [PubMed] [Google Scholar]
  17. Campbell-Meiklejohn D., Simonsen A., Frith C.D., Daw N.D. Independent neural computation of value from other people's confidence. The Journal of Neuroscience. 2017;37(3):673–684. doi: 10.1523/JNEUROSCI.4490-15.2017. [DOI] [PMC free article] [PubMed] [Google Scholar]
  18. Carpenter J., Sherman M.T., Kievit R.A., Seth A.K., Lau H., Fleming S.M. Domain-general enhancements of metacognitive ability through adaptive training. Journal of Experimental Psychology: General. 2019;148(1):51–64. doi: 10.1037/xge0000505. [DOI] [PMC free article] [PubMed] [Google Scholar]
  19. Carruthers P. How we know our own minds: The relationship between mindreading and metacognition. Behavioral and Brain Sciences. 2009;32(2):121–138. doi: 10.1017/S0140525X09000545. [DOI] [PubMed] [Google Scholar]
  20. Case P. Negotiating the domain of mental capacity: Clinical judgment or judicial diagnosis? Medical Law International. 2016;16(3-4):174–205. doi: 10.1177/0968533216674047. [DOI] [Google Scholar]
  21. Chib V.S., Rangel A., Shimojo S., Doherty J.P.O. Evidence for a common representation of decision values for dissimilar goods in human ventromedial prefrontal cortex. The Journal of Neuroscience. 2009;29(39):12315–12320. doi: 10.1523/JNEUROSCI.2575-09.2009. [DOI] [PMC free article] [PubMed] [Google Scholar]
  22. Christoff K., Gabrieli J.D.E. The frontopolar cortex and human cognition: Evidence for a rostrocaudal hierarchical organization within the human prefrontal cortex. Psychobiology. 2000;28(2):168–186. doi: 10.3758/BF03331976. [DOI] [Google Scholar]
  23. Cialdini R.B., Goldstein N.J. Social influence: Compliance and conformity. Annual Review of Psychology. 2004;55:591–621. doi: 10.1146/annurev.psych.55.090902.142015. [DOI] [PubMed] [Google Scholar]
  24. Constant A., Ramstead M.J.D., Veissière S.P.L., Friston K. Regimes of expectations: An active inference model of social conformity and human decision making. Frontiers in Psychology. 2019;10:679. doi: 10.3389/fpsyg.2019.00679. [DOI] [PMC free article] [PubMed] [Google Scholar]
  25. Crockett M.J. Models of morality. Trends in Cognitive Sciences. 2013;17(8):363–366. doi: 10.1016/j.tics.2013.06.005. [DOI] [PMC free article] [PubMed] [Google Scholar]
  26. David A.S. Insight and psychosis. British Journal of Psychiatry. 1990;156:798–808. doi: 10.1192/bjp.156.6.798. [DOI] [PubMed] [Google Scholar]
  27. David A.S., Bedford N., Wiffen B., Gilleen J. Failures of metacognition and lack of insight in neuropsychiatric disorders. Philosophical Transactions of the Royal Society B. 2012;267(1594):1379–1390. doi: 10.1098/rstb.2012.0002. [DOI] [PMC free article] [PubMed] [Google Scholar]
  28. De Martino B., Bobadilla-Suarez S., Nouguchi T., Sharot T., Love B.C. Social information is integrated into value and confidence judgments according to its reliability. The Journal of Neuroscience. 2017;37(25):6066–6074. doi: 10.1523/jneurosci.3880-16.2017. [DOI] [PMC free article] [PubMed] [Google Scholar]
  29. De Martino B., Fleming S.M., Garrett N., Dolan R.J. Confidence in value-based choice. Nature Neuroscience. 2013;16(1):105–110. doi: 10.1038/nn.3279.Confidence. [DOI] [PMC free article] [PubMed] [Google Scholar]
  30. Dehaene S., Posner M.I., Tucker D.M. Localization of a neural system for error detection and compensation. Psychological Science. 1994;5(5):303–305. [Google Scholar]
  31. Deutsch M., Gerard H.B. A study of normative and informational social influences upon individual judgment. Journal of Abnormal and Social Psychology. 1955;51(3):629–636. doi: 10.1037/h0046408. [DOI] [PubMed] [Google Scholar]
  32. Edelson M., Sharot T., Dolan R.J., Dudai Y. Following the crowd: Brain substrates of long-term memory conformity. Science. 2011;333(6038):108–111. doi: 10.1126/science.1203557. [DOI] [PMC free article] [PubMed] [Google Scholar]
  33. Fischer J.M., Ravizza M. Harcourt Brace Jovanovich College Publishers; Forth Worth, T. X: 1992. Responsibility, freedom, and reason. Ethics, problems and principles. [Google Scholar]
  34. Fleming S.M., Daw N.D. Self-evaluation of decision-making: A general Bayesian framework for metacognitive computation. 2017;124(1):91–114. doi: 10.1037/rev0000045. [DOI] [PMC free article] [PubMed] [Google Scholar]
  35. Fleming S.M., Dolan R.J. The neural basis of metacognitive ability. Philosophical Transactions of the Royal Society of London. Series B, Biological Sciences. 2012;367(1594):1338–1349. doi: 10.1098/rstb.2011.0417. [DOI] [PMC free article] [PubMed] [Google Scholar]
  36. Fleming S.M., Huijgen J., Dolan R.J. Prefrontal contributions to metacognition in perceptual decision making. Journal of Neuroscience. 2012;32(18):6117–6125. doi: 10.1523/JNEUROSCI.6489-11.2012. [DOI] [PMC free article] [PubMed] [Google Scholar]
  37. Fleming S.M., Lau H.C. How to measure metacognition. Frontiers in Human Neuroscience. 2014;8:1–9. doi: 10.3389/fnhum.2014.00443. [DOI] [PMC free article] [PubMed] [Google Scholar]
  38. Fleming S.M., van der Putten E.J., Daw N.D. Neural mediators of changes of mind about perceptual decisions. Nature Neuroscience. 2018;21(4):617–624. doi: 10.1038/s41593-018-0104-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
  39. Fleming S.M., Weil R.S., Nagy Z., Dolan R.J., Rees G. Relating introspective accuracy to individual differences in brain structure. Science. 2010;329(5998):1541–1543. doi: 10.1126/science.1191883. [DOI] [PMC free article] [PubMed] [Google Scholar]
  40. Folke T., Jacobsen C., Fleming S.M., De Martino B. Explicit representation of confidence informs future value-based decisions. Nature Human Behaviour. 2016;1:17–19. [Google Scholar]
  41. Foot P. Virtues and vices. Oxford review. Blackwell; 1967. The problem of abortion and the doctrine of the double effect. [Google Scholar]
  42. Frankfurt H.G. Freedom of the will and the concept of a person. The Journal of Philosophy. 1971 doi: 10.2307/2024717. [DOI] [Google Scholar]
  43. Freedman J.L., Fraser S.C. Compliance without pressure: The foot-in-the-door technique. Journal of Personality and Social Psychology. 1966;68(1):5–20. doi: 10.1037/h0023552. [DOI] [PubMed] [Google Scholar]
  44. Frith C.D. The role of metacognition in human social interactions. Philosophical Transactions of the Royal Society of London. Series B, Biological Sciences. 2012;367(1599):2213–2223. doi: 10.1098/rstb.2012.0123. [DOI] [PMC free article] [PubMed] [Google Scholar]
  45. Frith C.D., Frith U. Mechanisms of social cognition. Annual Review of Psychology. 2012;63:287–313. doi: 10.1146/annurev-psych-120710-100449. [DOI] [PubMed] [Google Scholar]
  46. Frith U., Frith C.D. Development and neurophysiology of mentalising. Philosophical Transactions of the Royal Society, B: Biological Sciences. 2003;358(1431):459–473. doi: 10.1098/rstb.2002.1218. [DOI] [PMC free article] [PubMed] [Google Scholar]
  47. Frith U., Happé F. Autism: Beyond “theory of mind”. Cognition. 1994;50(1–3):115–132. doi: 10.1016/0010-0277(94)90024-8. [DOI] [PubMed] [Google Scholar]
  48. Glimcher P.W. Understanding dopamine and reinforcement learning: The dopamine reward prediction error hypothesis. Proceedings of the National Academy of Sciences. 2011;108(3):15647–15654. doi: 10.1073/pnas.1014269108. [DOI] [PMC free article] [PubMed] [Google Scholar]
  49. Glimcher P.W., Fehr E. 2nd ed. 2013. Neuroeconomics: Decision making and the brain. [DOI] [Google Scholar]
  50. Gold J.I., Shadlen M.N. The neural basis of decision making. Annual Review of Neuroscience. 2007;30(1):535–574. doi: 10.1146/annurev.neuro.29.051605.113038. [DOI] [PubMed] [Google Scholar]
  51. Gold N., Pulford B.D., Colman A.M. Do as I say, Don't do as I do: Differences in moral judgments do not translate into differences in decisions in real-life trolley problems. Journal of Economic Psychology. 2015;47:50–61. doi: 10.1016/j.joep.2015.01.001. [DOI] [Google Scholar]
  52. Gomez-Beldarrain M., Harries C., Garcia-Monco J.C., Ballus E., Grafman J. Patients with right frontal lesions are unable to assess and use advice to make predictive judgments. Journal of Cognitive Neuroscience. 2004;16(1):74–89. doi: 10.1162/089892904322755575. [DOI] [PubMed] [Google Scholar]
  53. Gopnik A., Wellman H.M. Mapping the mind. 2010. The theory theory. [DOI] [Google Scholar]
  54. Greene J.D., Nystrom L.E., Engell A.D., Darley J.M., Cohen J.D. The neural bases of cognitive conflict and control in moral judgment. Neuron. 2004;44:389–400. doi: 10.1016/j.neuron.2004.09.027. [DOI] [PubMed] [Google Scholar]
  55. Greene J.D., Sommerville R.B., Nystrom L.E., Darley J.M., Cohen J.D. An fMRI investigation of emotional engagement in moral judgment. Science. 2001;239:2105–2108. doi: 10.1126/science.1062872. [DOI] [PubMed] [Google Scholar]
  56. Hanks T.D., Summerfield C. Perceptual decision making in rodents, monkeys, and humans. Neuron. 2017;93(1):15–31. doi: 10.1016/j.neuron.2016.12.003. [DOI] [PubMed] [Google Scholar]
  57. Harvey N., Fischer I. Taking advice: Accepting help, improving judgment, and sharing responsibility. Organizational Behavior and Human Decision Processes. 1997;70(2):117–133. [Google Scholar]
  58. Heekeren H.R., Marrett S., Bandettini P.A., Ungerleider L.G. A general mechanism for perceptual decision-making in the human brain. Nature. 2004;431(7010):859–862. doi: 10.1038/nature02966. [DOI] [PubMed] [Google Scholar]
  59. Hertz U., Palminteri S., Brunetti S., Olesen C., Frith C.D., Bahrami B. Neural computations underpinning the strategic management of influence in advice giving. Nature Communications. 2017;8 doi: 10.1038/s41467-017-02314-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
  60. Heyes C.M., Frith C.D. The cultural evolution of mind reading. Science. 2014;344(6190) doi: 10.1126/science.1243091. [DOI] [PubMed] [Google Scholar]
  61. Hilgenstock R., Weiss T., Witte O.W. You'd better think twice: Post-decision perceptual confidence. NeuroImage. 2014;99:323–331. doi: 10.1016/j.neuroimage.2014.05.049. [DOI] [PubMed] [Google Scholar]
  62. Holt D.J., Cassidy B.S., Andrews-Hanna J.R., Lee S.M., Coombs G., Goff D.C.…Moran J.M. An anterior-to-posterior shift in midline cortical activity in schizophrenia during self-reflection. Biological Psychiatry. 2011;69(5):415–423. doi: 10.1016/j.biopsych.2010.10.003. [DOI] [PMC free article] [PubMed] [Google Scholar]
  63. Hutcherson C.A., Bushong B., Rangel A. A Neurocomputational model of altruistic choice and its implications. Neuron. 2015;87(2):451–463. doi: 10.1016/j.neuron.2015.06.031. [DOI] [PMC free article] [PubMed] [Google Scholar]
  64. Izuma K., Adolphs R. Social manipulation of preference in the human brain. Neuron. 2013;78(3):563–573. doi: 10.1016/j.neuron.2013.03.023. [DOI] [PMC free article] [PubMed] [Google Scholar]
  65. Kennedy M., Dornan J., Rutledge E., O'Neill H., Kennedy H.G. Extra information about treatment is too much for the patient with psychosis. International Journal of Law and Psychiatry. 2009;32(6):369–376. doi: 10.1016/j.ijlp.2009.09.006. [DOI] [PubMed] [Google Scholar]
  66. Kiani R., Shadlen M.N. Representation of confidence associated with a decision by neurons in the parietal cortex. Science. 2009;324(5928):759–764. doi: 10.1126/science.1169405. [DOI] [PMC free article] [PubMed] [Google Scholar]
  67. Kim J., Shadlen M.N. Neural correlates of a decision in the dorsolateral prefrontal cortex of the macaque. Nature Neuroscience. 1999;2(2):176–185. doi: 10.1038/5739. [DOI] [PubMed] [Google Scholar]
  68. Koren D., Poyurovsky M., Seidman L.J., Goldsmith M., Wenger S., Klein E.M. The neuropsychological basis of competence to consent in first-episode schizophrenia: A pilot metacognitive study. Biological Psychiatry. 2005;57(6):609–616. doi: 10.1016/j.biopsych.2004.11.029. [DOI] [PubMed] [Google Scholar]
  69. Koren D., Seidman L.J., Poyurovsky M., Goldsmith M., Viksman P., Zichel S., Klein E. The neuropsychological basis of insight in first-episode schizophrenia: A pilot metacognitive study. Schizophrenia Research. 2004;70:195–202. doi: 10.1016/j.schres.2004.02.004. [DOI] [PubMed] [Google Scholar]
  70. Krajbich I., Armel C., Rangel A. Visual fixations and the computation and comparison of value in simple choice. Nature Neuroscience. 2010;13:1292–1298. doi: 10.1038/nn.2635. [DOI] [PubMed] [Google Scholar]
  71. Krajbich I., Hare T., Bartling B., Morishima Y., Fehr E. A common mechanism underlying food choice and social decisions. PLoS Computational Biology. 2015;11(10):1–24. doi: 10.1371/journal.pcbi.1004371. [DOI] [PMC free article] [PubMed] [Google Scholar]
  72. Luce D.R. Their role in Inferring Elementary Mental Organization. 1992. Response Times. (Oxford Psychology Series). [DOI] [Google Scholar]
  73. Manski C.F. The structure of random utility models. Theory and Decision. 1977 doi: 10.1007/BF00133443. [DOI] [Google Scholar]
  74. Mazaniceux A., Fleming S.M., Souchay C., Moulin C.J.A. 2018. Retrospective confidence judgments across tasks: Domain-general processes underlying metacognitive accuracy. [DOI] [Google Scholar]
  75. McCurdy, L. Y., Maniscalco, B., Metcalfe, J., Liu, K. Y., de Lange, F. P., & Lau, H. (2013). Anatomical coupling between distinct metacognitive systems for memory and visual perception. The Journal of Neuroscience: The Official Journal of the Society for Neuroscience, 33(5), 1897–1906. doi: 10.1523/JNEUROSCI.1890-12.2013. [DOI] [PMC free article] [PubMed]
  76. Milgram S. Behavioral study of obedience. Journal of Abnormal and Social Psychology. 1963;67:371–378. doi: 10.1037/h0040525. [DOI] [PubMed] [Google Scholar]
  77. Modinos G., Renken R., Ormel J., Aleman A. Self-reflection and the psychosis-prone brain: An fMRI study. Neuropsychology. 2011;25(3):295–305. doi: 10.1037/a0021747. [DOI] [PubMed] [Google Scholar]
  78. Morales J., Lau H., Fleming S.M. Domain-general and domain-specific patterns of activity supporting metacognition in human prefrontal cortex. The Journal of Neuroscience. 2018;38(14):3534–3546. doi: 10.1523/JNEUROSCI.2360-17.2018. [DOI] [PMC free article] [PubMed] [Google Scholar]
  79. Murphy P.R., Robertson I.H., Harty S., O'Connell R.G. Neural evidence accumulation persists after choice to inform metacognitive judgments. ELife. 2015 doi: 10.7554/elife.11946. [DOI] [PMC free article] [PubMed] [Google Scholar]
  80. O'Connell R.G., Dockree P.M., Kelly S.P. A supramodal accumulation-to-bound signal that determines perceptual decisions in humans. Nature Neuroscience. 2012;15(12):1729–1735. doi: 10.1038/nn.3248. [DOI] [PubMed] [Google Scholar]
  81. Olsen K., Roepstorff A., Bang D. Knowing whom to learn from: individual differences in metacognition and weighting of social information. PsyArXiv Preprints. 2019 doi: 10.31234/osf.io/jqheu. [DOI] [Google Scholar]
  82. Owen G.S., Freyenhagen F., Richardson G., Hotopf M. Mental capacity and decisional autonomy: An interdisciplinary challenge. Inquiry. 2009;52(1):79–107. doi: 10.1080/00201740802661502. [DOI] [Google Scholar]
  83. Owen G.S., Richardson G., David A.S., Szmukler G., Hayward P., Hotopf M. Mental capacity, diagnosis, and insight in psychiatric inpatients: A cross sectional study. Psychological Medicine. 2009;39(8):1389–1398. doi: 10.1017/S0033291708004637. [DOI] [PMC free article] [PubMed] [Google Scholar]
  84. Owen G.S., Freyenhagen F., Hotopf M., Martin W. Temporal inabilities and decision-making capacity in depression. Phenom Cogn Sci. 2015;14:163–182. doi: 10.1007/s11097-013-9327-x. [DOI] [Google Scholar]
  85. Peirce C.S., Jastrow J. On small differences in sensation. Memoirs of the National Academy of Science. 1884;3:75–83. doi: 10.5860/choice.35sup-462. [DOI] [Google Scholar]
  86. Pescetelli N., Yeung N. The role of decision confidence in advice-taking and trust formation. PsyArXiv. 2019 doi: 10.1037/xge0000960. [DOI] [PubMed] [Google Scholar]
  87. Petrova M., Dale J., Bill F. Values-based practice in primary care. British Journal of General Practice. 2006;56(530):703–709. [PMC free article] [PubMed] [Google Scholar]
  88. Platt M.L., Glimcher P.W. Neural correlates of decision variables in parietal cortex. Nature. 1999;400(6741):233–238. doi: 10.1038/22268. [DOI] [PubMed] [Google Scholar]
  89. Pousa E., Duñó R., Brebion G., David A.S., Ruiz A.I., Obiols J.E. Theory of mind deficits in chronic schizophrenia: Evidence for state dependence. Psychiatry Research. 2008;158(1):1–10. doi: 10.1016/j.psychres.2006.05.018. [DOI] [PubMed] [Google Scholar]
  90. Rabbitt P.M. Errors and error correction in choice-response tasks. Journal of Experimental Psychology. 1966;71(2):264–272. doi: 10.1037/h0022853. [DOI] [PubMed] [Google Scholar]
  91. Rangel A., Camerer C., Montague P.R. Neuroeconomics: The neurobiology of value-based decision-making. Nature Reviews. Neuroscience. 2008;9(7):545–556. doi: 10.1038/nrn2357.Neuroeconomics. [DOI] [PMC free article] [PubMed] [Google Scholar]
  92. Ratcliff R. A theory of memory retrieval. Psychological Review. 1978 doi: 10.1037/0033-295X.85.2.59. [DOI] [PubMed] [Google Scholar]
  93. Resulaj A., Kiani R., Wolpert D.M., Shadlen M.N. Changes of mind in decision-making. Nature. 2009;461(7261):263–266. doi: 10.1038/nature08275. [DOI] [PMC free article] [PubMed] [Google Scholar]
  94. Ridderinkhof K.R., van den Wildenberg W.P., Segalowitz S.J., Carter C.S. Neurocognitive mechanisms of cognitive control: The role of prefrontal cortex in action selection, response inhibition, performance monitoring, and reward-based learning. Brain and Cognition. 2004;56(2):129–140. doi: 10.1016/j.bandc.2004.09.016. [DOI] [PubMed] [Google Scholar]
  95. Roitman J.D., Shadlen M.N. Response of neurons in the lateral intraparietal area during a combined visual discrimination reaction time task. The Journal of Neuroscience. 2002;22(21):9475–9489. doi: 10.1523/JNEUROSCI.22-21-09475.2002. [DOI] [PMC free article] [PubMed] [Google Scholar]
  96. Rollwage M., Dolan R.J., Fleming S.M. Metacognitive failure as a feature of those holding radical beliefs. Current Biology. 2018;28(24):4014–4021.e8. doi: 10.1016/j.cub.2018.10.053. [DOI] [PMC free article] [PubMed] [Google Scholar]
  97. Rouault M., Mcwilliams A., Allen M.G., Fleming S.M. Human metacognition across domains : Insights from individual differences and neuroimaging. Personality Neuroscience. 2018;1(E17):1–35. doi: 10.1017/pen.2018.16. [DOI] [PMC free article] [PubMed] [Google Scholar]
  98. Rouault M., Seow T., Gillan C.M., Fleming S.M. Psychiatric symptom dimentions are associated with dissociable shifts in metacognition but not task performance. Biological Psychiatry. 2018;84(6):443–451. doi: 10.1016/j.biopsych.2017.12.017. [DOI] [PMC free article] [PubMed] [Google Scholar]
  99. Ruck Keene A., Kane N.B., Kim S.Y.H., Owen G.S. Taking capacity seriously? Ten years of mental capacity disputes before England’s Court of Protection. International Journal of Law and Psychiatry. 2019;62:56–72. doi: 10.1016/j.ijlp.2018.11.005. [DOI] [PMC free article] [PubMed] [Google Scholar]
  100. Sanfey A.G., Rilling J.K., Aronson J.A., Nystrom L.E., Cohen J.D. The neural basis of economic decision making in the ultimatum game. Science. 2013;300:1755–1758. doi: 10.1126/science.1082976. [DOI] [PubMed] [Google Scholar]
  101. Shea N., Boldt A., Bang D., Yeung N., Heyes C., Frith C.D. Supra-personal cognitive control and metacognition. Trends in Cognitive Sciences. 2014;18(4):186–193. doi: 10.1016/j.tics.2014.01.006. [DOI] [PMC free article] [PubMed] [Google Scholar]
  102. Shenhav A., Greene J.D. Integrative moral judgment: Dissociating the roles of the amygdala and ventromedial prefrontal cortex. The Journal of Neuroscience. 2014;34(13):4741–4749. doi: 10.1523/JNEUROSCI.3390-13.2014. [DOI] [PMC free article] [PubMed] [Google Scholar]
  103. Shimamura A.P., Squire L.R. Memory and metamemory: A study of the feeling-of-knowing phenomenon in amnesic patients. Journal of Experimental Psychology. Learning, Memory, and Cognition. 1986;12(3):452–460. doi: 10.1037/0278-7393.12.3.452. [DOI] [PubMed] [Google Scholar]
  104. Smith P.L., Ratcliff R. Psychology and neurobiology of simple decisions. Trends in Neurosciences. 2004;27(3):161–168. doi: 10.1016/j.tins.2004.01.006. [DOI] [PubMed] [Google Scholar]
  105. Sniezek J.A., Van Swol L.M. Trust, confidence, and expertise in a judge – Advisor system. Organizational Behavior and Human Decision Processes. 2001;84(2):288–307. doi: 10.1006/obhd.2000.2926. [DOI] [PubMed] [Google Scholar]
  106. Song C., Kanai R., Fleming S.M., Weil R.S., Schwarzkopf D.S., Rees G. Relating inter-individual differences in metacognitive performance on different perceptual tasks. Consciousness and Cognition. 2011;20(4):1782–1792. doi: 10.1016/j.concog.2010.12.011. [DOI] [PMC free article] [PubMed] [Google Scholar]
  107. Sowden S., Koletsi S., Lymberopoulos E., Militaru E., Catmur C., Bird G. Quantifying compliance and acceptance through public and private social conformity. Consciousness and Cognition. 2018;65:359–367. doi: 10.1016/j.concog.2018.08.009. [DOI] [PMC free article] [PubMed] [Google Scholar]
  108. Stephan K.E., Manjaly Z.M., Mathys C.D., Weber L.A.E., Paliwal S., Gard T.…Petzschner F.H. Allostatic self-efficacy: A metacognitive theory of Dyshomeostasis-induced fatigue and depression. Frontiers in Human Neuroscience. 2016;10 doi: 10.3389/fnhum.2016.00550. [DOI] [PMC free article] [PubMed] [Google Scholar]
  109. Talluri B.C., Urai A.E., Tsetsos K., Usher M., Donner T.H. Confirmation bias through selective overweighting of choice-consistent evidence. Current Biology. 2018;28:3128–3135. doi: 10.1016/j.cub.2018.07.052. [DOI] [PubMed] [Google Scholar]
  110. Thomson J.J. The trolley problem. The Yale Law Journal. 1985;94:1395–1415. doi: 10.1119/1.1976413. [DOI] [Google Scholar]
  111. Toelch U., Dolan R.J. Informational and normative in fl uences in conformity from a neurocomputational perspective. Trends in Cognitive Sciences. 2015;19(10):579–589. doi: 10.1016/j.tics.2015.07.007. [DOI] [PubMed] [Google Scholar]
  112. Valentin V.V., Dickinson A., Doherty J.P.O. Determining the neural substrates of goal-directed learning in the human brain. The Journal of Neuroscience. 2007;27(15):4019–4026. doi: 10.1523/JNEUROSCI.0564-07.2007. [DOI] [PMC free article] [PubMed] [Google Scholar]
  113. Vickers D. Academic Press; 1979. Decision processes in visual perception. [DOI] [Google Scholar]
  114. Wald A. Review: Theory of games and economic behavior. The Review of Economic Statistics. 1947 doi: 10.2307/1925651. [DOI] [Google Scholar]
  115. Wimmer H., Perner J. Beliefs about beliefs: Representation and constraining function of wrong beliefs in young children's understanding of deception. Cognition. 1983;13:103–128. doi: 10.1016/0010-0277(83)90004-5. [DOI] [PubMed] [Google Scholar]
  116. Zürcher T., Elger B., Trachsel M. The notion of free will and its ethical relevance for decision-making capacity. BMC Medical Ethics. 2019;20(31):3–10. doi: 10.1186/s12910-019-0371-0. [DOI] [PMC free article] [PubMed] [Google Scholar]

RESOURCES