Skip to main content
Biological Psychiatry Global Open Science logoLink to Biological Psychiatry Global Open Science
. 2022 Mar 8;3(2):179–186. doi: 10.1016/j.bpsgos.2022.02.010

Toward a Unifying Account of Dopamine’s Role in Cost-Benefit Decision Making

Alexander Soutschek a,, Alexander Jetter b, Philippe N Tobler c,d
PMCID: PMC10140448  PMID: 37124350

Abstract

Dopamine is thought to play a crucial role in cost-benefit decision making, but so far there is no consensus on the precise role of dopamine in decision making. Here, we review the literature on dopaminergic manipulations of cost-benefit decision making in humans and evaluate how well different theoretical accounts explain the existing body of evidence. Reduced D2 stimulation tends to increase the willingness to bear delay and risk costs (i.e., wait for later rewards, take riskier options), while increased D1 and D2 receptor stimulation increases willingness to bear effort costs. We argue that the empirical findings can best be explained by combining the strengths of two theoretical accounts: in cost-benefit decision making, dopamine may play a dual role both in promoting the pursuit of psychologically close options (e.g., sooner and safer rewards) and in computing which costs are acceptable for a reward at stake. Moreover, we identify several limiting factors in the study designs of previous investigations that prevented a fuller understanding of dopamine’s role in value-based choice. Together, the proposed theoretical framework and the methodological suggestions for future studies may bring us closer to a unifying account of dopamine in healthy and impaired cost-benefit decision making.

Keywords: Delay discounting, Dopamine, Effort discounting, Psychopharmacology, Reward, Risky decision making


In everyday life, we often face tradeoffs between desired goods (e.g., eating a chocolate cake, consuming drugs) and the nondesired costs one has to bear to obtain those goods (gaining weight, long-term consequences of drug abuse). These costs come in different forms such as delay of reward delivery, risk or ambiguity, mental or physical effort, or social costs such as inequity (1). Deficits in cost-benefit decision making belong to the core symptoms of several psychiatric disorders, including depression, schizophrenia, addiction, eating disorders, gambling disorders, or Parkinson’s disease (2, 3, 4, 5, 6). Substance addiction, for example, can be understood as maladaptive preference for immediate reinforcement (acute drug effects) under ignorance of negative long-term consequences (2). Insights into the neurochemical basis of cost-benefit decisions may therefore improve our understanding of the neural origins of decision-making deficits in psychiatric disorders as well as the effectiveness of pharmacological treatments.

A large body of evidence assigns a crucial role to the neurotransmitter dopamine in weighing costs against benefits [for a recent review, see (7)]. However, there is no agreement on the theoretical interpretation of these empirical findings, because several accounts with different hypothesized roles for dopamine in cost-benefit decision making have been postulated. Often, the predictions of different accounts are even in apparent conflict with each other: for example, while some accounts claim that dopamine increases the tolerance for delay costs (8,9), others posit that dopamine enhances the preference for immediate rewards (10,11). These conflicting assumptions hamper both the development of pharmacological treatments of decision-making deficits and the prediction of the impact of dopaminergic treatment on decision making. Thus, it is of major importance to develop a coherent account of dopamine’s function in cost-benefit decision making.

This review has three main goals. First, we assess whether prominent accounts of dopaminergic contributions to decision making can explain existing empirical findings. Each of the existing accounts can explain only a subset of the data, and we argue that none of them provides a coherent picture of the role of dopamine in cost-benefit weighing. Second, based on this, we aim to outline a model for dopamine in value-based choice, which combines the strengths of existing accounts. Third, we argue that the development of a coherent theoretical account is hampered by limitations in the design of decision neuroscience studies and make suggestions about which kind of future research can facilitate arbitration between different theoretical models.

Empirical Findings for Dopaminergic Manipulations in Decision Making

We first summarize the empirical literature on the influences of dopaminergic pharmacological manipulations on cost-benefit decision making in healthy volunteers. We separately discuss four different cost types: delay, risk, effort, and social costs. We also consider which dopamine receptor subtypes are targeted by a given pharmacological manipulation because some of the accounts described below (see Relating Empirical Findings to Existing Accounts of Dopamine Functioning in Value-Based Choice) suggest that different receptor subtypes play dissociable roles in cost-benefit decision making. In particular, dopamine receptors can be subdivided into D1-type receptor (D1R) and D2-type receptor (D2R) families. D1Rs are prevalent in the striatal direct “go” path, which is thought to energize behavior toward goals (12). D2Rs, in contrast, are expressed predominantly in the indirect “no go” path, which is involved in action inhibition. Because D2R activation suppresses activation in the inhibitory indirect path, stronger D2R signaling also energizes behavior by suppressing the inhibitory influence of the indirect path on cortical activity (13,14).

We first consider tradeoffs between reward magnitudes and temporal delays to reward delivery, as studied in intertemporal choices where agents typically have to choose between a smaller-sooner (e.g., 30 Swiss francs today) and a larger-later reward (e.g., 100 Swiss francs in 100 days) (Table S1). The majority of pharmacological studies used D2 antagonists or agents increasing both D1R and D2R activity (such as levodopa or d-amphetamine). Of the five studies employing selective D2 antagonists, four observed a stronger preference for delayed rewards after D2R blockade (15, 16, 17, 18), while one study observed no significant effects (10). Similarly, an unspecific decrease in dopamine transmission through acute phenylalanine and tyrosine depletion had no effect on intertemporal choice (19). Overall (and under the assumption that the used antagonists act post- rather than presynaptically), the data suggest that reduction of D2R activity reduces delay costs.

Compounds unspecifically increasing D1R and D2R activity showed mixed effects, i.e., either no impact on intertemporal decision making (20,21); a stronger preference for delayed rewards, at least at higher doses (8,9); or a reduced preference for delayed rewards (10). In addition, selective D1 or D2 agonists revealed no significant effects (22,23). The lack of consistent significant effects of dopamine agonists might be explained by inverted u–shaped response curves where dopamine agonists, particularly at higher doses, increase dopaminergic activity beyond the optimal performance level for some but not other individuals. Thus, the only relatively robust finding in the domain of waiting costs is that selective reduction of D2R neurotransmission decreases the preference for smaller-sooner rewards.

The literature on dopamine in risky decision making paints a picture that is somewhat harder to interpret (Table S2). An example for risky choices are lottery choices where agents decide between two lotteries with varying probabilities of winning or losing rewards. Dopamine agents stimulating both D1R and D2R show either no significant effects (8,20,24,25) or a higher risk tolerance, particularly in the gain domain (26, 27, 28, 29, 30). In contrast, one study with the combined (indirect) D1/D2 agonist levodopa observed a reduced preference for risky outcomes, but only in individuals with high baseline impulsivity (21). Moreover, high doses of a selective D1 agonist on average reduced preference for riskier but larger outcomes (22). D2 agonists showed no significant effects (23), increased propensity for risky choices (31), or increased sensitivity for gains combined with a lower sensitivity for losses in individuals with low reward sensitivity (32). D2R blockade showed either no significant effects (18,33) or reduced sensitivity to costs (34, 35, 36). Thus, if anything, reduced D2R neurotransmission increases the preference for high reward–high costs options, similar to the findings for delay costs.

Next, we consider tradeoffs between rewards and the costs of (mental or physical) effort (Table S3). For this cost type, pharmacological studies paint a relatively coherent picture. Dopamine agonists increase the willingness to exert effort for rewards (22,37, 38, 39, 40), with only one study showing no significant effects (41). Conversely, most dopamine antagonist studies report reduced willingness to exert effort (42, 43, 44). We note that two studies using a low dose of the D2 antagonists haloperidol and sulpiride (37,40) observed reduced effort discounting, consistent with the view that low doses of D2 antagonists increase dopaminergic activity via presynaptic effects. One study directly comparing the effects of methylphenidate and sulpiride observed that methylphenidate (stimulating both D1Rs and D2Rs) increased the sensitivity to rewards, whereas sulpiride (as selective D2 antagonist) attenuated the impact of costs on decision making (37). This is consistent with dissociable roles of D1Rs and D2Rs for benefit and cost processing, respectively. Taken together, pharmacological evidence on effort discounting suggests that dopamine enhances the motivation to work for rewards, with potentially dissociable roles for D1Rs and D2Rs.

Finally, we consider cost-benefit tradeoffs in social decision making (Table S4). Intuitively, one might assume that the cost of sharing goods with others opposes the goal of maximizing one’s selfish payoff, but many individuals perceive sharing as rewarding [“warm glow” effect (45,46)], and this effect seems to be stronger in women than in men (47). Correspondingly, reduced D2R neurotransmission shows gender-specific effects, with lower D2R activity reducing costly sharing in women and increasing it in men (16), particularly for close others. The same effect was observed for female participants under a D2 agonist (48), which may be reconciled with other findings (16) by assuming an inverted u–shaped dose-response curve [although one could alternatively assume that the D2 antagonist in (16) had stronger presynaptic than postsynaptic effects]. Levodopa increased selfishness in mixed samples of male and female participants (49) and in a male-only sample (50), the latter being consistent with our findings reported in (16). Methylphenidate showed no significant effect on prosocial giving in a sample of mixed female and male participants. Independently of gender, the D1/D2 agonist tolcapone increased the perceived costs of unequal outcomes (51), hinting to a role of dopamine in encoding social norms. Taken together, the effects in the social domain are difficult to interpret, particularly because it is hard to define subjective costs and benefits in social decision making (which may vary between individuals). However, under the assumption of the hypothesized gender-specific role of dopamine in social preferences, the majority of studies seem to suggest that dopamine strengthens the preference for higher valued, costly options.

Relating Empirical Findings to Existing Accounts of Dopamine Functioning in Value-Based Choice

In this section, we relate the empirical literature reviewed above to different theoretical accounts of the role of dopamine in cost-benefit decision making and consider how well these accounts explain the empirical data. An eminent account for the role of dopamine signaling in motivation is that dopamine energizes behavior toward goals (energization account; Figure 1A) (52,53). This account was based on findings that dopamine increases the willingness to work for rewards in rats (54, 55, 56, 57) and that (striatal) dopamine depletion abolishes speeding of reward-directed responses (58). It is also compatible with reports of increased dopamine release in the ventral striatum and ventromedial prefrontal cortex when rats respond more quickly because of higher reward rates (59). With regard to cost-benefit decision making, this account predicts that enhancing dopaminergic activity increases the preference for highly valued options despite the associated costs, while lower dopaminergic activity should decrease it. We note that different variants of this account have been formulated, some (60, 61, 62) also reconciling the motivational functions of dopamine with its involvement in reward learning (62,63) or ascribing dopamine a valence-independent role for approach behavior (64). Still, all the variants of this account link dopamine to choices of high reward–high cost options.

Figure 1.

Figure 1

Prominent theories of dopamine function in value-based choice. (A) The energization account assumes that increased dopamine activation enhances the preference for high benefit–high cost options over low benefit–low cost options (illustrated by a larger font size for the high-cost compared with the low-cost option). (B) Building upon this notion, the accelerator/brake model posits that D2 receptor (D2R) activation computes the costs that are acceptable for a reward, which determines how strong a D1 receptor (D1R)–mediated facilitatory signal needs to be to overcome D2R-mediated inhibition. Thus, while both D1R and D2R activation strengthen the preference for high benefit–high cost over low benefit–low cost options, they do so via dissociable effects on processing of benefits vs. action costs. D1R activation strengthens the processing of rewards (larger font size), whereas D2R activation reduces the sensitivity to costs (smaller font size). (C) According to the proximity account, the impact of dopamine on behavior depends on the proximity advantage of the more proximate (in many cases, also the low cost) option over the more distant (often high cost) option. If one option has a proximity advantage over the other, dopamine strengthens the preference for the proximate over the distant reward, as illustrated here. In contrast, if no option possesses a proximity advantage, dopamine favors the high-benefit option in a similar way as assumed by the energization account.

Regarding the reviewed literature, the energization account is consistent with the findings for effort-based choice, where increased dopaminergic activity enhances the willingness to tolerate effort costs. However, the energization account appears to be difficult to reconcile with the findings for D2 antagonists for risk and delay costs, because in these domains D2 antagonism increased, rather than decreased, the preference for high benefit–high cost options.

In contrast to the energization account, the accelerator/brake model (Figure 1B) distinguishes between the roles of different dopamine receptor subtypes and assumes that D1R and D2R neurotransmission enhance goal-directed behavior via separate computational mechanisms (65). While D1Rs in the direct path encode the benefit associated with an option (accelerator), D2Rs implement a cost control (brake) that needs to be overcome to suppress the inhibitory indirect path. If tonic dopamine levels are high (note that D2Rs are more sensitive to changes in tonic dopamine than D1Rs), even relatively small or costly rewards will elicit approach behavior. In contrast, a low tonus reduces the desirability of costly options, and only large above-average rewards will trigger reward seeking. This is also consistent with the view that tonic D2R-mediated dopaminergic activity encodes the background reward rate (66). Thus, the accelerator/brake account predicts D1R- versus D2R-targeting dopaminergic manipulations to have dissociable effects on the processing of benefits and costs in value-based decisions, with D1R activation increasing sensitivity to benefits and D2R activation lowering impact of action costs [similar to the assumptions of the opponent actor learning model (67)].

Evidence for dissociable roles of D1R and D2R is somewhat sparse, given that relatively few studies distinguished between effects on reward versus cost processing. Moreover, many studies analyze choice behavior with economic utility models, for example, hyperbolic delay discounting or prospect theory of risky choice. These models integrate rewards and action costs to quantify the utility of options and usually do not contain separate terms for costs and benefits. Among the studies distinguishing between cost and benefit processing, however, there is indeed some evidence that D1R activation shows stronger effects on reward than on cost processing, whereas manipulation of D2R neurotransmission changes cost sensitivity. As noted above, methylphenidate promotes decisions to engage in mental effort by strengthening the weight assigned to the potential benefits (37). While methylphenidate enhances both D1R and D2R activation, its impact on reward processing seems to be mediated primarily by D1R rather than D2R stimulation (68), suggesting that enhanced D1R activity may underlie the effects of methylphenidate on benefit processing in decision making. Evidence for D1R involvement in preferring costly higher rewards over cost-free lower rewards is also provided by a recent study with a selective D1 agonist (22). Relatedly, in a loss chasing task, the D2 antagonist pramipexole reduced costly attempts to recover losses (36), while methylphenidate increased sensitivity to high rewards (26). In addition, animal findings support the hypothesized role of D2R signaling for cost processing (69,70). These findings are consistent with the idea that D2R activity implements a cost control that determines how strong D1-mediated facilitatory signaling in the direct pathway needs to be to overcome D2-mediated indirect pathway suppression (65).

Finally, the proximity account (Figure 1C) highlights dopamine’s sensitivity to psychological proximity (71). According to this view, dopamine strengthens the preference for psychologically close (e.g., immediately available) over distant rewards. Evidence for proximity effects originally stems from animal findings showing enhanced firing rates of dopaminergic neurons for spatially close rewards (72), but Westbrook and Frank (71) extend the concept from physical proximity to psychological proximity. Although action costs and proximity may appear closely related, they can conceptually be distinguished: subjective costs are learned associations between actions and effort, waiting time, or risk and also vary depending on an agent’s internal state (e.g., available resources to exert effort). Proximity, in contrast, depends on situational factors such as the relative salience, familiarity, or concreteness of an option (71). Whereas in economic decision-making paradigms, costs and proximity are often confounded because less costly options are likely to be more proximate (e.g., outcomes in the near rather than far future are also more concrete), costs and proximity could in principle experimentally be distinguished (e.g., by making the future outcome more concrete; see Supplemental Discussion). The proximity account predicts that high dopaminergic activity strengthens the preference for proximate rewards compared with distant rewards. In contrast, if no option has a proximity advantage, dopamine favors high benefit–high cost actions, as also assumed by the energization and accelerator/brake models.

As tentative computational implementation of the proximity account (71), a proximity advantage shifts the starting point of the evidence accumulation in a drift diffusion model toward the more proximate option. Still, action costs and benefits can affect the drift rate (which captures the actual process of evidence accumulation) despite an initial starting bias, and the strength of these effects of costs and benefits on the accumulation process is mediated by receptor-specific dopaminergic activity. According to the proximity account (71), cost-benefit decisions depend on the interplay between dopaminergic effects on evidence accumulation for high benefit–high cost options and on an initial proximity advantage. Moreover, prefrontal mechanisms may increase thresholds, giving distant reward options more time to compete with options that are favored by a proximity advantage (73,74).

The proximity account is supported by intervention findings in intertemporal and risky choice where D2 antagonism increased the preference for delayed and risky outcomes. In addition, in the domain of social decisions, dopaminergic effects seem to differ between close and distant others (16,48), which may point to a potential proximity effect in prosocial giving where selfish rewards have a smaller proximity advantage over rewards shared with close than with distant others. In the domain of effort-based choice, however, there was no evidence for a proximity effect, and the result pattern rather consistently links dopamine with stronger preferences for high effort/high reward options. Accordingly, one may ask whether proximity effects are domain specific. In other words, why are delayed and risky rewards as well as rewards shared with strangers perceived as psychologically distant, whereas effortful rewards are not? A potential ad hoc explanation for this phenomenon is that in most effort-based decision paradigms, decision makers experience the required effort demands in a familiarization session where they learn that they can successfully deal even with high effort demands, such that all effort options are actually perceived as proximate (and potential confounds of effort aversion, such as risk aversion, are eliminated). Effortful rewards may therefore be perceived as just as psychologically close as effort-free rewards. Another possibility is that different dose-response curves underlie the mediating role of proximity in effort-based compared with intertemporal and risky choice. However, we emphasize that these explanations are speculative and depend on several assumptions (e.g., whether a given compound in a given dose acts primarily pre- or postsynaptically), such that they need to be tested by future studies.

Toward a Unifying Account of Dopamine in Value-Based Choice

Combining the accelerator/brake account with the proximity approach (Figure 2A) may provide a unified account of dopamine’s role in decision making. First, if a choice option possesses a proximity advantage (i.e., temporally close or low-risk rewards) over alternative options, higher tonic dopamine levels favor proximate over distant rewards. From the perspective of process models, the advantage corresponds to an initial bias toward the proximate option before the evidence accumulation process begins (71). Assuming that this bias is D2-mediated explains why D2 antagonists increase the willingness to tolerate risk and delay costs (15, 16, 17, 18,34, 35, 36). In the absence of consistent evidence that D1R-stimulating drugs increase the preference for proximate (i.e., risk-free or immediate) rewards, the reviewed literature (8, 9, 10,20,22,27) does not suggest a role of D1Rs in moderating proximity effects due to lack of consistent evidence that D1R-stimulating drugs increase the preference for proximate (i.e., risk-free or immediate) rewards. We therefore posit that the proximity bias is (preferentially) moderated by D2R rather than D1R activation, but note that additional empirical work is needed to directly test this assumption.

Figure 2.

Figure 2

Toward a unifying account of dopamine in value-based choice. (A) Combining the accelerator/brake with the proximity model (Figure 1), we propose that dopamine affects cost-benefit decision making both by enhancing the proximity advantage of proximate low-cost (e.g., sooner rewards) over distant high-cost options (e.g., later rewards) and by implementing a cost control. If no option possesses a proximity advantage over the other (upper panel), our account makes the same predictions as the accelerator/brake account, i.e., increasing both D1 receptor (D1R) (via enhanced benefit processing) and D2 receptor (D2R) activity (by increasing the acceptable costs) enhances the preference for high-benefit rewards. If the low-cost option is also more proximate than the high-cost option (lower panel; note that costs and proximity can conceptually be distinguished), low dopamine levels (by reducing the influence of the proximity advantage) and increased D1R activity strengthen the preference for more distant–high benefit rewards. In contrast, high D2R activity enhances both the proximity advantage of the more proximate (low cost) option and the acceptable costs. (B) Within the framework of a drift diffusion–style model of the choice process, D2R activation might increase the proximity advantage of low-cost options by shifting the starting point of the evidence accumulation process toward the boundary of the low-cost option (71). During evidence accumulation, D1R activation strengthens the impact of benefits on the velocity of the accumulation process (drift rate), while higher D2R activation lowers the sensitivity to action costs. Thus, if decision thresholds are low, stronger proximity effects under high D2R activation increase the likelihood of choosing the (proximate) low-cost option, whereas in cases of high decision thresholds, D2R effects on evidence accumulation will result in more choices of the high benefit–high cost option.

Second, D2R activation releasing the inhibitory impact of the indirect path may implement a cost control that determines whether a benefit is considered worth its costs (65). If tonic dopamine levels (and thus D2R activation) are low, the indirect path exerts a strong inhibitory influence on the cortex, such that only large rewards will lead to sufficient D1R-mediated direct path facilitation to overcome the D2R-mediated inhibition. In contrast, high tonic dopamine releases the inhibitory impact of the indirect path, such that relatively small rewards also become a worthy pursuit (22,26, 27, 28, 29, 30,37, 38, 39, 40). Computationally, this may be implemented through D1R and D2R activation affecting evidence accumulation rather than the starting bias. Once the evidence accumulation process has started (which trades the benefits against the costs of action alternatives), D1R and D2R mediate evidence accumulation for benefits and costs, respectively, associated with the action alternatives. In this way, dopamine can promote the choice of high benefit–high cost options despite an initial proximity advantage of low-cost rewards, particularly if decision thresholds are high and agents make cautious choices. The assumption that proximity effects are moderated by D2Rs rather than D1Rs makes the straightforward prediction that changes in decision thresholds may reverse the influence of manipulations of D2R activation on observed choices (Figure 2B). This account thus integrates several aspects of the empirical data that cannot be explained by existing accounts in isolation. The impact of D2 antagonists on intertemporal and risky choice cannot be explained by the energization or accelerator/brake accounts, unless one implausibly assumes that all administered D2 antagonists only have presynaptic rather than postsynaptic effects. Moreover, the dissociable effects of D1R and D2R activity on benefit versus cost processing were unspecified in the current formulation of the proximity model, which includes no explicit predictions for separate contributions of receptor types to the choice process (the focus seems to be on the role of tonic dopamine for moderating proximity effect). The proposed model therefore specifies the proximity account by more strongly emphasizing the dissociable roles of D1Rs and D2Rs for the choice process. As a caveat, we note that the focus of our account is primarily on striatal dopamine, but dopamine receptors are also expressed in other brain regions such as the prefrontal cortex, where D1Rs outnumber D2Rs (75). Because prefrontal D1R activation increases goal representations (76), striatal and prefrontal D1Rs may have similar effects on observable behavior. Nevertheless, given that the reviewed pharmacological manipulations are systemic, it remains open whether the drug-induced behavioral changes are mediated via effects on striatal or prefrontal receptor activity.

The proposed account predicts that the height of the decision threshold crucially influences the impact of dopamine on choice behavior: if decision thresholds are low (and agents make impetuous decisions), dopaminergic effects on the proximity advantage dominate, increasing choices of proximate, low-cost rewards. In contrast, under high decision thresholds (cautious decisions), dopaminergic effects on evidence accumulation may win out over initial proximity advantages, resulting in a higher likelihood of high benefit–high cost choices. Such a model of the role of dopamine for value-based choice allows integrating a large body of empirical evidence, which appeared partially inconsistent with previous existing accounts. The hypothesized crucial role of decision thresholds may also deepen our understanding of clinical decision-making deficits, given evidence from computational psychiatry for altered thresholds in some disorders (77,78). Increased thresholds in depression (77) lower the impact of proximity advantages on choices, which may explain why lower dopamine levels in depression lead to stronger delay discounting (79). In contrast, the positive correlation between delay discounting and dopamine levels in ADHD (80) can be explained by a lower decision threshold (81), which leads to a more influential role of dopamine effects on proximity advantages. Our account may thus explain why the link between dopamine and impulsivity varies between clinical and nonclinical populations (80). While details of the model (particularly its computational implementation) may change with growing empirical evidence, its principles represent a fruitful hypothesis that may put the heterogeneous field of dopaminergic studies in human decision making on more solid theoretical and computational grounds. We note that the computational role of dopamine in decision making has already been formalized previously (67,82,83). Therefore, a valuable enterprise for future empirical studies would be to directly test the predictions of these different accounts.

We close by noting that the lack of a unifying account of dopamine in decision making is at least partially caused by limitations in study designs and inherent properties of the biological substrate that hamper theory development. First, it is often unclear whether an effect of a pharmacological challenge reflects increased or decreased dopaminergic activation. For example, in studies using agonists, a drug-induced change in behavior may either reflect increased functioning relative to baseline or—assuming an inverted u–shaped dopamine-response curve (84)—reduced functioning due to increased dopaminergic activation beyond optimal dopamine levels. To make matters worse, evidence suggests that the shape of the dopamine-response curve is domain specific (22,84,85). To clarify the direction of effects in pharmacological intervention studies, we therefore recommend using multiple doses and measuring baseline dopamine levels, either directly (via positron emission tomography) or indirectly [e.g., with working memory performance (86) or trait impulsivity (87); see (21,22,85)]. An additional issue exists for D2 antagonists (such as amisulpride, sulpiride, or haloperidol), which, depending on the administered dose, can either increase or decrease dopaminergic activity through presynaptic or postsynaptic actions, respectively (88). For safety reasons, many studies administer a dose at the lower border for postsynaptic effects, which again hinders a straightforward interpretation of the direction of the effects.

Besides these methodological considerations, a further recommendation relates to statistical analysis where many previous studies did not explicitly distinguish between reward- versus cost-mediated effects. When assessing aggregated mean choice behavior or parameters from economic utility models integrating costs and rewards to a subjective value term, it is not possible to test for potential receptor-specific contributions to reward and cost processing. A way to address this issue is to analyze the effects of pharmacological interventions on decision making on attribute-wise comparisons of reward magnitudes and costs (e.g., the delay of the smaller-sooner and the larger-later option in intertemporal choice). Such attribute-wise comparisons were recently found to explain choice behavior better than comparisons between integrated subjective values (89,90). Moreover, they allow testing of the hypothesized attribute-specific contributions of different receptor subtypes. Finally, to test the hypothesis of an early dopamine-mediated proximity effect (71), we recommend employing process models such as the drift diffusion model, that provide insights into subcomponents of the decision process by integrating choice and decision time data (17,37).

Together, these recommendations for improving study designs and statistical analyses will conceptually advance research on dopamine and decision making by assessing central assumptions of the model we proposed (although evidently not all future studies need to fulfill all of these criteria to produce valuable insights). Such a unifying account is not only of theoretical value but will also deepen our understanding of the biological causes of decision-making deficits in psychiatric disorders and improve predictions for the behavioral effects of pharmacological treatments. Integrating the notions that dopamine implements a cost control and is sensitive to proximity differences between action options, the proposed model has the potential to advance and unify our mechanistic understanding of cost-benefit decision making in healthy and clinical populations.

Acknowledgments and Disclosures

PNT received financial support from Pfizer for conducting the D1 agonist study (Grant No. WI203648) and funding from the Swiss National Science Foundation (Grant Nos. 100019_176016, 100014_165884, and CRSII5_177277) and the Velux Foundation (Grant No. 981). AS received an Emmy Noether fellowship (Grant No. SO 1636/2-1) from the German Research Foundation.

The authors report no biomedical financial interests or potential conflicts of interest.

Footnotes

Supplementary material cited in this article is available online at https://doi.org/10.1016/j.bpsgos.2022.02.010.

Supplementary Material

Supplementary Material
mmc1.pdf (300.6KB, pdf)

References

  • 1.Soutschek A., Tobler P.N. Motivation for the greater good: Neural mechanisms of overcoming costs. Curr Opin Behav Sci. 2018;22:96–105. [Google Scholar]
  • 2.Monterosso J., Piray P., Luo S. Neuroeconomics and the study of addiction. Biol Psychiatry. 2012;72:107–112. doi: 10.1016/j.biopsych.2012.03.012. [DOI] [PubMed] [Google Scholar]
  • 3.Hasler G. Can the neuroeconomics revolution revolutionize psychiatry? Neurosci Biobehav Rev. 2012;36:64–78. doi: 10.1016/j.neubiorev.2011.04.011. [DOI] [PubMed] [Google Scholar]
  • 4.Hartmann M.N., Hager O.M., Reimann A.V., Chumbley J.R., Kirschner M., Seifritz E., et al. Apathy but not diminished expression in schizophrenia is associated with discounting of monetary rewards by physical effort. Schizophr Bull. 2015;41:503–512. doi: 10.1093/schbul/sbu102. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Perugini A., Ditterich J., Shaikh A.G., Knowlton B.J., Basso M.A. Paradoxical decision-making: A framework for understanding cognition in Parkinson’s disease. Trends Neurosci. 2018;41:512–525. doi: 10.1016/j.tins.2018.04.006. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Kyonka E.G.E., Schutte N.S. Probability discounting and gambling: A meta-analysis. Addiction. 2018;113:2173–2181. doi: 10.1111/add.14397. [DOI] [PubMed] [Google Scholar]
  • 7.Webber H.E., Lopez-Gamundi P., Stamatovich S.N., de Wit H., Wardle M.C. Using pharmacological manipulations to study the role of dopamine in human reward functioning: A review of studies in healthy adults. Neurosci Biobehav Rev. 2021;120:123–158. doi: 10.1016/j.neubiorev.2020.11.004. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.de Wit H., Enggasser J.L., Richards J.B. Acute administration of d-amphetamine decreases impulsivity in healthy volunteers. Neuropsychopharmacology. 2002;27:813–825. doi: 10.1016/S0893-133X(02)00343-3. [DOI] [PubMed] [Google Scholar]
  • 9.Kayser A.S., Allen D.C., Navarro-Cebrian A., Mitchell J.M., Fields H.L. Dopamine, corticostriatal connectivity, and intertemporal choice. J Neurosci. 2012;32:9402–9409. doi: 10.1523/JNEUROSCI.1180-12.2012. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Pine A., Shiner T., Seymour B., Dolan R.J. Dopamine, time, and impulsivity in humans. J Neurosci. 2010;30:8888–8896. doi: 10.1523/JNEUROSCI.6028-09.2010. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Volkow N.D., Baler R.D. NOW vs LATER brain circuits: Implications for obesity and addiction. Trends Neurosci. 2015;38:345–352. doi: 10.1016/j.tins.2015.04.002. [DOI] [PubMed] [Google Scholar]
  • 12.Kravitz A.V., Freeze B.S., Parker P.R.L., Kay K., Thwin M.T., Deisseroth K., Kreitzer A.C. Regulation of parkinsonian motor behaviours by optogenetic control of basal ganglia circuitry. Nature. 2010;466:622–626. doi: 10.1038/nature09159. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Lerner T.N., Kreitzer A.C. Neuromodulatory control of striatal plasticity and behavior. Curr Opin Neurobiol. 2011;21:322–327. doi: 10.1016/j.conb.2011.01.005. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Lovinger D.M. Neurotransmitter roles in synaptic modulation, plasticity and learning in the dorsal striatum. Neuropharmacology. 2010;58:951–961. doi: 10.1016/j.neuropharm.2010.01.008. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Weber S.C., Beck-Schimmer B., Kajdi M.E., Müller D., Tobler P.N., Quednow B.B. Dopamine D2/3- and μ-opioid receptor antagonists reduce cue-induced responding and reward impulsivity in humans. Transl Psychiatry. 2016;6 doi: 10.1038/tp.2016.113. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Soutschek A., Burke C.J., Raja Beharelle A., Schreiber R., Weber S.C., Karipidis I.I., et al. The dopaminergic reward system underpins gender differences in social preferences. Nat Hum Behav. 2017;1:819–827. doi: 10.1038/s41562-017-0226-y. [DOI] [PubMed] [Google Scholar]
  • 17.Wagner B., Clos M., Sommer T., Peters J. Dopaminergic modulation of human intertemporal choice: A diffusion model analysis using the D2-receptor antagonist haloperidol. J Neurosci. 2020;40:7936–7948. doi: 10.1523/JNEUROSCI.0592-20.2020. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Arrondo G., Aznárez-Sanado M., Fernández-Seara M.A., Goñi J., Loayza F.R., Salamon-Klobut E., et al. Dopaminergic modulation of the trade-off between probability and time in economic decision-making. Eur Neuropsychopharmacol. 2015;25:817–827. doi: 10.1016/j.euroneuro.2015.02.011. [DOI] [PubMed] [Google Scholar]
  • 19.Kelm M.K., Boettiger C.A. Effects of acute dopamine precursor depletion on immediate reward selection bias and working memory depend on catechol-O-methyltransferase genotype. J Cogn Neurosci. 2013;25:2061–2071. doi: 10.1162/jocn_a_00464. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Acheson A., de Wit H. Bupropion improves attention but does not affect impulsive behavior in healthy young adults. Exp Clin Psychopharmacol. 2008;16:113–123. doi: 10.1037/1064-1297.16.2.113. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Petzold J., Kienast A., Lee Y., Pooseh S., London E.D., Goschke T., Smolka M.N. Baseline impulsivity may moderate L-DOPA effects on value-based decision-making. Sci Rep. 2019;9:5652. doi: 10.1038/s41598-019-42124-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Soutschek A., Gvozdanovic G., Kozak R., Duvvuri S., de Martinis N., Harel B., et al. Dopaminergic D1 receptor stimulation affects effort and risk preferences. Biol Psychiatry. 2020;87:678–685. doi: 10.1016/j.biopsych.2019.09.002. [DOI] [PubMed] [Google Scholar]
  • 23.Hamidovic A., Kang U.J., de Wit H. Effects of low to moderate acute doses of pramipexole on impulsivity and cognition in healthy volunteers. J Clin Psychopharmacol. 2008;28:45–51. doi: 10.1097/jcp.0b013e3181602fab. [DOI] [PubMed] [Google Scholar]
  • 24.Evers E.A., Stiers P., Ramaekers J.G. High reward expectancy during methylphenidate depresses the dopaminergic response to gain and loss. Soc Cogn Affect Neurosci. 2017;12:311–318. doi: 10.1093/scan/nsw124. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Symmonds M., Wright N.D., Fagan E., Dolan R.J. Assaying the effect of levodopa on the evaluation of risk in healthy humans. PLoS One. 2013;8 doi: 10.1371/journal.pone.0068177. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Campbell-Meiklejohn D., Simonsen A., Scheel-Krüger J., Wohlert V., Gjerløff T., Frith C.D., et al. In for a penny, in for a pound: Methylphenidate reduces the inhibitory effect of high stakes on persistent risky choice. J Neurosci. 2012;32:13032–13038. doi: 10.1523/JNEUROSCI.0151-12.2012. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Rigoli F., Rutledge R.B., Chew B., Ousdal O.T., Dayan P., Dolan R.J. Dopamine increases a value-independent gambling propensity. Neuropsychopharmacology. 2016;41:2658–2667. doi: 10.1038/npp.2016.68. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Rutledge R.B., Skandali N., Dayan P., Dolan R.J. Dopaminergic modulation of decision making and subjective well-being. J Neurosci. 2015;35:9811–9822. doi: 10.1523/JNEUROSCI.0702-15.2015. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.White T.L., Lejuez C.W., de Wit H. Personality and gender differences in effects of d-amphetamine on risk taking. Exp Clin Psychopharmacol. 2007;15:599–609. doi: 10.1037/1064-1297.15.6.599. [DOI] [PubMed] [Google Scholar]
  • 30.Gross J., Faber N.S., Kappes A., Nussberger A.M., Cowen P.J., Browning M., et al. When helping is risky: The behavioral and neurobiological trade-off of social and risk preferences. Psychol Sci. 2021;32:1842–1855. doi: 10.1177/09567976211015942. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31.Riba J., Krämer U.M., Heldmann M., Richter S., Münte T.F. Dopamine agonist increases risk taking but blunts reward-related brain activity. PLoS One. 2008;3:e2479. doi: 10.1371/journal.pone.0002479. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Norbury A., Manohar S., Rogers R.D., Husain M. Dopamine modulates risk-taking as a function of baseline sensation-seeking trait. J Neurosci. 2013;33:12982–12986. doi: 10.1523/JNEUROSCI.5587-12.2013. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Zack M., Poulos C.X. A D2 antagonist enhances the rewarding and priming effects of a gambling episode in pathological gamblers. Neuropsychopharmacology. 2007;32:1678–1686. doi: 10.1038/sj.npp.1301295. [DOI] [PubMed] [Google Scholar]
  • 34.Burke C.J., Soutschek A., Weber S., Raja Beharelle A., Fehr E., Haker H., Tobler P.N. Dopamine receptor-specific contributions to the computation of value. Neuropsychopharmacol. 2018;43:1415–1424. doi: 10.1038/npp.2017.302. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35.Ojala K.E., Janssen L.K., Hashemi M.M., Timmer M.H.M., Geurts D.E.M., Ter Huurne N.P., et al. Dopaminergic drug effects on probability weighting during risky decision making. eNeuro. 2018;5 doi: 10.1523/ENEURO.0330-18.2018. ENEURO.0330-18.2018. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36.Campbell-Meiklejohn D., Wakeley J., Herbert V., Cook J., Scollo P., Ray M.K., et al. Serotonin and dopamine play complementary roles in gambling to recover losses. Neuropsychopharmacology. 2011;36:402–410. doi: 10.1038/npp.2010.170. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37.Westbrook A., van den Bosch R., Määttä J.I., Hofmans L., Papadopetraki D., Cools R., Frank M.J. Dopamine promotes cognitive effort by biasing the benefits versus costs of cognitive work. Science. 2020;367:1362–1366. doi: 10.1126/science.aaz5891. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 38.Wardle M.C., Treadway M.T., Mayo L.M., Zald D.H., de Wit H. Amping up effort: Effects of d-amphetamine on human effort-based decision-making. J Neurosci. 2011;31:16597–16602. doi: 10.1523/JNEUROSCI.4387-11.2011. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39.Zénon A., Devesse S., Olivier E. Dopamine manipulation affects response vigor independently of opportunity cost. J Neurosci. 2016;36:9516–9525. doi: 10.1523/JNEUROSCI.4467-15.2016. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40.Michely J., Viswanathan S., Hauser T.U., Delker L., Dolan R.J., Grefkes C. The role of dopamine in dynamic effort-reward integration. Neuropsychopharmacology. 2020;45:1448–1453. doi: 10.1038/s41386-020-0669-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41.Dean Z., Horndasch S., Giannopoulos P., McCabe C. Enhanced neural response to anticipation, effort and consummation of reward and aversion during bupropion treatment. Psychol Med. 2016;46:2263–2274. doi: 10.1017/S003329171600088X. [DOI] [PubMed] [Google Scholar]
  • 42.Korb S., Götzendorfer S.J., Massaccesi C., Sezen P., Graf I., Willeit M., et al. Dopaminergic and opioidergic regulation during anticipation and consumption of social and nonsocial rewards. Elife. 2020;9 doi: 10.7554/eLife.55797. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43.Ohmann H.A., Kuper N., Wacker J. A low dosage of the dopamine D2-receptor antagonist sulpiride affects effort allocation for reward regardless of trait extraversion. Personal Neurosci. 2020;3:e7. doi: 10.1017/pen.2020.7. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 44.Cawley E.I., Park S., aan het Rot M., Sancton K., Benkelfat C., Young S.N., et al. Dopamine and light: Dissecting effects on mood and motivational states in women with subsyndromal seasonal affective disorder. J Psychiatry Neurosci. 2013;38:388–397. doi: 10.1503/jpn.120181. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 45.Andreoni J. Impure altruism and donations to public goods: A theory of warm-glow giving. Econ J. 1990;100:464–477. [Google Scholar]
  • 46.Park S.Q., Kahnt T., Dogan A., Strang S., Fehr E., Tobler P.N. A neural link between generosity and happiness. Nat Commun. 2017;8:15964. doi: 10.1038/ncomms15964. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 47.Rand D.G., Brescoll V.L., Everett J.A.C., Capraro V., Barcelo H. Social heuristics and social roles: Intuition favors altruism for women but not for men. J Exp Psychol Gen. 2016;145:389–396. doi: 10.1037/xge0000154. [DOI] [PubMed] [Google Scholar]
  • 48.Oroz Artigas S., Liu L., Strang S., Burrasch C., Hermsteiner A., Münte T.F., Park S.Q. Enhancement in dopamine reduces generous behaviour in women. PLoS One. 2019;14 doi: 10.1371/journal.pone.0226893. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 49.Crockett M.J., Siegel J.Z., Kurth-Nelson Z., Ousdal O.T., Story G., Frieband C., et al. Dissociable effects of serotonin and dopamine on the valuation of harm in moral decision making. Curr Biol. 2015;25:1852–1859. doi: 10.1016/j.cub.2015.05.021. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 50.Pedroni A., Eisenegger C., Hartmann M.N., Fischbacher U., Knoch D. Dopaminergic stimulation increases selfish behavior in the absence of punishment threat. Psychopharmacology (Berl) 2014;231:135–141. doi: 10.1007/s00213-013-3210-x. [DOI] [PubMed] [Google Scholar]
  • 51.Sáez I., Zhu L., Set E., Kayser A., Hsu M. Dopamine modulates egalitarian behavior in humans. Curr Biol. 2015;25:912–919. doi: 10.1016/j.cub.2015.01.071. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 52.Salamone J.D., Correa M. The mysterious motivational functions of mesolimbic dopamine. Neuron. 2012;76:470–485. doi: 10.1016/j.neuron.2012.10.021. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 53.Robbins T.W., Everitt B.J. Functions of dopamine in the dorsal and ventral striatum. Semin Neurosci. 1992;4:119–127. [Google Scholar]
  • 54.Salamone J.D., Correa M., Farrar A., Mingote S.M. Effort-related functions of nucleus accumbens dopamine and associated forebrain circuits. Psychopharmacology (Berl) 2007;191:461–482. doi: 10.1007/s00213-006-0668-9. [DOI] [PubMed] [Google Scholar]
  • 55.Nowend K.L., Arizzi M., Carlson B.B., Salamone J.D. D1 or D2 antagonism in nucleus accumbens core or dorsomedial shell suppresses lever pressing for food but leads to compensatory increases in chow consumption. Pharmacol Biochem Behav. 2001;69:373–382. doi: 10.1016/s0091-3057(01)00524-x. [DOI] [PubMed] [Google Scholar]
  • 56.Salamone J.D., Farrar A.M., Font L., Patel V., Schlar D.E., Nunes E.J., et al. Differential actions of adenosine A1 and A2A antagonists on the effort-related effects of dopamine D2 antagonism. Behav Brain Res. 2009;201:216–222. doi: 10.1016/j.bbr.2009.02.021. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 57.Koch M., Schmid A., Schnitzler H.U. Role of muscles accumbens dopamine D1 and D2 receptors in instrumental and Pavlovian paradigms of conditioned reward. Psychopharmacology (Berl) 2000;152:67–73. doi: 10.1007/s002130000505. [DOI] [PubMed] [Google Scholar]
  • 58.Brown V.J., Robbins T.W. Simple and choice reaction time performance following unilateral striatal dopamine depletion in the rat. Impaired motor readiness but preserved response preparation. Brain. 1991;114:513–525. doi: 10.1093/brain/114.1.513. [DOI] [PubMed] [Google Scholar]
  • 59.Mohebi A., Pettibone J.R., Hamid A.A., Wong J.M.T., Vinson L.T., Patriarchi T., et al. Dissociable dopamine dynamics for learning and motivation [published correction appears in Nature 2019; 571:E3] Nature. 2019;570:65–70. doi: 10.1038/s41586-019-1235-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 60.Beeler J.A. Thorndike’s law 2.0: Dopamine and the regulation of thrift. Front Neurosci. 2012;6:116. doi: 10.3389/fnins.2012.00116. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 61.Beeler J.A., Frazier C.R.M., Zhuang X. Putting desire on a budget: Dopamine and energy expenditure, reconciling reward and resources. Front Integr Neurosci. 2012;6:49. doi: 10.3389/fnint.2012.00049. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 62.Stauffer W.R., Lak A., Schultz W. Dopamine reward prediction error responses reflect marginal utility. Curr Biol. 2014;24:2491–2500. doi: 10.1016/j.cub.2014.08.064. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 63.Schultz W., Dayan P., Montague P.R. A neural substrate of prediction and reward. Science. 1997;275:1593–1599. doi: 10.1126/science.275.5306.1593. [DOI] [PubMed] [Google Scholar]
  • 64.Guitart-Masip M., Duzel E., Dolan R., Dayan P. Action versus valence in decision making. Trends Cogn Sci. 2014;18:194–202. doi: 10.1016/j.tics.2014.01.003. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 65.Beeler J.A., Mourra D. To do or not to do: Dopamine, affordability and the economics of opportunity. Front Integr Neurosci. 2018;12:6. doi: 10.3389/fnint.2018.00006. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 66.Niv Y., Daw N.D., Joel D., Dayan P. Tonic dopamine: Opportunity costs and the control of response vigor. Psychopharmacology (Berl) 2007;191:507–520. doi: 10.1007/s00213-006-0502-4. [DOI] [PubMed] [Google Scholar]
  • 67.Collins A.G.E., Frank M.J. Opponent actor learning (OpAL): Modeling interactive effects of striatal dopamine on reinforcement learning and choice incentive. Psychol Rev. 2014;121:337–366. doi: 10.1037/a0037015. [DOI] [PubMed] [Google Scholar]
  • 68.Tye K.M., Tye L.D., Cone J.J., Hekkelman E.F., Janak P.H., Bonci A. Methylphenidate facilitates learning-induced amygdala plasticity. Nat Neurosci. 2010;13:475–481. doi: 10.1038/nn.2506. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 69.Zalocusky K.A., Ramakrishnan C., Lerner T.N., Davidson T.J., Knutson B., Deisseroth K. Nucleus accumbens D2R cells signal prior outcomes and control risky decision-making. Nature. 2016;531:642–646. doi: 10.1038/nature17400. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 70.Hikida T., Kimura K., Wada N., Funabiki K., Nakanishi S. Distinct roles of synaptic transmission in direct and indirect striatal pathways to reward and aversive behavior. Neuron. 2010;66:896–907. doi: 10.1016/j.neuron.2010.05.011. [DOI] [PubMed] [Google Scholar]
  • 71.Westbrook A., Frank M. Dopamine and proximity in motivation and cognitive control. Curr Opin Behav Sci. 2018;22:28–34. doi: 10.1016/j.cobeha.2017.12.011. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 72.Morrison S.E., Nicola S.M. Neurons in the nucleus accumbens promote selection bias for nearer objects. J Neurosci. 2014;34:14147–14162. doi: 10.1523/JNEUROSCI.2197-14.2014. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 73.Cavanagh J.F., Wiecki T.V., Cohen M.X., Figueroa C.M., Samanta J., Sherman S.J., Frank M.J. Subthalamic nucleus stimulation reverses mediofrontal influence over decision threshold. Nat Neurosci. 2011;14:1462–1467. doi: 10.1038/nn.2925. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 74.Wiecki T.V., Frank M.J. A computational model of inhibitory control in frontal cortex and basal ganglia. Psychol Rev. 2013;120:329–355. doi: 10.1037/a0031542. [DOI] [PubMed] [Google Scholar]
  • 75.Lidow M.S., Goldman-Rakic P.S., Gallager D.W., Rakic P. Distribution of dopaminergic receptors in the primate cerebral cortex: Quantitative autoradiographic analysis using [3H] raclopride, [3H] spiperone and [3H] SCH23390. Neuroscience. 1991;40:657–671. doi: 10.1016/0306-4522(91)90003-7. [DOI] [PubMed] [Google Scholar]
  • 76.Robbins T.W. Chemistry of the mind: Neurochemical modulation of prefrontal cortical function. J Comp Neurol. 2005;493:140–146. doi: 10.1002/cne.20717. [DOI] [PubMed] [Google Scholar]
  • 77.Lawlor V.M., Webb C.A., Wiecki T.V., Frank M.J., Trivedi M., Pizzagalli D.A., Dillon D.G. Dissecting the impact of depression on decision-making. Psychol Med. 2020;50:1613–1622. doi: 10.1017/S0033291719001570. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 78.O’Callaghan C., Hall J.M., Tomassini A., Muller A.J., Walpola I.C., Moustafa A.A., et al. Visual hallucinations are characterized by impaired sensory evidence accumulation: Insights from hierarchical drift diffusion modeling in Parkinson’s disease. Biol Psychiatry Cogn Neurosci Neuroimaging. 2017;2:680–688. doi: 10.1016/j.bpsc.2017.04.007. [DOI] [PubMed] [Google Scholar]
  • 79.Amlung M., Marsden E., Holshausen K., Morris V., Patel H., Vedelago L., et al. Delay discounting as a transdiagnostic process in psychiatric disorders A meta-analysis. JAMA Psychiatry. 2019;76:1176–1186. doi: 10.1001/jamapsychiatry.2019.2102. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 80.Castrellon J.J., Seaman K.L., Crawford J.L., Young J.S., Smith C.T., Dang L.C., et al. Individual differences in dopamine are associated with reward discounting in clinical groups but not in healthy adults. J Neurosci. 2019;39:321–332. doi: 10.1523/JNEUROSCI.1984-18.2018. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 81.Weigard A., Huang-Pollock C. A diffusion modeling approach to understanding contextual cueing effects in children with ADHD. J Child Psychol Psychiatry. 2014;55:1336–1344. doi: 10.1111/jcpp.12250. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 82.Gershman S.J., Uchida N. Believing in dopamine. Nat Rev Neurosci. 2019;20:703–714. doi: 10.1038/s41583-019-0220-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 83.Friston K., Schwartenbeck P., FitzGerald T., Moutoussis M., Behrens T., Dolan R.J. The anatomy of choice: Dopamine and decision-making. Philos Trans R Soc Lond B Biol Sci. 2014;369:20130481. doi: 10.1098/rstb.2013.0481. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 84.Floresco S.B. Prefrontal dopamine and behavioral flexibility: Shifting from an “inverted-U” toward a family of functions. Front Neurosci. 2013;7:62. doi: 10.3389/fnins.2013.00062. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 85.Soutschek A., Kozak R., de Martinis N., Howe W., Burke C.J., Fehr E., et al. Activation of D1 receptors affects human reactivity and flexibility to valued cues. Neuropsychopharmacology. 2020;45:780–785. doi: 10.1038/s41386-020-0617-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 86.Cools R., Gibbs S.E., Miyakawa A., Jagust W., D’Esposito M. Working memory capacity predicts dopamine synthesis capacity in the human striatum. J Neurosci. 2008;28:1208–1212. doi: 10.1523/JNEUROSCI.4475-07.2008. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 87.Buckholtz J.W., Treadway M.T., Cowan R.L., Woodward N.D., Li R., Ansari M.S., et al. Dopaminergic network differences in human impulsivity. Science. 2010;329:532. doi: 10.1126/science.1185778. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 88.Schoemaker H., Claustre Y., Fage D., Rouquier L., Chergui K., Curet O., et al. Neurochemical characteristics of amisulpride, an atypical dopamine D2/D3 receptor antagonist with both presynaptic and limbic selectivity. J Pharmacol Exp Ther. 1997;280:83–97. [PubMed] [Google Scholar]
  • 89.Reeck C., Wall D., Johnson E.J. Search predicts and changes patience in intertemporal choice. Proc Natl Acad Sci U S A. 2017;114:11890–11895. doi: 10.1073/pnas.1707040114. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 90.Amasino D.R., Sullivan N.J., Kranton R.E., Huettel S.A. Amount and time exert independent influences on intertemporal choice. Nat Hum Behav. 2019;3:383–392. doi: 10.1038/s41562-019-0537-2. [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Supplementary Material
mmc1.pdf (300.6KB, pdf)

Articles from Biological Psychiatry Global Open Science are provided here courtesy of Elsevier

RESOURCES