Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2022 Jan 1.
Published in final edited form as: Neurosci Biobehav Rev. 2020 Nov 14;120:123–158. doi: 10.1016/j.neubiorev.2020.11.004

Using pharmacological manipulations to study the role of dopamine in human reward functioning: A review of studies in healthy adults

Heather E Webber a, Paula Lopez-Gamundi b,c, Sydney N Stamatovich d, Harriet de Wit e, Margaret C Wardle f,*
PMCID: PMC7855845  NIHMSID: NIHMS1652278  PMID: 33202256

Abstract

Dopamine (DA) plays a key role in reward processing and is implicated in psychological disorders such as depression, substance use, and schizophrenia. The role of DA in reward processing is an area of highly active research. One approach to this question is drug challenge studies with drugs known to alter DA function. These studies provide good experimental control and can be performed in parallel in laboratory animals and humans. This review aimed to summarize results of studies using pharmacological manipulations of DA in healthy adults. ‘Reward’ is a complex process, so we separated ‘phases’ of reward, including anticipation, evaluation of cost and benefits of upcoming reward, execution of actions to obtain reward, pleasure in response to receiving a reward, and reward learning. Results indicated that i) DAergic drugs have different effects on different phases of reward; ii) the relationship between DA and reward functioning appears unlikely to be linear; iii) our ability to detect the effects of DAergic drugs varies depending on whether subjective, behavioral, imaging measures are used.

Keywords: dopamine, reward functioning, pharmacological challenge, anticipation, execution, invigoration, evaluation of costs, motivation, effort, learning, pleasure

I. Introduction

A. Rationale for review

The reward system is key to adaptive functioning, as it helps us identify, respond to, and remember stimuli, such as food, that aid in survival. Dopamine (DA) is thought to play a critical role in the processing of rewards. Understanding the role of DA in reward-related functioning is clinically important, as hypo- and hyper-activity of the DA system is thought to contribute to reward-related pathologies in addiction, depression, and schizophrenia (Barch et al., 2014; Davis et al., 1991; Garfield et al., 2013; Leyton, 2010; Treadway and Zald, 2011a; Volkow et al., 2011, 2007). Much of our understanding of DA function is derived from studies with animal models, which can test highly complex brain circuits and receptor mechanisms involved in motivated behaviors in detail, with strong experimental control (Bamford et al., 2018; Berke, 2018a; Schultz, 2016a; Volkow et al., 2017). Although these studies have unquestionably advanced our understanding of how DA is involved in behavior, limitations in translating findings from animals to humans make it critical to confirm the role of DA in reward processes in humans (Stephens et al., 2011). Clinical studies have investigated reward functioning in patients with psychiatric disorders associated with DA dysfunction, such as depression and addiction (Hamilton et al., 2012; McGuire et al., 2008; Volkow et al., 2011). However, these studies provide only indirect evidence for the role of DA, because our understanding of the pathophysiology of these disorders is incomplete, and because it is difficult to rule out confounding factors related to the natural history of the disorder. Thus, pharmacological challenge studies with agents that increase or decrease DA function offer a valuable translational bridge, as they can be conducted in animals, healthy humans, and psychiatric patients.

In particular, drug challenge studies with healthy volunteers may provide critical information about the DA-related processes that mediate and maintain reward-motivated behavior. Drug challenge studies provide good experimental control, often use behavioral tasks with high translational fidelity to the animal literature, and can be combined with imaging to provide at least some degree of neuroanatomical specificity. Studying healthy adults allows investigators to examine the effect of DA manipulations in individuals whose responses are not confounded by co-existing psychiatric disorders or extensive exposure to drugs. Thus, in this paper, we focus particularly on what can be learned about reward functioning in humans from pharmacological challenge studies with dopaminergic drugs, conducted in healthy volunteers. The goal of the review is to summarize key conclusions that can be drawn from these studies about how DA mediates reward functioning in humans.

B. Scope and organization of review

i. Defining “reward”.

The term “reward” is used in many ways, from experimental procedures and events (e.g., the food reward was delivered), to purported brain processes (e.g., neural response to reward). The brain processes involved in reward are extremely complex, including memory, sensory and motor processing, motivation, and decision-making circuits. In this review we separate reward processing into five behavioral components, corresponding to five phases of processing a single reward, as derived from theoretical perspectives (Assadi et al., 2009; Berridge and Kringelbach, 2015; Berridge and Robinson, 1998; Cannon and Bseikri, 2004; Niv et al., 2007; Salamone, 2006; Schultz, 2016a; Schultz et al., 1997; Zald and Treadway, 2017). These phases are: 1) “Anticipation” - hedonic impact of anticipating future reward, 2) “Evaluation” – evaluation of the costs and benefits of an upcoming reward, 3) “Execution” – execution of behavior to obtain reward, 4) “Pleasure” - hedonic responses to receipt of a reward, and 5) “Learning” - learning from rewarding outcomes (see Figure 1). We realize this taxonomy may not perfectly segment reward-related processing and that there is likely overlap between phases. However, we felt that grouping by behavioral components was necessary to 1) directly compare studies using similar measures and 2) make the large literature manageable. Further, studies in animals indicate that these different behavioral components engage different neuroanatomical processes (Assadi et al., 2009; Berridge and Kringelbach, 2015; Berridge and Robinson, 1998; Cannon and Bseikri, 2004; Niv et al., 2007; Salamone, 2006; Schultz, 2016a; Schultz et al., 1997; Zald and Treadway, 2017). Within each phase, we further group studies based on the type of measure employed, reviewing self-report, behavioral, and imaging findings in separate sections. Finally, within measurement types, we group findings according to whether the drug administered was intended to augment or reduce DA functioning. Studies with multiple outcomes or drugs thus appear in multiple sections. The studies reviewed reflect an extensive search of key terms related to reward (e.g., anticipation, evaluation, pleasure, etc.), DA function (e.g., DA agonist, antagonist, etc.) and specific dopaminergic agents. Only studies performed on healthy adults using a pharmacological manipulation of DA were included, totaling over 100 articles included in the final review.

Figure 1.

Figure 1.

Organizational taxonomy of reward-related processes

To provide context for interpreting study findings, we first briefly overview DA receptor types and signaling dynamics, and typical pharmacological manipulations used to study the role of DA in reward in humans. We then devote one section to each reward phase.

C. Brief overview of relevant dopaminergic neuropharmacology

i. DA receptor types

Different DA receptors have been implicated in different components of reward-related behavior in animal models. In humans, even though some drugs vary in their affinities for DA receptor types, it is difficult to delineate the role of receptor subtypes in observed behaviors. Nevertheless, because some human drug challenge studies have interpreted patterns of results based on receptor subtypes, we briefly review these subtypes to help the reader evaluate these explanations.

Five different DA receptor types have been identified (D1–D5), but in general, the focus is on two subtypes: D1-like (D1 and D5) and D2-like (D2–D4). Subtypes differ in their affinities, G-protein coupling, which intracellular pathways they stimulate, and their presumed functions (Beaulieu and Gainetdinov, 2011; Tritsch and Sabatini, 2012). D2-like receptors are present both pre- and post-synaptically, while D1-like are mostly post-synaptic (Chen et al., 2005; Leriche et al., 2008). Pre-synaptic D2-like autoreceptors have 10× to 100× greater affinity than D1-types, and thus can be activated in low-DA states, while D1 receptors require higher concentrations of DA (Cooper et al., 2003; Grace, 1991). D1 receptors are widespread in the brain and abundant in the striatum, NAcc, olfactory tubercle, and prefrontal cortex, whereas D2 receptors are more abundant in the striatum (Camps et al., 1990; Lidow et al., 1991). These differences become important when considered in context of DA signaling dynamics reviewed in the next section.

ii. DA signaling dynamics

Variations in DA signaling dynamics and levels, which can be observed directly in preclinical studies, have long been important to theories of reward-related functions. These include different modes of DA release, and non-linear responses to stimulation seen in dynamic systems.

A long-standing theory proposes that two major types of DA signaling exist in the ventral tegmental area (VTA) where DA neurons originate, each with different time courses. This theory distinguishes between phasic DA release -- stereotyped, transient bursts of activation that are brief, but release large DA concentrations sufficient to activate post-synaptic, lower affinity D1 receptors (Grace, 2000, 1991; Schultz, 2013) – and tonic DA release -- low levels of single spike firing in DA neurons which diffuse DA out of the synapse, contributing to extracellular DA, and primarily activating more sensitive D2 presynaptic autoreceptors (Floresco et al., 2003; Grace, 2000, 1991). These two types of DA signaling were hypothesized to be relevant to different reward functions, such that phasic DA was proposed to underpin reward learning, while tonic DA was thought to be involved in effort exerted to obtain a reward (Niv et al., 2007; Salamone and Correa, 2012; Schultz et al., 2017). Finally, phasic and tonic DA were thought to be related through homeostatic mechanisms, with higher levels of tonic DA thought to decrease phasic DA transmission through autoreceptors (Grace, 2000, 1991). This theory has strongly influenced the human literature, from selection of DA manipulations to interpretation of results, so we describe it here and will reference it elsewhere in the paper when it has been used by investigators to explain results. However, it is important to note that the idea of a slow, tonic DA signal has been challenged. Newer results from the animal literature suggest instead a distinction between “burst” DA releases originating in the VTA and traveling “long distance” to the NAcc and elsewhere, still thought to be critically involved in the acquisition of new learning, and “local” DA release from neurons in in the NAcc, which is thought to influence motivation and effort (Berke, 2018b; Mohebi et al., 2019). Both of these processes are now thought to operate on fast time scales, and it is unclear whether a homeostatic relationship is expected.

Responses to DA-related drugs may also be non-linear, and interact with baseline levels of DA. In particular, the “inverted-U” hypothesis suggests there is an optimal level of DA for various reward functions, and that increases beyond this level will impair function (Vaillancourt et al., 2013). Thus, increasing DA could either enhance or impair reward functioning, depending both on whether the organism’s baseline DA levels are low or high, and on the optimal level of DA for that particular function (Cools, 2019; Cools and D’Esposito, 2011; Kroener et al., 2009). However, there is also an alternate possibility, which could be characterized as, “the rich get richer, while the poor get poorer”. In this hypothesis, individuals with good baseline DA receptor capacity might benefit more from DA augmentation (Swart et al., 2017), and conversely, the effects of DA reduction might only be evident in individuals with poor baseline functioning (Kelm and Boettiger, 2013). Under either of these theories, individual differences in baseline DA could moderate the effects of DA challenges, so we note when baseline differences are explored as possible moderators.

Potential baseline differences in DA have been measured in several ways in the literature we will review. PET imaging is the “gold standard”, but the expense of this approach means that in the reviewed articles, baseline functioning has more often been imputed from genetic polymorphisms or behavioral proxy measures. In genetics, the COMT Val148Met polymorphism has been most commonly examined. Homozygote carriers of the val allele are thought to have less prefrontal DA at baseline (Nicholson, 2002); however, some research suggests that the val allele may also produce higher tonic DA in the midbrain (Corral-Frías et al., 2016), complicating interpretations of interactions between COMT genotype and drug challenges, as systemic challenges are not regionally specific. Behavioral measures that have been used as proxies of baseline DA levels include working memory (Cools et al., 2008; Landau et al., 2009), impulsivity (Cools et al., 2008), and extraversion (Wacker and Gatt, 2010).

D. Pharmacological challenge strategies

The drugs used to alter DA function in humans fall into two broad classes: those intended to increase DA function and those intended to reduce it. Drugs used to augment DA include receptor agonists (e.g. bromocriptine), re-uptake blockers (e.g. d-amphetamine), and precursors (e.g. levodopa; L-DOPA), and drugs that impair metabolism of DA (e.g. tolcapone). Drugs used to reduce DA function include receptor antagonists such as haloperidol, or the acute phenylalanine/tyrosine depletion procedure (APTD), which limits availability of DA precursor amino acids (Moja et al. 1996; Leyton et al. 2000). NMDA receptor antagonists (e.g., memantine) are also thought to decrease “phasic” or “burst” DA transmission (Jocham et al., 2014). “Low doses” of D2 agonists have also been used as a DA reduction strategy, based on the idea that activating more sensitive presynaptic autoreceptors will decrease phasic DA transmission (e.g., Pizzagalli et al. 2008a). Conversely, “low doses” of D2 antagonists have been used to increase DA, based on their presumed ability to increase phasic DA (e.g., Mueller et al., 2014). However, there is no clear agreement in the field on what constitutes a “low dose”, or when autoreceptor vs. post-synaptic effects dominate. In the absence of a pharmacologically-based cut-point, for purposes of this review, we classify studies as augmentation or reduction based on the authors’ stated hypothesis about how that drug and dose should affect DA. Occasionally this results in studies using the same drug and dose with opposing purposes. To address this, we highlight when a “low dose” strategy is being used, and note drug and dose for every study, so that readers may take this into account.

II. Anticipation of Reward

A. Anticipation of reward - Definition

We label the first phase of reward processing, as outlined in Figure 1, “Anticipation”. We define anticipation as the hedonic impact of expecting a reward, e.g. subjective excitement about future or possible reward. Anticipation starts as soon as a future reward is perceived and continues up until the reward has been received. Anticipation is also clearly related to evaluation of benefits in reward-related decision-making, but measures involving explicit decision-making are discussed in “Evaluation”. Similarly, anticipation is intimately involved in predicting the value of a reward in reward-related learning, but studies that measure changes in behavior over repeated trials are discussed in “Learning”. Here we focus on studies that simply examine anticipation without measuring impacts on decision-making or future behavior.

B. Anticipation of reward – Predictions based on putative neurobiology

Neurobiological studies suggest that DA manipulations should affect anticipation in humans. In laboratory animals, the areas associated most strongly with reward anticipation are the amygdala, VTA, NAcc, and ventral pallidum (Berridge and Robinson, 2003). DA increases in the VTA and NAcc of rats during presentation of a conditioned stimulus associated with reward, prior to receipt of the reward (Schultz, 2016b). fMRI studies in humans show that anticipation of reward activates subcortical limbic and prefrontal brain regions, including the NAcc, thalamus, striatum, prefrontal cortex, and ACC (Oldham et al., 2018; Schott et al., 2008; Schultz, 1998). Although fMRI is not pharmacologically specific, these areas are in known dopaminergic pathways.

Behavioral studies in animals also support the idea that DA is involved in anticipation. Although anticipation is more difficult to measure in animals, rats emit ultra-sonic vocalizations when rewards are expected that are thought to indicate anticipation (Knutson et al., 1998). Systemic administration of dopamine antagonists reduces these calls in male rats anticipating sexual contact with a female, consistent with the idea that systemic manipulations of DA should alter anticipation (Ciucci et al., 2007).

Based on the putative neurobiology of reward anticipation, we hypothesize that pharmacological challenges that increase DA in humans will generally increase anticipation of reward, while reduction of DA will decrease anticipation.

C. Anticipation of reward - Measures

Anticipation of reward in humans has been assessed using self-report, behavioral measures and neuroimaging. Self-reports have been administered before the reward has been received, retrospectively after the reward is received, or when imagining future hypothetical rewards (Sharot et al., 2009), and typically consist of simple happiness or excitement ratings. One study examined a behavioral measure, skin conductance, in response to anticipation of reward. However, the most common method is neuroimaging, with most studies using fMRI during cues signaling the possibility of reward.

One task, the Monetary Incentive Delay (MID) task (Knutson et al., 2000), has been most often used for imaging anticipation, so we describe it in more detail. The MID requires participants to a) view a cue that corresponds to a reward value; b) view a fixation cross during a delay period; c) respond quickly to a target by pressing a button; d) view a fixation cross during a delay period; and e) get feedback on whether they received a reward. Receipt of reward is based on the speed of the participants’ response to the target, with the required speed titrated to produce a certain percentage of rewarded trials. Depending on the study, anticipation may be measured during the cue, during the delay before the participant completes the action to receive a reward, or after the individual completes the task but before feedback on the outcome. In this review, reaction time in response to the cue is considered to represent “Execution”, and is discussed in that section.

Studies on anticipation have used a range of reward types including real and hypothetical monetary rewards and pleasant/unpleasant tastes, so reward type will be noted as a possible moderator.

D. Anticipation of reward - Summary of Studies

i. Anticipation of reward – Self-report measures

Three studies have tested the effect of drugs intended to increase DA on self-reported anticipation. Only one was consistent with our hypothesis that increasing DA would increase anticipation. In this study, L-DOPA (100mg) enhanced participants’ estimates of the pleasure they would derive from hypothetical future vacations (Sharot, Shiner, Brown, Fan, & Dolan, 2009). However, in another study that used “real time” ratings before actual events, the DA agonist bupropion (150 mg/day for 7 days, experiment occurred on day 7) had no effect. In this study, participants viewed pictures of either chocolate or moldy food, which signaled delivery of either a rewarding pleasant taste (chocolate) or a punishing unpleasant taste (mixture of chocolate and beetroot juice), respectively. Bupropion had no effect on reports of how much participants “wanted” the pleasant taste, measured just prior to taste delivery (Dean et al., 2016). A final study produced results directly contrary to our hypothesis. In this study, participants either received amphetamine or placebo while performing the MID task, and retrospectively reported “excitement” during cues that signaled upcoming monetary reward. d-Amphetamine (.25 mg/kg), produced a non-significant trend to report less excitement while viewing a reward cue (+$5.00), and significantly increased reports of excitement while viewing loss cue (−$5.00). Thus, DA augmentation seemed to “flatten” anticipation to reward and loss (Knutson et al., 2004). In summary, drugs that increase DA function have had mixed effects on self-reported anticipation. Of note, each study used a different reward type, drug, time frame for assessment and scale, making it difficult to assess contributions of design differences.

Only one study has examined the effect of a manipulation intended to decrease DA on self-reported anticipation. Depletion of DA via APTD did not change retrospective reports of excitement during monetary reward cues on the MID (Bjork et al., 2014). This is also inconsistent with our hypothesis.

ii. Anticipation of reward – Behavioral measures

Only one study has examined the effects of a dopaminergic modulation on a behavioral response during anticipation of reward. This study examined skin conductance, a physiological measure of arousal, during the anticipation period of the MID (Ferreri et al., 2019). Increasing DA with L-DOPA (100mg) increased electrodermal responses, compared to reducing DA with risperdone (2mg; although a placebo was included, placebo comparisons were not reported). Importantly, these changes were not evident during no-reward trials, indicating they were not a physiological artifact of the drugs. This result is consistent with our hypothesis.

ii. Anticipation of reward – Neuroimaging measures

Fifteen studies have examined the effects of drugs intended to increase DA function on brain activity during reward anticipation. The most commonly reported results are for the striatum, including the ventral striatum, which encompasses the NAcc, and the dorsal striatum or caudate, so we will focus on these.

Seven studies found that drugs intended to increase DA function increased striatal activity, consistent with our hypothesis. In the Dean et al (2016) study already described above, bupropion (150mg/day for 7 days) increased BOLD activity in the caudate, pregenual ACC/ventromedial prefrontal cortex and lateral orbitofrontal cortex during anticipation of pleasant tastes (Dean et al., 2016). The drug also increased caudate activity during anticipation of unpleasant tastes, suggesting increasing DA may also increase anticipation of punishment/loss. Six other studies examined the effects of DA-increasing drugs during anticipation of monetary rewards. Methylphenidate (40mg) increased BOLD activity in the ventral striatum (Evers et al., 2017), modafinil (200mg) increased bilateral NAcc activation (Funayama et al., 2014), and bupropion (150mg) increased right Nacc activation (Ikeda et al., 2019), all during anticipation of monetary reward. The agonist bromocriptine (1.5mg) also increased activity in the right NAcc, but only in individuals with the A1A1 DRD2 genotype (Kirsch et al., 2006). This genotype is related to lower baseline levels of DA, so these findings are consistent with an inverted-U relationship in which individuals low in DA at baseline experience a greater boost in anticipatory activity from DA-increasing drugs. Finally, a paradigm designed to parallel animal studies of sensitization, a phenomenon in which repeated administration of stimulants results in increased DA release at subsequent administrations, administered d-amphetamine (20mg) or placebo every 48 hours for three sessions, and then again a final time after a two-week abstinence period. At the final administration, BOLD activity in the caudate was increased in the d-amphetamine condition during anticipation of monetary reward (O’Daly et al., 2014).

While each of the above studies examined anticipation when individuals were preparing to make a response, one additional study used a task in which participants were also asked to withhold responses to get reward. DA is thought to play a dual role in both anticipation of reward and invigoration actions taken to gain reward (this will be further discussed in the “Execution” section). Based on this dual role, Guitart-Masip and colleagues argued that increasing DA might preferentially enhance anticipation of actions taken to gain reward (“go to win” actions), but not actions withheld to gain reward (“no-go to win”), or actions taken to avoid punishment (“no-go to avoid punishment”; Guitart-Masip et al., 2012a). In an fMRI paradigm that dissociated anticipation from preparation for action by crossing go and no-go response requirements with reward and punishment, these authors found that L-DOPA (150mg) increased activity in the striatum and substantia nigra/VTA only during anticipation of “go to win” trials, not “go to avoid punishment” or “no-go to win” trials. Taken together, these seven studies suggest that DA augmenting drugs boost activity in the striatum (broadly defined) during anticipation.

Null results were seen in five other studies. Benzylpiperazine, a “designer” drug thought to have primarily DAergic effects (Curley et al., 2013), did not affect activity in the striatum during reward anticipation, although it decreased activity in inferior frontal gyrus, insula, and mid-occipital regions. However, it is not clear that this study assessed anticipation of reward per se, as it used a task in which participants did not know whether they were expecting a reward or loss (50/50 chance). Three studies using the more typical MID task showed no effect of bupropion (150ng) or atomoxetine (40mg) within the NAcc (Graf et al., 2016a; Suzuki et al., 2019), or of L-DOPA (100mg) on broader striatal activity during reward anticipation (Wittmann et al., 2015). However, one of these studies (Wittmann et al., 2015) did see an increase in striatal activity when participants anticipated monetary losses. A final study examined effects of L-DOPA (150mg) in a MID-like task with auditory, rather than visual, cues. In this study the drug did not affect responses in the striatum during anticipation, but increased responding in the left auditory cortex, left inferior frontal gyrus, and ACC (Weis et al., 2012).

Three imaging studies suggested that DA increasing drugs decrease activity in the striatum during anticipation of reward. In one study methylphenidate (0.5mg.kg) decreased activity in the caudate, left thalamus and right cerebellum during anticipation of reward in an “anticipation, reward, conflict” task (Ivanov et al., 2014). In another study d-amphetamine (0.25 mg/kg) “flattened” levels of ventral striatal activity during the MID by reducing maximal activation during anticipation of gains, but increasing activation during anticipation of losses (Knutson et al., 2004). However, despite the reduction in maximal activation during reward anticipation, d-amphetamine also prolonged striatal activity during this phase. The authors suggested this might reflect increased tonic DA activity blunting phasic DA activity via autoreceptors. Unfortunately, no future studies have examined duration of activation to further test this hypothesis. A final study used MEG to study reward function after treatment with L-DOPA (150mg). L-DOPA reduced ERF amplitude over the occipital areas, specifically to cues predicting high probability, high magnitude reward (Apitz and Bunzeck, 2014). The ERF effects were accompanied by decreased power in the beta band over the frontal areas during anticipation. ERFs during anticipation are sensitive to outcome valence (Doñamayor et al., 2011; HajiHosseini et al., 2012; Marco-Pallares et al., 2008), and beta power is thought to reflect the linking of stimulus-response associations in the dorsolateral prefrontal cortex (HajiHosseini and Holroyd, 2015; Melgari et al., 2014). Thus, these results also suggest that L-DOPA may decrease anticipation of reward in both visual and frontal areas. However, these findings are difficult to compare with fMRI studies that examined primarily striatal regions.

In sum, we hypothesized that drugs that increase DA function would enhance striatal activity during reward anticipation, and a plurality of studies supports this. Of note, although responses to punishment are not a focus of this article, findings from several studies suggest DA increasing drugs may also increase striatal responses to anticipated punishment.

Seven studies have explored the effects of DA reduction on brain activity during anticipation, and of these, five found decreases in activity, consistent with our hypothesis. Four of these studies utilized fMRI. In one, the antagonist olanzapine (5mg) decreased activity during reward anticipation in the MID task in VS, ACC, and inferior frontal cortex (Abler et al., 2007). In two other studies, APDT reduced reward-related corticostriatal activation during the anticipation period of a rewarded pattern-matching task and the MID task, particularly in the ventral striatum and NAcc (Bjork et al., 2014; Nagano-Saito et al., 2012). Another study using risperidone (0.5mg) found decreases in dorsal striatum, and orbitofrontal cortex during anticipation of reward in a task requiring application of precise amount of force to gain a reward -- although it should be noted that this study omitted a “no reward” condition (Fiore et al., 2018).

In addition to these fRMI studies, one study recorded EEG during a Go/No-Go task. The authors measured the reward-related positivity, probability-related positivity, and P300 waveforms in response to a cue that indicated a button press needed to be made to obtain reward (Schutte et al., 2020a). These ERPs occur in response to rewards or probability of upcoming reward, and thus may be DA-dependent. On each trial, participants were told they could win 0 or 5 Euros and the probability of winning given the correct button press was either 50% or 98%. The reward-related positivity and P300, but not the probability-related positivity, were attenuated by haloperidol, suggesting that DA is involved in the anticipation of rewarding events rather than estimating the probability an event will occur. This pattern was more pronounced in individuals with a higher spontaneous eye blink rate, thought to reflect higher baseline dopamine, consistent with an inverted-U relationship between DA and anticipation.

One study utilizing a-methylparatyrosine (1.5g), a competitive inhibitor of tyrosine hydroxylase, found null results in striatum, but increased BOLD activation in the left cingulate gyrus during anticipation on the MID task (da Silva Alves et al., 2011). Finally there was one contradictory result, in which “low dose” pramipexole (0.5mg) increased NAcc activity during the anticipation phase of the MID, but also weakened connectivity between the NAcc and prefrontal cortex, another key dopaminergic area (Ye et al., 2011). The authors speculated that this pattern could represent opposing effects of pramipexole on tonic and phasic DA, although this possibility cannot be directly tested in humans. Overall, studies on DA reduction generally add support to the idea that there is a relationship between DA levels and striatal activation during anticipation of reward.

E. Anticipation of reward - Summary, conclusions, recommendations

Our hypothesis, based on the putative neurobiology of anticipation, was that pharmacological challenges intended to increase DA would increase anticipation of a rewarding outcome, while conversely, decreasing DA would dampen anticipation of a rewarding outcome. fMRI findings were fairly consistent with this hypothesis, with a plurality of studies supporting the idea that increasing DA enhances striatal activity during anticipation, and reducing DA reduces striatal activity during anticipation. The single existing behavioral study also supported our hypothesis. However, these results are difficult to interpret in the presence of highly mixed self-report findings. It is certainly the case that there were few self-report studies, and most that existed were not directed at investigating self-report outcomes, and thus may not have been powered to examine these outcomes. Indeed, the sole study that primarily focused on self-report found that DA augmentation increased anticipated pleasure (Sharot et al., 2009). Thus, a primary recommendation in this area is to include more sensitive measures of subjective effects, and power studies accordingly.

One other feature of this literature is that the majority of studies of anticipation use the MID. This enables easier comparison across studies, but it also limits the conclusions that can be drawn about reward types. There are also variations among MID tasks that may affect outcomes, such as the timing of measures of “anticipation”, use of gain cues only vs. both loss and gain cues, and size of rewards. Finally, one study indicates that the action requirements of a task (i.e. taking vs. withholding an action) may critically moderate the effects of DA, a suggestion that should be replicated and examined further.

III. Evaluation of Rewards

A. Evaluation of rewards – Definition

We label the second stage of reward processing, as outlined in Figure 1, “Evaluation”. We define evaluation as the process of appraising the relative costs and benefits of a given reward, ending with the formation of a preference (Assadi et al., 2009; Bailey et al., 2016; Zald and Treadway, 2017). Evaluation is also referred to in the literature as reward-related, or cost-benefit, decision-making. As evidence indicates that different brain areas are involved when reward is weighed against costs involving time vs. risk vs. effort (Bailey et al., 2016; Winstanley and Floresco, 2016), we further divide the evaluation section into sub-sections on time, risk and effort-related decision-making.

C. Evaluation of rewards – Predictions based on putative neurobiology

Neurobiological studies suggest that DA manipulations should affect evaluation of reward in humans. Studies in laboratory animals indicate there is both a core DAergic system engaged in all evaluations of reward, and areas uniquely attuned to time, probability and effort costs. Animal studies suggest the “core system” for reward-related decision-making comprises the VTA, NAcc and basolateral amygdala (Bailey et al., 2016). However, the lateral orbitofrontal cortex, prelimbic and infralimbic areas of the medial prefrontal cortex and subthalamic nucleus appear to play more prominent roles in decisions involving delay, while the medial orbitofrontal cortex has a more critical role in decisions about risk, and effort-related decision-making appears to rely more on the ACC (Bailey et al., 2016). Imaging studies in humans converge to a certain extent with these findings in animals. For example, they similarly indicate that the ACC is more heavily implicated in effort-related decisions (Dreher, 2013; Massar et al., 2015; Prévost et al., 2010; Vassena et al., 2014). Human imaging studies also indicate that the vmPFC and VS (which encompasses the NAcc) are sensitive to probability and delay costs (Dreher, 2013; Prévost et al., 2010), although their role in weighing effort costs remains an area of active debate (Chong et al., 2017; Westbrook et al., 2019), thus raising the possibility of divergence from the “core system” identified in animals. These findings emphasize the importance of confirming animal findings in humans, and the need to test the impact of DA challenges on these different costs separately.

Consistent with the idea that the DAergic system is involved in cost-benefit evaluation, the literature in animals indicates that drugs that increase DA generally increase willingness to endure delays, risks, and effort to gain reward; while drugs that decrease DA have the opposite effect (Floresco et al., 2008; Koffarnus et al., 2011; Salamone et al., 2018, 2012, 2009; St Onge and Floresco, 2009; van Gaalen et al., 2006; Wade et al., 2000; Winstanley et al., 2003). However, consistent with findings that there are also cost-specific areas, evidence from animal studies suggests that optimal DA levels may vary across different cost-benefit calculations. For example, high doses of amphetamine decreased willingness to exert effort for reward (consistent with an inverted-U effect), but did not affect willingness to endure delay (Floresco et al., 2008).

Based on the putative neurobiology of reward evaluation, including existence of a general DAergic “core system”, our overall prediction is that pharmacological challenges that increase DA in humans will promote preferences for larger, more costly (more delayed/higher risk/more effortful) rewards, while reducing DA will promote a preference for smaller, less costly rewards. However, in recognition that there may also be cost-specific circuitry and dose-response curves, we review evidence for each type of cost-benefit decision-making separately.

B. Evaluation of rewards - Measures

Reward evaluation in humans has been assessed using self-report, behavioral, and neuroimaging measures. Self-report measures include Visual Analogue Scales (VAS) measuring motivation to perform a rewarded task (Baranski et al., 2004; Volkow et al., 2006). Behavioral measures include discounting tasks, in which individuals repeatedly select between low cost/low reward vs. higher cost/higher reward options. These tasks produce discounting rates, which indicate how quickly the reward is devalued as the cost rises; thus, “increased discounting” indicates less willingness to accept delay, risk or effort to gain the reward (Acheson and de Wit, 2008; de Wit et al., 2002; Kelm and Boettiger, 2013; Massar et al., 2015; Pine et al., 2010). There are also many tasks specific to evaluating decisions about risk vs. reward. One is the Balloon Analog Risk task, in which participants get a small reward for each “pump” added to a balloon, while each pump also increases the risk of the balloon exploding. If a balloon explodes, no reward is given (Acheson and de Wit, 2008; White et al., 2007). Gambling tasks are also commonly used to study the evaluation of risk vs. reward (Ojala et al., 2018; Rigoli et al., 2016b; Rutledge et al., 2015). There are also behavioral measures of effort-related decision-making, which require physical effort (e.g. button-pressing and hand grip strength), or cognitive effort (e.g. set-switching) to gain reward (Kurniawan et al., 2011; Lopez-Gamundi and Wardle, 2018; Reddy et al., 2015). A few behavioral tasks simultaneously manipulate multiple costs, such as the Effort Expenditure for Rewards Task (EEfRT) developed by Treadway et al. (2009), which requires participants to choose between a low effort/low reward option and a high effort/high reward option, both of which vary in probability of a reward being obtained (Treadway et al., 2009). Finally, neuroimaging of the evaluation phase has generally used fMRI to measure BOLD signals during cues that signal the time, risk or effort required to obtain rewards (Botvinick et al., 2009; Croxson et al., 2009; Guitart-Masip et al., 2011; Hauser et al., 2017; Kurniawan et al., 2013, 2010). Until now, most studies on the effects of DA manipulations on reward have used money as the reward. Whether these findings will generalize to other forms of reward remains to be determined.

D. Evaluation of rewards - Summary of studies

i. Evaluation of time vs. reward - Behavioral measures

Five studies have tested the effects of DA augmentation on decisions involving weighing reward against time costs, with conflicting findings. Two studies found decreases in delay discounting, consistent with our hypothesis that DA augmentation would make individuals more willing to endure delays to gain reward. In the first, augmenting DA using d-amphetamine (20mg) decreased delay discounting (de Wit et al. 2002b). In the second, the COMT inhibitor tolcapone (200mg) decreased delay discounting, particularly in individuals with high baseline impulsivity, which was used as a proxy of low baseline DA functioning (Kayser et al. 2012). This would be consistent with an inverted-U relationship between DA and delay discounting.

However, there were also three null findings. In the study that found effects of d-amphetamine on a traditional delay discounting task (which measures the effect of delays up to many months long), d-amphetamine did not affect discounting in an “experiential” task in which participants experienced short delays and small rewards in the lab (de Wit et al. 2002b). An additional study also found no significant effects of either d-amphetamine (20mg) or bupropion (150mg or 300mg) on the traditional delay discounting task (Acheson & de Wit, 2008). This null study tested the same dose of d-amphetamine as the de Wit et al. 2002 study, in a similar number of participants, providing directly contradictory information. Another study found no effect of increasing doses of the novel D1 agonist PF-06412562 (6mg, 15mg, or 30mg) on delay discounting (Soutschek et al., 2020). Finally, a single study found effects opposite to our hypothesis -- increased delay discounting when augmenting DA using L-DOPA (150mg) (Pine et al., 2010). Thus, taken together the literature suggests highly inconsistent effects of DA augmentation on delay discounting, without a clear pattern of moderation by other variables.

Six studies have tested effects of reducing DA on delay discounting, and of these, none found overall increased delay discounting consistent with our prediction. The one study that reported increased discounting in some participants used APTD, and studied individuals who varied in COMT genotype. In that study, the depletion procedure increased discounting only in participants with the val/val COMT genotype (Kelm and Boettiger, 2013). This drug/genotype interaction suggests the importance of genotype, but is otherwise difficult to interpret. The results could be consistent with a “poor get poorer” phenomenon, in which effects of DA depletion are only evident in val/val individuals, who are thought to have with lower prefrontal DA functioning at baseline. However, as noted in the introduction, the val/val COMT genotype may also increase DAergic tone in the midbrain (Corral-Frías et al., 2016), in which case this finding could also be consistent with an inverted-U. Given the small sample size and complexity of interpretation, replication and extension with imaging is needed to fully interpret this result.

Two studies failed to detect any effects of DA reducing drugs: Neither the DA receptor antagonist haloperidol (1.5 mg) (Pine et al., 2010) nor “low dose” pramipexole (0.25 and 0.5mg; Hamidovic, Kang, & de Wit, 2008) altered delay discounting.

Finally, three studies found that DA antagonists actually decreased delay discounting, contradictory to our prediction. The D2 antagonist, metoclopramide (10mg) decreased delay discounting. Although in this study delay was confounded with probability -- delayed rewards were also always more certain than immediate rewards -- computational modeling indicated that metoclopramide primarily reduced the impact of delay on decisions (Arrondo et al., 2015). The D2/D3 antagonist amisulpride (400mg) also decreased delay discounting overall in one study (Weber et al., 2016), and in another decreased delay discounting in individuals with lower BMI (and thus presumably higher amisulpride levels) (Soutschek et al., 2017). In summary, the plurality of studies were inconsistent with our general hypothesis that decreasing DA should lead individuals to prefer less costly, sooner rewards. Indeed in some cases, particularly when D2 antagonists were used, DA reduction appeared to make people more willing to wait for reward.

ii. Evaluation of time vs. reward - Neuroimaging measures

Two studies have used neuroimaging to study the effects of DA augmentation on delay discounting. In one study, the researchers first used computational modeling to identify brain areas associated with delay discounting parameters, including the “discounting factor”, i.e. the extent to which delay devalued rewards, and the “net utility”, i.e. the overall value of an option after discounting. Regions associated with the discounting factor included the striatum, insula, subgenual cingulate, and lateral orbitofrontal cortices, while net utility was related to activity in caudate, insula, and lateral inferior frontal regions. Augmenting DA with L-DOPA (150mg) increased activity in areas associated with the discounting factor, while regions associated with net utility of an option became less active. These imaging results were consistent with behavioral findings from the same study showing that L-DOPA increased delay discounting (Pine et al., 2010). Another study compared regions within a pre-determined frontostriatal mask, and augmented DA with the COMT-inhibitor tolcapone (Kayser et al., 2012). This study found that tolcapone increased activity in similar regions as L-DOPA, including striatum, pregenual cingulate and anterior insula, dorsolateral and medial frontal cortex. However, these increases took place in the context of opposite behavioral findings – a decrease in delay discounting. Taken together, although brain activity changes were fairly consistent, they are difficult to interpret in the presence of opposite behavioral results. The authors of the tolcapone study suggested that because COMT preferentially regulates DA tone in the prefrontal cortex, effects of tolcapone might be specific to the mesocortical DA system. Although regional specificity could explain different behavioral results for tolcapone vs. L-DOPA, this hypothesis is difficult to reconcile with the similar regions identified by imaging in both studies. A direct comparison of different augmentation strategies in the same paradigm and participants is needed to resolve these questions.

Regarding reductions in DA, two studies have examined delay-discounting in conjunction with imaging. The first tested the effect of haloperidol (1.5 mg) on the same brain areas associated with the “discounting factor”, and “net utility” as described above. This study found no effect on BOLD (Pine et al., 2010), consistent with null behavioral results for haloperidol in the same study. Another study examined the effect of reducing DA using the antagonist metoclopramide, using a whole brain approach (Arrondo et al., 2015). This study showed reduced activity in the post-central gyrus in the context of a decrease in delay discounting. Thus, although D2 antagonism has generally produced behavioral reductions in discounting, these behavioral results have not been accompanied by similarly consistent imaging findings.

iii. Evaluation of risk vs. reward - Behavioral measures

Twelve studies have used behavioral measures to investigate the role of DA in evaluation of probability costs vs. rewards. These broadly group into null studies using probability discounting and the Balloon Analog Risk Task (BART), vs. gambling tasks, which show more complex findings.

Examining probability discounting and the BART, there have been three studies of DA augmentation, all with null results. Two studies examining probability discounting (Acheson and de Wit, 2008; de Wit et al., 2002) with two different DA augmentation strategies (d-amphetamine 10mg and 20mg, bupropion 150mg and 300mg) found no effects. One of those studies also investigated the BART with similar null results (d-amphetamine 20mg, bupropion 150mg and 300mg) (Acheson and de Wit 2008b). Another study found null effects of d-amphetamine (20mg) on the BART overall, but these were moderated by both gender and trait reward sensitivity, such that d-amphetamine increased risk-taking in males with high reward sensitivity, decreased risk-taking in males with low reward sensitivity, and did not affect risk-taking in women (White et al., 2007). If reward sensitivity is interpreted as a proxy of baseline DA, these results could be partly consistent with a “rich get richer” phenomenon, but the further interaction with gender is not straightforwardly explainable.

In contrast to these generally null results for discounting paradigms and the BART, gambling tasks show more complex effects of DA increasing drugs. Seven studies have examined gambling tasks with DA increasing drugs. Two found increases in risk-taking when enhancing DA using L-DOPA (both 150mg), which is consistent with our overall hypothesis, although in one study this effect was only evident in individuals with low body weight (Rigoli et al., 2016b; Rutledge et al., 2015). Two found null results, such that methylphenidate (40mg) and L-DOPA (100mg) did not change risk-taking (Evers et al., 2017; Symmonds et al., 2013). Contrary to our hypothesis, one study found that increasing doses of the selective D1 receptor agonist PF-06412562 (6, 15, or 30 mg) reduced preferences for high-risk gambles (Soutschek et al., 2020).

The final two studies found more complex effects not accounted for by our hypothesis. In one study, augmenting DA using methylphenidate (20mg) increased risky “double or nothing” choices for high stakes bets, but decreased them for low stakes bets (Campbell-Meiklejohn et al., 2012). The authors suggest this pattern represents insensitivity to loss, which could reduce “loss chasing” for smaller losses, but also reduce the perceived risk of higher value “double or nothing” gambles. The second study assessed the effects of cabergoline (1.5mg) on an “experimental gamble” task, in which participants selected repeatedly between a control gamble, with 50:50 odds of 10 point gain vs. a 10 point loss and an “experimental gamble”, with odds ranging from 40–60% and gains and losses ranging from 30 to 70 points. It is difficult to interpret selection of the experimental option as risk taking, given that at times the experimental option had a higher probability of reward. However, cabergoline did decrease the extent to which losses impacted subsequent choice of the experimental option. This is broadly consistent with the prior loss chasing in indicating that increasing DA decreases the impact of loss. Cabergoline also increased selection of higher probability gambles, while decreasing selection of lower probability gambles (Norbury et al., 2013). This is consistent with a reduction in “probability distortion”, a common psychological error in which individuals generally underestimate the likelihood of high probability events, and overestimate the likelihood of low probability events.

In summary, the effects of augmenting DA on probability costs appear to depend heavily on type of task and parameters examined. Probability discounting paradigms and the BART show generally null results. Gambling tasks show mixed evidence. Some studies report increased risk taking, which may be driven by decreased sensitivity to potential losses. This is consistent with our general hypothesis that DA augmentation would reduce sensitivity to probability costs. However, gambling tasks also indicate that DA augmentation may increase use of probability information in decision-making, a different effect than predicted.

Seven studies have examined the impact of reductions in DA on evaluation of risk vs. reward. Consistent with null results for probability discounting tasks and DA augmentation, two studies using probability discounting paradigms with DA reduction also showed null results. The first, Arrondo et al. (2015) was discussed previously under delay discounting, as in this study delay was confounded with probability. Computational modeling of this task indicated that metoclopramide (10mg) affected delay, but not probability parameters. In the second study, “low dose” pramipexole (0.25mg and 0.50mg) did not affect probability discounting or the BART (Hamidovic et al., 2008).

Examining the five studies that tested gambling tasks and DA reduction again shows more complex picture, with similar results to DA augmentation. None of the results were consistent with our general hypothesis that DA reduction should produce risk-aversion. One study had null results, such that haloperidol (3mg) did not alter betting behavior on a slot machine game (Zack and Poulos, 2007). The other four showed either results directly contradictory to our hypothesis, or more complex effects. One study using “low dose” pramipexole (0.5mg) found increased risk taking, particularly when the previous trial resulted in a large gain -- a situation that usually elicits conservative decision-making (Riba et al., 2008). A further study with “low dose” pramipexole (0.5mg) found that pramipexole increased risky “double or nothing” choices for high stakes bets, but decreased them for low stakes bets (Campbell-Meiklejohn et al., 2011). This pattern was similar to results for methylphenidate on the same task, and the authors interpreted these findings as suggestive of a similar insensitivity to losses. Two additional studies indicated that reducing DA using D2/D3 antagonists also reduces “probability distortion”, similar to results found for the agonist cabergoline. In one study, reducing DA using amisulpride (400mg) increased general preference for risky options in a gambling task, while also decreasing probability distortion (Burke et al., 2018). Another study using a similar antagonist, dose and modeling parameters (sulpiride, 400 mg) did not find alterations in risk preference overall, but did find similar reductions in probability distortion (Ojala et al., 2018). In summary, reducing DA does not appear to consistently decrease risk taking. Instead, there is evidence for a potential increase, and for more subtle effects on impact of loss and use of probability information. The similarity of these findings to results for DA augmentation may be explainable in two ways. First, it is possible that the either the putative augmentation or reduction strategies described above are actually engaging pre-synaptic mechanisms, producing similar results. Second, it may be the result of an inverted-U relationship between DA and these outcomes. Either possibility strongly indicates the need for more dose-response studies in humans.

iv. Evaluation of effort vs. reward – Self-report measures

One study has tested the effect of augmenting DA on self-reported willingness to exert effort for a reward. In this study, methylphenidate (20 mg) increased self-reported motivation to complete a rewarded cognitive effort task (Volkow et al., 2004), consistent with our hypothesis.

v. Evaluation of effort vs. reward – Behavioral measures

Four out of the five studies examining the effect of DA enhancement on behavioral measures of evaluation of effort were consistent with our hypothesis that increasing DA should increase willingness to exert effort for reward. Increasing DA using d-amphetamine (20 mg) increased preference for high effort/high reward choices on the EEfRT task (Wardle et al., 2011). Of note, this increase was most evident under low probability of reward, and the high effort task also took longer to execute, so contributions of probability- and time-related decision-making cannot be ruled out. However, enhancing DA using L-DOPA (125 mg), this time in a task where time to complete the task was fixed and rewards were not probabilistic, also increased willingness to engage in physical effort for reward (Zenon et al., 2016). Similarly, augmenting DA using a D1 agonist (PF-06412562; 6mg, 15mg, 30mg) increased willingness to exert physical effort for reward (Soutschek et al., 2020). Finally, both methylphenidate (20mg) and “low dose” sulpride (400mg) increased willingness to exert cognitive effort for reward in an effort discounting paradigm (Westbrook et al., 2020). However, these effects were only evident in individuals with low baseline DA synthesis capacity (as measured by PET), suggesting an inverted-U relationship. Only one study, using L-DOPA (150mg), found no changes in willingness to exert grip-strength effort for reward (Michely et al., 2020).

Finally, four studies involving reduction of DA provide limited support for the complementary idea that depletion of DA reduces willingness to exert effort for reward. In one study women with subsyndromal seasonal affective disorder exhibited lower progressive ratio (PR) break points after APTD (Cawley et al., 2013). APTD also reduced the willingness of light smokers to work for cigarettes (Venugopalan et al., 2011) and for healthy controls to work for exercise (O’Hara et al., 2016) in a PR task. In another study (Michely et al., 2020) haloperidol (1.5mg) failed to reduce willingness to exert effort, but instead narrowed the gap in effort exerted on high vs. low reward trials, seemingly reducing sensitivity to differences in reward magnitude. In summary, the number of studies in this area is smaller, but findings suggest that modulating DA affects decisions about whether to put forth effort for reward.

vi. Evaluation of effort vs. reward – Neuroimaging measures

To our knowledge, no studies have conducted neuroimaging during evaluation of effort costs with a dopaminergic manipulation. However, PET measures of methylphenidate-induced increases in striatal DA during anticipation of a rewarded cognitive effort task were positively associated with self-report measures of motivation to engage in effort, suggesting that increases in DA relate to reward motivation (Volkow et al., 2004).

E. Evaluation of reward - Summary, conclusions, recommendations

We hypothesized that augmenting DA would generally increase willingness to endure time, risk and effort to gain reward, while reducing DA would reduce willingness to accept these costs in order to gain reward. However, we examined reward-related decisions involving tradeoffs for time, risk, and effort separately, as they are thought to have partially separable neural bases. Consistent with the idea that these domains of decision-making are distinct, we saw different patterns of results for decisions about time, risk and effort.

In the context of evaluating time vs. reward, the literature did not support our hypothesis. Augmenting DA had extremely mixed results, and a slim plurality of studies suggested that reducing DA actually had the opposite of the predicted effect, increasing the willingness of individuals to spend time waiting for rewards. Imaging studies in this area were too few and disparate in approach reach definitive conclusions.

In the context of evaluations of risk vs. reward, probability discounting procedures and the BART generally yielded null results. The apparent insensitivity of these tasks to DA manipulation may be explained by recent work in the animal literature showing that DA plays a differential role in choices involving reward omission vs. loss (Orsini et al., 2015). Discounting and BART procedures did not involve possibility of loss, while the gambling tasks generally did. Future studies in humans should attend to this aspect of probability tasks when using DA manipulations.

Gambling tasks revealed more complex effects of DA manipulation, suggesting that both DA augmentation and reduction may increase risk taking, reduce the impact of losses on choice, and ameliorate a common cognitive distortion in which individuals overestimate the probability of low chance events and underestimate the probability of high chance events. The similarity of results for augmentation and reduction may be due to either paradoxical pre-synaptic effects of DA augmentation/reduction, or represent an inverted-U relationship between DA and evaluation of probability costs. Attending to more subtle effects of DA on constructs such as loss sensitivity, along with incorporating the most recent research from the animal literature on the modulatory impact of the possibility of loss, could help clarify findings.

Finally, there was more general support for a direct relationship between DA levels and willingness to exert effort to gain reward, with the exception of two contradictory results. However, this area of work also contained the fewest studies, and is most in need of replication and extension.

IV. Execution of Actions to Attain Reward

A. Execution - Definition

We label the third stage of reward processing, as outlined in Figure 1, “Execution”. We define execution as the mobilization of resources to obtain a selected reward (Bailey et al., 2016). Execution has also been referred to as invigoration or response vigor (Kurniawan et al., 2011; Salamone et al., 2018; Treadway et al., 2009). Execution begins when a preference is generated by the evaluation phase and ends with attainment (or failure to attain) of the selected reward. Thus, in this section we cover measures taken after an explicit choice has been made, or that do not require explicit choice. There are three related theories about the role of DA in execution: 1. “Simple invigoration” – This hypothesis states that DA increases the speed or vigor of actions executed to obtain reward. Reward-related invigoration may be difficult to separate from simple locomotor effects (Bailey et al., 2016). Thus, we focus on studies that include both reward and no reward/punishment trials, and omit studies where these comparison conditions are unavailable (Wardle et al., 2011), or not analyzed (Frank and O’Reilly, 2006). 2. “Average reward rate” – This hypothesis states that DA levels represent the average rate of rewards that the organism has experienced in the environment. In this hypothesis, increased DA increases vigor by signaling that there will be lost opportunities if the organism does not quickly execute and “move on” from the current task to gain other available rewards (Niv et al., 2007). 3. “Go to win” - As described briefly in the “Anticipation” section, this theory states that DA facilitates a prepared, Pavlovian association between taking action and reward, or “go to win” actions (Guitart-Masip et al., 2011; Swart et al., 2017), as opposed to the inhibition of action for reward (i.e., “no-go to win”). Evidence related to these three theories will be covered in separate sub-sections.

B. Execution - Predictions based on putative neurobiology

Neurobiological studies suggest that DA manipulations should affect execution. (Niv et al., 2007; Salamone, 1992; Salamone and Correa, 2012). DA projections from the VTA to the NAcc appear critical to the vigor of responding for a reward (de Jong et al., 2015; Salamone et al., 2018, 2016), with some evidence suggesting that DA in the NAcc core, as opposed to NAcc shell, has the more critical role (Bailey et al., 2016). In humans, the VTA/NAcc circuit has also been implicated in execution of reward-related behaviors, with activations in the NAcc and VTA predicting degree of effort exerted (Rigoli et al., 2016a). In addition to the VTA and NAcc, the ACC, and particularly the dorsal ACC, has been implicated in sequencing, initiating and maintaining goal-directed behaviors (Assadi et al., 2009; Kurniawan et al., 2011). Indeed, it has been hypothesized that the “lower tier” of the dorsal ACC, which connects to the NAcc and hypothalamus, is particularly important in the execution phase of reward (Assadi et al., 2009). Regarding DA dynamics, the “average reward rate” hypothesis suggests that low, tonic levels of accumbens DA specifically signal the reward rate (Assadi et al., 2009; Niv et al., 2007).

Regarding the effect of DA manipulations, computational re-analyses of studies in which animals must exert effort to gain reward (Niv et al., 2007) suggest dopamine depletion primarily reduce vigor, as indicated by slower response speeds. These manipulations include specific lesions, but also systemic administration of haloperidol.

Based on the putative neurobiology of execution, we expect that augmenting DA will generally invigorate reward-related actions and increase “go” responding, whereas depleting DA will slow reward-related actions and decrease “go” responding. However, in recognition that the “simple invigoration”, “average reward rate” and “go-to-win” hypotheses presented differ somewhat in exact focus and predictions, we discuss each in separate sections.

C. Execution - Measures

Execution in humans has been assessed using self-report, behavioral and neuroimaging measures. Self-report measures consist of VAS measuring perceptions of how much effort was actually exerted on tasks (as distinct from measures of motivation or willingness to exert effort, which were summarized in “Evaluation”; Baranski et al. 2004; Volkow et al. 2006). Behaviorally, execution is generally operationalized as speed of response on a task offering a reward (Beierholm et al., 2013; Croxson et al., 2009; Guitart-Masip et al., 2014, 2012a, 2011). For example, we consider reaction times (RTs) on the MID task a measure of execution. Studies that investigate neural correlates of execution use fMRI to measure BOLD response during execution of tasks required to obtain reward (Croxson et al., 2009; Dean et al., 2016; Schmidt et al., 2012).

D. Execution - Summary of studies

i. Execution - “simple invigoration” – Self-report measures

Although dopaminergic stimulants generally increase self-reported overall feelings of “vigor”, we examined only self-reports of vigor directed at attaining a reward. Two studies have tested such measures. In one, consistent with our hypothesis, modafinil (20mg) increased retrospective self-reported ratings of the effort individuals put into a rewarded task (Funayama et al., 2014). However, in a second, bupropion (150mg) had no effect on a similar retrospective rating of effort (Ikeda et al., 2019).

ii. Execution - “simple invigoration” – Behavioral measures

The nine studies that have utilized behavioral measures generally do not support the hypothesis that increasing DA should increase speed or vigor of responses made for reward. Only one study was wholly consistent. In this study, participants needed to stay above a certain grip strength threshold to earn a reward (Michely et al., 2020). The authors measured “excess” grip strength applied beyond the threshold, on the basis that this represents implicit, unconscious invigoration, rather than explicit decisions about how to allocate effort. L-DOPA (150mg) increased “excess” grip strength in high reward trials, as compared to low or uncertain reward trials. In contrast, seven studies reported no effect of increasing DA on reaction time. In seven studies reporting raw reaction times on the MID or a similar task, modafinil (20 mg), L-DOPA (150mg and 100mg), methamphetamine (20mg), methylphenidate (0.5 mg/kg) and bupropion (150mg) did not affect the speed of responses on reward trials vs. non-reward or loss trials (Apitz and Bunzeck, 2014; Funayama et al., 2014; Graf et al., 2016b; Ivanov et al., 2014; Mayo et al., 2019; Weis et al., 2012; Wittmann et al., 2015). In a sub-acute study by Dean et al. (2016), augmenting DA over the course of a week (bupropion, 150 mg daily) also did not alter response speed in a button-pressing task to either obtain a pleasant taste or avoid an aversive taste. A single study suggests that some of these null results could be due to failure to account for baseline levels of DA. This study found that augmenting DA using bromocriptine (1.5mg) sped performance on rewarded trials of a reaction time task for individuals with the A1A1 DRD2 genotype, which is associated with reduced DA transmission at baseline, while impairing performance in all other genotypes (Kirsch et al., 2006). In summary, the effects of DA augmentation are generally null, although one study suggests a more complex inverted-U pattern in which only individuals with low baseline DA may be invigorated by DA increasing drugs.

The seven studies examining the effect of reducing DA similarly were dominated by null results. Only one was consistent with our broad hypothesis that reducing DA should reduce speed of rewarded responses. In this study, reducing DA using an atypical antipsychotic (olanzapine, 5mg) slowed reaction times to rewarded trials specifically (Abler et al. 2007). In contrast, three studies reducing DA using amino acid mixtures (AMPT, APTD) found no effects on speed of rewarded responses (Bjork et al., 2014; da Silva Alves et al., 2011; Nagano-Saito et al., 2012). Risperidone (0.5mg) also did not change the speed with which participants reached a pre-specified grip strength required to obtain reward (Fiore et al., 2018). Haloperidol (1.5mg) did not significantly reduce “excess” grip strength in the same grip strength paradigm described above (Michely et al., 2020). Haloperidol (2mg) also did not slow responses to rewarded vs. non-rewarded trials on a go/no-go task (Schutte et al., 2020b). This evidence suggests that augmenting and reducing DA in healthy adults has limited effects on simple vigor of responses for reward.

iii. Execution - “simple invigoration” – Neuroimaging measures

Two pharmaco-fMRI studies have investigated brain activity during execution of reward-related behaviors. As described above, Dean et al. (2016) used sub-acute administration of bupropion (150mg daily) over the course of a week, and required participants to complete either an easy or hard task to either gain a pleasant taste or avoid an unpleasant one. Bupropion increased BOLD signal in the caudate, ventromedial prefrontal cortex, dorsal ACC/paracingulate gyrus and putamen during easy trials to obtain a pleasant taste, compared to hard trials to obtain a pleasant taste. The selected contrast, between high and low effort reward trials, makes these results somewhat difficult to interpret, but broadly, increasing DA levels did potentiate activity in DAergic and reward-related areas when required effort was low (Dean et al., 2016).

Regarding DA reduction, Fiore et al. (2018) examined activity in the internal and external globus pallidus, and substantia nigra pars reticulata during the exertion of grip for reward, under risperidone (0.5mg). They found that risperidone decreased activity during low reward (vs. high reward) trials in the external globus pallidus, but increased activity to low reward (vs. high reward) trials in the substantia nigra. This complex regional interaction effect suggests that finer parsing of the striatum may be necessary to understand the effects of DA on execution of actions to gain reward.

iv. Execution - “average reward rate” – Behavioral measures

As noted above, Niv et al. suggested that DA levels represent the average local reward rate as an indicator of opportunity cost, rather than being related to the reward available for a given trial (Niv et al., 2007). This would predict a differential effect of DA manipulations based on average rate of reward available/opportunity cost, rather than the reward available for a given trial. Only two studies have tested this hypothesis, both using DA augmentation. The first was consistent with this hypothesis. In Beierholm et al. (2013), participants were presented with a varying reward and required to respond to a simple cognitive effort task within 500ms to attain the reward. Computational modeling of responses found that average reward rate was positively related to response times, and this relationship was stronger after augmenting DA with L-DOPA (150 mg). However, these findings contrast with those reported by Zenon et al. (2016). In this study, participants were required to complete two types of handgrip tasks in exchange for monetary reward. In one, exerting more effort allowed faster completion, and thus effort was related to opportunity cost. In the other, the time it took to complete the task was fixed, but amount of reward received was proportional to the amount of effort exerted, so effort was unrelated to opportunity cost. Contrary to what would be expected based on the theory of Niv et al., augmenting DA (L-DOPA, 125mg) increased effort on the task where more effort directly related to more reward, but did not affect effort when exerting more effort allowed the person to save time (Zenon et al., 2016). The authors interpreted these findings as indicating that DA may change evaluation of the cost of effort vs. reward (and these results were thus also discussed in “Evaluation”), but does not alter vigor. These studies differed in a number of ways (task complexity, cognitive vs. physical effort requirements) making them difficult to compare, but based on this small literature, there are conflicting findings regarding whether DA manipulation in humans affects vigor by enhancing a DAergic signal representing average rate of reward.

v. Execution - “go to win” – Behavioral measures

As described above, the possibility that DA controls a Pavlovian, prepared “go to win” response can be investigated in orthogonalized tasks that contain both “go” and “no-go” trials to both obtain reward and avoid loss. Two studies have investigated such tasks with null results. In one, augmenting DA with L-DOPA (150 mg) sped response times to all go trials on an orthogonalized go/no-go task, regardless of outcome valence, suggesting a simple psychomotor effect rather than “go to win” enhancement (Guitart-Masip et al., 2012a). In another, Swart et al. found that methylphenidate (20mg) did not speed reaction times overall, or on to “go to win” trials specifically. Of note, the task used by Swart et al. also required learning of action and outcome contingencies, which may have affected results (learning outcomes from this task will be discussed in “Learning”).

Finally, one study examined DA reduction. This study used a Pavlovian-to-instrumental transfer design to investigate the effect of DA on conditioned associations between reward cues and action (Weber et al., 2016). Following classic animal paradigms, healthy adults were first trained to press a button for a chocolate candy reward. They then completed Pavlovian conditioning in which one visual stimulus (the CS+) was associated with delivery of the chocolate reward, and one visual stimulus (the CS−) was presented the same number of times without reward delivery. Finally, the two visual stimuli were re-presented during a period in which participants were free to press the button, but no rewards were delivered. During this period, there was less button pressing to the CS+ under amisulpride (400mg), consistent with the idea reducing DA impairs Pavlovian cue-action associations. In summary, there are comparatively few studies evaluating the “go to win” theory, and the disparate results and outcome measures used (reaction times, accuracy, Pavlovian-to-instrumental transfer) make it difficult to draw conclusions.

E. Execution - Summary, conclusions, recommendations

Overall, there is less evidence for the role of DA in the execution and invigoration of goal-directed behaviors. Little self-report work exists. There is also a lack of studies using imaging to elucidate the neurobiology of response vigor in response to DAergic manipulations. Those that do, suggest that finer anatomical resolution may be needed to capture the complex effects of DA on execution for reward. Thus, most of the evidence comes from behavioral studies. Regarding the “simple invigoration” theory, the majority of behavioral findings are null, suggesting research should focus on the more sophisticated “average reward rate” and “go to win” hypothesis. However, the available behavioral evidence testing whether vigor is linked to a DAergic signal capturing average reward rate yields conflicting results, as do the few behavioral studies investigating the “go to win” hypothesis.

There are notable weaknesses in existing behavioral evidence. First, the majority of behavioral studies reviewed under the “simple invigoration” theory were providing secondary results in the context of fMRI, so these tasks might not be particularly sensitive to behavioral changes. Second, we compared behavioral results from a variety of measures with different types of response requirements and structures. The “go-to-win” theory suggests that different types of response requirements may not be comparable – in particular, performing an action may differ qualitatively from having to withhold an action to gain a desired outcome.

V. Pleasure

A. Pleasure - Definition

We label the fourth stage of reward processing, as outlined in Figure 1, “Pleasure”. We define pleasure as the hedonic impact of receiving a reward, which has also been called “liking” or “consummatory reward” (Berridge, 1996). The hedonic impact of anticipating a future reward was covered in “Anticipation”, and adjustments made to behavior based on receiving a rewarding outcome are covered in “Learning”.

B. Pleasure - Predictions based on putative neurobiology

Unlike the other processes covered in this review, the known neurobiology of pleasure does not suggest a strong role for DA. Although DA was originally thought to play an important role in pleasure (Olds and Milner, 1954), current evidence shows that opioid “hot spots” in the ventral striatum, pons, and orbitofrontal cortex, are responsible for the hedonic effects of rewards (Barbano and Cador, 2006; Berridge and Kringelbach, 2015; Colasanti et al., 2012). Although these hot spots are co-located with DA reward-related circuitry, and may have mutual influence with DA circuits, they are opioidergic (Castro and Berridge, 2014; Ho and Berridge, 2013).

Consistent with this, in animals, opioid agonists such as morphine increase the hedonic liking reaction to sweet and even bitter tastes (Barbano and Cador, 2006; Doyle et al., 1993; Murphy et al., 1990), and this occurs in the hedonic hot spots in the NAcc and ventral pallidum (Smith and Berridge, 2007). However, dopaminergic drugs do not affect taste palatability in rodents – for example, pimozide had no effect on liking reactions in rats (Peciña et al., 1997) and tetrabenzadine similarly did not affect hedonic responses (Pardo et al., 2015). These findings suggest that dopaminergic drugs may have less effect on liking or pleasure compared to other aspects of reward (Berridge and Kringelbach, 2015).

Based on the putative neurobiology of pleasure, we expect that DA augmentation and reduction should have comparatively less effect on pleasure, in contrast to other reward-related functions. However, we note that in typical fMRI paradigms it is difficult to separate the hedonic impact of receiving a reward from other processes with DAergic elements, including reward learning. Therefore, we will note when imaging studies included measures of self-reported pleasure, and particularly when these were correlated with imaging results.

C. Pleasure - Measures

Pleasure in humans has been assessed using self-report, behavioral and neuroimaging measures. Self-report measures include the Profile of Mood States (POMS) subscales of elation, positive mood, or total mood disturbance (McNair et al., 1971), and Addiction Research Center Inventory (ARCI), which measures prototypical effects of abused drugs (Martin et al., 1971). In this review, the ARCI generally appears in studies where the goal is to measure how one drug moderates the pleasurable effects of another drug, and the two subscales of most interest are the amphetamine (A) scale and morphine-benzedrine (MBG) scale, which both capture euphoric drug effects. Pleasure can also be measured by VAS ratings after receiving a reward, e.g. of happiness, liking and enjoyment. Behaviorally, electromyography (EMG) of the facial muscles (i.e., zygomatic “smile” muscle and corrugator “frown” muscle”) can measure small facial movements in response to receiving rewards. Neuroimaging of pleasure has generally used fMRI to measure BOLD response at the time of receiving a reward. ERPs are also commonly used to measure brain activity during reward receipt; however, most ERP studies employ reward-learning paradigms comparing receipt to anticipation, rather than focusing on primarily on activity at the time of receipt of reward, and thus are covered in the “Learning” section.

While many studies of pleasure use monetary outcomes, other rewards have also been used, e.g., sweet tastes, food pictures, erotic images, drugs of abuse, so type of reward will be noted as a potential moderator.

D. Pleasure - Summary of studies

i. Pleasure – Self-report measures

While it is well documented that many DA-releasing drugs induce euphoria (Boileau et al., 2007; Brauer and de Wit, 1995; Childs and de Wit, 2009; Clatworthy et al., 2009; de Wit et al., 2002; Leyton et al., 2002; Sevak et al., 2009; Weafer et al., 2017; White et al., 2006), here we focus on how dopaminergic drugs affect pleasure in response to other rewards. There have been five studies examining effects of DA augmenting drugs on self-reported pleasure in response to rewards. Only one showed the null effects we would predict. In this study, liking of pleasant and unpleasant tastes was unchanged by 150mg bupropion (Dean et al., 2016). In contrast, two other studies showed a “flattening effect” -- under placebo, pleasure (i.e., excitement or happiness) was higher after receiving large monetary rewards compared to receiving small rewards or losing (Knutson et al., 2004; Rutledge et al., 2015), but under d-amphetamine (.25mg/kg) or L-DOPA (150mg), pleasure was similar after receiving small or large outcomes (and in fact, after wins vs. losses). In both cases, DA increasing drugs increased pleasure/excitement for events that were not previously pleasurable, while perhaps slightly decreasing responses to previously highly pleasurable events. Similarly, in another study d-amphetamine (20mg) increased pleasantness ratings equally across positive, neutral and negative emotional images. This is not exactly a “flattening” pattern, but again suggests (in a different reward type) that DA augmentation might boost the pleasurable effect of stimuli not typically considered pleasant (Wardle and de Wit, 2012). Finally, one study reported that L-DOPA (100mg) increased liking ratings in response to listening to music (Ferreri et al., 2019). Thus, the modal finding was suggestive of “flattening” of pleasure.

Results from 14 studies examining DA reduction were more supportive of our overall hypothesis that DA manipulations should not strongly impact pleasure. These results were also consistent across a variety of reward types. DA reduction with both APTD and haloperidol (3mg) did not significantly affect positive mood and elation in response to monetary wins in gambling tasks (Bjork et al., 2014; Zack and Poulos, 2007). Null results were also seen in two studies examining the effects of DA antagonists on the euphoric effects of other drugs. Pimozide (4mg) and fluphenazine (3mg and 6 mg) did not reduce the euphoric effects of d-amphetamine (Brauer and de Wit, 1995) and haloperidol (3mg) had no effect on the positive mood effects of methamphetamine (Wachtel et al., 2002). Similarly, three studies reported no effects of ATPD on subjective responses to d-amphetamine (Leyton et al., 2007), alcohol (Barrett et al., 2008), and cocaine (Leyton et al., 2005) when administered to individuals without substance use disorders. ATPD also did not alter responses to nicotine (Venugopalan et al., 2011). Reduction of DA (via APTD, 400mg sulpiride, and .25mg pramipexole) also had no effect on liking ratings of pleasant and unpleasant food (Frank et al., 2016; Hardman et al., 2012; McCabe et al., 2013, 2011); although, some of these null self-report effects were accompanied by significant changes in brain responses. One study reported that haloperidol (3mg) had no effect on affective ratings of sexual stimuli. Only one study reported an effect. Risperidone (2mg) decreased pleasure in response to listening to music (Ferreri et al., 2019). To summarize, DA reduction appears to have a small, if any, effect on self-reported pleasure in response to rewards, based on 13 out of 14 studies reporting null results.

ii. Pleasure – Behavioral measures

Only two studies have investigated the effects of DA drugs on pleasure using behavioral measures. In contrast to our hypotheses, both suggested that DA augmentation modulates physiological responses to reward. Specifically, d-amphetamine increased zygomatic (smile muscle) activity (at 10 mg) and decreased corrugator (frown muscle) activity (at 20 mg) to positive images, indicating enhancement of positive responses (Wardle and de Wit, 2012). More research using facial muscle activity is needed to confirm these results. Another study looked at the effect of L-DOPA (100mg) and risperidone (2mg) on electrodermal activity while listening to music. L-DOPA increased, while risperidone decreased, electrodermal activity during highly pleasing portions of music, compared to more neutral portions (Ferreri et al., 2019). These behavioral measures suggest DA manipulations may have a stronger effect on physiological or behavioral measures rather than self-report measures of pleasure, but more research is needed.

iii. Pleasure – Neuroimaging measures

DA augmenting drugs appear to have mixed effects in the eight neuroimaging studies that have investigated responses during receipt of rewards. Only one of these found null results consistent with our hypothesis. d-Amphetamine (.25mg/kg) had no significant effect on BOLD activity in any brain area in response to receiving monetary rewards (Knutson et al., 2004). Two studies reported results consistent with the “flattening” effect seen in self-reported pleasure. The first reported that bupropion (150mg) increased BOLD activity in the orbitofrontal cortex to both pleasant and unpleasant tastes, but reduced BOLD in the caudate to pleasant tastes (Dean et al., 2016). Ventral striatal and amygdala activity were also increased in response to unpleasant tastes only. No DA effects on self-reported liking were found in this study and the task contained multiple phases of reward; therefore, it is unclear whether this brain activity reflects pleasure or other task-related processes. In the second, DA augmentation (d-amphetamine; .5mg/kg) decreased dorsal ACC activity to sucrose, compared to water (Melrose et al., 2016). Yet while dorsal ACC activity was “flattened”, insular activity showed the opposite effect, increasing to sucrose compared to water under d-amphetamine.

Three studies using a variety of DA increasing drugs and rewards have reported results consistent with enhancement of brain activity during receipt of reward. Increasing DA with bupropion (150mg) increased activity in the NAcc, posterior midcingulate, mediodorsal thalamus, and amygdala in response to erotic videos (Abler et al., 2012, 2011), d-amphetamine (20mg) increased responses to receiving a monetary reward in the amygdala (O’Daly et al., 2014), and L-DOPA (100mg) increased responses to erotic images in the NAcc, and dorsal ACC (N. Y. L. Oei, Rombouts, Soeter, van Gerven, & Both, 2012). Of these studies, only Abler et al. (2011) reported relationships with subjective variables. Abler et al. (2011) found no correlation between subjective sexual functioning and the increase in response to erotic videos seen with bupropion.

Finally, one study reported decreased responses to reward. Methylphenidate (0.5 mg/kg) decreased activity in the mid occipital and left inferior parietal cortices to receipt of monetary reward (Ivanov et al., 2014).

Thus, for DA increasing drugs, imaging produces mixed results, with the brain areas and direction of effect inconsistent across studies. Further, because of task-related factors, such as the inclusion of an effort phase in Dean et al. (2016) or expectation and conflict manipulations in Ivanov et al., (2014), and the comparatively small number of studies investigating subjective responses in conjunction with imaging, it is difficult to discern if brain effects relate to pleasure or other processes. Thus, there is limited evidence that drug intended to increase DA enhance pleasure via changes in ventral striatal and/or orbitofrontal cortex activity.

Nine studies have examined the effect of DA reduction on brain responses to receipt of rewards. Three of these studies were consistent with our hypothesis in reporting no significant changes in BOLD response in any brain areas after DA reduction (Bjork et al., 2014; Graf et al., 2015; Sescousse et al., 2016), and one of these was accompanied by similarly null self-report findings (Bjork et al., 2014). Two of the null studies found that DA reduction using APTD and sulpiride (400mg) had no effect on BOLD activity in response to receiving monetary rewards (Bjork et al., 2014; Sescousse et al., 2016), while the third found no significant effect of amisulpride (200mg/day over 7 days) on BOLD to an erotic video (Graf et al., 2015). However, five of the remaining studies generally favor decreased activation, most commonly in the ventral striatum and prefrontal cortex, across a variety of interventions including APTD, “low dose” pramipexole (0.25 and 0.5mg), sulpride (400mg) and haloperidol (3mg), across food, monetary and sexual stimuli (Frank et al., 2016; McCabe et al., 2013, 2011; Oei et al., 2012; Riba et al., 2008). Notably, all three studies using food stimuli saw decreases, while other stimuli showed more mixed effects. However, the three food studies also saw no accompanying effects on self-report. Finally, one study reported that DA reduction with “low dose” pramipexole (0.5mg) actually increased activation of the NAcc in response to rewards (Ye et al., 2011). To summarize, modal results suggest DA reduction decreases activity in the striatum and frontal cortex during receipt of reward. Of note, the reward type most consistently accompanied by decreased BOLD response with DA reduction was food reward. Studies investigating other reward types had more inconsistent results. It is notable that reductions in brain activation in response to rewards do not always correspond to reductions in the hedonic experience of receiving a reward. Although more research is needed on the relationship between neural responses to reward and subjective experience, the findings to date suggested that these measures reflect different underlying processes, both involving DA. It is also possible that different outcomes are related to methodological factors because many processes are involved in both reward valuation (e.g., evaluation and learning) and in self-report measures (e.g., expectancies).

E. Pleasure - Summary, conclusions, recommendations

In conclusion, DA’s role in the hedonic impact of receiving a reward is still unclear. Self-report results indicated that increasing DA may “flatten” subjective pleasure, by increasing it in response to otherwise non-rewarding stimuli, and possibly slightly decreasing responses to highly rewarding stimuli. In contrast, decreasing DA appears to have little effect on self-reported pleasure. In general, both DA enhancement and reduction modify brain activity in response to rewards, most often in the NAcc, amygdala, dorsal ACC and orbitofrontal cortex, but the direction of effect is sometimes inconsistent. Both manipulation and reward type appear to have an impact here, as studies of DA reduction with food rewards had the most consistent findings of reduced activation. Further, it is often unclear if brain activity in response to receiving a reward actually reflects hedonic impact, as many studies do not report correlations between self-report ratings of pleasure and brain activity in response to receiving a reward, and those that do had largely inconsistent results when comparing self-report to imaging.

This summary suggests several recommendations for future research. First, researchers should include both subjective and objective measures when assessing pleasure in response to a reward. Studies that include both subjective and objective responses to rewards are sparse, and those that do, typically do not report the correlation between the two. One advantage of DA manipulation studies in humans is the ability to measure subjective impact of the underlying processes of interest. Second, more studies are needed that measure pleasure in response to a reward without confounding factors, such as manipulations of expectations. Simpler designs, especially when measuring brain activity, would help parse out brain responses to other processes such as learning and prediction. Third, varying levels of reward and loss should be included to assess whether effects on self-reported pleasure or brain responses change with the magnitude of reward/loss. Studies like this will allow for a more nuanced understanding of DA’s role in pleasure and may confirm or deny the “flattening” result reported above.

VI. Learning from Reward

A. Learning from reward - Definition

We label the fifth stage of reward processing, as outlined in Figure 1, “Learning from Reward”. We define learning as the process in which associations between actions and rewarding outcomes are established and expectations and behaviors are adjusted when outcomes differ from expectations (Schultz, 2016a). We also include here studies investigating “reward prediction errors” or RPEs, which refer to the difference between predictions about rewards and actual obtained rewards, as these are thought to be integral “teaching” signals for reward learning (Rescorla and Wagner, 1972; Sutton, 1988). Note that while we list learning last in Figure 1, learning occurs over time, typically over several trials, and can occur at any point in the reward process when new information becomes available (Sutton and Barto, 1998).

B. Learning from reward - Predictions based on putative neurobiology

Studies of the neurobiology of learning suggests that DA manipulations should affect learning in humans. Evidence suggests that phasic DA signaling in the VTA and NAcc represents a biological instantiation of the reward prediction error. Midbrain DA neurons initially fire in phasic bursts during receipt of rewarding stimuli such as food, sexual contact, and drugs of abuse (Bassareo et al., 2007; Hernandez and Hoebel, 1988; Kiyatkin and Rebec, 2001; Pfaus et al., 1990); however, when the reward is predicted by another stimulus, dopaminergic firing shifts backward in time to the predictive stimulus (i.e., cue), and instead of occurring when the reward is delivered. When a reward is unexpected or larger than predicted by a cue, DA signaling increases at the time of the received reward, indicating that the outcome was better than expected. When a reward is predicted, but then not delivered, the DA neurons suppress firing at the time of the expected reward, signaling the outcome is worse than expected. Thus, the phasic firing of midbrain DA neurons appears to code RPEs (Cohen et al., 2012; Pan et al., 2005; Zaghloul et al., 2009).

Consistent with this neurobiology, DA antagonists appear to blunt both Pavlovian and instrumental learning in animals (Wise, 2004). For example, animals that have already learned stimulus-reward associations will reduce their responses to the rewarding stimulus over several minutes or days after being treated with DA antagonists (McFarland and Ettenberg, 1995; Yokel et al., 1975). Animals who have not yet learned these associations and are treated with DA antagonists do not learn to press a lever for rewards and do not develop conditioned place preferences (Wise and Schwartz, 1981). Similarly, in a rat probabilistic reward task, response bias toward rewarded cues is increased after administration of d-amphetamine and decreased after “low dose” pramipexole (Der-Avakian et al., 2013). Taken together, these findings suggest that systemic administration of DA drugs should affect reward learning in humans.

Based on the putative neurobiology of reward learning, our hypothesis is that increasing DA should increase reward learning and promote RPE coding in the midbrain, striatum, and associated frontal cortex areas, while reducing or blocking DA should impair reward learning and disrupt neural coding of RPEs.

C. Learning from reward - Measures

Reward learning in humans has been assessed using behavioral measures and neuroimaging. Behavioral measures generally consist of reinforcement learning tasks, in which a pair of stimuli is presented and the participant must learn which stimulus is more likely to lead to reward, with percentage of correct choices (or accuracy) indicating learning (e.g., Pessiglione, Seymour, Flandin, Dolan, & Frith, 2006). Variants on this paradigm include reversal learning, in which reward rules change, and learning is quantified by the mean error rate, time to reach a specified learning criterion, probability of correct choice following a reversal, or probability of switching after receiving non-reward feedback. Another variant is learning Go/No-Go tasks, in which participants must learn which stimulus requires a response or withholding of a response to receive or avoid losing a reward (Guitart-Masip et al., 2012b), with variables of interest including the probability of correct choice for go trials and probability of correct omission for no-go trials. Other popular tasks include the probabilistic reward task (PRT; Pizzagalli, Jahn, & O’Shea, 2005; Tripp & Alsop, 1999), which assesses development of response bias for a more highly rewarded stimulus utilizing signal detection theory, and time estimation tasks, which require adjusting responding according to feedback (e.g., Zirnheld et al., 2004). Measures from the above designs will be referred to as “raw” learning measures, as they are direct measures of task metrics and do not reflect a theoretically driven computational model.

Computational modeling, which models the fit of observed data to reinforcement learning algorithms, is often used to analyze effects of DA manipulations on these tasks (Rescorla and Wagner, 1972; Sutton and Barto, 1990). Here we briefly review the most frequently examined learning parameters in these models. The main assumption of a standard reinforcement-learning model is that the goal of learning is to predict the sum of future rewards based on the perceptual stimuli available. Reward prediction errors are available immediately after each reward trial and help estimate the actual sum of future rewards. In the standard model, the prediction error δ(t) is the difference between actual and expected outcomes, and serves as a teaching signal for the next trial: δ(t) = Rt + V(St) − V(St1), where Rt is the actual value of the present outcome (at time t) and V(St) and V(St1) are predictions about upcoming rewards at time t and time t-1 (the prediction in the previous state)1. The estimate V(t) of all future rewards is updated using the prediction error δ(t) and a constant learning rate α to scale the size of the update: V(t)new = V(t)old + α*δ(t). In another model referred to as a basic Q learning algorithm, prediction errors are also used to update the expected value of choosing a particular action (e.g., Samejima, Ueda, Doya, & Kimura, 2005). In this case, the model estimates the expected value of choosing from two actions or stimuli Qa or Qb (which typically begin at 0 on trial 1) and updates the expected value by the prediction error: Qa(t+1) = Qa(t)+ α*δ(t). The probability P of choosing a or b can then be modeled by using the estimated expected values for Qa and Qb. Typically, the softmax rule is employed such that:

Pa(t)=exp(Qa(t)/β)(exp(Qa(t)/β)+exp(Qb(t)/β)).

β is another constant, referred to as the temperature, which represents the participant’s biases towards random choice or choice of the highest value. Both learning rate and temperature are free model parameters that can either be set or compared between conditions. These models are modified to test specific predictions about how reinforcement learning networks function. Then, model fit is assessed to test if the model explains the observed data well. For the purpose of this review, we will focus primarily on whether DA manipulations affect α and β, as these are the most commonly examined parameters with the best accumulation of evidence.

Reward prediction errors have also been operationalized as brain activity in response to rewards on tasks that involve learning or making predictions about upcoming rewards. There are two primary ways RPEs are operationalized in imaging: 1) though BOLD activity that is correlated with the prediction error δ(t) estimated from the computational model and 2) through the error-related (ERN) or feedback-related (FRN) negativity ERP components measured using EEG. The ERN and FRN occur 0–150 ms and 250–250 ms post behavioral error or error feedback/unexpected outcomes, respectively (Gehring and Fencsik, 2001; Miltner et al., 1997; Scheffers et al., 1996). It is hypothesized that the ACC generates the ERN/FRN (Dehaene et al., 1994). Finally, one MEG study measured both event-related fields (ERFs) and oscillatory responses to reward feedback following differing reward expectations.

D. Learning from reward - Summary of studies

i. Learning from reward – “Raw” behavioral measures

We hypothesized that increasing DA should generally improve reward learning, but results from 18 studies are highly mixed, with only two studies supporting the most straightforward version of this hypothesis. In one, L-DOPA (100mg) improved learning of an association between a stimulus and monetary reward (but only when compared to a DA antagonist, not compared to placebo; Pessiglione et al. 2006). Similarly, in another study where participants had to make judgments about somatosensory stimuli and were rewarded when those judgments were correct, L-DOPA (100 mg) increased accuracy overall compared to placebo and haloperidol (Pleger et al., 2009). Thus, only two studies found that increasing DA with L-DOPA increased reward learning.

Seven studies indicated that administration of a DA-enhancing drug did not significantly affect “raw” learning rates on reversal learning or probabilistic learning tasks (Bellebaum et al., 2017; Bernacer et al., 2013; Diederen et al., 2017; Dodds et al., 2008; Jocham et al., 2011; Weis et al., 2012). These studies employed a variety of medications (i.e., modafinil, methylphenidate, methamphetamine, and “low dose” amisulpride) and a variety of tasks, including reinforcement learning, reversal learning, and probabilistic selection tasks.

Finally, two studies found results directly contradictory to our hypothesis. One reported that the agonist bromocriptine (1.25mg) produced worse reversal learning, although this was only significant compared to the antagonist sulpride, not placebo (Van Der Schaaf et al., 2012). The second reported worse learning on a reversal learning task under L-DOPA (100mg) compared to placebo, which the authors interpreted as consistent with an inverted-U hypothesis, suggesting “too much” DA can be detrimental, although baseline differences were not explicitly tested in this study (Vo et al., 2016).

Indeed, our straightforward hypothesis -- that increasing DA should increase reward learning -- does not take into account possible baseline differences or inverted-U effects, which have been commonly examined in this area. Five studies have considered the role of baseline DA in DAergic modulation of learning, with three supportive of an inverted-U effect and two supportive of a “rich get richer” effect. Supportive of the inverted-U hypothesis, both methylphenidate (60mg) and bromocriptine (1.25mg) improved reward learning in those with less baseline DA in the caudate and ventral striatum (as measured by PET), and impaired learning in those with more baseline DA (Clatworthy et al., 2009; Cools et al., 2009). Similarly, “low dose” sulpiride (200mg) impaired learning from feedback in extraverts compared to introverts (Mueller et al., 2014b), also broadly supportive of an inverted-U model although employing a less precise proxy measure. Yet others have reported potential “rich get richer” effects. In one study, methylphenidate increased reward learning in those with high baseline working memory and decreased it in those with low working memory (Van Der Schaaf et al., 2013). Another study found that learning rate was increased under cabergoline (1.5 mg), but only for those with the A1- DRD2 Taq1a genotype associated with increased DA (Cohen et al., 2007).

Finally, three studies have investigated the more nuanced hypothesis that DA may specifically facilitate learning associations between taking action and gaining reward, or “go-to-win” learning (Guitart-Masip et al., 2014). Only one was supportive -- “low dose” haloperidol (2mg) increased go and decreased no-go learning. However, this effect was seen only in those with low working memory (thought to be indicative of lower DA tone; Frank and O’Reilly 2006), suggesting an inverted-U effect. Of note, this same dose of haloperidol (2mg), employed as an intended reduction strategy, has also reduced reward learning (these findings will be discussed in the upcoming section). One study found null results of methylphenidate (20mg) on “go to win” learning (Swart et al., 2017). However, although learning was unaffected, Swart et al. (2017) did find that methylphenidate increased a general “go-to-win” bias in responses, particularly in those with high working memory span (Swart et al., 2017). These authors suggested that this might represent DA increasing the extent to which learning relies on prepared, Pavlovian (i.e., “go-to-win”) systems while not strongly affecting instrumental learning. Finally, one study found contrasting results. Contrary to the hypothesis that DA should enhance “go-to-win” learning, L-DOPA (150mg) instead reduced the bias towards “go-to-win” responding in the learning version of the go/no-go task (Guitart-Masip et al., 2014). These authors suggested that this might represent DA reducing the extent to which learning relies on prepared, Pavlovian (i.e., “go-to-win”) systems and enhancing the influence of instrumental, executive control systems, in contrast with the interpretation of Swart and colleagues. To summarize, “raw” results on reward learning tasks with drugs intended to increase DA levels have produced highly mixed results and divergent interpretations.

In contrast to the mixed effects for drugs intended to increase DA, DA reduction seems to more consistently impair reward learning. A total of 19 studies have investigated the effect of DA reduction on reward learning, with eleven supporting the hypothesis that DA reduction decreases learning. For example, administration of haloperidol (1mg) led to less money won and less rewarding choices on an instrumental learning task, although this effect was not significantly different from placebo, only from L-DOPA (Pessiglione et al., 2006). Similarly, during a rewarded somatosensory judgement task, haloperidol (2mg) decreased accuracy overall and eliminated magnitude of reward effects on accuracy (Pleger et al., 2009). Reducing DA (3 mg haloperidol) also decreases accuracy on time estimation tasks (Forster et al., 2017; Zirnheld et al., 2004), reversal learning tasks (400–800 mg sulpiride; Janssen et al., 2015) reward prediction tasks (600mg sulpiride; Diederen et al., 2017), and rewarded go/no-go tasks (APTD; Leyton et al., 2007). Consistent results have been reported for amisulpride (400mg), memantine (20mg; Jocham, Klein, & Ullsperger, 2014) and “low dose” pramipexole (.5mg; Pizzagalli et al. 2008b; Santesso et al. 2009b). One study measured slot-machine gambling behavior and found that under placebo, healthy adults showed a consistent positive relationship between payoff and bet size on the subsequent trial. Under haloperidol (3mg), this relationship was unstable, with the correlation between pay off and bet size reducing over time (Tremblay et al., 2011). While slot machines differ from typical learning designs, as there is no real relationship between winning and betting, these results suggest that DA reduction affected instrumental learning aspects of gambling behavior.

However, seven studies also reported null effects. One theme is that while DA antagonists appear to impair reward learning fairly consistently, DA depletion through APTD does not seem as effective. Apart from one study (Leyton et al., 2007), the majority of APTD studies report no effect on reversal learning, instrumental learning, investment learning tasks, or gambling tasks (de Wit et al., 2012; Robinson et al., 2010; Sevy et al., 2006; Tobia et al., 2014). In addition to the null findings with APTD, two studies reported no effects of antagonists (sulpride 400mg and 800mg) and one no effect of “low dose” (0.25 and 0.5mg) pramipexole on reward learning for monetary outcomes (Brom et al., 2016; Dodds et al., 2008; Eisenegger et al., 2014; Hamidovic et al., 2008). Finally, one study measured the effect of haloperidol (3mg) on conditioned sexual arousal and found no effects on reward learning (Brom et al., 2016). The significant results reported for antagonists above were all with monetary reward, so this null finding also suggests these effects may not generalize to non-monetary rewards.

The above studies did not measure baseline individual differences in DA levels or putative proxies such as working memory capacity. The two studies that did, reported that DA reduction (“low dose” cabergoline, 1.25mg; sulpride 400mg) actually increased reward learning in those with greater working memory capacity (Frank and O’Reilly, 2006; Van Der Schaaf et al., 2012), consistent with an inverted-U hypothesis.

In summary, antagonists appear to consistently decrease reward learning, while agonists and dietary manipulations have mixed to null results. A single study suggests that this effect of antagonists may be limited to monetary rewards, but more investigation of reward type is needed. Two studies that investigated baseline differences also provided support for a non-linear, inverted-U relationship.

ii. Learning from reward - Computational modeling measures

As computational models are flexible, we attempt to synthesize studies that have employed similar algorithms, such as standard reinforcement learning models. Here, we summarize studies that measured the effect of a DA enhancing or reducing drug on the learning rate (α) parameter and the temperature (β) parameter. It should be noted that other studies exist that employ reinforcement learning models, but hold α and β constant, and as such, those will not be reviewed here.

Six studies have measured the effect of a DA enhancing drug on α and none of them support the hypothesis that increasing DA should increase α. Four studies that looked at α reported no effects with DA enhancement (Diederen et al., 2017; Guitart-Masip et al., 2014; Jocham et al., 2011; Wunderlich et al., 2012). Notably, Diederen et al (2017), Guitart-Masip et al. (2014), and Wunderlich et al (2012) utilized modified versions of reinforcement learning algorithms to specifically test other model parameters, which could explain the null results on α. One study reported that increasing DA with bromocriptine (1.25mg) increased α in those with low striatal DA (measured with PET) and low working memory; this drug effect decreased as DA synthesis capacity increased, suggesting an inverted-U shape relationship (Cools et al., 2009). One study found contradictory results. Bernacer et al. (2013) reported that methamphetamine (.3mg/kg) decreased α, and further α was restored to normal when methamphetamine was taken with the DA antagonist amisulpride (400mg). Thus, taken together, it appears unlikely that increasing DA improves α. In terms of the temperature parameter β, all six studies reported null results (Bernacer et al., 2013; Diederen et al., 2017; Guitart-Masip et al., 2014; Jocham et al., 2011; Wunderlich et al., 2012). Therefore, we conclude that dopamine enhancing drugs also seem to have little to no effect on noise in responding.

DA reduction also has not shown strong effects on α, as all four studies reported null effects of DA reduction on model-derived learning rate (Diederen et al., 2017; Eisenegger et al., 2014; Jocham et al., 2014; Tobia et al., 2014). While DA reduction may not have a strong effect on α, it may alter β instead. Eisenegger et al. (2014) reported that the temperature (β) was higher under sulpiride (800mg), suggesting that decreasing DA simply increases the noise in the participant’s responding such that choices are less correlated with subjective value. However, this was the only study that reported an effect with β, the remaining three studies reported null results (Diederen et al., 2017; Jocham et al., 2014; Tobia et al., 2014). To summarize, the modal results is that DA enhancing and reducing drugs have little to no effect on α and β when using reinforcement learning models.

iii. Learning from reward – Neuroimaging measures

As it is thought that RPEs are DA dependent, augmenting DA functioning would also be expected to increase the brain activity thought to represent RPEs. Across the studies reported here, RPE activity is most often observed in the striatum (most commonly ventral), frontal cortex (ventromedial prefrontal cortex/orbitofrontal cortex and ACC), amygdala, and midbrain (Bernacer et al., 2013; Diederen et al., 2017; Jocham et al., 2011; Pleger et al., 2009). A total of ten studies have investigated the effects of drugs intended to increase DA on these brain areas during reward learning. Four of these are supportive of our hypothesis. In one study, increasing DA via L-DOPA (100mg) increased RPE activity within the ventral striatum, which was accompanied by increased reward learning (Pessiglione et al., 2006). Similarly, L-DOPA (100mg) increased activity in the OFC and VS in response to monetary reward signaling a correct responses on a somatosensory judgement task (Pleger et al., 2009). “Low dose” amisulpride (200mg), thought to boost D1 signaling, also enhanced RPE coding in the striatum and frontal cortex, with ventromedial prefrontal cortex activity specifically correlating with overall learning (Jocham et al., 2011). Finally, another study found that methylphenidate (60mg) decreased striatal BOLD during switch after negative feedback trials and increased frontal BOLD during stay after positive feedback, suggesting changes in DA facilitates behavior change (Dodds et al., 2008).

Four studies also reported null results. Two with bromocriptine (1.25mg/2.5mg), one with L-DOPA (100mg), and one with bupropion (150mg) found that DA manipulation had no effect on brain activity to rewarding feedback (Diederen et al., 2017; Graf et al., 2016a; Van Der Schaaf et al., 2012; Weis et al., 2013). Of note, in the study by Van der Schaaf et al (2012), the bromocriptine condition was associated with less activity in the VS in comparison to a sulpiride condition; however, bromocriptine did not differ from placebo.

Conversely, two studies reported that DA augmentation decreased RPE activity. One study reported that methamphetamine (0.3mg/kg) reduced model-derived RPE activity in the ventral striatum; however, in this study administration of methamphetamine was accompanied by mild psychotic symptoms, potentially contributing to the differing patterns of brain activity (Bernacer et al., 2013). In another study, methylphenidate (40mg) also decreased RPE BOLD activity in the ventral striatum. In this case the authors speculated that methylphenidate might have increased tonic DA, leading to a decrease in phasic DA (Evers et al., 2017). To summarize, four studies support the hypothesis that DA augmentation increases RPE brain activity, but results were mixed with four showing null effects and two showing contradictory effects.

If DA is necessary for RPEs, then as a corollary, reduction of DA should lead to reductions in brain activity thought to represent RPEs. Results from almost every study (six studies) in this section are consistent with this hypothesis. DA reduction (sulpiride 600mg; amisulpride 400mg; haloperidol 2mg; APTD) disrupts RPE activity most commonly in the ventral striatum and caudate (Diederen et al., 2017; Jocham et al., 2014; Pessiglione et al., 2006; Tobia et al., 2014). DA reduction (sulpiride 600mg; APTD) may also disrupt RPE activity in the amygdala and midbrain (Diederen et al., 2017; Tobia et al., 2014). However, one combination drug study reported that sulpiride (400mg) did not alter the complex pattern of methylphenidate (60mg) effects seen in the same study (Dodds et al., 2008).

A sole study indicated that effects of DA reduction on fMRI outcomes may also depend on working memory (again used as a proxy for baseline levels of DA). In this single study, those with higher working memory showed an increase in striatal activation in response to rewards after administration of a DA antagonist (along with an increase in learning, as described in the behavioral section). This is consistent with the inverted-U hypothesis, suggesting individuals with high working memory were potentially moving from a hyperactive DA state to a more optimal state (Van Der Schaaf et al., 2012). However, this is the only study in this section that reported on baseline working memory; therefore, the modal result is that DA reduction simply disrupts RPEs.

We also expected that DA should modulate the ERN and FRN, which are thought to represent RPEs in the ACC (Holroyd and Coles, 2002; Proudfit, 2015). One study using d-amphetamine (15mg) suggested that DA agonists straightforwardly increase the ERN/FRN (de Bruijn et al., 2004). However, three studies using “low dose” sulpiride (200mg) suggest an inverted-U relationship. Of note, two of these examined the same participants using different proxies of DA functioning. All found “low dose” sulpiride enhanced the FRN in individuals with presumably low baseline DA levels (COMT Val/Val genotype, introverts – although see note above about possible conflicting effects of COMT genotype in different brain areas), but decreased it in individuals with high baseline DA levels (COMT Met allele carriers, extraverts) (Mueller et al., 2014b, 2014a, 2011). Finally, one study reported null effects of L-DOPA on ERFs (Apitz and Bunzeck, 2014).

Five studies have investigated the effect of DA reduction on the ERN/FRN, with three in support of our hypothesis. Antagonists decrease the amplitude of the ERN/FRN (de Bruijn et al., 2006), with accompanying effects on learning (Santesso et al., 2009; Zirnheld et al., 2004). Still, two papers have reported null effects of dopaminergic drugs on the FRN (APTD; Larson et al. 2015; 3mg haloperidol; Forster et al. 2017), one employing APTD and the other administering haloperidol. Notably, the APTD study did not use a learning task per se, so this could explain the null result.

E. Learning from reward - Summary, conclusions, recommendations

Across “raw” learning metrics, computational modeling outcomes and imaging measures of the RPE, drugs intended to increase DA and decreasing DA via precursor depletion appears to have mixed effects, while DA reduction using antagonists more consistently affect reward learning. Regarding the potential inverted-U relationship between DA and learning, evidence for this is difficult to evaluate due to characteristics of the literature. Specifically, few studies utilized the same drug manipulations, making it difficult to compare doses across studies, and no studies incorporated the multiple doses required to establish a dose-response curve. Further, studies differed significantly in how they established baseline DA functioning. While a minority used PET, more used working memory or other behavioral or self-report indicators. Within studies that used similar “proxy” measures such as working memory, some detected inverted-U patterns while others did not, but it is difficult to know whether to attribute this to the inexactness of the proxy measures. Relatively more consistently, the neuroimaging literature supports the role of DA in the generation of reward prediction errors, with augmentation increasing coding of RPEs and reduction decreasing RPEs in general.

However, complicating the picture, while computational modeling shows support for DA’s role in producing reward prediction errors in humans, there is limited evidence that DA modulation consistently affects model-derived learning rate. In response to this, authors have suggested (and calculated) additional model parameters that may be affected by DA, such as the balance between Pavlovian and instrumental learning (Frank and O’Reilly, 2006; Guitart-Masip et al., 2014; Santesso et al., 2009; Swart et al., 2017), PE scaling (Diederen et al., 2017), among others (Cools et al., 2009; Jocham et al., 2014; Tobia et al., 2014; Wunderlich et al., 2012). Because of the comparatively small number of papers examining each of these additional parameters, we have not reviewed these results here. However, it is possible that the unclear effect of DA augmentation on learning rate could potentially be due to failure to capture these more complex parameters. Computational modeling shows strong promise in highlighting the more complex ways DA may interact with learning, but we also note that comparability has been strongly hampered by a multiplicity of models. Calculating and reporting drug effects on both the simpler (and thus more likely comparable across studies) models along with more complex ones is recommended.

VI. Conclusions

A. Final Summary

The primary goal of this review was to answer the question: What can drug challenge studies in healthy human adults tell us about the role of DA in human reward processing? We draw three main conclusions from the review: i) DAergic drugs have markedly different effects on different phases of reward processing; ii) the relationship between DA and any given reward function appears unlikely to be linear; iii) the effects of DAergic drugs on reward-related behaviors vary depending on the type of outcome measure used (i.e., self-report, behavioral, and neural).

First, we found that the effects of DAergic drugs vary across different phases of reward processing. DAergic manipulations appear clearly to impact anticipation of rewards, and particularly neural measures of anticipation. DA also appears involved in the evaluation of risk-related costs vs. benefits when there is the possibility of loss, and the evaluation of effort costs vs. benefits. However, DAergic drugs had little consistent impact on reward-related decision-making when the cost involved was time. We found, as expected, that decreasing DA had little effect on self-reported pleasure, although increasing DA in some cases appeared to lessen the distinction in responses between pleasant and unpleasant stimuli. We found little evidence that DAergic drugs affected the speed or vigor of actions taken to attain reward, although many studies considered may not have been powered to measure behavioral vigor, as most were incidental measures taken during imaging studies. Finally, the role of DA reward learning appears complex, and may depend on the manner in which learning is assessed. DA antagonists in particular decreased behavioral measures of learning and neural correlates of prediction errors (both with fMRI and EEG) in the striatum and sometimes in the anterior cingulate, but DA agonists did not have a consistent effect.

These findings indicating both the distinctness of different reward functions, and the pattern of greater and lesser DAergic influence on these functions may help us to understand deficits in psychiatric conditions. For example, our findings are consistent with theories suggesting that “anhedonia”, as experienced in depression and schizophrenia, may have several distinct neurobehavioral components (Kring and Barch, 2014; Treadway and Zald, 2011b). Theorists have posited that anhedonia may involve reductions in anticipation of reward and willingness to exert effort for reward, as distinct from changes in ability to experience pleasure, and further suggested that anticipation and effort-related decision-making are more DAergically based (Treadway and Zald, 2011b). The idea that these functions are distinct and that anticipation and effort-related decision-making are more DAergically-based is quite consistent with our synthesis of findings here. These findings may also be relevant to understanding the effect of DAergic treatments, such as those commonly used for ADHD. Our synthesis would suggest that DAergic drugs may be more effective at ameliorating motivational aspects of ADHD rather than aspects of impulsivity that involve difficulty waiting for rewards (although note that problems with delay of gratification are only one potential aspect of impulsivity). Again, this is highly consistent with theorizing about ADHD as an overall pathology of reward (Luman et al., 2010), and recent findings in the clinical literature about the effects of stimulants in individuals with ADHD (Addicott et al., 2019). Thus, these findings have bearing on our understanding of psychiatric disorders that relate to reward functions.

A second major implication of this review is that changes in DA appear unlikely to linearly relate to any given reward function. Simplistically, if DA is important for reward functioning, then increasing DA should increase that function and decreasing it should decrease that function, as our general hypothesis stated. The studies reviewed here indicate that this is often not the case. In some areas, the effect appeared asymmetrical. For example, antagonists appeared to more consistently affect learning than agonists. In other areas, agonists and antagonists had converging effects. For example, on measures of risk-related decision-making, both increasing and decreasing DA decreased the impact of loss, and reduced the impact of probability on decision-making. Finally, throughout this review, analyses that examined individual differences showed potential U-shaped patterns and modulation of drug effects by baseline functioning. Our findings suggest that both basic and clinical studies must take into consideration the possibility of nonlinear relationships between drugs, DA levels, and performance. For example, recent neuroimaging evidence shows negative correlations between DA levels and reward pathology in addiction, yet positive correlations between DA and reward pathology in ADHD, consistent with this synthesis (Castrellon et al., 2019).

A third major implication is that our ability to detect the effects of DAergic drugs may depend in part on whether the outcome measures are self-report, behavioral, or imaging measures. DAergic drug manipulations appeared to affect neuroimaging measures more consistently than subjective or behavioral measures. For example, DA drugs had little effect on subjective anticipation of upcoming rewards, but produced strong effects on the striatum during anticipation of rewards. Similarly, DA drugs had little effect on self-reported pleasure, but at least some studies indicate that they did affect neural reactivity during receipt of reward. Quite a number of studies indicate that reducing DA function consistently reduces behavioral reward learning as well as brain activity in the ventral striatum and anterior cingulate. Yet, studies that used computational modeling have reported little effect of DA drugs on model-derived learning rate. However, these findings also raise interesting questions about the meaning of brain responses to an event that are not related to behavior or self-report. Ultimately, although imaging appears most sensitive to DA, information about self-report, behavioral and imaging studies will need to be integrated to understand reward processing deficits, and develop medications for associated psychiatric disorders.

B. Recommendations

There are several recommendations that arise from this review. First, there is an urgent need for dose-response studies examining the relationship between DA and reward functioning in humans. As seen in this review, results with single doses can be interpreted post-hoc in multiple ways, and researchers should be cognizant of the possibility that responses are not linearly related to dose. This can only be resolved by using multiple doses to establish a dose-response curve. Second, there is a need to study multiple drugs that manipulate the DA system in the same measures and participants. Any single drug is likely to have actions on several other receptor systems, and so does not provide a pure effect of increasing or decreasing DA function. Converging findings with several drugs with shared actions on the same measures would provide more confidence in results. Further, studies comparing multiple drugs in the same participants and measures would strongly assist our ability to evaluate interpretations that depend on theorizing about differences in the actions or affinities of individual drugs. Third, there is a need to utilize multiple outcome measures to characterize reward function, and examine multiple reward-related processes in the same studies. As we have shown here, outcomes depend strongly on the measures used. Further, although this review strongly suggests different sensitivities to DA across different functions, there are few studies directly and explicitly comparing effects on multiple reward functions for the same doses in the same participants. These recommendations reflect the complexity of measuring the role of DA in studies of reward function.

A further recommendation is that researchers be cognizant of possible baseline differences in DA function, which may affect responses to drugs. For example, variation in baseline dopamine levels could make findings especially difficult to interpret when the dose response is hypothesized to be an inverted-U shape. One approach to addressing this issue is to measure baseline dopamine levels using PET (Westbrook et al., 2020). However, PET studies are difficult and expensive. An alternative approach is to examine the effects of the drug in relation to individual differences in “baseline functioning” assessed behaviorally in the lab. Here we strongly recommend standardizing the behavioral measures that are used to assess baseline functioning, and tentatively recommend working memory capacity as a rough proxy for baseline DA function. Even though other neurotransmitters are involved in working memory, this measure has some evidence for relationships to DA, and already is in more widespread usage than any other, assisting comparability (Cools et al., 2008).

A third recommendation concerns reward type. The reward most commonly used in human studies is money. Whether these findings would generalize to other reward types, such as food, sex, emotional, or social reward is largely unknown. For example, there was some evidence that pleasure from food reward was more strongly influenced by DA function, but there were no studies directly comparing this reward type with another. Only one study in this review compared DA effects on more than one reward in the same participants, contrasting positive music and money (Ferreri et al., 2019). Future studies should compare responses across rewards to determine generalizability of the findings, and to permit better translation to studies with laboratory animals, which generally use food rewards. While there is evidence of a common neurobiology for the processing of all reward types (Volkow et al., 2013), it seems possible base don our findings that DA challenges may have different effects on different types of rewards, so more studies of this kind are needed.

Fourth, drugs that reduce DA function (especially receptor antagonists) appear to have more pronounced effects than drugs that enhance DA function (precursors, reuptake inhibitors, or agonists). This may be because healthy adults are likely to already have close to optimal DAergic functioning and optimal performance, making it difficult to detect improvements. Therefore, reducing DA levels in healthy adults might be a more powerful way to study the effects of DA on reward-related functions. Notably, the only function on which DA antagonists had little effect was pleasure, consistent with the hypothesis that DA is less involved in this process. Therefore, reduction strategies may be more appropriate when the sample involves healthy controls.

Finally, while only healthy human studies were presented here, the studies reviewed set the stage for future translational research. An important advantage of drug challenge studies is that some methods can be performed in both humans and laboratory animals. Future work should aim a-priori to employ similar pharmacological manipulations across laboratory animals, healthy humans, and patients to fully understand reward processing and the ways it can go awry (e.g. Pergadia et al., 2014). In addition to similar pharmacological manipulations, future work should look to utilizing tasks that better translate preclinical to human studies. For example, the Probabilistic Reward Task has been useful in measuring reward responsiveness across species and appears sensitive to DA manipulations (Der-Avakian et al., 2013; Pergadia et al., 2014). Development and use of similar tasks for other reward functions is critical to bridge the gap between basic preclinical research and the role of DA in human disorders.

C. Conclusions

The goal of this review was to summarize what can be learned about the role of DA in reward functioning by looking at drug challenge studies in healthy human subjects. Drug challenge studies have good experimental control, utilize tasks that translate well to animal behavioral studies, and can be combined with neuroimaging. Drug challenge studies in healthy humans reduce confounds that are present when studying psychiatric populations and can therefore provide stronger causal evidence for how DA mediates reward functioning. Here we found that the body of evidence suggests different sensitivity to DA manipulations across different phases of reward processing, non-linear relationships between DA and reward functioning, and different sensitivity of different measures to these effects. Results from these reviewed studies are critical for understanding how deficits in reward functioning are related to psychiatric disorders and for guiding medication development for disorders where anticipation of reward, evaluation of effort or gambling costs, or reward learning is altered, such as depression, substance use, and schizophrenia (Keiflin and Janak, 2015; Smoski et al., 2009; Strauss et al., 2011; Treadway et al., 2012).

Table 1.

Anticipation

Subjective Findings
Dopamine Enhancement
First Author Year Drug Dose N Measure(s) Result(s) Notes
Sharot 2009 L-DOPA 100mg N Drug = 29 N Placebo = 32 Anticipated pleasure of vacation destination Increase
Dean 2016 bupropion 150mg/day N Within = 17 Wanting to cues of pleasant & unpleasant No effect Sub-acute dosing over 7 days
Knutson 2004 d-amphetamine 0.25mg/kg N Within = 8 Excitement to cues of monetary reward/loss Trend toward “flattening” -decrease to reward cues; Increase to loss cues
Dopamine Reduction
First Author Year Drug Dose N Measure(s) Result(s) Notes
Bjork 2014 APTD -- N Within = 16 Excitement to cues of monetary reward No effect
Behavioral Findings
Dopamine Enhancement
First Author Year Drug Dose N Measure(s) Result(s) Notes
Ferreri 2019 L-DOPA 100mg N Within = 27 Skin conductance during anticipation of monetary reward Increase Compared to risperidone
Behavioral Findings
Dopamine Reduction
First Author Year Drug Dose N Measure(s) Result(s) Notes
Ferreri 2019 risperidone 2mg N Within = 27 Skin conductance during anticipation of monetary reward Decrease Compared to L-DOPA
Neuroimaging Findings
Dopamine Enhancement
First Author Year Drug Dose N Measure(s) Result(s) Notes
Dean 2016 bupropion 150mg/day N Within = 17 BOLD during anticipation of pleasant & unpleasant taste Increase in caudate during both pleasant & unpleasant taste anticipation; Increase during pleasant taste anticipation in pregenual ACC/vmPFC & lateral OFC No effect on self-report; sub-acute dosing over 7 days
Evers 2017 methylphenidate 40mg N Within = 24 BOLD during during anticipation of monetary reward/loss Increase in ventral striatum during reward anticipation
Funayama 2014 modafinil 200mg N Within = 20 BOLD during anticipation of monetary reward/loss Increase in bilateral NAc, left middle frontal & precentral gyri, & left cuneus during reward anticipation
Ikeda 2019 bupropion 150mg N Within = 15 BOLD during anticipation of monetary reward/loss Increase in right NAc
Kirsch 2006 bromocriptine 1.5mg N Within = 24 BOLD during anticipation of monetary reward/no reward Increase in right NAcc Strongest in individuals with the A1A1 DRD2 genotype - consistent with Inverted-U
O’Daly 2014 d-amphetamine 20mg, 4 times N Drug = 11 N Placebo = 11 BOLD during anticipation during anticipation of monetary reward/loss Increase in right caudate during reward anticipation “Sensitization” paradigm with repeated dose of amphetamine
Curley 2013 benzylpiperazine 200mg N Within = 13 BOLD during anticipation of monetary reward/loss No effect in striatum; Decrease in inferior frontal gyrus, insula, mid-occipital regions Participants did not know if they were anticipating reward or loss
Graf 2016 bupropion 150mg N Within = 17 BOLD during anticipation of monetary reward/loss No effect in NAcc
Suzuki 2019 atomoxetine 40mg N Within = 14 BOLD during anticipation of monetary reward/no reward No effect in NAcc
Weis 2012 L-DOPA 100mg N Drug = 27 N Placebo = 27 BOLD during anticipation of monetary reward/no reward No effect in striatum; Increase in left auditory cortex, left inferior frontal gyrus, anterior cingulate cortex Auditory adaptation of MID
Wittman 2015 L-DOPA 100mg N Drug = 14 N Placebo = 14 BOLD during anticipation of monetary reward/loss No effect in striatum during reward anticipation; Increase in striatum during loss anticipation
Apitz 2014 L-DOPA 150mg N Drug = 20 N Placebo = 18 ERFs & beta oscillatory power (MEG) during cues of high/low monetary reward Decrease in impact of reward probability on ERF magnitude, Decrease in beta power
Ivanov 2014 methylphenidate 0.5mg/kg N Within = 16 BOLD during anticipation of monetary reward/no reward Decrease in left insula, left caudate, left thalamus, & right cerebellum
Knutson 2004 d-amphetamine 0.25mg/kg N Within = 8 BOLD during anticipation of monetary reward/loss Decrease in VS during anticipation, but prolonged activation of VS during reward anticipation; Increase in VS during loss anticipation
Guitart-Masip 2012 L-DOPA 150mg N Drug = 16 N Placebo = 20 BOLD during orthogonalized go/no-go task Increase in striatum & SN/VTA “Go-to-win” paradigm
Dopamine Reduction
First Author Year Drug Dose N Measure(s) Result(s) Notes
Abler 2007 olanzapine 5mg N Within = 8 BOLD during cues of monetary reward/no reward Decrease in VS, anterior cingulate & inferior frontal cortex
Bjork 2014 APTD -- N Within = 16 BOLD during cues of monetary reward/no reward Decrease in NAcc
Fiore 2018 risperidone 0.5mg N Within = 20 BOLD during cues of monetary reward Decrease in dorsal striatum and right orbitofrontal cortex
Nagano-Saito 2012 APTD -- N Within = 17 BOLD during cues of monetary reward/no reward Decrease in areas associated with mesocortical system including NAcc
Schutte 2017 haloperidol 2mg N Within = 23 ERPs (RRP, PRP, & P300) to cue during go/no-go task RRP and P300 reduced; no effect on PRP More evident in those with higher eye blink rate (proxy of higher DA function), consistent with Inverted-U
da Silva Alves 2010 AMPT 1.5g N Within = 10 BOLD during cues of monetary reward/no reward No effect in striatum, Increase in left cingulate gyrus
Ye 2011 pramipexole 0.5mg N Within = 16 BOLD during cues of monetary reward/no reward Increase in right NAcc; Weakening of interaction between NAcc & prefrontal cortex

Table 2.

Evaluation

Time vs. Reward
Behavioral Findings
Dopamine Enhancement
First Author Year Drug Dose N Measure(s) Result(s) Notes
de Wit 2002 d-amphetamine 10 & 20mg N Within = 36 Delay discounting Decrease in traditional task, no effect in “experiential” task “Experiential” task allowed participants to experience small rewards/short time delays in lab
Kayser 2012 tolcapone 200mg N Within = 23 Delay discounting Decrease Greater effects in individuals with high baseline impulsivity, consistent with Inverted-U
Acheson 2008 d-amphetamine/bupropion 20/150mg & 300mg N Within = 32 Delay discounting No effect Included smokers - smoking status did not affect responses
Soutschek 2020 PF-06412562 6mg, 15mg, 30mg N Placebo = 30, N Drug (each dose) = 30 Delay discounting No effect
Pine 2010 L-DOPA 150mg N Within = 13 Delay discounting Increase
Dopamine Reduction
First Author Year Drug Dose N Measure(s) Result(s) Notes
Kelm 2013 APTD -- N Within = 15 Delay discounting Poor get poorer or Inverted-U Increase in COMT gene val/val carriers, decrease in met carriers - possible conflicting effects of COMT in different brain regions complicate interpretation
Hamidovic 2008 pramipexole 0.25mg or 0.50mg N Within = 10 Delay discounting No effect
Pine 2010 haloperidol 1.5mg N Within = 13 Delay discounting No effect
Arrondo 2015 metoclopramide 10mg N Within = 14 Delay discounting Decrease Delay confounded with probability - computational modeling indicated change in delay parameters
Soutschek 2017 amisulpride 400mg N Within = 55 Delay discounting Decrease Only in lower BMI subjects
Weber 2016 amisulpride 400mg N Drug = 41 N Placebo = 40 Delay discounting Decrease
fMRI Findings
Dopamine Enhancement
First Author Year Drug Dose N Measure(s) Result(s) Notes
Kayser 2012 tolcapone 200mg N Within = 23 BOLD during delay discounting Increase in striatum, pregenual cingulate & anterior insula, dorsolateral & medial frontal cortex In context of decrease in delay discounting behavior
Pine 2010 L-DOPA 150mg N Within = 13 BOLD during delay discounting Increase in striatum, insula, subgenual cingulate, & lateral OFC; Decrease in caudate, insula, & lateral inferior frontal regions In context of increase in delay discounting behavior
Dopamine Reduction
First Author Year Drug Dose N Measure(s) Result(s) Notes
Pine 2010 haloperidol 1.5mg 13 BOLD during delay discounting No effect
Arrondo 2015 metoclopramide 10mg N Within = 14 BOLD during delay discounting Decrease in post-central gyms
Risk vs. Reward
Behavioral Findings
Dopamine Enhancement
First Author Year Drug Dose N Measure(s) Result(s) Notes
Rigoli 2016 L-DOPA 150mg N Within = 32 Gambling Increase Only in lower BMI subjects
Rutledge 2015 L-DOPA 150mg N Within = 30 Gambling Increase
Acheson 2008 d-amphetamine/bupropion 20mg/150mg & 300mg N Within = 32 Probability discounting/ BART pumps No effect Included smokers - smoking status did not affect responses
de Wit 2002 d-amphetamine 10&20mg N Within = 36 Probability discounting No effect
Evers 2017 methylphenidate 40mg N Within = 24 Gambling No effect
Symmonds 2013 L-DOPA 100mg N Within = 20 in two studies Gambling No effect
White 2007 d-amphetamine 20mg N Within = 37 BART pumps No main effect Increase in males with high reward sensitivity, decrease in males with low reward sensitivity (consistent with “rich get richer”), but no effect in women
Soutschek 2020 PF-06412562 6mg, 15mg, 30mg N Placebo = 30, N Drug (each dose) = 30 Gambling Decrease
Campbell-Meiklejohn 2012 methylphenidate 20mg N Drug = 20 N Placebo = 20 Loss-chasing Increase for high-stakes bets, decrease for low-stakes bets Consistent with decreased impact of loss
Norbury 2013 cabergoline 1.5mg N Within = 20 Gambling Decrease in “probability distortion”, decreased effect of prior loss Consistent with decreased impact of loss
Dopamine Reduction
First Author Year Drug Dose N Measure(s) Result(s) Notes
Arrondo 2015 metoclopramide 10mg N Within = 14 Probability discounting No effect Delay confounded with probability - computational modeling indicated no change in probability parameters
Hamidovic 2008 pramipexole 0.25mg & 0.50mg N Within = 10 Probability discounting/ BART pumps No effect
Zack 2007 Haloperidol 3mg N Within = 18 Gambling No effect
Burke 2018 amisulpride 400mg N Drug = 45 N Placebo = 48 Gambling Overall increase, decrease in “probability distortion”
Campbell-Meiklejohn 2011 pramipexole 0.176mg N Drug = 20 N Placebo Loss-chasing Increase for high-stakes bets, decrease for low-stakes bets Consistent with decreased impact of loss
Ojala 2018 sulpiride 400mg N Within = 21 Gambling No overall effect, decrease in “probability distortion”
Riba 2008 pramipexole 0.5mg N Within = 15 Gambling Increase Only when previous trial resulted in large gain
Effort vs. Reward
Subjective Findings
Dopamine Enhancement
First Author Year Drug Dose N Measure(s) Result(s) Notes
Volkow 2004 methylphenidate 20mg N Within = 16 Motivation for effortful task Increase
Behavioral Findings
Dopamine Enhancement
First Author Year Drug Dose N Measure(s) Result(s) Notes
Soutschek 2020 PF-06412562 6mg, 15mg, 30mg N Placebo = 30, N Drug (each dose) = 30 Choice of effortful task Increase
Wardle 2011 d-amphetamine 10 or 20mg N Within = 17 Choice of effortful task Increase Strongest effect under low probability of reward, high effort confounded with time to complete
Westbrook 2020 methylphenidate/suplride 20mg/400mg N Within = 50 Choice of effortful task Increase Greater increase in individuals with lower baseline DA synthesis capacity, consistent with Inverted-U
Zenon 2016 L-DOPA 125mg N Within = 19 Grip force on effortful task Increase
Michely 2019 L-DOPA 150mg N Within = 20 Perseverance in effortful grip task No change
Dopamine Reduction
First Author Year Drug Dose N Measure(s) Result(s) Notes
Cawley 2013 APTD -- N Within = 32 Progressive ratio breakpoint Decrease Subjects were women with subsyndromal seasonal affective disorder
Venugopalan 2011 APTD -- N Within = 47 Progressive ratio breakpoint Decrease Participants had varying levels of nicotine use
Michely 2019 haloperidol 1.5mg N Within = 20 Perseverance in effortful grip task Decreased impact of reward magnitude
fMRI Findings
Dopamine Enhancement
First Author Year Drug Dose N Measure(s) Result(s) Notes
Volkow 2004 methylphenidate 20mg N Within = 16 Striatal DA release under methylphenidate Correlated with increased self-reported motivation

Table 3.

Execution

“Simple Invigoration” Theory
Subjective Findings
Dopamine Enhancement
First Author Year Drug Dose N Measure(s) Result(s) Notes
Funayama 2014 modafinil 20mg N Within = 20 Exertion of effort Increase
Ikeda 2019 bupropion 150mg N Within = 15 Exertion of effort No effect
Behavioral Findings
Dopamine Enhancement
First Author Year Drug Dose N Measure(s) Result(s) Notes
Michely 2019 L-DOPA 150mg N Within = 20 Grip strength in rewarded task Increase
Apitz 2014 L-DOPA 150mg N Drug = 20 N Placebo = 18 Speed in rewarded task No effect
Dean 2016 bupropion 150mg/day N Within = 17 Speed in rewarded task No effect Sub-acute dosing over 7 days
Funayama 2014 modafinil 20mg N Within = 20 Speed in rewarded task No effect
Graf 2016 bupropion 150mg N Within = 17 Speed in rewarded task No effect
Ivanov 2014 methylphenidate 5mg/kg N Within = 16 Speed in rewarded task No effect
Mayo 2019 methamphetamine 20mg N Within = 73 Speed in rewarded task No effect
Weis 2012 L-DOPA 100mg N Drug = 27 N Placebo = 28 Speed in rewarded task, conditioning paradigm No effect
Wittman 2015 L-DOPA 100mg N Drug = 14 Speed in rewarded task No effect
Kirsch 2006 bromocriptine 1.5mg N Within = 24 Speed in rewarded task Inverted-U Increase in individuals with the A1A1 DRD2 genotype, decrease
for other genotypes, consistent with Inverted-U
Dopamine Reduction
First Author Year Drug Dose N Measure(s) Result(s) Notes
Abler 2007 olanzapine 5mg N Within = 8 Speed in rewarded task Decrease
Bjork 2014 APTD -- N Within = 16 Speed in rewarded task No effect
da Silva Alves 2014 AMPT -- N Within = 10 Speed in rewarded task No effect
Fiore 2018 risperidone 0.5mg N Within = 20 Time to reach required grip strength in rewarded task No effect
Michely 2019 haloperidol 1.5mg N Within = 20 Grip strength in rewarded task No effect
Nagano-Saito 2012 APTD -- N Within = 17 Speed in rewarded task No effect
Schutte 2017 haloperidol 2mg N Within = 23 Speed in rewarded task No effect
Neuroimaging Findings
Dopamine Enhancement
First Author Year Drug Dose N Measure(s) Result(s) Notes
Dean 2016 bupropion 150mg/day for 7 days N Within = 17 BOLD during execution of hard or easy task to gain pleasant or avoid unpleasant task Increase in caudate, vmPFC, dACC/paracingulate gyrus & putamen Contrast was between easy/pleasant & hard/pleasant trials; sub-acute dosing over 7 days
Dopamine Reduction
Fiore 2018 risperidone 0.5mg N Within = 20 BOLD during execution of grip to gain high vs. low reward Decrease in external globus pallidus to high reward, increase in substantia nigra to low reward
“Average Rate of Reward” Theory
Behavioral Findings
Dopamine Enhancement
First Author Year Drug Dose N Measure(s) Result(s) Notes
Beierholm 2013 L-DOPA 150mg N Drug = 30 N Placebo = 30 Computational model of influence of average rate of reward Increase
Zenon 2016 L-DOPA 125mg N Within = 20 Grip force on effortful task, when force was related to time to complete task No effect
“Go to Win” Theory
Behavioral Findings
Dopamine Enhancement
First Author Year Drug Dose N Measure(s) Result(s) Notes
Guitart-Masip 2012 L-DOPA 150mg N Drug = 16 N Placebo = 20 Speed on orthogonalized go/no-go task Increase for all go trials, regardless of win/loss type
Swart 2017 methylphenidate 20mg N Within = 99 Speed on orthogonalized go/no-go task No effect Task required contingencies learning
Dopamine Reduction
First Author Year Drug Dose N Measure(s) Result(s) Notes
Weber 2016 amisulpride 400mg N Drug = 41 N Placebo = 40 Pavlovian-to-instrumental transfer Decrease

Table 4.

Pleasure

Subjective Findings
Dopamine Enhancement
First Author Year Drug Dose n Measure(s) Result(s) Notes
Dean 2016 bupropion 150mg N Within = 17 Liking ratings of food No effect Sub-acute dosing over 7 days
Knutson 2004 d-amphetamine 0.25mg/kg N Within = 8 Excitement ratings to small vs. large monetary reward “Flattened”
Rutledge 2015 L-DOPA 150mg N Within = 30 Happiness ratings to small vs. large monetary reward “Flattened”
Wardle 2012 d-amphetamine 10mg & 20mg N Within = 36 Positive ratings of pictures Increase/’Flattened” Enhanced positive ratings of both pleasant & unpleasant pictures
Ferreri 2018 L-DOPA 100mg N Within = 27 Liking ratings of music Increase Compared to risperidone
Dopamine Reduction
First Author Year Drug Dose n Measure(s) Result(s) Notes
Bjork 2014 APTD -- N Within = 16 POMS (total mood disturbance) to monetary reward No effect
Brauer 1995 pimozide 4mg N Within = 12 POMS (elated, positive mood) ARCI (A, MBG) to d-amphetamine No effect
Brauer 1995 fluphenazine 3mg & 6mg N Within = 12 POMS (elated, positive mood) ARCI (A, MBG) to d-amphetamine No effect
Brom 2016 haloperidol 3mg N Drug = 29 N Placebo = 19 Affective ratings of sexual pictures No effect
Frank 2016 APTD -- N Within = 34 Liking ratings of food No effect
Hardman 2012 APTD -- N Within = 17 Liking ratings of food No effect
Leyton 2007 APTD 0.3mg N Within = 14 POMS (elated) VAS (euphoria) to d-amphetamine No effect
Barrett 2008 APTD -- N Within = 18 VAS (euphoric) to alcohol No effect
Leyton 2005 APTD -- N Within = 8 VAS (euphoric) to cocaine No effect
Venugopalan 2011 APTD -- N Within = 47 VAS (euphoria) to nicotine No effect Participants had varying levels of nicotine use
McCabe 2011 sulpiride 400mg N Drug = 15 N Placebo = 15 Liking ratings of mood No effect
McCabe 2012 pramipexole .2 mg N Within = 16 Liking ratings of food POMS (elated, positive mood) No effect
Wachtel 2002 haloperidol 3mg N Within = 18 VAS (euphoria) ACRI (A, GB) to methamphetamine No effect
Zack 2007 haloperidol 3mg N Within = 18 VAS (enjoyment, excitement) to gambling No effect
Ferreri 2018 risperidone 2mg N Within = 27 Liking ratings of music Decrease Compared to L-DOPA
Behavioral Findings
Dopamine Enhancement
First Author Year Drug Dose n Measure(s) Result(s) Notes
Ferreri 2018 L-DOPA 100mg N Within = 27 EDA to music Increase Compared to risperidone
Wardle 2012 d-amphetamine 10mg & 20mg 36 N Within = 36 EMG to pictures Increase for zygomatic; Decrease for corrugator Zygomatic is “smile”, corrugator is “frown”
Behavioral Findings
Dopamine Reduction
First Author Year Drug Dose n Measure(s) Result(s) Notes
Ferreri 2018 risperidone 2mg N Within = 27 EDA to music Decrease Compared to L-DOPA
Neuroimaging Findings
Dopamine Enhancement
First Author Year Drug Dose n Measure(s) Result(s) Notes
Knutson 2004 d-amphetamine .25mg/kg N Within = 8 BOLD to monetary reward No effect
Dean 2016 bupropion 150mg N Within = 17 BOLD to food “Flattened”- Increase in medial OFC to pleasant & unpleasant; Increase in amygdala & VS to unpleasant; Decrease in caudate to pleasant No effect on self-report; sub-acute dosing over 7 days
Melrose 2016 d-amphetamine .5mg/kg N Within = 11 BOLD to sucrose “Flattened” - Increase in right middle insula to sucrose; Decrease in dorsal ACC to sucrose
Abler 2011 bupropion 150mg N Within = 18 BOLD to erotic video Increase in posterior midcingulate, mediodorsal thalamus, amygdala Not related to subjective sexual functioning; sub-acute dosing over 7 days
Abler 2012 bupropion 150mg N Within = 18 BOLD to erotic video Increase in NAcc; decreased interaction between NAcc and avPFC Only compared to paroxetine
O’Daly 2014 d-amphetamine 2 mg, 4 times N Drug = 11 N Placebo = 11 BOLD to monetary reward Increase in amygdala “Sensitization” paradigm with repeated doses of amphetamine
Oei 2012 L-DOPA 100mg N Drug = 18 N Placebo = 17 BOLD to erotic images Increase in NAcc & dorsal ACC
Ivanov 2014 methylphenidate .5mg/kg N Within = 16 BOLD to monetary reward Decrease in mid-occipital & left inferior parietal cortex; Decrease in left insula & pallidum
Dopamine Reduction
First Author Year Drug Dose n Measure(s) Result(s) Notes
Bjork 2014 APTD -- N Within = 16 BOLD to monetary reward in MID No effect No effect on self-report
Sescousse 2016 sulpiride 400mg N Within = 22 BOLD to near-misses in a gambling task No effect
Frank 2016 APTD -- N Within = 34 BOLD to food Decrease in striatum; Increase in superior frontal gyrus No effect on self-report
McCabe 2011 sulpiride 400mg N Drug = 15 N Placebo = 15 BOLD to food Decrease in VS & ACC No effect on self-report
McCabe 2012 pramipexole 0.25mg N Within = 16 BOLD to food Decrease in VS & dorsal ACC to pleasant food; Decrease in inferior OFC & insula to unpleasant food No effect on self-report
Oei 2012 haloperidol 3mg N Drug = 18 N Placebo BOLD to erotic images Decrease in NAcc & dorsal ACC
Riba 2008 pramipexole 0.5mg N Within = 15 BOLD to monetary reward in gambling task Decrease in rostral basal ganglia, including VS
Graf 2015 amisulpride 200mg/d N Within = 19 BOLD to erotic video No effect
Ye 2011 pramipexole 0.5mg N Within = 16 BOLD to monetary reward in MID Increase in right NAcc

Table 5.

Reward Learning

“Raw” Behavioral Findings
Dopamine Enhancement
First Author Year Drug Dose n Measure(s) Result(s) Notes
Pessiglione 2006 L-DOPA 100mg N Drug = 13 N Placebo = 13 Money won & % of rewarding choices on instrumental learning task Increase L-DOPA only different from haloperidol, not placebo
Pleger 2009 L-DOPA 100mg N Drug = 10 N Placebo = 10 Accuracy on a somatosensory judgement task Increase
Bellebaum 2016 modafinil 200mg N Drug = 18 N Placebo = 21 % correct choice on probabilistic selection task during learning phase No effect
Bernacer 2013 methamphetamine 0.3mg/kg N Within = 18 Accuracy on reinforcement learning task No effect
Bernacer 2013 methamphetamine + amisulpride 0.3mg/kg 400 mg N Within = 18 Accuracy on reinforcement learning task No effect
Diederen 2017 bromocriptine 2.5mg N Drug = 20 N Placebo = 21 Accuracy in predicting reward magnitude based on reward history No effect Reduced accuracy at a trend level
Dodds 2008 methylphenidate 60mg N Within = 17 # of errors preceding a switch & probability of switching after error feedback No effect
Jocham 2011 amisulpride 200mg N Within = 16 % correct choice on probabilistic selection task during learning phase No effect
Weis 2013 L-DOPA 100mg N Drug = 27 N Placebo = 28 Learning rate on instrumental conditioning paradigm No effect
van der Schaff 2012 bromocriptine 1.25mg N Within = 22 Proportion of correct responses following unexpected outcomes on reversal learning task Decrease Compared to sulpiride, not placebo
Vo 2006 L-DOPA 100mg N Within = 26 Accuracy on reversal learning task Decrease
Clatworthy 2009 methylphenidate 60mg N Within = 10 Accuracy on reversal learning task Inverted-U Increased learning in those with small displacements & impaired learning in those with larger displacements in caudate nucleus
Cools 2009 bromocriptine 1.25mg N Within = 11 Relative reversal learning scores on reversal learning task Inverted-U Improved learning in those with lower dopamine synthesis in the striatum measured by PET
Mueller 2014 sulpiride 200mg N Drug = 43 N Placebo = 43 % of catch trials on ball catching task Inverted-U Worse task performance in extraverts compared to introverts
Cohen 2007 cabergoline 1.5mg N Within = 22 Speed to reach learning criteria on reversal learning task Rich get richer Sped up learning in A- (increased DA) genotype; slowed learning in A+ (reduced DA) genotype
van der Schaff 2013 methylphenidate 20mg N Within = 19 Proportion of correct responses following unexpected outcomes on reversal learning task Rich get richer Increased learning in greater working memory & impaired in low working memory
Frank 2006 haloperidol 2mg N Within = 28 Accuracy to positive stimulus & accuracy in withholding response to most negative stimulus Inverted-U “Go-to-win” paradigm; haloperidol increased go learning & decreased no-go in low working memory
Swart 2017 methylphenidate 20mg N Within = 99 Probability of correct choice for go to win No effect “Go-to-win” paradigm
Guitart-Masip 2014 L-DOPA 150mg N Drug = 30 N Placebo = 29 Probability of correct choice for go & correct omission for no-go Decrease “Go-to-win” paradigm
Dopamine Reduction
First Author Year Drug Dose n Measure(s) Result(s) Notes
Diederen 2017 sulpiride 600mg N Drug = 22 N Placebo = 21 Accuracy in predicting reward magnitude based on reward history Decrease
Forster 2017 haloperidol 3mg N Drug = 18 N Placebo = 21 Accuracy on a time estimation task Decrease Sample from Zirnheld et al. 2004
Janssen 2015 sulpiride 400mg 22 Accuracy on reversal trials of reversal learning task Decrease
Jocham 2014 amisulpride 400mg N Within = 22 Probability of selecting rewarding option on instrumental learning task Decrease Interaction was marginally significant & results reflect a post hoc test
Jocham 2014 memantine 20mg N Within = 22 Probability of selecting rewarding option on instrumental learning task Decrease Approach learning was decreased, avoidance learning was unaffected
Leyton 2007 APTD -- N Within = 13 Accuracy on rewarding-correct trials on go/no-go task Decrease
Pessiglione 2006 haloperidol 1mg N Drug = 13 N Placebo = 13 Money won on instrumental learning task Decrease Haloperidol group won less money compared to L-DOPA, not placebo
Pizzagali 2008 pramipexole 5mg N Drug = 11 N Placebo = 13 Reward bias over blocks & win-stay strategy on probabilistic reward task Decrease
Pleger 2009 haloperidol 2mg N Drug = 10 N Placebo = 10 Accuracy on a somatosensory judgement task Decrease
Tremblay 2011 haloperidol 3mg N Within = 20 Relationship between payoff & bet size on slot machine gambling task Decrease
Zirnheld 2004 haloperidol 3mg N Drug = 17 N Placebo = 18 Accuracy on a time estimation task Decrease
Brom 2016 haloperidol 3mg 58 Conditioned sexual arousal No effect
de Wit 2012 APTD -- N Drug = 14 N Placebo = 14 Accuracy on instrumental learning No effect
Dodds 2008 sulpiride 400mg N Within = 17 # of consecutive errors preceding a switch & probability of switching after error feedback No effect
Eisenegger 2014 sulpiride 800mg N Drug = 41 N Placebo = 35 Probability of choosing the correct symbol No effect
Hamidovic 2008 pramipexole .025 & 0.5mg N Within = 10 Errors of omission & commission on go/no-go & performance on card preservation task No effect
Robinson 2010 APTD -- N Within = 27 Proportion of errors for each type of trial on reversal learning No effect
Tobia 2014 APTD -- N Drug = 25 N Placebo = 30 Reaching most lucrative state in strategic sequential investment No effect
Frank 2006 cabergoline 1.25mg N Within = 28 Accuracy when most positive stimulus is presented & accuracy in withholding response to most negative stimulus Inverted-U Increased avoidance learning relative to approach learning in high working memory
van der Schaff 2012 sulpiride 400mg N Within = 22 Proportion of correct responses following unexpected outcomes on reversal learning task Inverted-U Greater working memory associated with greater sulpiride-induced increases in learning
Computational Modeling Findings
Dopamine Enhancement
First Author Year Drug Dose n Measure(s) Result(s) Notes
Diederen 2017 bromocriptine 2.5mg N Drug = 20 N Placebo = 21 Model fit across several computational models; α, PE scaling & decay in α as predicted by an adaptive Pearce-Hall computational model No effect on α or β No effect on PE scaling No effect on decay
Guitart-Masip 2014 L-DOPA 150mg N Drug = 30 N Placebo = 29 α, β, Pavlovian & bias parameters as predicted by modified Rescorla-Wagner computational models No effect on α or β Decrease in Pavlovian parameter No effect on bias parameter
Jocham 2011 amisulpride 200mg N Within = 16 α & β as predicted by standard Q-leaning computational model No effect on α or β
Wunderluch 2012 L-DOPA 150mg N Within = 18 α, β, & model-based & model-free parameters on hybrid model No effect on α or β
Cools 2009 bromocriptine 1.25mg N Within = 11 α predicted by an updated Q-learning model that included a rule to update the alternative choice in the opposite direction Inverted-U Increased α in low DA synthesis, decreased α in high DA synthesis; similar results found with working memory
Bernacer 2013 methamphetamine alone & methamphetamine + amisulpride 0.03mg/kg & 0.3mg/kg + 400mg N Within = 18 α & β as predicted by standard Q-learning computational model Increase in α No effect on β
Dopamine Reduction
First Author Year Drug Dose n Measure(s) Result(s) Notes
Diederen 2017 sulpiride 600mg N Drug = 22 N Placebo = 21 α, PE scaling & decay in α as predicted by an adaptive Pearce-Hall computational model No effect on α or β No effect on decay Decrease in scaling parameter
Eisenegger 2014 sulpiride 800mg N Drug = 41 N Placebo = 35 α & β as predicted by standard Q-learning computational model No effect on α Increased β Only in A1 carriers of the TaqlA polymorphism
Jocham 2014 amisulpride 400mg N Within = 22 Model fit; α, β, & performance as predicted by updated (relative) Q-learning computational model No effect on α or β Decreased performance
Jocham 2014 memantine 20mg N Within = 22 Model fit; α, β, & performance as predicted by updated (relative) Q-learning computational model No effect on α or β No effect on performance
Tobia 2014 APTD -- N Drug = 25 N Placebo = 30 Model fit, α, & β predicted by standard Q-learning & fictive prediction error (FPE) models on strategic sequential investment task No effect on α No effect on β
Neuroimaging Findings - fMRI
Dopamine Enhancement
First Author Year Drug Dose n Measure(s) Result(s) Notes
Dodds 2008 methylphenidate 60mg N Within = 17 BOLD to switch after negative & stay after positive feedback Increase in ACC & PFC during positive feedback; Decrease in VS during switch
Jocham 2011 amisulpride 200mg N Within = 16 BOLD correlates of prediction errors Increase in VS & frontal cortex vmPFC correlated with overall learning under amisulpride
Pessiglione 2006 L-DOPA 100mg N Drug = 13 N Placebo = 13 BOLD correlates of prediction errors Increase in VS VS activity correlated with RPE
Pleger 2009 L-DOPA 100mg N Drug = 10 N Placebo = 10 BOLD to monetary reward Increase in VS & OFC
Diederen 2017 bromocriptine 2.5mg N Drug = 20 N Placebo = 21 BOLD correlates of prediction errors No effect
van der Schaff 2012 bromocriptine 1.25mg N Within = 22 BOLD to rewards & punishments No effect Compared to sulpiride, did not differ from placebo
Weis 2013 L-DOPA 100mg N Drug = 27 N Placebo = 28 BOLD to reward feedback No effect
Graf 2016 bupropion 150mg N Within = 17 BOLD to reward outcome on MID task No effect
Bernacer 2013 methamphetamine alone & methamphetamine + amisulpride 0.03mg/kg & 0.3mg/kg + 400mg N Within = 18 BOLD correlates of prediction errors Decrease in left NAcc Not restored by amisulpride
Evers 2017 methylphenidate 40mg N Within = 24 BOLD correlates of prediction errors Decrease in VS
Dopamine Reduction
First Author Year Drug Dose n Measure(s) Result(s) Notes
Diederen 2017 sulpiride 600mg N Drug = 22 N Placebo = 21 BOLD correlates of prediction errors Decrease in VS & midbrain
Dodds 2008 methylphenidate + sulpiride 60mg 400mg N Within = 17 BOLD to switch after negative & stay after positive feedback Decrease in VS during switch; increase in ACC & PFC during positive feedback Sulpiride did not change methylphenidate effects
Jocham 2014 amisulpride 400mg N Within = 22 BOLD correlates of prediction errors Decrease in VS & caudate
Jocham 2014 memantine 20mg N Within = 22 BOLD correlates of prediction errors Decrease in VS & caudate
Pessiglione 2006 haloperidol 1mg N Drug = 13 N Placebo = 13 BOLD correlates of prediction errors Decrease in striatum
Pleger 2009 haloperidol 2mg N Drug = 10 N Placebo = 10 BOLD to monetary reward Decrease in VS & OFC
Tobia 2014 APTD -- N Drug = 25 N Placebo = 30 BOLD correlates of prediction errors Decrease in thalamus & amygdala
van der Schaff 2012 sulpiride 400mg N Within = 22 BOLD to monetary reward Inverted-U Sulpiride-induced increases in VS correlated with reversal learning in greater working memory
Neuroimaging Findings - EEG
Dopamine Enhancement
First Author Year Drug Dose n Measure(s) Result(s) Notes
de Bruijn 2004 d-amphetamine 15mg N Within = 12 ERN amplitude to errors on Flanker task Increase ERN more negative
Mueller 2011 sulpiride 200mg N Drug = 90 N Placebo = 79 ERN amplitude to errors on Flanker task Inverted-U Sulpride enhanced ERN in Val/Val genotype, reduced ERN in Met carriers
Mueller 2014a sulpiride 200mg N Drug = 45 N Placebo = 41 FRN amplitude to errors on ball catching task Inverted-U Same participants as Mueller 2014b; Decreased only in extraverts compared to introverts
Mueller 2014b sulpiride 200mg N Drug = 45 N Placebo = 38 FRN amplitude to errors on ball catching task Inverted-U Same participants as Mueller 2014a; Val carriers had enhanced FRNs; under low dose sulpiride this effect was reversed
Apitz 2014 L-DOPA 150mg N Drug = 20 N Placebo = 18 ERFs to monetary reward No effect ERFs potentially represent the MEG counterpart to the FRN
Dopamine Reduction
First Author Year Drug Dose n Measure(s) Result(s) Notes
de Bruijn 2006 haloperidol 2.5mg N Within = 14 ERN amplitude to errors on Flanker task Decrease ERN was attenuated (less negative)
Santesso 2009 pramipexole 5mg N Drug = 7 N Placebo = 13 FRN to correct feedback Decrease Subset of participants from Pizzagali 2008
Zirnheld 2004 haloperidol 3mg N Drug = 17 N Placebo = 18 ERN amplitude to errors on Flanker task Decrease
Forster 2017 haloperidol 3mg N Drug = 18 N Placebo = 21 FRN amplitude to negative feedback on a temporal interval learning task No effect Same participants from Zirnheld et al. 2004
Larson 2015 APTD -- N Within = 12 ERN amplitude on Stroop task No effect

Highlights.

  • The role of dopamine in human reward functioning is highly researched

  • Drug challenge studies are translational and have good experimental control

  • Dopamine drug challenge studies in healthy adults were summarized by reward phase

  • Dopamine drugs have differing effects on various reward phases

  • Drug effects are likely nonlinear and behavioral effects depend on outcome measure

Acknowledgements

This work was funded in part by the National Institute on Drug Abuse F32DA048542 (HES) and K08DA040006 (MCW) and the “la Caixa” Foundation, ID 100010434, fellowship code LCF/BQ/DI19/11730047 (PLG). The sponsors had no role in design of this research or the decision to submit the article for publication.

Footnotes

Publisher's Disclaimer: This is a PDF file of an unedited manuscript that has been accepted for publication. As a service to our customers we are providing this early version of the manuscript. The manuscript will undergo copyediting, typesetting, and review of the resulting proof before it is published in its final form. Please note that during the production process errors may be discovered which could affect the content, and all legal disclaimers that apply to the journal pertain.

1

Some equations have been simplified for the purposes of the review.

References

  1. Abler B, Erk S, Walter H, 2007. Human reward system activation is modulated by a single dose of olanzapine in healthy subjects in an event-related, double-blind, placebo-controlled fMRI study. Psychopharmacology (Berl). 191, 823–833. 10.1007/s00213-006-0690-y [DOI] [PubMed] [Google Scholar]
  2. Abler B, Grön G, Hartmann A, Metzger C, Walter M, 2012. Modulation of frontostriatal interaction aligns with reduced primary reward processing under serotonergic drugs. J. Neurosci 32, 1329–1335. 10.1523/JNEUROSCI.5826-11.2012 [DOI] [PMC free article] [PubMed] [Google Scholar]
  3. Abler B, Seeringer A, Hartmann A, Grön G, Metzger C, Walter M, Stingl J, 2011. Neural correlates of antidepressant-related sexual dysfunction: A placebo-controlled fmri study on healthy males under subchronic paroxetine and bupropion. Neuropsychopharmacology 36, 1837–1847. 10.1038/npp.2011.66 [DOI] [PMC free article] [PubMed] [Google Scholar]
  4. Acheson A, de Wit H, 2008. Bupropion improves attention but does not affect impulsive behavior in healthy young adults. Exp. Clin. Psychopharmacol 16, 113–123. 10.1037/1064-1297.16.2.113 [DOI] [PMC free article] [PubMed] [Google Scholar]
  5. Addicott MA, Schechter JC, Sapyta JJ, Selig JP, Kollins SH, Weiss MD, 2019. Methylphenidate increases willingness to perform effort in adults with ADHD. 10.1016/j.pbb.2019.06.008 [DOI] [PMC free article] [PubMed]
  6. Apitz T, Bunzeck N, 2014. Early Effects of Reward Anticipation Are Modulated by Dopaminergic Stimulation. PLoS One 9, e108886 10.1371/journal.pone.0108886 [DOI] [PMC free article] [PubMed] [Google Scholar]
  7. Arrondo G, Aznárez-Sanado M, Fernández-Seara MA, Goñi J, Loayza FR, Salamon-Klobut E, Heukamp FH, Pastor MA, 2015. Dopaminergic modulation of the trade-off between probability and time in economic decision-making. Eur. Neuropsychopharmacol 25, 817–827. 10.1016/J.EURONEURO.2015.02.011 [DOI] [PubMed] [Google Scholar]
  8. Assadi SM, Yücel M, Pantelis C, 2009. Dopamine modulates neural networks involved in effort-based decision-making, Neuroscience and Biobehavioral Reviews. Pergamon 10.1016/j.neubiorev.2008.10.010 [DOI] [PubMed] [Google Scholar]
  9. Bailey MR, Simpson EH, Balsam PD, 2016. Neural substrates underlying effort, time, and risk-based decision making in motivated behavior. Neurobiol. Learn. Mem 133, 233–256. 10.1016/J.NLM.2016.07.015 [DOI] [PMC free article] [PubMed] [Google Scholar]
  10. Bamford NS, Wightman RM, Sulzer D, 2018. Dopamine’s Effects on Corticostriatal Synapses during Reward-Based Behaviors. Neuron 10.1016/j.neuron.2018.01.006 [DOI] [PMC free article] [PubMed] [Google Scholar]
  11. Baranski JV, Pigeau R, Dinich P, Jacobs I, 2004. Effects of modafinil on cognitive and meta-cognitive performance. Hum. Psychopharmacol. Clin. Exp 19, 323–332. 10.1002/hup.596 [DOI] [PubMed] [Google Scholar]
  12. Barbano MF, Cador M, 2006. Differential Regulation of the Consummatory, Motivational and Anticipatory Aspects of Feeding Behavior by Dopaminergic and Opioidergic Drugs. Neuropsychopharmacology 31, 1371–1381. 10.1038/sj.npp.1300908 [DOI] [PubMed] [Google Scholar]
  13. Barch DM, Treadway MT, Schoen N, 2014. Effort, anhedonia, and function in schizophrenia: Reduced effort allocation predicts amotivation and functional impairment. J. Abnorm. Psychol 123, 387–397. 10.1037/a0036299 [DOI] [PMC free article] [PubMed] [Google Scholar]
  14. Barrett SP, Pihl RO, Benkelfat C, Brunelle C, Young SN, Leyton M, 2008. The role of dopamine in alcohol self-administration in humans: Individual differences. Eur. Neuropsychopharmacol 18, 439–447. 10.1016/j.euroneuro.2008.01.008 [DOI] [PubMed] [Google Scholar]
  15. Bassareo V, De Luca MA, Di Chiara G, 2007. Differential impact of pavlovian drug conditioned stimuli on in vivo dopamine transmission in the rat accumbens shell and core and in the prefrontal cortex. Psychopharmacology (Berl). 191, 689–703. 10.1007/s00213-006-0560-7 [DOI] [PubMed] [Google Scholar]
  16. Beaulieu J-M, Gainetdinov RR, 2011. The physiology, signaling, and pharmacology of dopamine receptors. Pharmacol. Rev 63, 182–217. 10.1124/pr.110.002642 [DOI] [PubMed] [Google Scholar]
  17. Beierholm U, Guitart-Masip M, Economides M, Chowdhury R, Düzel E, Dolan R, Dayan P, Duzel E, Dolan R, Dayan P, 2013. Dopamine modulates reward related vigor. Neuropsychopharmacology 38, 1495–503. 10.1038/npp.2013.48 [DOI] [PMC free article] [PubMed] [Google Scholar]
  18. Bellebaum C, Kuchinke L, Roser P, 2017. Modafinil alters decision making based on feedback history – a randomized placebo-controlled double blind study in humans. J. Psychopharmacol 31, 243–249. 10.1177/0269881116668591 [DOI] [PubMed] [Google Scholar]
  19. Berke JD, 2018a. What does dopamine mean? Nat. Neurosci 21, 787–793. 10.1038/s41593-018-0152-y [DOI] [PMC free article] [PubMed] [Google Scholar]
  20. Berke JD, 2018b. What does dopamine mean? Nat. Neurosci 10.1038/s41593-018-0152-y [DOI] [PMC free article] [PubMed] [Google Scholar]
  21. Bernacer J, Corlett PR, Ramachandra P, McFarlane B, Turner DC, Clark L, Robbins TW, Fletcher PC, Murray GK, 2013. Methamphetamine-induced disruption of Frontostriatal reward learning signals: Relation to psychotic symptoms. Am. J. Psychiatry 170, 1326–1334. 10.1176/appi.ajp.2013.12070978 [DOI] [PubMed] [Google Scholar]
  22. Berridge KC, 1996. Food reward: Brain substrates of wanting and liking. Neurosci. Biobehav. Rev 20, 1–25. 10.1016/0149-7634(95)00033-B [DOI] [PubMed] [Google Scholar]
  23. Berridge KC, Kringelbach ML, 2015. Pleasure Systems in the Brain. Neuron 10.1016/j.neuron.2015.02.018 [DOI] [PMC free article] [PubMed] [Google Scholar]
  24. Berridge KC, Robinson TE, 2003. Parsing reward. Trends Neurosci. 26, 507–513. [DOI] [PubMed] [Google Scholar]
  25. Berridge KC, Robinson TE, 1998. What is the role of dopamine in reward: Hedonic impact, reward learning, or incentive salience? Brain Res. Rev 28, 309–369. 10.1016/S0165-0173(98)00019-8 [DOI] [PubMed] [Google Scholar]
  26. Bjork JM, Grant SJ, Chen G, Hommer DW, 2014. Dietary Tyrosine/Phenylalanine Depletion Effects on Behavioral and Brain Signatures of Human Motivational Processing. Neuropsychopharmacology 39, 595–604. 10.1038/npp.2013.232 [DOI] [PMC free article] [PubMed] [Google Scholar]
  27. Boileau I, Dagher A, Leyton M, Welfeld K, Booij L, Diksic M, Benkelfat C, 2007. Conditioned dopamine release in humans: A positron emission tomography [¬π¬πC]raclopride study with amphetamine. J. Neurosci 27, 3998–4003. 10.1523/jneurosci.4370-06.2007 [DOI] [PMC free article] [PubMed] [Google Scholar]
  28. Botvinick MM, Huffstetler S, McGuire JT, 2009. Effort discounting in human nucleus accumbens. Cogn. Affect. Behav. Neurosci 9, 16–27. 10.3758/CABN.9.1.16 [DOI] [PMC free article] [PubMed] [Google Scholar]
  29. Brauer LH, de Wit H, 1995. Role of dopamine in d-amphetamine-induced euphoria in normal, healthy volunteers. Exp. Clin. Psychopharmacol 3, 371–381. 10.1037/1064-1297.3.4.371 [DOI] [Google Scholar]
  30. Brom M, Laan E, Everaerd W, Spinhoven P, Trimbos B, Both S, 2016. The effect of a dopamine antagonist on conditioning of sexual arousal in women. Psychopharmacology (Berl). 233, 1179–1189. 10.1007/s00213-015-4201-x [DOI] [PMC free article] [PubMed] [Google Scholar]
  31. Burke CJ, Soutschek A, Weber S, Raja Beharelle A, Fehr E, Haker H, Tobler PN, 2018. Dopamine Receptor-Specific Contributions to the Computation of Value. Neuropsychopharmacology 43, 1415–1424. 10.1038/npp.2017.302 [DOI] [PMC free article] [PubMed] [Google Scholar]
  32. Campbell-Meiklejohn D, Simonsen A, Scheel-Krüger J, Wohlert V, Gjerløff T, Frith CD, Rogers RD, Roepstorff A, Møller A, 2012. In for a penny, in for a pound: methylphenidate reduces the inhibitory effect of high stakes on persistent risky choice. J. Neurosci 32, 13032–8. 10.1523/JNEUROSCI.0151-12.2012 [DOI] [PMC free article] [PubMed] [Google Scholar]
  33. Campbell-Meiklejohn D, Wakeley J, Herbert V, Cook J, Scollo P, Ray MK, Selvaraj S, Passingham RE, Cowen P, Rogers RD, 2011. Serotonin and dopamine play complementary roles in gambling to recover losses. Neuropsychopharmacology 36, 402–10. 10.1038/npp.2010.170 [DOI] [PMC free article] [PubMed] [Google Scholar]
  34. Camps M, Kelly PH, Palacios JM, 1990. Autoradiographic localization of dopamine D1 and D2 receptors in the brain of several mammalian species. J. Neural Transm 80, 105–127. 10.1007/BF01257077 [DOI] [PubMed] [Google Scholar]
  35. Cannon CM, Bseikri MR, 2004. Is dopamine required for natural reward? Physiol. Behav 81, 741–748. 10.1016/j.physbeh.2004.04.020 [DOI] [PubMed] [Google Scholar]
  36. Castrellon JJ, Seaman KL, Crawford JL, Young JS, Smith CT, Dang LC, Hsu M, Cowan RL, Zald DH, Samanez-Larkin GR, 2019. Individual differences in dopamine are associated with reward discounting in clinical groups but not in healthy adults. J. Neurosci 39, 321–332. 10.1523/JNEUROSCI.1984-18.2018 [DOI] [PMC free article] [PubMed] [Google Scholar]
  37. Castro DC, Berridge KC, 2014. Opioid Hedonic Hotspot in Nucleus Accumbens Shell: Mu, Delta, and Kappa Maps for Enhancement of Sweetness “Liking” and “Wanting.” J. Neurosci 34, 4239–4250. 10.1523/JNEUROSCI.4458-13.2014 [DOI] [PMC free article] [PubMed] [Google Scholar]
  38. Cawley E, Park S, aan het Rot M, Sancton K, Benkelfat C, Young S, Boivin D, Leyton M, 2013. Dopamine and light: dissecting effects on mood and motivational states in women with subsyndromal seasonal affective disorder. J. Psychiatry Neurosci 38, 388–397. 10.1503/jpn.120181 [DOI] [PMC free article] [PubMed] [Google Scholar]
  39. Chen YCI, Choi JK, Andersen SL, Rosen BR, Jenkins BG, 2005. Mapping dopamine D2/D3 receptor function using pharmacological magnetic resonance imaging. Psychopharmacology (Berl). 180, 705–715. 10.1007/s00213-004-2034-0 [DOI] [PubMed] [Google Scholar]
  40. Childs E, de Wit H, 2009. Amphetamine-Induced Place Preference in Humans. Biol. Psychiatry 65, 900–904. [DOI] [PMC free article] [PubMed] [Google Scholar]
  41. Chong TT-J, Apps M, Giehl K, Sillence A, Grima LL, Husain M, 2017. Neurocomputational mechanisms underlying subjective valuation of effort costs. PLOS Biol. 15, e1002598 10.1371/journal.pbio.1002598 [DOI] [PMC free article] [PubMed] [Google Scholar]
  42. Ciucci MR, Ma ST, Fox C, Kane JR, Ramig LO, Schallert T, 2007. Qualitative changes in ultrasonic vocalization in rats after unilateral dopamine depletion or haloperidol: A preliminary study. Behav. Brain Res. 182, 284–289. 10.1016/j.bbr.2007.02.020 [DOI] [PMC free article] [PubMed] [Google Scholar]
  43. Clatworthy PL, Lewis SJG, Brichard L, Hong YT, Izquierdo D, Clark L, Cools R, Aigbirhio FI, Baron J-C, Fryer TD, Robbins TW, 2009. Dopamine Release in Dissociable Striatal Subregions Predicts the Different Effects of Oral Methylphenidate on Reversal Learning and Spatial Working Memory. J. Neurosci 29, 4690–4696. 10.1523/JNEUROSCI.3266-08.2009 [DOI] [PMC free article] [PubMed] [Google Scholar]
  44. Cohen JY, Haesler S, Vong L, Lowell BB, Uchida N, 2012. Neuron-type-specific signals for reward and punishment in the ventral tegmental area. Nature 482, 85–88. 10.1038/nature10754 [DOI] [PMC free article] [PubMed] [Google Scholar]
  45. Cohen MX, Krohn-Grimberghe A, Elger CE, Weber B, 2007. Dopamine gene predicts the brain’s response to dopaminergic drug. Eur. J. Neurosci 26, 3652–3660. 10.1111/j.1460-9568.2007.05947.x [DOI] [PubMed] [Google Scholar]
  46. Colasanti A, Searle GE, Long CJ, Hill SP, Reiley RR, Quelch D, Erritzoe D, Tziortzi AC, Reed LJ, Lingford-Hughes AR, Waldman AD, Schruers KRJJ, Matthews PM, Gunn RN, Nutt DJ, Rabiner EA, 2012. Endogenous Opioid Release in the Human Brain Reward System Induced by Acute Amphetamine Administration. Biol. Psychiatry 72, 371–377. 10.1016/j.biopsych.2012.01.027 [DOI] [PubMed] [Google Scholar]
  47. Cools R, 2019. Chemistry of the Adaptive Mind: Lessons from Dopamine. Neuron 10.1016/j.neuron.2019.09.035 [DOI] [PubMed] [Google Scholar]
  48. Cools R, D’Esposito M, 2011. Inverted-U-shaped dopamine actions on human working memory and cognitive control. Biol. Psychiatry 69, e113–e125. 10.1016/j.biopsych.2011.03.028 [DOI] [PMC free article] [PubMed] [Google Scholar]
  49. Cools R, Frank MJ, Gibbs SE, Miyakawa A, Jagust W, D’Esposito M, 2009. Striatal dopamine predicts outcome-specific reversal learning and its sensitivity to dopaminergic drug administration. J. Neurosci 29, 1538–43. 10.1523/JNEUROSCI.4467-08.2009 [DOI] [PMC free article] [PubMed] [Google Scholar]
  50. Cools R, Gibbs SE, Miyakawa A, Jagust W, Esposito MD’, 2008. Brief Communications Working Memory Capacity Predicts Dopamine Synthesis Capacity in the Human Striatum. 10.1523/JNEUROSCI.4475-07.2008 [DOI] [PMC free article] [PubMed]
  51. Cooper JR, Bloom FE, Roth R, 2003. The Biochemical Basis of Neuropharmacology, 8th ed. Oxford University Press, Oxford: 10.1037/000623 [DOI] [Google Scholar]
  52. Corral-Frías NS, Pizzagalli DA, Carré JM, Michalski LJ, Nikolova YS, Perlis RH, Fagerness J, Lee MR, Conley ED, Lancaster TM, Haddad S, Wolf A, Smoller JW, Hariri AR, Bogdan R, 2016. COMT Val 158 Met genotype is associated with reward learning: a replication study and meta-analysis. Genes, Brain Behav. 15, 503–513. 10.1111/gbb.12296 [DOI] [PMC free article] [PubMed] [Google Scholar]
  53. Croxson PL, Walton ME, O’Reilly JX, Behrens TEJ, Rushworth MFS, 2009. Effort-based cost-benefit valuation and the human brain. J. Neurosci 29, 4531–41. 10.1523/JNEUROSCI.4515-08.2009 [DOI] [PMC free article] [PubMed] [Google Scholar]
  54. Curley LE, Kydd RR, Kirk IJ, Russell BR, 2013. Differential responses to anticipation of reward after an acute dose of the designer drugs benzylpiperazine (BZP) and trifluoromethylphenylpiperazine (TFMPP) alone and in combination using functional magnetic resonance imaging (fMRI). Psychopharmacology (Berl). 10.1007/s00213-013-3128-3 [DOI] [PubMed] [Google Scholar]
  55. da Silva Alves F, Schmitz N, Figee M, Abeling N, Hasler G, van der Meer J, Nederveen A, de Haan L, Linszen D, van Amelsvoort T, 2011. Dopaminergic modulation of the human reward system: a placebo-controlled dopamine depletion fMRI study. J. Psychopharmacol 25, 538–549. 10.1177/0269881110367731 [DOI] [PubMed] [Google Scholar]
  56. Davis KL, Kahn RS, Ko G, Davidson M, 1991. Dopamine in schizophrenia: a review and reconceptualization. Am. J. Psychiatry 148, 1474–1486. 10.1176/ajp.148.11.1474 [DOI] [PubMed] [Google Scholar]
  57. de Bruijn ERA, Hulstijn W, Verkes RJ, Ruigt GSF, Sabbe BGC, 2004. Drug-induced stimulation and suppression of action monitoring in healthy volunteers. Psychopharmacology (Berl). 177, 151–160. 10.1007/s00213-004-1915-6 [DOI] [PubMed] [Google Scholar]
  58. de Bruijn ERA, Sabbe BGC, Hulstijn W, Ruigt GSF, Verkes RJ, 2006. Effects of antipsychotic and antidepressant drugs on action monitoring in healthy volunteers. Brain Res. 1105, 122–129. 10.1016/J.BRAINRES.2006.01.006 [DOI] [PubMed] [Google Scholar]
  59. de Jong JW, Roelofs TJM, Mol FMU, Hillen AEJ, Meijboom KE, Luijendijk MCM, van der Eerden HAM, Garner KM, Vanderschuren LJMJ, Adan RAH, 2015. Reducing Ventral Tegmental Dopamine D2 Receptor Expression Selectively Boosts Incentive Motivation. Neuropsychopharmacology 40, 2085–2095. 10.1038/npp.2015.60 [DOI] [PMC free article] [PubMed] [Google Scholar]
  60. de Wit H, Enggasser JL, Richards JB, 2002. Acute Administration of d-Amphetamine Decreases Impulsivity in Healthy Volunteers. Neuropsychopharmacology 27, 814–823. 10.1016/S0893-133X(02)00343-3 [DOI] [PubMed] [Google Scholar]
  61. de Wit S, Standing HR, DeVito EE, Robinson OJ, Ridderinkhof KR, Robbins TW, Sahakian BJ, 2012. Reliance on habits at the expense of goal-directed control following dopamine precursor depletion. Psychopharmacology (Berl). 219, 621–631. 10.1007/s00213-011-2563-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
  62. Dean Z, Horndasch S, Giannopoulos P, McCabe C, 2016. Enhanced neural response to anticipation, effort and consummation of reward and aversion during bupropion treatment. Psychol. Med 46, 2263–2274. 10.1017/S003329171600088X [DOI] [PubMed] [Google Scholar]
  63. Dehaene S, Posner MI, Tucker DM, 1994. LOCALIZATION OF A NEURAL SYSTEM FOR ERROR DETECTION AND COMPENSATION. Psychol. Sci 5, 303–305. [Google Scholar]
  64. Der-Avakian A, D’Souza MS, Pizzagalli DA, Markou A, 2013. Assessment of reward responsiveness in the response bias probabilistic reward task in rats: implications for cross-species translational research. Transl. Psychiatry 3, e297 10.1038/tp.2013.74 [DOI] [PMC free article] [PubMed] [Google Scholar]
  65. Diederen KMJ, Ziauddeen H, Vestergaard MD, Spencer T, Schultz W, Fletcher PC, 2017. Dopamine Modulates Adaptive Prediction Error Coding in the Human Midbrain and Striatum. J. Neurosci 37, 1708–1720. 10.1523/JNEUROSCI.1979-16.2016 [DOI] [PMC free article] [PubMed] [Google Scholar]
  66. Dodds CM, Müller U, Clark L, van Loon A, Cools R, Robbins TW, 2008. Methylphenidate has differential effects on blood oxygenation level-dependent signal related to cognitive subprocesses of reversal learning. J. Neurosci 28, 5976–5982. 10.1523/JNEUROSCI.1153-08.2008 [DOI] [PMC free article] [PubMed] [Google Scholar]
  67. Doñamayor N, Marco-Pallarés J, Heldmann M, Schoenfeld MA, Münte TF, 2011. Temporal dynamics of reward processing revealed by magnetoencephalography. Hum. Brain Mapp 32, 2228–2240. 10.1002/hbm.21184 [DOI] [PMC free article] [PubMed] [Google Scholar]
  68. Doyle TG, Berridge KC, Gosnell BA, 1993. Morphine enhances hedonic taste palatability in rats. Pharmacol. Biochem. Behav 46, 745–749. 10.1016/0091-3057(93)90572-B [DOI] [PubMed] [Google Scholar]
  69. Dreher J-C, 2013. Neural coding of computational factors affecting decision making. Prog. Brain Res 202, 289–320. 10.1016/B978-0-444-62604-2.00016-2 [DOI] [PubMed] [Google Scholar]
  70. Eisenegger C, Naef M, Linssen A, Clark L, Gandamaneni PK, Müller U, Robbins TW, 2014. Role of dopamine D2 receptors in human reinforcement learning. Neuropsychopharmacology 39, 2366–2375. 10.1038/npp.2014.84 [DOI] [PMC free article] [PubMed] [Google Scholar]
  71. Evers EAL, Stiers P, Ramaekers JG, 2017. High reward expectancy during methylphenidate depresses the dopaminergic response to gain and loss. Soc. Cogn. Affect. Neurosci 12, 311–318. 10.1093/scan/nsw124 [DOI] [PMC free article] [PubMed] [Google Scholar]
  72. Ferreri L, Mas-Herrero E, Zatorre RJ, Ripollés P, Gomez-Andres A, Alicart H, Olivé G, Marco-Pallarés J, Antonijoan RM, Valle M, Riba J, Rodriguez-Fornells A, 2019. Dopamine modulates the reward experiences elicited by music. Proc. Natl. Acad. Sci. U. S. A 116, 3793–3798. 10.1073/pnas.1811878116 [DOI] [PMC free article] [PubMed] [Google Scholar]
  73. Fiore VG, Nolte T, Rigoli F, Smittenaar P, Gu X, Dolan RJ, 2018. Value encoding in the globus pallidus: fMRI reveals an interaction effect between reward and dopamine drive. Neuroimage 173, 249–257. 10.1016/j.neuroimage.2018.02.048 [DOI] [PMC free article] [PubMed] [Google Scholar]
  74. Floresco SB, Tse MTL, Ghods-Sharifi S, 2008. Dopaminergic and Glutamatergic Regulation of Effort- and Delay-Based Decision Making. Neuropsychopharmacology 33, 1966–1979. 10.1038/sj.npp.1301565 [DOI] [PubMed] [Google Scholar]
  75. Floresco SB, West AR, Ash B, Moorel H, Grace AA, Moore H, Grace AA, 2003. Afferent modulation of dopamine neuron firing differentially regulates tonic and phasic dopamine transmission. Nat. Neurosci 6, 968–973. 10.1038/nn1103 [DOI] [PubMed] [Google Scholar]
  76. Forster SE, Zirnheld P, Shekhar A, Steinhauer SR, O’Donnell BF, Hetrick WP, 2017. Event-related potentials reflect impaired temporal interval learning following haloperidol administration. Psychopharmacology (Berl). 234, 2545–2562. 10.1007/s00213-017-4645-2 [DOI] [PubMed] [Google Scholar]
  77. Frank MJ, O’Reilly RC, 2006. A mechanistic account of striatal dopamine function in human cognition: Psychopharmacological studies with cabergoline and haloperidol. Behav. Neurosci 120, 497–517. 10.1037/0735-7044.120.3.497 [DOI] [PubMed] [Google Scholar]
  78. Frank S, Veit R, Sauer H, Enck P, Friederich H-CC, Unholzer T, Bauer U-MM, Linder K, Heni M, Fritsche A, Preissl H, 2016. Dopamine depletion reduces food-related reward activity independent of BMI. Neuropsychopharmacology 41, 1551–1559. 10.1038/npp.2015.313 [DOI] [PMC free article] [PubMed] [Google Scholar]
  79. Funayama T, Ikeda Y, Tateno A, Takahashi H, Okubo Y, Fukayama H, Suzuki H, 2014. Modafinil augments brain activation associated with reward anticipation in the nucleus accumbens. Psychopharmacology (Berl). 231, 3217–3228. 10.1007/s00213-014-3499-0 [DOI] [PubMed] [Google Scholar]
  80. Garfield JBB, Lubman DI, Yücel M, 2013. Anhedonia in substance use disorders: A systematic review of its nature, course and clinical correlates. Aust. N. Z. J. Psychiatry 48, 36–51. 10.1177/0004867413508455 [DOI] [PubMed] [Google Scholar]
  81. Gehring WJ, Fencsik DE, 2001. Functions of the medial frontal cortex in the processing of conflict and errors. J. Neurosci 21, 9430–7. 10.1523/JNEUROSCI.21-23-09430.2001 [DOI] [PMC free article] [PubMed] [Google Scholar]
  82. Grace AA, 2000. The tonic/phasic model of dopamine system regulation and its implications for understanding alcohol and psychostimulant craving. Addiction 95, 119–128. 10.1080/09652140050111690 [DOI] [PubMed] [Google Scholar]
  83. Grace AA, 1991. Phasic versus tonic dopamine release and the modulation of dopamine system responsivity: A hypothesis for the etiology of schizophrenia. Neuroscience 41, 1–24. 10.1016/0306-4522(91)90196-U [DOI] [PubMed] [Google Scholar]
  84. Graf H, Metzger CD, Walter M, Abler B, 2016a. Serotonergic antidepressants decrease hedonic signals but leave learning signals in the nucleus accumbens unaffected. Neuroreport 27, 18–22. 10.1097/WNR.0000000000000487 [DOI] [PubMed] [Google Scholar]
  85. Graf H, Metzger CD, Walter M, Abler B, 2016b. Serotonergic antidepressants decrease hedonic signals but leave learning signals in the nucleus accumbens unaffected. Neuroreport 27, 18–22. 10.1097/WNR.0000000000000487 [DOI] [PubMed] [Google Scholar]
  86. Graf H, Wiegers M, Metzger CD, Walter M, Grön G, Abler B, 2015. Erotic stimulus processing under Amisulpride and Reboxetine: A Placebo-controlled fMRI study in healthy subjects. Int. J. Neuropsychopharmacol 18, 1–8. 10.1093/ijnp/pyu004 [DOI] [PMC free article] [PubMed] [Google Scholar]
  87. Guitart-Masip M, Chowdhury R, Sharot T, Dayan P, Duzel E, Dolan RJ, 2012a. Action controls dopaminergic enhancement of reward representations. Proc. Natl. Acad. Sci. U. S. A 109, 7511–6. 10.1073/pnas.1202229109 [DOI] [PMC free article] [PubMed] [Google Scholar]
  88. Guitart-Masip M, Economides M, Huys QJMJM, Frank MJ, Chowdhury R, Duzel E, Dayan P, Dolan RJ, 2014. Differential, but not opponent, effects of l-DOPA and citalopram on action learning with reward and punishment. Psychopharmacology (Berl). 231, 955–966. 10.1007/s00213-013-3313-4 [DOI] [PMC free article] [PubMed] [Google Scholar]
  89. Guitart-Masip M, Fuentemilla L, Bach DR, Huys QJM, Dayan P, Dolan RJ, Duzel E, 2011. Action dominates valence in anticipatory representations in the human striatum and dopaminergic midbrain. J. Neurosci 31, 7867–75. 10.1523/JNEUROSCI.6376-10.2011 [DOI] [PMC free article] [PubMed] [Google Scholar]
  90. Guitart-Masip M, Huys QJM, Fuentemilla L, Dayan P, Duzel E, Dolan RJ, 2012b. Go and no-go learning in reward and punishment: Interactions between affect and effect. Neuroimage 62, 154–166. 10.1016/J.NEUROIMAGE.2012.04.024 [DOI] [PMC free article] [PubMed] [Google Scholar]
  91. HajiHosseini A, Holroyd CB, 2015. Reward feedback stimuli elicit high-beta EEG oscillations in human dorsolateral prefrontal cortex. Sci. Rep 5, 13021 10.1038/srep13021 [DOI] [PMC free article] [PubMed] [Google Scholar]
  92. HajiHosseini A, Rodríguez-Fornells A, Marco-Pallarés J, 2012. The role of beta-gamma oscillations in unexpected rewards processing. Neuroimage 60, 1678–1685. 10.1016/J.NEUROIMAGE.2012.01.125 [DOI] [PubMed] [Google Scholar]
  93. Hamidovic A, Kang UJ, de Wit H, 2008. Effects of low to moderate acute doses of pramipexole on impulsivity and cognition in healthy volunteers. J. Clin. Psychopharmacol 28, 45–51. 10.1097/jcp.0b013e3181602fab [DOI] [PubMed] [Google Scholar]
  94. Hamilton JP, Etkin A, Furman DJ, Lemus MG, Johnson RF, Gotlib IH, 2012. Functional neuroimaging of major depressive disorder: A meta-analysis and new integration of baseline activation and neural response data. Am. J. Psychiatry 10.1176/appi.ajp.2012.11071105 [DOI] [PMC free article] [PubMed] [Google Scholar]
  95. Hardman CA, Herbert VMB, Brunstrom JM, Munafò MR, Rogers PJ, 2012. Dopamine and food reward: Effects of acute tyrosine/phenylalanine depletion on appetite. Physiol. Behav 105, 1202–1207. 10.1016/J.PHYSBEH.2011.12.022 [DOI] [PubMed] [Google Scholar]
  96. Hauser TU, Eldar E, Dolan RJ, 2017. Separate mesocortical and mesolimbic pathways encode effort and reward learning signals. Proc. Natl. Acad. Sci 114, E7395–E7404. 10.1073/pnas.1705643114 [DOI] [PMC free article] [PubMed] [Google Scholar]
  97. Hernandez L, Hoebel BG, 1988. Food reward and cocaine increase extracellular dopamine in the nucleus accumbens as measured by microdialysis. Life Sci. 42, 1705–1712. 10.1016/0024-3205(88)90036-7 [DOI] [PubMed] [Google Scholar]
  98. Ho C-Y, Berridge KC, 2013. An Orexin Hotspot in Ventral Pallidum Amplifies Hedonic ‘Liking’ for Sweetness. Neuropsychopharmacology 38, 1655–1664. 10.1038/npp.2013.62 [DOI] [PMC free article] [PubMed] [Google Scholar]
  99. Holroyd CB, Coles MGH, 2002. The neural basis of human error processing: Reinforcement learning, dopamine, and the error-related negativity. Psychol. Rev 109, 679–709. 10.1037/0033-295X.109.4.679 [DOI] [PubMed] [Google Scholar]
  100. Ikeda Y, Funayama T, Tateno A, Fukayama H, Okubo Y, Suzuki H, 2019. Bupropion increases activation in nucleus accumbens during anticipation of monetary reward. Psychopharmacology (Berl). 236, 3655–3665. 10.1007/s00213-019-05337-6 [DOI] [PubMed] [Google Scholar]
  101. Ivanov I, Liu X, Clerkin S, Schulz K, Fan J, Friston K, London ED, Schwartz J, Newcorn JH, 2014. Methylphenidate and brain activity in a reward/conflict paradigm: role of the insula in task performance. Eur. Neuropsychopharmacol 24, 897–906. 10.1016/j.euroneuro.2014.01.017 [DOI] [PubMed] [Google Scholar]
  102. Janssen LK, Sescousse G, Hashemi MM, Timmer MHM, Ter Huurne NP, Geurts DEM, Cools R, 2015. Abnormal modulation of reward versus punishment learning by a dopamine D2-receptor antagonist in pathological gamblers. Psychopharmacology (Berl). 232, 3345–3353. 10.1007/s00213-015-3986-y [DOI] [PMC free article] [PubMed] [Google Scholar]
  103. Jocham G, Klein TA, Ullsperger M, 2014. Differential Modulation of Reinforcement Learning by D2 Dopamine and NMDA Glutamate Receptor Antagonism. J. Neurosci 34, 13151–13162. 10.1523/JNEUROSCI.0757-14.2014 [DOI] [PMC free article] [PubMed] [Google Scholar]
  104. Jocham G, Klein TA, Ullsperger M, 2011. Dopamine-Mediated Reinforcement Learning Signals in the Striatum and Ventromedial Prefrontal Cortex Underlie Value-Based Choices. J. Neurosci 31, 1606–1613. 10.1523/JNEUROSCI.3904-10.2011 [DOI] [PMC free article] [PubMed] [Google Scholar]
  105. Kayser AS, Allen DC, Navarro-Cebrian A, Mitchell JM, Fields HL, 2012. Dopamine, Corticostriatal Connectivity, and Intertemporal Choice. J. Neurosci 32, 9402–9409. 10.1523/JNEUROSCI.1180-12.2012 [DOI] [PMC free article] [PubMed] [Google Scholar]
  106. Keiflin R, Janak PH, 2015. Dopamine Prediction Errors in Reward Learning and Addiction: From Theory to Neural Circuitry. Neuron. 10.1016/j.neuron.2015.08.037 [DOI] [PMC free article] [PubMed] [Google Scholar]
  107. Kelm MK, Boettiger CA, 2013. Effects of acute dopamine precusor depletion on immediate reward selection bias and working memory depend on catechol-O-methyltransferase genotype. J. Cogn. Neurosci 25, 2061–71. 10.1162/jocn_a_00464 [DOI] [PMC free article] [PubMed] [Google Scholar]
  108. Kirsch P, Reuter M, Mier D, Lonsdorf T, Stark R, Gallhofer B, Vaitl D, Hennig J, 2006. Imaging gene–substance interactions: The effect of the DRD2 TaqIA polymorphism and the dopamine agonist bromocriptine on the brain activation during the anticipation of reward. Neurosci. Lett 405, 196–201. 10.1016/J.NEULET.2006.07.030 [DOI] [PubMed] [Google Scholar]
  109. Kiyatkin EA, Rebec GV, 2001. Impulse activity of ventral tegmental area neurons during heroin self-administration in rats. Neuroscience 102, 565–580. 10.1016/S0306-4522(00)00492-9 [DOI] [PubMed] [Google Scholar]
  110. Knutson B, Bjork JM, Fong GW, Hommer D, Mattay VS, Weinberger DR, 2004. Amphetamine modulates human incentive processing. Neuron 43, 261–269. 10.1016/j.neuron.2004.06.030 [doi]S0896627304004180 [pii] [DOI] [PubMed] [Google Scholar]
  111. Knutson B, Burgdorf J, Panksepp J, 1998. Anticipation of play elicits high-frequency ultrasonic vocalizations in young rats. J. Comp. Psychol 112, 65–73. 10.1037/0735-7036.112.1.65 [DOI] [PubMed] [Google Scholar]
  112. Knutson B, Westdorp A, Kaiser E, Hommer D, 2000. FMRI visualization of brain activity during a monetary incentive delay task. Neuroimage 12, 20–27. 10.1006/nimg.2000.0593 [DOI] [PubMed] [Google Scholar]
  113. Koffarnus MN, Newman AH, Grundt P, Rice KC, Woods JH, 2011. Effects of selective dopaminergic compounds on a delay-discounting task. Behav. Pharmacol 22, 300–11. 10.1097/FBP.0b013e3283473bcb [DOI] [PMC free article] [PubMed] [Google Scholar]
  114. Kring AM, Barch DM, 2014. The motivation and pleasure dimension of negative symptoms: Neural substrates and behavioral outputs. Eur. Neuropsychopharmacol 24, 725–736. 10.1016/J.EURONEURO.2013.06.007 [DOI] [PMC free article] [PubMed] [Google Scholar]
  115. Kroener S, Chandler LJ, Phillips PEM, Seamans JK, 2009. Dopamine Modulates Persistent Synaptic Activity and Enhances the Signal-to-Noise Ratio in the Prefrontal Cortex. PLoS One 4, e6507 10.1371/journal.pone.0006507 [DOI] [PMC free article] [PubMed] [Google Scholar]
  116. Kurniawan IT, Guitart-Masip M, Dayan P, Dolan RJ, 2013. Effort and Valuation in the Brain: The Effects of Anticipation and Execution. J. Neurosci 33, 6160. [DOI] [PMC free article] [PubMed] [Google Scholar]
  117. Kurniawan IT, Guitart-Masip M, Dolan RJ, 2011. Dopamine and effort-based decision making. Front. Neurosci 5, 81 10.3389/fnins.2011.00081 [DOI] [PMC free article] [PubMed] [Google Scholar]
  118. Kurniawan IT, Seymour B, Talmi D, Yoshida W, Chater N, Dolan RJ, 2010. Choosing to Make an Effort: The Role of Striatum in Signaling Physical Effort of a Chosen Action. J. Neurophysiol 104, 313–321. 10.1152/jn.00027.2010 [DOI] [PMC free article] [PubMed] [Google Scholar]
  119. Landau SM, Lal R, O’Neil JP, Baker S, Jagust WJ, 2009. Striatal Dopamine and Working Memory. Cereb. Cortex 19, 445–454. 10.1093/cercor/bhn095 [DOI] [PMC free article] [PubMed] [Google Scholar]
  120. Larson MJ, Clayson PE, Primosch M, Leyton M, Steffensen SC, 2015. The effects of acute dopamine precursor depletion on the cognitive control functions of performance monitoring and conflict processing: An event-related potential (ERP) study. PLoS One 10, e0140770 10.1371/journal.pone.0140770 [DOI] [PMC free article] [PubMed] [Google Scholar]
  121. Leriche L, Bezard E, Gross C, Guillin O, Foll B, Diaz J, Sokoloff P, 2008. The Dopamine D3 Receptor: A Therapeutic Target for the Treatment of Neuropsychiatric Disorders. CNS Neurol. Disord. - Drug Targets 5, 25–43. 10.2174/187152706784111551 [DOI] [PubMed] [Google Scholar]
  122. Leyton M, 2010. The Neurobiology of Desire: Dopamine and Regulation of Mood and Motivational States in Humans. Pleasures of the brain 222–243. [Google Scholar]
  123. Leyton M, Boileau I, Benkelfat C, Diksic M, Baker G, Dagher A, 2002. Amphetamine-induced increases in extracellular dopamine, drug wanting, and novelty seeking: A PET/[11C]raclopride study in healthy men. Neuropsychopharmacology 27, 1027–1035. 10.1016/S0893-133X(02)00366-4 [DOI] [PubMed] [Google Scholar]
  124. Leyton M, Casey KF, Delaney JS, Kolivakis T, Benkelfat C, 2005. Cocaine craving, euphoria, and self-administration: A preliminary study of the effect of catecholamine precursor depletion. Behav. Neurosci 119, 1619–1627. 10.1037/0735-7044.119.6.1619 [DOI] [PubMed] [Google Scholar]
  125. Leyton M, Young SN, Pihl RO, Etezadi S, Lauze C, Blier P, Baker GB, Benkelfat C, 2000. Effects on mood of acute phenylalanine/tyrosine depletion in healthy women. Neuropsychopharmacology 22, 52–63. 10.1016/S0893-133X(99)00086-X [DOI] [PubMed] [Google Scholar]
  126. Leyton Marco, Marije Rot, Booij Linda, Baker Glen, 2007. Mood-elevating effects of d-amphetamine and incentive salience: the. J. Psychiatry Neurosci. 32. [PMC free article] [PubMed] [Google Scholar]
  127. Lidow MS, Goldman-Rakic PS, Gallager DW, Rakic P, 1991. Distribution of dopaminergic receptors in the primate cerebral cortex: Quantitative autoradiographic analysis using [3H]raclopride, [3H]spiperone and [3H]SCH23390. Neuroscience 40, 657–671. 10.1016/0306-4522(91)90003-7 [DOI] [PubMed] [Google Scholar]
  128. Lopez-Gamundi P, Wardle MC, 2018. The cognitive effort expenditure for rewards task (CEEfRT): A novel measure of willingness to expend cognitive effort. Psychol. Assess 30, 1237–1248. 10.1037/pas0000563 [DOI] [PubMed] [Google Scholar]
  129. Luman M, Tripp G, Scheres A, 2010. Identifying the neurobiology of altered reinforcement sensitivity in ADHD: A review and research agenda. Neurosci. Biobehav. Rev 10.1016/j.neubiorev.2009.11.021 [DOI] [PubMed] [Google Scholar]
  130. Marco-Pallares J, Cucurell D, Cunillera T, García R, Andrés-Pueyo A, Münte TF, Rodríguez-Fornells A, 2008. Human oscillatory activity associated to reward processing in a gambling task. Neuropsychologia 46, 241–248. 10.1016/J.NEUROPSYCHOLOGIA.2007.07.016 [DOI] [PubMed] [Google Scholar]
  131. Martin WR, Sloan JW, Sapira JD, Jasinski DR, 1971. Physiologic, subjective, and behavioral effects of amphetamine, methamphetamine, ephedrine, phenmetrazine, and methylphenidate in man. Clin. Pharmacol. Ther 12, 245–258. 10.1002/cpt1971122part1245 [DOI] [PubMed] [Google Scholar]
  132. Massar SAA, Libedinsky C, Weiyan C, Huettel SA, Chee MWL, 2015. Separate and overlapping brain areas encode subjective value during delay and effort discounting. Neuroimage 120, 104–113. 10.1016/J.NEUROIMAGE.2015.06.080 [DOI] [PubMed] [Google Scholar]
  133. Mayo LM, Paul E, DeArcangelis J, Van Hedger K, de Wit H, 2019. Gender differences in the behavioral and subjective effects of methamphetamine in healthy humans. Psychopharmacology (Berl). 236, 2413–2423. 10.1007/s00213-019-05276-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
  134. McCabe C, Harwood J, Brouwer S, Harmer CJ, Cowen PJ, 2013. Effects of pramipexole on the processing of rewarding and aversive taste stimuli. Psychopharmacology (Berl). 10.1007/s00213-013-3033-9 [DOI] [PubMed] [Google Scholar]
  135. McCabe C, Huber A, Harmer CJ, Cowen PJ, 2011. The D2 antagonist sulpiride modulates the neural processing of both rewarding and aversive stimuli in healthy volunteers. Psychopharmacology (Berl). 217, 271–278. 10.1007/s00213-011-2278-4 [DOI] [PMC free article] [PubMed] [Google Scholar]
  136. McFarland K, Ettenberg A, 1995. Haloperidol differentially affects reinforcement and motivational processes in rats running an alley for intravenous heroin. Psychopharmacology (Berl). 122, 346–350. 10.1007/BF02246264 [DOI] [PubMed] [Google Scholar]
  137. McGuire P, Howes OD, Stone J, Fusar-Poli P, 2008. Functional neuroimaging in schizophrenia: diagnosis and drug discovery. Trends Pharmacol. Sci 10.1016/j.tips.2007.11.005 [DOI] [PubMed] [Google Scholar]
  138. McNair DM, Lorr M, Droppelman LF, 1971. Manual for the Profile of Mood States. Educational and Industrial Testing Services, San Diego, CA. [Google Scholar]
  139. Melgari J-M, Curcio G, Mastrolilli F, Salomone G, Trotta L, Tombini M, di Biase L, Scrascia F, Fini R, Fabrizio E, Rossini PM, Vernieri F, 2014. Alpha and beta EEG power reflects L-dopa acute administration in parkinsonian patients. Front. Aging Neurosci 6, 302 10.3389/fnagi.2014.00302 [DOI] [PMC free article] [PubMed] [Google Scholar]
  140. Melrose AJ, Bailer U, Wierenga CE, Bischoff-Grethe A, Paulus MP, Kaye WH, 2016. Amphetamine alters neural response to sucrose in healthy women. Psychiatry Res. Neuroimaging 252, 19–25. 10.1016/J.PSCYCHRESNS.2016.04.017 [DOI] [PubMed] [Google Scholar]
  141. Michely J, Viswanathan S, Hauser TU, Delker L, Dolan RJ, Grefkes C, 2020. The role of dopamine in dynamic effort–reward integration. Neuropsychopharmacology 1–7. 10.1038/s41386-020-0669-0 [DOI] [PMC free article] [PubMed] [Google Scholar]
  142. Miltner WHR, Braun CH, Coles MGH, 1997. Event-Related Brain Potentials Following Incorrect Feedback in a Time-Estimation Task: Evidence for a “Generic” Neural System for Error Detection. J. Cogn. Neurosci 9, 788–798. 10.1162/jocn.1997.9.6.788 [DOI] [PubMed] [Google Scholar]
  143. Mohebi A, Pettibone JR, Hamid AA, Wong J-MT, Vinson LT, Patriarchi T, Tian L, Kennedy RT, Berke JD, 2019. Dissociable dopamine dynamics for learning and motivation. Nature 570, 65–70. 10.1038/s41586-019-1235-y [DOI] [PMC free article] [PubMed] [Google Scholar]
  144. Moja EA, Lucini V, Benedetti F, Lucca A, 1996. Decrease in plasma phenylalanine and tyrosine after phenylalanine-tyrosine free amino acid solutions in man. Life Sci. 58, 2389–2395. 10.1016/0024-3205(96)00242-1 [DOI] [PubMed] [Google Scholar]
  145. Mueller EM, Burgdorf C, Chavanon M-LL, Schweiger D, Hennig J, Wacker J, Stemmler G, 2014a. The COMT Val158Met polymorphism regulates the effect of a dopamine antagonist on the feedback-related negativity. Psychophysiology 51, 805–809. 10.1111/psyp.12226 [DOI] [PubMed] [Google Scholar]
  146. Mueller EM, Burgdorf C, Chavanon M-LL, Schweiger D, Wacker J, Stemmler G, 2014b. Dopamine modulates frontomedial failure processing of agentic introverts versus extraverts in incentive contexts. Cogn. Affect. Behav. Neurosci 14, 756–768. 10.3758/s13415-013-0228-9 [DOI] [PubMed] [Google Scholar]
  147. Mueller EM, Makeig S, Stemmler G, Hennig J, Wacker J, 2011. Dopamine Effects on Human Error Processing Depend on Catechol-O-Methyltransferase VAL158MET Genotype. J. Neurosci 31, 15818–15825. 10.1523/JNEUROSCI.2103-11.2011 [DOI] [PMC free article] [PubMed] [Google Scholar]
  148. Murphy MR, Checkley SA, Seckl JR, Lightman SL, 1990. Naloxone inhibits oxytocin release at orgasm in man. J. Clin. Endocrinol. Metab 71, 1056–1058. 10.1210/jcem-71-4-1056 [DOI] [PubMed] [Google Scholar]
  149. Nagano-Saito A, Cisek P, Perna AS, Shirdel FZ, Benkelfat C, Leyton M, Dagher A, 2012. From anticipation to action, the role of dopamine in perceptual decision making: an fMRI-tyrosine depletion study. J. Neurophysiol 108, 501–512. 10.1152/jn.00592.2011 [DOI] [PubMed] [Google Scholar]
  150. Nicholson S, 2002. Socialization in the “virtual hallway”: Instant messaging in the asynchronous Web-based distance education classroom. Internet High. Educ 5, 363–372. 10.1016/S1096-7516(02)00127-6 [DOI] [Google Scholar]
  151. Niv Y, Daw ND, Joel D, Dayan P, 2007. Tonic dopamine: Opportunity costs and the control of response vigor. Psychopharmacology (Berl). 191, 507–520. 10.1007/s00213-006-0502-4 [DOI] [PubMed] [Google Scholar]
  152. Norbury A, Manohar S, Rogers RD, Husain M, 2013. Dopamine modulates risk-taking as a function of baseline sensation-seeking trait. J. Neurosci 33, 12982–6. 10.1523/JNEUROSCI.5587-12.2013 [DOI] [PMC free article] [PubMed] [Google Scholar]
  153. O’Daly OG, Joyce D, Tracy DK, Azim A, Stephan KE, Murray RM, Shergill SS, O’Daly OG, Joyce D, Tracy DK, Azim A, Stephan KE, Murray RM, Shergill SS, O’Daly OG, Joyce D, Tracy DK, Azim A, Stephan KE, Murray RM, Shergill SS, 2014. Amphetamine sensitization alters reward processing in the human striatum and amygdala. PLoS One 9, e93955 10.1371/journal.pone.0093955 [DOI] [PMC free article] [PubMed] [Google Scholar]
  154. O’Hara CB, Keyes A, Renwick B, Leyton M, Campbell IC, Schmidt U, 2016. The Effects of Acute Dopamine Precursor Depletion on the Reinforcing Value of Exercise in Anorexia Nervosa. PLoS One 11, e0145894 10.1371/journal.pone.0145894 [DOI] [PMC free article] [PubMed] [Google Scholar]
  155. Oei NY, Rombouts SA, Soeter RP, van Gerven JM, Both S, 2012. Dopamine modulates reward system activity during subconscious processing of sexual stimuli. Neuropsychopharmacology 37, 1729–1737. 10.1038/npp.2012.19 [DOI] [PMC free article] [PubMed] [Google Scholar]
  156. Ojala KE, Janssen LK, Hashemi MM, Timmer MHM, Geurts DEM, Ter Huurne NP, Cools R, Sescousse G, 2018. Dopaminergic Drug Effects on Probability Weighting during Risky Decision Making. eNeuro 5, 330–18. 10.1523/ENEURO.0330-18.2018 [DOI] [PMC free article] [PubMed] [Google Scholar]
  157. Oldham S, Murawski C, Fornito A, Youssef G, Yücel M, Lorenzetti V, 2018. The anticipation and outcome phases of reward and loss processing: A neuroimaging meta-analysis of the monetary incentive delay task. Hum. Brain Mapp 39, 3398–3418. 10.1002/hbm.24184 [DOI] [PMC free article] [PubMed] [Google Scholar]
  158. Olds J, Milner P, 1954. Positive reinforcement produced by electrical stimulation of septal area and other regions of rat brain. J. Comp. Physiol. Psychol 47, 419–427. 10.1037/h0058775 [DOI] [PubMed] [Google Scholar]
  159. Orsini CA, Moorman DE, Young JW, Setlow B, Floresco SB, 2015. Neural mechanisms regulating different forms of risk-related decision-making: Insights from animal models. Neurosci. Biobehav. Rev 58, 147–167. 10.1016/J.NEUBIOREV.2015.04.009 [DOI] [PMC free article] [PubMed] [Google Scholar]
  160. Pan W-X, Schmidt R, Wickens JR, Hyland BI, 2005. Dopamine Cells Respond to Predicted Events during Classical Conditioning: Evidence for Eligibility Traces in the Reward-Learning Network. J. Neurosci 25, 6235–6242. 10.1523/JNEUROSCI.1478-05.2005 [DOI] [PMC free article] [PubMed] [Google Scholar]
  161. Pardo M, López-Cruz L, Miguel NS, Salamone JD, Correa M, 2015. Selection of sucrose concentration depends on the effort required to obtain it: studies using tetrabenazine, D1, D2, and D3 receptor antagonists. Psychopharmacology (Berl). 232, 2377–2391. 10.1007/s00213-015-3872-7 [DOI] [PubMed] [Google Scholar]
  162. Peciña Susana, Berridge Kent C, Parker LA, Peciña S, Berridge KC, Pimozide LAP, 1997. Pimozide Does Not Shift Palatability: Separation of Anhedonia from Sensorimotor Suppression by Taste Reactivity, Pharmacology Biochemistry and Behavior. [DOI] [PubMed]
  163. Pergadia ML, Der-Avakian A, D’Souza MS, Madden PAF, Heath AC, Shiffman S, Markou A, Pizzagalli DA, 2014. Association between nicotine withdrawal and reward responsiveness in humans and rats. JAMA Psychiatry 71, 1238–1245. 10.1001/jamapsychiatry.2014.1016 [DOI] [PMC free article] [PubMed] [Google Scholar]
  164. Pessiglione M, Seymour B, Flandin G, Dolan RJ, Frith CD, 2006. Dopamine-dependent prediction errors underpin reward-seeking behaviour in humans. Nature 442, 1042–1045. https://doi.org/nature05051 [pii] 10.1038/nature05051 [doi] [DOI] [PMC free article] [PubMed] [Google Scholar]
  165. Pfaus JG, Damsma G, Nomikos GG, Wenkstern DG, Blaha CD, Phillips AG, Fibiger HC, 1990. Sexual behavior enhances central dopamine transmission in the male rat. Brain Res. 530, 345–348. 10.1016/0006-8993(90)91309-5 [DOI] [PubMed] [Google Scholar]
  166. Pine A, Shiner T, Seymour B, Dolan RJ, 2010. Dopamine, time, and impulsivity in humans. J. Neurosci 30, 8888–96. 10.1523/JNEUROSCI.6028-09.2010 [DOI] [PMC free article] [PubMed] [Google Scholar]
  167. Pizzagalli DA, Evins AE, Schetter EC, Frank MJ, Pajtas PE, Santesso DL, Culhane M, 2008. Single dose of a dopamine agonist impairs reinforcement learning in humans: behavioral evidence from a laboratory-based measure of reward responsiveness. Psychopharmacology (Berl). 196, 221–32. 10.1007/s00213-007-0957-y [DOI] [PMC free article] [PubMed] [Google Scholar]
  168. Pizzagalli DA, Jahn AL, O’Shea JP, 2005. Toward an objective characterization of an anhedonic phenotype: A signal-detection approach. Biol. Psychiatry 57, 319–327. 10.1016/J.BIOPSYCH.2004.11.026 [DOI] [PMC free article] [PubMed] [Google Scholar]
  169. Pleger B, Ruff CC, Blankenburg F, Klöppel S, Driver J, Dolan RJ, 2009. Influence of dopaminergically mediated reward on somatosensory decision-making. PLoS Biol. 7, e1000164 10.1371/journal.pbio.1000164 [DOI] [PMC free article] [PubMed] [Google Scholar]
  170. Prévost C, Pessiglione M, Météreau E, Cléry-Melin M-L, Dreher J-C, 2010. Separate valuation subsystems for delay and effort decision costs. J. Neurosci 30, 14080–90. 10.1523/JNEUROSCI.2752-10.2010 [DOI] [PMC free article] [PubMed] [Google Scholar]
  171. Proudfit GH, 2015. The reward positivity: From basic research on reward to a biomarker for depression. Psychophysiology 52, 449–459. 10.1111/psyp.12370 [DOI] [PubMed] [Google Scholar]
  172. Reddy LF, Horan WP, Barch DM, Buchanan RW, Dunayevich E, Gold JM, Lyons N, Marder SR, Treadway MT, Wynn JK, Young JW, Green MF, 2015. Effort-Based Decision-Making Paradigms for Clinical Trials in Schizophrenia: Part 1—Psychometric Characteristics of 5 Paradigms. Schizophr. Bull 41, 1045–54. 10.1093/schbul/sbv089 [DOI] [PMC free article] [PubMed] [Google Scholar]
  173. Rescorla RA, Wagner AR, 1972. A theory of Pavlovian conditioning: Variations in the effectiveness of reinforcement and nonreinforcement. Class. Cond. II Curr. Res. Theory 21, 64–99. 10.1101/gr.110528.110 [DOI] [Google Scholar]
  174. Riba J, Kramer UM, Heldmann M, Richter S, Munte TF, Krämer UM, Heldmann M, Richter S, Münte TF, 2008. Dopamine agonist increases risk taking but blunts reward-related brain activity. PLoS One 3, e2479 10.1371/journal.pone.0002479 [DOI] [PMC free article] [PubMed] [Google Scholar]
  175. Rigoli F, Chew B, Dayan P, Dolan RJ, 2016a. The dopaminergic midbrain mediates an effect of average reward on pavlovian vigor. J. Cogn. Neurosci 28, 1303–1317. 10.1162/jocn_a_00972 [DOI] [PubMed] [Google Scholar]
  176. Rigoli F, Rutledge RB, Chew B, Ousdal OT, Dayan P, Dolan RJ, 2016b. Dopamine Increases a Value-Independent Gambling Propensity. Neuropsychopharmacology 41, 2658–67. 10.1038/npp.2016.68 [DOI] [PMC free article] [PubMed] [Google Scholar]
  177. Robinson OJ, Standing HR, DeVito EE, Cools R, Sahakian BJ, 2010. Dopamine precursor depletion improves punishment prediction during reversal learning in healthy females but not males. Psychopharmacology (Berl). 211, 187–195. 10.1007/s00213-010-1880-1 [DOI] [PMC free article] [PubMed] [Google Scholar]
  178. Rutledge RB, Skandali N, Dayan P, Dolan RJ, 2015. Dopaminergic Modulation of Decision Making and Subjective Well-Being. J. Neurosci 35, 9811. [DOI] [PMC free article] [PubMed] [Google Scholar]
  179. Salamone JD, 2006. Will the last person who uses the term “reward” please turn out the lights? Comments on processes related to reinforcement, learning, motivation and effort. Addict. Biol 10.1111/j.1369-1600.2006.00011.x [DOI] [PubMed] [Google Scholar]
  180. Salamone JD, 1992. Complex motor and sensorimotor functions of striatal and accumbens dopamine: involvement in instrumental behavior processes. Psychopharmacology (Berl). 107, 160–74. [DOI] [PubMed] [Google Scholar]
  181. Salamone JD, Correa M, 2012. The Mysterious Motivational Functions of Mesolimbic Dopamine. Neuron 76, 470–485. 10.1016/j.neuron.2012.10.021 [DOI] [PMC free article] [PubMed] [Google Scholar]
  182. Salamone JD, Correa M, Farrar AM, Nunes EJ, Pardo M, 2009. Dopamine, Behavioral Economics, and Effort. Front. Behav. Neurosci 3, 13 10.3389/neuro.08.013.2009 [DOI] [PMC free article] [PubMed] [Google Scholar]
  183. Salamone JD, Correa M, Nunes EJ, Randall PA, Pardo M, 2012. The Behavioral Pharmacology of Effort-related Choice Behavior: Dopamine, Adenosine and Beyond. J. Exp. Anal. Behav 97, 125–146. 10.1901/jeab.2012.97-125 [DOI] [PMC free article] [PubMed] [Google Scholar]
  184. Salamone JD, Correa M, Yang J-H, Rotolo R, Presby R, 2018. Dopamine, Effort-Based Choice, and Behavioral Economics: Basic and Translational Research. Front. Behav. Neurosci 12, 52 10.3389/fnbeh.2018.00052 [DOI] [PMC free article] [PubMed] [Google Scholar]
  185. Salamone JD, Yohn SE, López-Cruz L, San Miguel N, Correa M, 2016. Activational and effort-related aspects of motivation: neural mechanisms and implications for psychopathology. Brain 139, 1325–1347. 10.1093/brain/aww050 [DOI] [PMC free article] [PubMed] [Google Scholar]
  186. Samejima K, Ueda Y, Doya K, Kimura M, 2005. Representation of action-specific reward values in the striatum. Science (80-. ). 310, 1337–1340. 10.1126/science.1115270 [DOI] [PubMed] [Google Scholar]
  187. Santesso DL, Evins AE, Frank MJ, Schetter EC, Bogdan R, Pizzagalli DA, 2009. Single dose of a dopamine agonist impairs reinforcement learning in humans: evidence from event-related potentials and computational modeling of striatal-cortical function. Hum. Brain Mapp 30, 1963–1976. 10.1002/hbm.20642 [DOI] [PMC free article] [PubMed] [Google Scholar]
  188. Scheffers MK, Coles MGH, Bernstein P, Gehring WJ, Donchin E, 1996. Event-related brain potentials and error-related processing: An analysis of incorrect responses to go and no-go stimuli. Psychophysiology 33, 42–53. 10.1111/j.1469-8986.1996.tb02107.x [DOI] [PubMed] [Google Scholar]
  189. Schmidt L, Lebreton M, Cléry-Melin M-L, Daunizeau J, Pessiglione M, 2012. Neural Mechanisms Underlying Motivation of Mental Versus Physical Effort. PLOS Biol. 10, e1001266–. [DOI] [PMC free article] [PubMed] [Google Scholar]
  190. Schott BH, Minuzzi L, Krebs RM, Elmenhorst D, Lang M, Winz OH, Seidenbecher CI, Coenen HH, Heinze H-J, Zilles K, Duzel E, Bauer A, 2008. Mesolimbic Functional Magnetic Resonance Imaging Activations during Reward Anticipation Correlate with Reward-Related Ventral Striatal Dopamine Release. J. Neurosci 10.1523/JNEUROSCI.2058-08.2008 [DOI] [PMC free article] [PubMed] [Google Scholar]
  191. Schultz W, 2016a. Dopamine reward prediction-error signalling: a two-component response. Nat. Rev. Neurosci 17, 183–95. 10.1038/nrn.2015.26 [DOI] [PMC free article] [PubMed] [Google Scholar]
  192. Schultz W, 2016b. Dopamine reward prediction error coding. Dialogues Clin. Neurosci 18, 23–32. [DOI] [PMC free article] [PubMed] [Google Scholar]
  193. Schultz W, 2013. Updating dopamine reward signals. Curr. Opin. Neurobiol 23, 229–238. 10.1016/J.CONB.2012.11.012 [DOI] [PMC free article] [PubMed] [Google Scholar]
  194. Schultz W, 1998. Predictive Reward Signal of Dopamine Neurons. J. Neurophysiol 10.1152/jn.1998.80.1.1 [DOI] [PubMed] [Google Scholar]
  195. Schultz W, Dayan P, Montague PR, 1997. A neural substrate of prediction and reward. Science (80-.). 275, 1593–1599. 10.1126/science.275.5306.1593 [DOI] [PubMed] [Google Scholar]
  196. Schultz W, Stauffer WR, Lak A, 2017. The phasic dopamine signal maturing: from reward via behavioural activation to formal economic utility. Curr. Opin. Neurobiol 43, 139–148. 10.1016/J.CONB.2017.03.013 [DOI] [PubMed] [Google Scholar]
  197. Schutte I, Deschamps PKH, van Harten PN, Kenemans JL, 2020a. Dopaminergic and noradrenergic manipulation of anticipatory reward and probability event-related potentials. Psychopharmacology (Berl). 1–12. 10.1007/s00213-020-05515-x [DOI] [PMC free article] [PubMed] [Google Scholar]
  198. Schutte I, Deschamps PKH, van Harten PN, Kenemans JL, 2020b. Dopaminergic and noradrenergic manipulation of anticipatory reward and probability event-related potentials. Psychopharmacology (Berl). 1–12. 10.1007/s00213-020-05515-x [DOI] [PMC free article] [PubMed] [Google Scholar]
  199. Sescousse G, Janssen LK, Hashemi MM, Timmer MHMM, Geurts DEMM, Ter Huurne NP, Clark L, Cools R, 2016. Amplified Striatal Responses to Near-Miss Outcomes in Pathological Gamblers. Neuropsychopharmacology 41, 2614–23. 10.1038/npp.2016.43 [DOI] [PMC free article] [PubMed] [Google Scholar]
  200. Sevak R, Stoops WW, Hays L, Rush CR, 2009. Discriminative stimulus and self-reported effects of methylphenidate, d-amphetamine, and triazolam in methylphenidate-trained humans. J. Pharmacol. Exp. Ther 328, 1007–1018. 10.1037/1064-1297.13.1.56 [DOI] [PMC free article] [PubMed] [Google Scholar]
  201. Sevy S, Hassoun Y, Bechara A, Yechiam E, Napolitano B, Burdick K, Delman H, Malhotra A, 2006. Emotion-based decision-making in healthy subjects: Short-term effects of reducing dopamine levels. Psychopharmacology (Berl). 188, 228–235. 10.1007/s00213-006-0450-z [DOI] [PMC free article] [PubMed] [Google Scholar]
  202. Sharot T, Shiner T, Brown A, Fan J, Dolan RJ, 2009. Dopamine enhances expectation of pleasure in humans. Curr. Biol 19, 2077–2080. [DOI] [PMC free article] [PubMed] [Google Scholar]
  203. Smith KS, Berridge KC, 2007. Opioid Limbic Circuit for Reward: Interaction between Hedonic Hotspots of Nucleus Accumbens and Ventral Pallidum. J. Neurosci 27, 1594–1605. 10.1523/JNEUROSCI.4205-06.2007 [DOI] [PMC free article] [PubMed] [Google Scholar]
  204. Smoski MJ, Felder J, Bizzell J, Green SR, Ernst M, Lynch TR, Dichter GS, 2009. fMRI of alterations in reward selection, anticipation, and feedback in major depressive disorder. J. Affect. Disord 118, 69–78. 10.1016/j.jad.2009.01.034 [DOI] [PMC free article] [PubMed] [Google Scholar]
  205. Soutschek A, Burke CJ, Raja Beharelle A, Schreiber R, Weber SC, Karipidis II, Ten Velden J, Weber B, Haker H, Kalenscher T, Tobler PN, 2017. The dopaminergic reward system underpins gender differences in social preferences. Nat. Hum. Behav 10.1038/s41562-017-0226-y [DOI] [PubMed] [Google Scholar]
  206. Soutschek A, Gvozdanovic G, Kozak R, Duvvuri S, de Martinis N, Harel B, Gray DL, Fehr E, Jetter A, Tobler PN, 2020. Dopaminergic D1 Receptor Stimulation Affects Effort and Risk Preferences. Biol. Psychiatry 87, 678–685. 10.1016/j.biopsych.2019.09.002 [DOI] [PubMed] [Google Scholar]
  207. St Onge JR, Floresco SB, 2009. Dopaminergic Modulation of Risk-Based Decision Making. Neuropsychopharmacology 34, 681–697. 10.1038/npp.2008.121 [DOI] [PubMed] [Google Scholar]
  208. Stephens DN, Crombag HS, Duka T, 2011. The Challenge of studying parallel behaviors in humans and animal models. Curr. Top. Behav. Neurosci 13, 611–645. 10.1007/7854_2011_133 [DOI] [PubMed] [Google Scholar]
  209. Strauss GP, Frank MJ, Waltz JA, Kasanova Z, Herbener ES, Gold JM, 2011. Deficits in positive reinforcement learning and uncertainty-driven exploration are associated with distinct aspects of negative symptoms in schizophrenia. Biol. Psychiatry 69, 424–431. 10.1016/j.biopsych.2010.10.015 [DOI] [PMC free article] [PubMed] [Google Scholar]
  210. Sutton RS, 1988. Learning to Predict by the Methods of Temporal Differences. Mach. Learn 3, 9–44. [Google Scholar]
  211. Sutton RS, Barto AG, 1990. Time-derivative models of Pavlovian reinforcement, in: Learning and Computational Neuroscience: Foundations of Adaptive Networks. The MIT Press, Cambridge, MA, US, pp. 497–537. [Google Scholar]
  212. Suzuki C, Ikeda Y, Tateno A, Okubo Y, Fukayama H, Suzuki H, 2019. Acute atomoxetine selectively modulates encoding of reward value in ventral medial prefrontal cortex. J. Nippon Med. Sch 86, 98–107. 10.1272/jnms.JNMS.2019_86-205 [DOI] [PubMed] [Google Scholar]
  213. Swart JC, Froböse MI, Cook JL, Geurts DE, Frank MJ, Cools R, den Ouden HE, 2017. Catecholaminergic challenge uncovers distinct Pavlovian and instrumental mechanisms of motivated (in)action. Elife 6, e22169 10.7554/eLife.22169 [DOI] [PMC free article] [PubMed] [Google Scholar]
  214. Symmonds M, Wright ND, Fagan E, Dolan RJ, 2013. Assaying the effect of levodopa on the evaluation of risk in healthy humans. PLoS One 8, e68177 10.1371/journal.pone.0068177 [DOI] [PMC free article] [PubMed] [Google Scholar]
  215. Tobia MJ, Guo R, Schwarze U, Boehmer W, Gläscher J, Finckh B, Marschner A, Büchel C, Obermayer K, Sommer T, 2014. Neural systems for choice and valuation with counterfactual learning signals. Neuroimage 89, 57–69. 10.1016/j.neuroimage.2013.11.051 [DOI] [PubMed] [Google Scholar]
  216. Treadway MT, Bossaller NA, Shelton RC, Zald DH, 2012. Effort-based decision-making in major depressive disorder: A translational model of motivational anhedonia. J. Abnorm. Psychol 121, 553–558. 10.1037/a0028813 [DOI] [PMC free article] [PubMed] [Google Scholar]
  217. Treadway MT, Buckholtz JW, Schwartzman AN, Lambert WE, Zald DH, 2009. Worth the ‘EEfRT’? The Effort Expenditure for Rewards Task as an Objective Measure of Motivation and Anhedonia. PLoS One 4, e6598 10.1371/journal.pone.0006598 [DOI] [PMC free article] [PubMed] [Google Scholar]
  218. Treadway MT, Zald DH, 2011a. Reconsidering anhedonia in depression: Lessons from translational neuroscience. Neurosci. Biobehav. Rev 35, 537–555. 10.1016/j.neubiorev.2010.06.006 [DOI] [PMC free article] [PubMed] [Google Scholar]
  219. Treadway MT, Zald DH, 2011b. Reconsidering anhedonia in depression: Lessons from translational neuroscience. Neurosci. Biobehav. Rev 35, 537–555. 10.1016/j.neubiorev.2010.06.006 [DOI] [PMC free article] [PubMed] [Google Scholar]
  220. Tremblay A-M, Desmond RC, Poulos CX, Zack M, 2011. Haloperidol modifies instrumental aspects of slot machine gambling in pathological gamblers and healthy controls. Addict. Biol 16, 467–484. 10.1111/j.1369-1600.2010.00208.x [DOI] [PubMed] [Google Scholar]
  221. Tripp G, Alsop B, 1999. Sensitivity to Reward Frequency in Boys With Attention Deficit Hyperactivity Disorder. J. Clin. Child Psychol 28, 366–375. 10.1207/S15374424jccp280309 [DOI] [PubMed] [Google Scholar]
  222. Tritsch NX, Sabatini BL, 2012. Dopaminergic Modulation of Synaptic Transmission in Cortex and Striatum. Neuron 76, 33–50. 10.1016/J.NEURON.2012.09.023 [DOI] [PMC free article] [PubMed] [Google Scholar]
  223. Vaillancourt DE, Schonfeld D, Kwak Y, Bohnen NI, Seidler R, 2013. Dopamine overdose hypothesis: Evidence and clinical implications. Mov. Disord 28, 1920–1929. 10.1002/mds.25687 [DOI] [PMC free article] [PubMed] [Google Scholar]
  224. Van Der Schaaf ME, Fallon SJ, Ter Huurne N, Buitelaar J, Cools R, 2013. Working memory capacity predicts effects of methylphenidate on reversal learning. Neuropsychopharmacology 38, 2011–2018. 10.1038/npp.2013.100 [DOI] [PMC free article] [PubMed] [Google Scholar]
  225. Van Der Schaaf ME, Van Schouwenburg MR, Geurts DEM, Schellekens AFA, Buitelaar JK, Verkes RJ, Cools R, 2012. Establishing the dopamine dependency of human striatal signals during reward and punishment reversal learning. Cereb. Cortex 24, 633–642. 10.1093/cercor/bhs344 [DOI] [PubMed] [Google Scholar]
  226. van Gaalen MM, van Koten R, Schoffelmeer ANM, Vanderschuren LJMJ, 2006. Critical Involvement of Dopaminergic Neurotransmission in Impulsive Decision Making. Biol. Psychiatry 60, 66–73. 10.1016/J.BIOPSYCH.2005.06.005 [DOI] [PubMed] [Google Scholar]
  227. Vassena E, Silvetti M, Boehler CN, Achten E, Fias W, Verguts T, 2014. Overlapping Neural Systems Represent Cognitive Effort and Reward Anticipation. PLoS One 9, e91008 10.1371/journal.pone.0091008 [DOI] [PMC free article] [PubMed] [Google Scholar]
  228. Venugopalan VV, Casey KF, O’Hara C, O’Loughlin J, Benkelfat C, Fellows LK, Leyton M, 2011. Acute phenylalanine/tyrosine depletion reduces motivation to smoke cigarettes across stages of addiction. Neuropsychopharmacology 36, 2469–2476. 10.1038/npp.2011.135 [DOI] [PMC free article] [PubMed] [Google Scholar]
  229. Vo A, Seergobin KN, Morrow SA, MacDonald PA, 2016. Levodopa impairs probabilistic reversal learning in healthy young adults. Psychopharmacology (Berl). 233, 2753–2763. 10.1007/s00213-016-4322-x [DOI] [PubMed] [Google Scholar]
  230. Volkow ND, Fowler JS, Wang GJ, Swanson JM, Telang F, 2007. Dopamine in drug abuse and addiction: Results of imaging studies and treatment implications. Arch. Neurol 10.1001/archneur.64.11.1575 [DOI] [PubMed] [Google Scholar]
  231. Volkow ND, Wang G-J, Fowler JS, Telang F, Maynard L, Logan J, Gatley SJ, Pappas N, Wong C, Vaska P, Zhu W, Swanson JM, 2004. Evidence That Methylphenidate Enhances the Saliency of a Mathematical Task by Increasing Dopamine in the Human Brain. Am. J. Psychiatry 161, 1173–1180. 10.1176/appi.ajp.161.7.1173 [DOI] [PubMed] [Google Scholar]
  232. Volkow ND, Wang G-J, Fowler JS, Tomasi D, Telang F, 2011. Addiction: Beyond dopamine reward circuitry. Proc. Natl. Acad. Sci 108, 15037–15042. 10.1073/pnas.1010654108 [DOI] [PMC free article] [PubMed] [Google Scholar]
  233. Volkow ND, Wang G-J, Ma Y, Fowler JS, Wong C, Jayne M, Telang F, Swanson JM, 2006. Effects of expectation on the brain metabolic responses to methylphenidate and to its placebo in non-drug abusing subjects. Neuroimage 32, 1782–1792. 10.1016/J.NEUROIMAGE.2006.04.192 [DOI] [PubMed] [Google Scholar]
  234. Volkow ND, Wang GJ, Tomasi D, Baler RD, 2013. Obesity and addiction: Neurobiological overlaps. Obes. Rev 10.1111/j.1467-789X.2012.01031.x [DOI] [PMC free article] [PubMed] [Google Scholar]
  235. Volkow ND, Wise RA, Baler R, 2017. The dopamine motive system: Implications for drug and food addiction. Nat. Rev. Neurosci 10.1038/nrn.2017.130 [DOI] [PubMed] [Google Scholar]
  236. Wachtel SR, Ortengren A, de Wit H, 2002. The effects of acute haloperidol or risperidone on subjective responses to methamphetamine in healthy volunteers. Drug Alcohol Depend. 68, 23–33. 10.1016/S0376-8716(02)00104-7 [DOI] [PubMed] [Google Scholar]
  237. Wacker J, Gatt JM, 2010. Resting posterior versus frontal delta/theta EEG activity is associated with extraversion and the COMT VAL158MET polymorphism. Neurosci. Lett 478, 88–92. 10.1016/J.NEULET.2010.04.071 [DOI] [PubMed] [Google Scholar]
  238. Wade TR, de Wit H, Richards JB, 2000. Effects of dopaminergic drugs on delayed reward as a measure of impulsive behavior in rats. Psychopharmacology (Berl). 150, 90–101. [DOI] [PubMed] [Google Scholar]
  239. Wardle MC, Treadway MT, Mayo LM, Zald DH, de Wit H, 2011. Amping up effort: effects of d-amphetamine on human effort-based decision-making. J. Neurosci 31, 16597–16602. 10.1523/JNEUROSCI.4387-11.2011 [DOI] [PMC free article] [PubMed] [Google Scholar]
  240. Wardle MCMC, de Wit H, 2012. Effects of amphetamine on reactivity to emotional stimuli. Psychopharmacology (Berl). 220, 143–153. 10.1007/s00213-011-2498-7 [DOI] [PMC free article] [PubMed] [Google Scholar]
  241. Weafer J, Gorka SM, Hedeker D, Dzemidzic M, Kareken DA, Phan KL, de Wit H, 2017. Associations Between Behavioral and Neural Correlates of Inhibitory Control and Amphetamine Reward Sensitivity. Neuropsychopharmacology 42, 1905–1913. 10.1038/npp.2017.61 [DOI] [PMC free article] [PubMed] [Google Scholar]
  242. Weber SC, Beck-Schimmer B, Kajdi M-E, Müller D, Tobler PN, Quednow BB, 2016. Dopamine D2/3- and μ-opioid receptor antagonists reduce cue-induced responding and reward impulsivity in humans. Transl. Psychiatry 6, e850 10.1038/tp.2016.113 [DOI] [PMC free article] [PubMed] [Google Scholar]
  243. Weis T, Brechmann A, Puschmann S, Thiel CM, 2013. Feedback that confirms reward expectation triggers auditory cortex activity. J. Neurophysiol 110, 1860–1868. 10.1152/jn.00128.2013 [DOI] [PubMed] [Google Scholar]
  244. Weis T, Puschmann S, Brechmann A, Thiel CM, 2012. Effects of L-dopa during auditory instrumental learning in humans. PLoS One 7, e52504 10.1371/journal.pone.0052504 [DOI] [PMC free article] [PubMed] [Google Scholar]
  245. Westbrook A, Lamichhane B, Braver T, 2019. The Subjective Value of Cognitive Effort is Encoded by a Domain-General Valuation Network. J. Neurosci 39, 3934–3947. 10.1523/JNEUROSCI.3071-18.2019 [DOI] [PMC free article] [PubMed] [Google Scholar]
  246. Westbrook A, van den Bosch R, Määttä JI, Hofmans L, Papadopetraki D, Cools R, Frank MJ, 2020. Dopamine promotes cognitive effort by biasing the benefits versus costs of cognitive work. Science (80-.). 367, 1362–1366. 10.1126/science.aaz5891 [DOI] [PMC free article] [PubMed] [Google Scholar]
  247. White TL, Lejuez CW, de Wit H, 2007. Personality and gender differences in effects of d-amphetamine on risk taking. Exp. Clin. Psychopharmacol 15, 599–609. 10.1037/1064-1297.15.6.599 [DOI] [PubMed] [Google Scholar]
  248. White TL, Lott DC, de Wit H, 2006. Personality and the subjective effects of acute amphetamine in healthy volunteers. Neuropsychopharmacology 31, 1064–1074. 10.1038/sj.npp.1300939 [DOI] [PubMed] [Google Scholar]
  249. Winstanley CA, Dalley JW, Theobald DEH, Robbins TW, 2003. Global 5-HT depletion attenuates the ability of amphetamine to decrease impulsive choice on a delay-discounting task in rats. Psychopharmacology (Berl). 170, 320–331. 10.1007/s00213-003-1546-3 [DOI] [PubMed] [Google Scholar]
  250. Winstanley CA, Floresco SB, 2016. Deciphering Decision Making: Variation in Animal Models of Effort- and Uncertainty-Based Choice Reveals Distinct Neural Circuitries Underlying Core Cognitive Processes. J. Neurosci 36, 12069–12079. 10.1523/JNEUROSCI.1713-16.2016 [DOI] [PMC free article] [PubMed] [Google Scholar]
  251. Wise RA, 2004. Dopamine, learning and motivation. Nat. Rev. Neurosci 10.1038/nrn1406 [DOI] [PubMed] [Google Scholar]
  252. Wise RA, Schwartz HV, 1981. Pimozide attenuates acquisition of lever-pressing for food in rats. Pharmacol. Biochem. Behav 15, 655–656. 10.1016/0091-3057(81)90225-2 [DOI] [PubMed] [Google Scholar]
  253. Wittmann BC, D’Esposito M, D’Esposito M, 2015. Levodopa administration modulates striatal processing of punishment-associated items in healthy participants. Psychopharmacology (Berl). 232, 135–144. 10.1007/s00213-014-3646-7 [DOI] [PMC free article] [PubMed] [Google Scholar]
  254. Wunderlich K, Smittenaar P, Dolan RJJ, 2012. Dopamine enhances model-based over model-free choice behavior. Neuron 75, 418–24. 10.1016/j.neuron.2012.03.042 [DOI] [PMC free article] [PubMed] [Google Scholar]
  255. Ye Z, Hammer A, Camara E, Munte TF, Münte TF, 2011. Pramipexole modulates the neural network of reward anticipation. Hum. Brain Mapp 32, 800–811. 10.1002/hbm.21067 [DOI] [PMC free article] [PubMed] [Google Scholar]
  256. Yokel R, Wise R, deWit H, Gerberg G, 1975. Increased lever pressing for amphetamine after pimozide in rats: implications for a dopamine theory of reward. Science (80-.). 187, 547–549. 10.1126/science.1114313 [DOI] [PubMed] [Google Scholar]
  257. Zack M, Poulos CX, 2007. A D2 Antagonist Enhances the Rewarding and Priming Effects of a Gambling Episode in Pathological Gamblers. Neuropsychopharmacology 32, 1678–1686. 10.1038/sj.npp.1301295 [DOI] [PubMed] [Google Scholar]
  258. Zaghloul KA, Blanco JA, Weidemann CT, McGill K, Jaggi JL, Baltuch GH, Kahana MJ, 2009. Human Substantia Nigra Neurons Encode Unexpected Financial Rewards. Science (80-.). 323. [DOI] [PMC free article] [PubMed] [Google Scholar]
  259. Zald DH, Treadway MT, 2017. Reward Processing, Neuroeconomics, and Psychopathology. Annu. Rev. Clin. Psychol 13, 471–95. 10.1146/annurev-clinpsy-032816 [DOI] [PMC free article] [PubMed] [Google Scholar]
  260. Zenon A, Devesse S, Olivier E, 2016. Dopamine Manipulation Affects Response Vigor Independently of Opportunity Cost. J. Neurosci 36, 9516–9525. 10.1523/JNEUROSCI.4467-15.2016 [DOI] [PMC free article] [PubMed] [Google Scholar]
  261. Zirnheld PJ, Carroll CA, Kieffaber PD, O’Donnell BF, Shekhar A, Hetrick WP, 2004. Haloperidol impairs learning and error-related negativity in humans. J. Cogn. Neurosci 16, 1098–1112. 10.1162/0898929041502779 [DOI] [PubMed] [Google Scholar]

RESOURCES