Abstract
Both unexpected pain and unexpected pain absence can drive avoidance learning, but whether they do so via shared or separate neural and neurochemical systems is largely unknown. To address this issue, we combined an instrumental pain-avoidance learning task with computational modeling, functional magnetic resonance imaging (fMRI), and pharmacological manipulations of the dopaminergic (100 mg levodopa) and opioidergic (50 mg naltrexone) systems (N = 83). Computational modeling provided evidence that untreated participants learned more from received than avoided pain. Our dopamine and opioid manipulations negated this learning asymmetry by selectively increasing learning rates for avoided pain. Furthermore, our fMRI analyses revealed that pain prediction errors were encoded in subcortical and limbic brain regions, whereas no-pain prediction errors were encoded in frontal and parietal cortical regions. However, we found no effects of our pharmacological manipulations on the neural encoding of prediction errors. Together, our results suggest that human pain-avoidance learning is supported by separate threat- and safety-learning systems, and that dopamine and endogenous opioids specifically regulate learning from successfully avoided pain.
Research organism: Human
Introduction
Learning to avoid actions that cause damage to our body is critical for health and survival. The experience of pain is an important teaching signal in this learning process, such as when a child learns to avoid touching a hot stove, or when a patient who underwent knee surgery learns to avoid bending his or her knee. However, the absence of otherwise expected pain can be an equally important teaching signal. When, for example, some weeks after surgery a patient realizes that bending his or her knee is not painful anymore, this suggests that particular movements are safe again and no longer need to be avoided. Adaptive behavior in situations associated with pain thus requires an optimal balance between threat and safety learning when confronted with, respectively, the unexpected presence and absence of pain.
Previous studies have made considerable progress in our understanding of the neural basis of passive cue–pain association learning (Ploghaus et al., 2000; Seymour et al., 2004; Seymour et al., 2005) and—more recently—active pain-avoidance and -relief learning (Roy et al., 2014; Eldar et al., 2016; Zhang et al., 2018) in humans. A key aspect of these studies was the application of reinforcement-learning models to the analysis of neuroimaging data. According to reinforcement-learning theory, learning is driven by prediction errors, which signal the difference between the actual and expected outcomes of an action (Sutton and Barto, 1998). Thus, actions that result in the unexpected presence vs. absence of pain yield oppositely signed prediction errors which, respectively, increase and decrease the aversive value associated with that action. However, whether these opponent teaching signals drive learning via one underlying brain system, or via separate ones, is still largely unknown.
One possibility is that prediction errors elicited by the unexpected presence and absence of pain are encoded as opposite activity patterns in the same brain regions (i.e., one learning system). If this is the case, we may also expect that—at the neurochemical level—learning from these two outcomes is supported by the same neuromodulator(s). Furthermore, these two outcomes may then be equally effective in driving learning, such that they are associated with the same learning rate. Most previous studies, including our own, have assumed that this is the case. For example, in a previous functional magnetic resonance imaging (fMRI) study, we identified brain activity encoding general aversive prediction errors, signaling the degree to which both received- and avoided-pain outcomes are relatively worse (or less good) than expected (Roy et al., 2014). Another possibility, however, is that learning from received and successfully avoided pain are subserved by two separate brain systems. In this case, learning from these two outcomes may also be supported by different neuromodulatory systems, and associated with different learning rates. Although the notion of two systems is broadly consistent with theoretical accounts of avoidance learning, such as two-factor theory (Mowrer, 1951), there is not much empirical evidence for this idea, especially not in humans.
Regarding the role of neuromodulators, there is a wealth of evidence that reward prediction errors, which are thought to drive reward-pursuit behaviors, are signaled by phasic activity of midbrain dopamine neurons (Schultz et al., 1997) that project to the ventral striatum (O’Doherty et al., 2003, Rutledge et al., 2010). Specifically, unexpected rewards trigger a phasic increase (burst) in dopamine activity, while the unexpected absence of reward triggers a phasic decrease (dip) in dopamine activity. Whether the dopamine system also has a role in aversive learning is controversial, but several hypotheses have been proposed, mostly based on animal studies (Palminteri and Pessiglione, 2017). According to one prominent account, the same dopaminergic prediction-error response that reinforces actions associated with reward also reinforces actions associated with the omission of aversive outcomes (Mowrer, 1956; Dinsmoor, 2001; Moutoussis et al., 2008). This account thus predicts that learning from successfully avoided pain is supported by phasic dopamine bursts. However, animal studies have provided mixed evidence for this idea (Josselyn et al., 2005; Oleson et al., 2012; Wietzikoski et al., 2012; Dombrowski et al., 2013; Fernando et al., 2013; Salinas-Hernández et al., 2018; Wenzel et al., 2018; Stelly et al., 2019) and evidence in humans is scarce (Raczka et al., 2011). In addition, it has been proposed that negative punishment learning is mediated by phasic dips or pauses in dopamine activity, via an effect on D2 receptors (Frank, 2005; Maia and Frank, 2011). Evidence for his account in humans is largely based on studies using secondary reinforcers (e.g., monetary losses), and whether phasic dopamine dips also mediate learning from unexpected pain is yet unknown.
The endogenous opioid system is another interesting candidate neuromodulatory system for pain-avoidance learning. Pavlovian fear conditioning studies in animals have provided evidence for a causal role of opioidergic activity—specifically via μ-opioid receptors in the periaqueductal gray (PAG)—in both fear conditioning (McNally and Cole, 2006; Cole and McNally, 2007; Cole and McNally, 2009; McNally et al., 2011) and fear-extinction learning (McNally and Westbrook, 2003; McNally et al., 2004; McNally et al., 2005; Kim and Richardson, 2009; Parsons et al., 2010). These findings suggest that the endogenous μ-opioid system may mediate learning from received and/or avoided pain, although this remains to be studied in instrumental learning tasks and in humans. In humans, endogenous opioids are well known to mediate pain relief (Levine et al., 1978; Eippert et al., 2009) as well as pleasure and liking responses (Berridge and Kringelbach, 2015; Nummenmaa and Tuominen, 2018), but it is currently unknown whether the roles of the opioid system in affective valuation also influence pain-related learning. One interesting possibility—that remains to be explored—is that endogenous opioids support pain-avoidance learning by enhancing the pleasant feeling of relief following successful pain avoidance (Sirucek et al., 2021).
In the present study, we aimed to dissociate learning from received and avoided pain in terms of behavior (learning rates), neural encoding of prediction errors, and the roles of dopamine and endogenous opioids. To this end, we combined an instrumental pain-avoidance learning task with computational modeling, fMRI, and pharmacological manipulations of the dopamine and opioid systems, in a randomized, double-blind, between-subject design. Specifically, participants completed the learning task under one of three drug conditions: 100 mg levodopa (a dopamine precursor), 50 mg naltrexone (an opioid antagonist, with highest affinity for the μ-opioid receptor), or placebo. PET studies in humans suggest that levodopa increases phasic dopamine bursts (Floel et al., 2008). Thus, if dopamine bursts support learning from successfully avoided pain, we expect levodopa to enhance learning rates and neural prediction-error signaling when pain is avoided. Additionally, although levodopa is thought to primarily increase phasic dopamine bursts (Floel et al., 2008; Black et al., 2015), it may increase tonic dopamine levels as well, which could in turn interfere with dopamine dips (Breitenstein et al., 2006; Moustafa et al., 2013; Poletti and Bonuccelli, 2013). Thus, if phasic dopamine dips support learning from unexpected pain, we expect levodopa to suppress learning rates and neural prediction-error signaling when pain is received. Naltrexone blocks the majority of μ-opioid receptors in the brain (Lee et al., 1988; Preston and Bigelow, 1993; Schuh et al., 1999; Weerts et al., 2013). Thus, if μ-opioid-receptor activity supports learning from received- and/or avoided-pain outcomes, we expect naltrexone to reduce learning rates and neural prediction-error signaling for the corresponding outcome(s).
Results
Eighty-three healthy human participants completed a pain-avoidance learning task during fMRI, under one of three treatment conditions (levodopa, naltrexone, or placebo). On each of 144 trials of the pain-avoidance learning task, participants chose between two options (Figure 1A), each probabilistically associated with the delivery of painful heat (49 or 50°C, 1.9 s duration) to their left lower leg. Pain probabilities for each option were governed by two independently varying random walks (Figure 1B). Thus, participants had to track the changing reinforcement values, and kept experiencing prediction errors throughout the task.
There were no treatment effects on subjective state (alertness, calmness, or contentment; Figure 1—figure supplement 1) or self-reported heat pain during a separate pain-rating task immediately preceding the avoidance-learning task (Figure 1—figure supplement 2).
Nine participants were excluded from the fMRI (but not the behavioral) analyses because of excessive head movement. Thus, the behavioral analyses included 83 participants (26–29 per treatment group), and the fMRI analyses included 74 participants (24–26 per treatment group).
No drug effects on model-independent measures of task performance
On average, participants received pain on 57.1 of the 144 trials (standard deviation [SD] = 8.6). As expected, participants switched to the other choice option more frequently after receiving pain (46.6% of those trials, SD = 21.2) than after avoiding pain (5.4% of those trials, SD = 6.9; t(82) = 18.6, p < 0.001, Cohen’s d = 2.0). The effect of previous pain outcomes on switching also decayed exponentially over time, in all treatment groups (Figure 1C; p < 0.001, Cohen’s d > 2.2 for 1 trial back, p < 0.002, Cohen’s d > 0.7 for 2 trials back, and p > 0.047, Cohen’s d < 0.4 for 3–6 trials back, in all groups).
The three treatment groups did not differ in the number of received pain stimuli (F(2,80) = 0.56, p = 0.57, = 0.01), frequency of switching following pain outcomes (F(2,80) = 0.03, p = 0.97, = 0.0007), or frequency of switching following no-pain outcomes (F(2,80) = 1.18, p = 0.31, = 0.03). Thus, our pharmacological manipulations did not affect basic measures of task performance.
Choice reaction times (RTs) were faster when participants stayed with the same choice option following a no-pain outcome (no pain/stay choices; mean = 729 ms) than when participants either stayed or switched following a pain outcome (mean = 778 and 795, respectively; no pain/stay vs. pain/stay: t(82) = 5.0, p < 0.001, Cohen’s d = 0.54; no pain/stay vs. pain/switch: t(82) = 6.9, p < 0.001, Cohen’s d = 0.76). Pain/stay and pain/switch RTs did not differ from each other (t(82) = 1.7, p = 0.10, Cohen’s d = 0.18). The faster no pain/stay choices likely reflect that no-pain outcomes usually indicate that participants are on the right track, which makes staying with that option a relatively simple decision. The interpretation of pain outcomes is less straightforward, as these outcomes could either indicate a change in outcome probabilities—in which case a switch to the other option is warranted—or an occasional ‘unlucky’ outcome due to the probabilistic task nature. There was no treatment effect on RT for no pain/stay, pain/stay, or pain/switch trials (F(2,80) = 0.35, p = 0.70, = 0.01; F(2,80) = 0.68, p = 0.51, = 0.02; F(2,80) = 0.93, p = 0.40, = 0.02, respectively).
Computational modeling
To formalize and quantify the latent learning and decision processes thought to underlie participants’ choice behavior, we applied two candidate reinforcement-learning models to the choice data, and compared their goodness-of-fit. Both models update the expected pain probability for the chosen option on each trial, in proportion to the prediction error (Rescorla and Wagner, 1972). The two models differ in that Model 1 uses a single learning rate, , for all outcomes, whereas Model 2 uses separate learning rates for received and avoided pain: and , respectively (see Methods for model equations and parameter-estimation details). If Model 2 is better able to explain the choice data than Model 1, this could be taken as initial support for the idea that learning from received and avoided pain is subserved by different learning systems.
Both learning models were combined with a softmax decision function that translates expected pain probabilities into choice probabilities. Inverse-temperature parameter controls the degree of choice randomness, such that the likelihood that the model chooses the option with the lowest expected pain probability increases as increases.
We estimated the model parameters using a hierarchical Bayesian approach. This approach assumes that every participant has a different set of model parameters, which are drawn from group-level distributions. In this way, the information in the individual data is aggregated, while still respecting individual differences (Gelman, 2014). Each group-level distribution is in turn governed by a group-level mean and a group-level SD parameter. These group-level parameters were estimated separately for the placebo, levodopa, and naltrexone groups. As we are primarily interested in differences between treatment groups, our primary variables of interest are the parameters governing the means of the group-level distributions, which we denote with overbars (e.g., refers to the group-level mean of ).
Model comparison: evidence for asymmetric learning from received and avoided pain
In all treatment groups, the model with separate learning rates for received and avoided pain (Model 2) outperformed the model with a single learning rate (Model 1), providing initial evidence for the presence of two learning systems. The Watanabe–Akaike information criterion (WAIC) for Model 1 vs. Model 2 was 3109 vs. 2959, 2958 vs. 2941, and 3164 vs. 3106 for the placebo, levodopa, and naltrexone groups, respectively.
Parameter estimates: levodopa and naltrexone increase learning rates for avoided pain
We next examined the parameter estimates of the best-fitting model. Figure 2A shows the posterior distributions of the group-level mean parameters, per treatment group. The corresponding 95% highest density intervals (HDIs) are reported in Figure 2—source data 1. The 95% HDIs of each individual participant’s learning-rate posteriors are shown in Figure 2—figure supplement 1.
Figure 2. Model parameters.
(A) Posterior distributions of the parameters’ group-level means for each group (left and right panels). Parameters and are learning rates for avoided and received pain outcomes, respectively; parameter is the inverse-temperature parameter. The middle panels are joint density plots of and (dots are samples from the Markov chain Monte Carlo [MCMC]), showing that is reliably greater than in the placebo group only. (B) The difference between the posterior distributions for each drug group vs. the placebo group, showing that is greater and is smaller in both drug groups compared to the placebo group. Red lines indicate 95% highest density intervals (HDIs).
Figure 2—figure supplement 1. The 95% highest density intervals (HDIs) of the posterior distributions of each participant’s learning rate for pain ( ) and no-pain () outcomes.
Figure 2—figure supplement 2. Modeling results from an independent sample of untreated participants from a previous study (N = 23), replicating the asymmetric learning rates ( > ) found in our placebo group.
In the placebo group, the posterior distribution of (median = 0.72) was considerably higher than the posterior distribution of (median = 0.32; > for 99.6% of the Markov chain Monte Carlo [MCMC] samples), indicative of stronger expectation updating when pain was received than avoided. In contrast, in both drug groups, the posterior distributions of and were highly similar, due to a specific increase in relative to the placebo group. In the levodopa group, the posterior medians of and were, respectively, 0.66 and 0.66 ( > for 50% of the MCMC samples). In the naltrexone group, the posterior medians of and were, respectively, 0.72 and 0.76 ( > for 38% of the MCMC samples). Note that the best-fitting model for both drug groups contained separate learning rates for received and avoided pain. Combined with the finding that the group-level mean learning-rate parameters for these two outcomes were highly similar, this suggests that some participants in each drug group learned more from received than avoided pain while others showed the opposite bias, but that there was no systematic learning asymmetry (at the individual level, was higher than for 50% of the levodopa and 41% of the naltrexone participants).
Thus, at the group level, both levodopa and naltrexone, as compared to placebo, increased learning rates for avoided pain, while not affecting learning from received pain (Figure 2A, left and middle panels). To test the significance of these group differences, we computed the difference between the posterior distributions of the group-level mean parameters for each drug group vs. the placebo group (Figure 2B). For , 99.7% of the difference distribution for levodopa vs. placebo, and 99.9% of the difference distribution for naltrexone vs. placebo, lay above 0. In contrast, did not differ between the drug and placebo groups: 34% and 49% of the difference distributions for levodopa vs. placebo and naltrexone vs. placebo, respectively, lay above 0. Thus, both drugs selectively increased learning rates for avoided pain.
The posterior distribution of inverse-temperature parameter was higher for the placebo group (median = 8.8) than the levodopa and naltrexone group (median = 5.3 and 5.7, respectively), as well. Specifically, 98.8% of the difference distribution for levodopa vs. placebo, and 98.1% for naltrexone vs. placebo, lay below 0, indicating that was reliably lower for both drug groups compared to the placebo group. This suggests that participants in the two drug groups, as compared to the placebo group, were less prone to choose the option with the lowest expected pain probability (i.e., more stochastic choice behavior).
Together, the parameter estimates suggest that (1) untreated (placebo group) participants updated their expectations more rapidly following received than avoided pain, (2) levodopa and naltrexone negated this learning asymmetry by selectively increasing learning rates for avoided pain, and (3) levodopa and naltrexone also increased choice stochasticity, possibly reflecting a more exploratory or risky choice strategy.
Replication of learning-rate asymmetry in an independent group of untreated participants
To examine the robustness of our finding of asymmetric learning rates in the placebo group, we applied our hierarchical Bayesian modeling approach to the choice data of 23 untreated participants from our previous pain-avoidance learning fMRI study (Roy et al., 2014), and tested whether the higher learning rate for received than avoided pain found in our placebo group was replicated in this previous dataset (Figure 2—figure supplement 2). In this previous dataset, (median = 0.62) was indeed higher than (median = 0.44), resembling the placebo-group results from our current study, although the learning-rate asymmetry was somewhat smaller in the previous dataset ( > for 90% of the MCMC samples). The posterior median of in the previous dataset was 7.1 (95% HDI = 4.7–9.9). The finding of a higher learning rate for pain than no-pain outcomes in this independent dataset corroborates the idea that people normally (in the absence of a pharmacological manipulation) update their expectations more rapidly following received than avoided pain.
Data simulation and parameter recovery
The fact that our pharmacological manipulations increased (negating the learning asymmetry found in the placebo group) and reduced (increasing choice stochasticity) seems at odds with the absence of drug effects on model-independent performance measures. A possible explanation for these results is that the observed performance measures (e.g., switch/stay behavior) reflect a combination of several underlying variables—including the learning rates for received and avoided pain and the degree of choice stochasticity—and that the drug effects on learning rate and choice stochasticity canceled each other out. Specifically, symmetric learning rates for received and avoided pain (as found in the drug groups) likely resulted in more accurate pain-probability estimates than asymmetric learning rates (as found in the placebo group). This beneficial effect of a symmetric learning process in the drug groups, however, may have been counteracted by the detrimental effect of a more stochastic choice process, such that the combination of these two computational effects resulted in no net performance difference between the placebo and drug groups. However, since and are generally negatively correlated to one another (Cools et al., 2011), it is also possible that the observed tradeoff between and reflected an artifact of the parameter-optimization procedure rather than ‘true’ effects of the pharmacological manipulations.
To rule out the possibility of an artifactual consequence of the parameter-optimization procedure, we simulated two sets of choice data on our task using the parameter values found in the placebo and drug groups, respectively, and performed parameter-recovery analyses (Appendix 1). In sum, our modeling procedure was able to correctly recover the simulated parameters, suggesting that it is unlikely that the parameters observed in the different pharmacological-manipulation conditions are artifacts of our parameter-optimization procedure. Moreover, the simulated datasets also produced similar model-independent performance measures (number of received pain stimuli, switch frequencies) in the drug and placebo groups, despite the fact that the simulated choices were produced by different sets of parameters. These findings suggest that levodopa and naltrexone really had two computational effects—an increased learning rate for avoided pain, and an increased degree of choice stochasticity—which combination yielded no significant effects on model-independent performance measures.
fMRI analyses
To address the question whether learning from received and avoided pain is supported by two separate brain systems, we sought to identify brain activation encoding outcome-specific prediction-error signals. We focused on the first second of the outcome period as this is when prediction errors are triggered. We modeled drug effects using two second-level regressors (levodopa vs. placebo and naltrexone vs. placebo). fMRI results are thresholded at q < 0.05, false discovery rate (FDR) corrected for multiple comparisons across the whole brain (gray matter masked). Unthresholded t maps can be found on https://neurovault.org/collections/RIVRRMAK/. When referring to subdivisions of the anterior cingulate cortex (ACC), we use the terms rostral and dorsal ACC, as is common in neuroimaging studies (Bush et al., 2000). Note that these two regions have also been referred to as ACC and midcingulate cortex, respectively (Palomero-Gallagher et al., 2009).
We also performed an axiomatic analysis to identify brain activation encoding general aversive prediction errors (i.e., activation encoding the degree to which both pain and no-pain outcomes are relatively worse, or less good, than expected). In our previous study, this analysis revealed a general aversive prediction-error signal in the PAG (Roy et al., 2014). Here, we examined whether we could replicate this finding. As this analysis is not directly linked to the research questions addressed in the present study, we report it in Appendix 2.
Outcome-specific prediction-error signals
Regions encoding the unexpectedness or surprise evoked by received pain should respond stronger to pain outcomes when pain was less expected (i.e., a negative correlation with expected pain probability on pain trials). In contrast, regions encoding the surprise evoked by avoided pain should respond stronger to no-pain outcomes when pain was more expected (i.e., a positive correlation with expected pain probability on no-pain trials; Figure 3A, left plots). To identify regions that show a stronger surprise response for received than avoided pain, we thus specified the following contrast: negative correlation with expected pain probability on pain trials > positive correlation with expected pain probability on no-pain trials. This contrast revealed activation in the vmPFC and rostral ACC, posterior cingulate cortex, insula extending into the left parahippocampal gyrus, cerebellum, and a brainstem region encompassing part of the PAG (Figure 3A, yellow regions). We also found extensive clusters that showed the opposite effect—that is positive correlation with expected pain probability on no-pain trials > negative correlation with expected pain probability on pain trials—in sensorimotor cortex, parietal and occipital cortex, and the bilateral frontal poles (Figure 3A, blue regions), suggesting that these regions show a stronger surprise response for avoided than received pain.
Figure 3. Outcome-specific prediction-error signals (N = 74).
(A) Activation tracking surprise more for received than avoided pain (yellow) and vice versa (blue). Note that this includes activation that tracks expected pain probability across both outcomes. Expected P(pain) = expected pain probability. (B) Activation tracking surprise for both received and avoided pain (i.e., absolute prediction error). Activation maps in A and B are thresholded at q < 0.05, false discovery rate (FDR) corrected for multiple comparisons across the whole brain. (C) Regions encoding surprise more for received than avoided pain, which cannot be explained by a general sensitivity to expected pain probability. These regions showed positive activation for both the first (A) and second (B) contrast, each thresholded at q < 0.05, FDR corrected. (D) Regions encoding surprise more for avoided than received pain, which cannot be explained by a general sensitivity to expected pain probability. These regions showed negative activation for the first, and positive activation for the second contrast, each thresholded at q < 0.05, FDR corrected. The line plots show the mean activity extracted from the brainstem and right amygdala (C) and left dlPFC and parietal (D) clusters per quartile of expected pain probability, illustrating the encoding of outcome-specific prediction errors in these regions. Error bars are standard errors.
We also sought to identify activation encoding the surprise elicited by both received and avoided pain, that is, activation encoding absolute prediction errors. To this end, we specified a second contrast that tested for a negative correlation with expected pain probability on pain trials and a positive correlation with expected pain probability on no-pain trials (Figure 3B). This contrast revealed extensive activation clusters in the dorsal ACC extending into the supplementary motor cortex, insula, sensorimotor cortex, thalamus, part of the brainstem and cerebellum, suggesting that these regions encoded absolute prediction error (Figure 3B, yellow regions). In addition, a few smaller clusters in left sensorimotor cortex, right dlPFC, and left frontopolar cortex showed the opposite effect, suggesting that these regions encoded the overall expectedness of outcomes (Figure 3B, blue regions).
Finally, note that a caveat of the first contrast reported above (Figure 3A) is that it also identifies activation that is stronger when the expected pain probability is lower (i.e., activation encoding expected safety), regardless of the outcome. This likely explains the vmPFC activation, which has been found to represent positive affective value in many domains (Roy et al., 2012; Bartra et al., 2013; Rich and Wallis, 2016). The second contrast (for absolute prediction errors; Figure 3B), on the other hand, will not detect activation encoding expected safety regardless of the outcome, as the correlation with expected pain probability is specified in opposite directions for pain and no-pain outcomes. Thus, we reasoned that regions identified by both of the contrasts reported in Figure 3A, B encode outcome-specific prediction errors (Figure 3A) unconfounded by outcome-nonspecific expected pain probability (Figure 3B). Therefore, we next examined the conjunction of these two contrasts. Specifically, we masked the activation identified by the first contrast (separately for the positive and negative activation) by the positive activation identified by the second contrast. Positive activation for both contrasts was found in a set of mostly subcortical and limbic regions, including a large cluster in the brainstem (not covering the PAG), bilateral insula extending into the amygdala, rostral ACC, posterior cingulate cortex, bilateral supramarginal gyrus, and cerebellum (Figure 3C). These regions thus encoded surprise more for received than avoided pain (pain-specific prediction errors), which could not be explained by a general sensitivity to expected pain probability. Negative activation for the first contrast and positive activation for the second contrast, on the other hand, was found in several cortical regions, including the supplementary motor cortex, left parietal cortex (supramarginal gyrus), bilateral somatosensory cortex (postcentral gyrus), and left dlPFC (middle and superior frontal gyrus) (Figure 3D). These regions thus encoded surprise more for avoided than received pain (no-pain-specific prediction errors), which could not be explained by a general sensitivity to expected pain probability. Together, these results provide evidence that prediction errors evoked by pain and no-pain outcomes are encoded in largely distinct brain regions.
No drug effects on surprise-related brain activation
Because levodopa and naltrexone specifically increased learning rates for no-pain outcomes, we expected these drugs to increase prediction-error-related brain activation for no-pain outcomes as well. However, we found no differences between the levodopa and placebo group, or between the naltrexone and placebo group, for any of the contrasts reported above (whole-brain FDR corrected). Drug effects were virtually absent at lower, uncorrected, thresholds as well (see unthresholded t maps on https://neurovault.org/collections/RIVRRMAK/).
Discussion
Our results provide novel evidence that unexpectedly received and avoided pain—signaling threat and safety, respectively—drive human pain-avoidance learning via different learning systems. First, computational modeling suggested that participants’ choices were best explained by a model with separate learning rates for received and avoided pain, and that untreated participants learned more from received than avoided pain. Second, levodopa and naltrexone selectively increased learning rates for avoided pain, suggesting a role for the dopamine and endogenous-opioid systems in safety, but not threat, learning. Third, our fMRI analyses revealed that different brain circuits encode prediction errors elicited by received and avoided pain, providing evidence for two dissociable learning systems at the neural level as well. Somewhat surprisingly, however, we found no drug effects on fMRI activity for any of the contrasts we examined. We discuss each of these findings below.
Learning rates for received vs. successfully avoided pain
The higher learning rate for pain than no-pain outcomes in the placebo group suggests that people normally (in the absence of a pharmacological manipulation) update their expectations more following received than avoided pain. A similar learning-rate asymmetry was found in a recent aversive reversal-learning study (Wise et al., 2019). Interestingly, however, reward-learning studies using secondary reinforcers (e.g., monetary gains and losses) have provided evidence for the opposite asymmetry: higher learning rates for favorable than unfavorable outcomes. This has been attributed to an optimistic learning bias (Sharot and Garrett, 2016; Lefebvre et al., 2017) and a tendency to learn preferentially from information that confirms one’s current action (Palminteri et al., 2017). The opposite learning bias in pain-avoidance learning tasks may be due to the intrinsically aversive nature of pain, arguably rendering unexpected pain a more salient teaching signal than unexpected pain absence. Relatedly, the experience of pain may trigger a reflexive tendency to change one’s course of action (Huys et al., 2012), expressed in elevated learning rates for pain outcomes. Thus, higher learning rates for received than avoided pain may reflect a Pavlovian influence on choice which operates in parallel to the instrumental learning system. Alternatively, the seemingly opposite learning asymmetries in reward-learning and pain-avoidance learning tasks may also reflect a cognitive process related to the framing of the task. That is, participants who are instructed to maximize reward vs. minimize pain may pay most attention to—and hence learn most from—reward vs. pain outcomes, respectively.
The presence and direction of learning asymmetries may also depend on the specific task demands. For example, a previous fMRI study that used a more complex pain-avoidance learning task (pain probabilities of three different options were learned in parallel, in an indirect manner), and included a risk-taking component, found no systematic difference in learning rates for received and avoided pain (Eldar et al., 2016). Interestingly, in contrast to our present and previous (Roy et al., 2014) fMRI findings, the PAG in that study positively encoded expected pain probability on no-pain trials (one of the axioms for appetitive prediction errors), and did not encode expected pain probability on pain trials. These findings suggest that the task used in that previous study (Eldar et al., 2016) evoked different learning processes than our simpler task. Indeed, learning rates in that previous study were much lower than those found in our task as well. Further examination of the degree to which avoidance-learning processes and their neural implementation generalize across different learning tasks is an important objective for future work (Yarkoni, 2020).
Effects of levodopa and naltrexone on learning parameters
Levodopa and naltrexone selectively increased learning rates for successfully avoided pain, consistent with a role for dopamine and endogenous opioids in safety learning. Previous studies have associated levodopa-induced increases in phasic dopamine activity with enhanced learning from (secondary) rewards (Frank et al., 2004; Pessiglione et al., 2006). Combined with these previous findings, our levodopa results suggest that phasic dopamine activity may signal the degree to which outcomes are ‘better than expected’ across both reward and punishment domains. We are aware of one previous study that provided correlational evidence for a role of dopamine in human safety learning in a Pavlovian fear-conditioning task (Raczka et al., 2011). That study found that individual differences in fear-extinction learning rates were associated with genetic variation in the dopamine transporter gene, which presumably affects phasic striatal dopamine release. Our levodopa results are consistent with this result, and provide the first causal evidence for a role of dopamine in human safety learning in an instrumental learning task. Levodopa did not affect learning rates for received pain; hence our results do not support the idea that dopamine supports learning from aversive outcomes. Instead, our results suggest a selective role for the human dopamine system in learning from successfully avoided pain.
Regarding the endogenous opioid system, we expected that if pain-avoidance learning relies on μ-opioid activity, naltrexone—which blocks this activity—would suppress learning from received and/or avoided pain. However, we found the opposite effect for avoided-pain outcomes: like levodopa, naltrexone increased learning rates for avoided pain. This finding counterintuitively suggests that μ-opioid activity normally suppresses learning from avoided pain and that naltrexone countered this effect, which seems to contradict findings that μ-opioid-receptor antagonists impair fear-extinction learning in rats (McNally and Westbrook, 2003; McNally et al., 2004; McNally et al., 2005; Kim and Richardson, 2009; Parsons et al., 2010). Obviously, these animal studies differed from our study in several ways—such as the nature of the learning task (Pavlovian vs. instrumental), outcome measure (freezing behavior vs. learning-rate estimates), opioid-receptor antagonist (naloxone vs. naltrexone), and drug administration (injection into the PAG vs. oral administration)—each of which may explain the apparently contradictory results. Thus, our results show that untreated participants learn more rapidly from received than avoided pain (replicated in our previous study) and that both levodopa and naltrexone negate this learning asymmetry, but the neurobiological mechanisms underlying the naltrexone effect remain to be elucidated. One informative approach for future studies would be to directly compare effects of opioid-receptor agonists and antagonists on pain-avoidance learning parameters.
The levodopa and naltrexone groups also showed a higher degree of choice stochasticity than the placebo group, suggesting that participants in both drug groups were less prone to choose the option with the lowest expected pain probability. Importantly, our parameter-recovery analysis indicated that the drug effects on learning rate for avoided pain and choice stochasticity could be independently and correctly retrieved. The increased choice stochasticity in the levodopa group may reflect a positive association between dopamine activity and risk preference (Voon et al., 2006; Gallagher et al., 2007; St Onge and Floresco, 2009; Chew et al., 2019) or exploration (Beeler et al., 2010; Kayser et al., 2015; Gershman and Tzovaras, 2018). It is also broadly consistent with recent evidence that levodopa reduces the impact of valence on information seeking (Vellani et al., 2020). We are not aware of previous studies that associated the endogenous opioid system with choice stochasticity, risk taking, or exploration. It is possible that the similar effects of levodopa and naltrexone on choice stochasticity were due to common general side effects of both drugs on participants’ attention or motivation, which disrupted the decision-making process. However, we believe this is unlikely because (1) the drugs did not affect subjective state (alertness, calmness, or contentment) and (2) general side effects cannot easily explain the specific increase in learning rates for avoided pain.
Interestingly, the drug effects on learning rate for avoided pain and choice stochasticity were not accompanied by drug effects on basic performance measures (number of received pain outcomes, pain-switch or avoid-switch behavior). In a similar vein, previous studies have reported effects of pharmacological manipulations and dopaminergic genotype on model parameters despite a lack of significant effects on behavioral measures (Raczka et al., 2011; Chakroun et al., 2020), illustrating the added value of computational models. Our simulation results suggested that the drug effects on learning rate and choice stochasticity in our study canceled each other out. Specifically, symmetric learning rates for received and avoided pain (found in the drug groups) result in more accurate pain-probability estimates than asymmetric learning rates (found in the placebo group). This beneficial effect of a symmetric learning process in the drug groups was, however, counteracted by the detrimental effect of a more stochastic choice process, resulting in no net performance difference between the placebo and drug groups.
Separate brain circuits support learning from received and avoided pain
Our fMRI results provided evidence for two dissociable learning systems at the brain level as well. Pain-specific prediction errors were predominantly encoded in subcortical (brainstem, cerebellum) and limbic (insula, amygdala, rostral ACC) regions that are typically associated with emotional and affective processes, including fear conditioning (Phillips and LeDoux, 1992) and affective responses to errors (Bush et al., 2000). In contrast, no-pain-specific prediction errors were encoded in frontal and parietal cortical areas typically associated with higher-order cognitive processing (Ptak et al., 2017).
Prediction errors for no-pain outcomes were not represented in the ventral striatum, which is typically found for reward prediction errors (O’Doherty et al., 2003, Rutledge et al., 2010). Instead, prediction errors for no-pain outcomes were associated with frontoparietal activity. This activity is unlikely to reflect a reward signal, but possibly reflected increased attention on trials in which pain was expected but not received. That is, the unexpected absence of pain may have prompted participants to carefully monitor the thermode’s temperature in order to verify whether pain was really avoided or still to come, as reflected in increased frontoparietal activity. The absence of a striatal ‘reward-like’ prediction-error response for no-pain outcomes suggests that, in terms of neural processing, learning from avoided pain is not comparable to learning from rewards. However, the lack of a detectable reward-like prediction-error response may also be related to our task design. Specifically, pain outcomes in our task involved a change in sensory input (a rise in temperature) whereas no-pain outcomes did not (maintenance of the baseline temperature), which may have caused a more prominent neural prediction-error response for the pain outcomes. One way to examine this issue would be to use a task in which choices result in either an increase or a decrease in painful stimulus intensity from a tonic pain level, such that aversive and appetitive-like outcomes are associated with similar changes in sensory input (Seymour et al., 2005). Such a task would examine pain-relief rather than pain-avoidance learning. The current task design, however, more closely resembles the situation of a patient recovering from injury or surgery, who is pain free as long as he or she is resting but expects that physical activity may result in pain. It is an interesting speculation that, in such situations, the stronger subcortical and limbic (‘emotional’) responses for pain than for no-pain prediction errors, as found in our study, may foster a behavioral state that favors inactivity and rest, which could promote tissue healing and recovery.
No effects of levodopa and naltrexone on fMRI activation
Unexpectedly, we found no effects of levodopa or naltrexone on any of our fMRI prediction-error contrasts. This may indicate that the dopamine and opioid systems are not involved in pain-avoidance learning, although this is inconsistent with the drug effects on learning rates for avoided pain. It is also possible that our pharmacological manipulations did affect prediction-error-related dopamine and/or opioid activity, but that we did not have enough statistical power to detect these effects due to our moderate sample size and between-subject design (24–26 participants per treatment group). We opted for a between-subject design because we expected people’s previous (possibly drug-related) experiences with our task to affect their motivation and behavior during subsequent sessions. For example, drug-induced impairments in avoidance learning could result in a perceived lack of control over the task outcomes—or learned helplessness (Seligman and Maier, 1967)—which could carry over to subsequent sessions. However, a disadvantage of our choice for a between-subject design is that it comes with less statistical power than a within-subject design.
The absence of drug effects on fMRI activation at the group level may also be explained by individual differences in drug responses. Several studies have shown that effects of dopaminergic drugs on cognitive function depend on an individual’s baseline level of performance or dopamine activity reviewed in Cools and D’Esposito, 2011, Frank and Fossella, 2011, consistent with the idea that the relationship between dopamine activity and neurocognitive function follows an inverted U-shape function (Cools and Robbins, 2004). Thus, levodopa may have affected pain-avoidance learning and related brain function in opposite directions in different participants, depending on their baseline level of dopamine activity. Such baseline-dependent drug effects could have been studied in within-subject designs (in which each participant completes both a placebo and a drug session), ideally combined with a measure of participants’ baseline dopamine/opioid level. In addition, future studies could use more than one drug dose to sample the putative inverted U-shape function.
Finally, it is important to note that drug effects on prediction-error-related dopamine and/or opioid activity may not always produce corresponding changes in the BOLD signal (Knutson and Gibbs, 2007; Brocka et al., 2018). This last possibility is clearly discouraging for pharmacological fMRI studies, but is conceivable given the absence of consistent levodopa effects on reward prediction-error signals in previous fMRI studies: One study reported an increased reward prediction-error signal in an levodopa group compared to a haloperidol group (although neither of these groups differed from a placebo group) (Pessiglione et al., 2006), but two recent studies using within-subject designs found no levodopa effects on neural reward prediction-error signals (Kroemer et al., 2019; Chakroun et al., 2020). To better understand the degree to which drug-induced changes in human neuromodulatory (e.g., dopamine) activity are detectable using fMRI, future research could directly compare drug effects on local changes in neuromodulator activity—for example, using molecular imaging procedures such as PET—with effects of these same drugs on the BOLD signal.
Limitations and directions for future research
As mentioned above, a lack of statistical power due to our moderate sample size and between-subject design may have prevented the detection of drug effects in our fMRI analyses. Another limitation of our study is the absence of a pharmacological manipulation check, for example via blood samples and/or autonomic or behavioral measures that are known to be influenced by dopamine or opioid activity, such as eye blink rate for dopamine (Jongkees and Colzato, 2016). Our drug doses were based on previous pharmacological studies with between-subject designs which reported significant drug effects (Pessiglione et al., 2006; Eippert et al., 2008; Guitart-Masip et al., 2012; Oei et al., 2012; Beierholm et al., 2013; Bunzeck et al., 2014; Guitart-Masip et al., 2014; Wittmann and D’Esposito, 2015), and the drug effects on learning parameters in the present study suggests that our pharmacological interventions were effective as well. However, given the absence of a manipulation check, it is unknown whether our null findings are due to a true lack of dopaminergic/opioidergic regulation of pain-avoidance learning, an ineffectiveness of our pharmacological manipulations (e.g., because the doses were too low), or a lack of statistical power. Thus, the absence of drug effects on our fMRI results and model-independent performance measures should be taken with caution.
A limitation of our experimental design is that we did not acquire pain ratings during the pain-avoidance learning task. Therefore, we cannot exclude the possibility that our pharmacological manipulations affected participants’ sensitivity to the pain outcomes during this task. This is especially relevant for naltrexone as the μ-opioid system is known to be involved in nociceptive processing, and blockade of this system had been associated with pain inhibition. However, we believe that it is unlikely that naltrexone affected participants’ pain sensitivity because (1) we found no effects of naltrexone on pain ratings immediately prior to the learning task, and (2) previous studies suggest that opioid antagonists rarely affect pain perception in experimental pain paradigms (Grevert and Goldstein, 1977; Eippert et al., 2008; Werner et al., 2015; Sirucek et al., 2021).
Our fMRI data suffered from signal dropout in inferior parts of the prefrontal cortex (including the orbitofrontal cortex), which is a common problem in fMRI studies using echo-planar imaging (Ojemann et al., 1997; Deichmann et al., 2003). Therefore, our results are agnostic with respect to the contribution of ventral prefrontal areas to pain-avoidance learning, and their potential modulation by levodopa or naltrexone. In addition, we did not collect physiological data during fMRI scanning; hence could not remove potential artifacts related to cardiac and respiratory processes. Correcting for physiological noise may have resulted in a higher signal-to-noise ratio and hence increase the significance of the results, especially in brainstem regions (Linnman et al., 2012; Brooks et al., 2013). Importantly, it is unlikely that physiological noise was synchronized with our experimental design (e.g., the application of pain outcomes) and thereby produced ‘false positive’ fMRI results because (1) frequencies of cardiac and respiration cycles are much higher than the frequency of our pain outcomes, and (2) we used jittered interstimulus intervals. Furthermore, even if physiological changes in respiratory and cardiac cycle did correlate with the timing of our pain outcomes, this would be unlikely to systematically affect our prediction-error signals as these were based on parametric-modulator regressors that were orthogonal to the main outcome-onset regressors.
Finally, regarding the role of neuromodulators, we focused on dopamine and endogenous opioids, but other neuromodulators are almost certainly involved in pain-avoidance learning as well. In particular, future work may focus on the serotonergic system, which has traditionally been associated with aversive processing, behavioral inhibition, and ‘fight or flight’ responses in rodents (Deakin, 1983; Soubrié, 1986, Deakin and Graeff, 1991). More recent pharmacological and genetic studies, which mostly used tasks with secondary reinforcers, have provided evidence for a role of the human serotonergic system in various aspects of aversive learning (Chamberlain et al., 2006; Cools et al., 2008; Crockett et al., 2012; Hindi Attar et al., 2012; Robinson et al., 2012; den Ouden et al., 2013). Based on these findings, it has been proposed that serotonin acts as an opponent to dopamine by mediating behavioral inhibition in response to punishment (Daw et al., 2002; Dayan and Huys, 2009; Boureau and Dayan, 2011; Cools et al., 2011). When generalizing these findings to the domain of pain-avoidance learning, we may expect a specific role for the serotonergic system learning from received pain (threat learning). However, other studies have found effects of serotonin manipulations on both reward and punishment learning (Palminteri et al., 2012; Guitart-Masip et al., 2014), and a study in which participants simultaneously learned probabilities of monetary rewards and painful shocks suggested that serotonin selectively modulates reward processing (Seymour et al., 2012). Taken together, previous work suggests that the serotonergic system has highly intricate and multifaceted roles in affective learning, likely due to its large number of receptor types, widespread projections, and interactions with other neuromodulators (Dayan and Huys, 2009). Future studies are required to further clarify the functional role(s) of serotonin in pain-avoidance learning.
Conclusion
In sum, our results suggest that received and avoided pain drive human pain-avoidance learning via two different learning systems, in terms of both learning rates and the neural encoding of prediction errors. In addition, our computational-modeling results provide evidence for a causal role of the dopamine and endogenous opioid systems in learning from avoided, but not received, pain. Future studies are needed to elucidate the neural mechanisms via which our dopamine and opioid manipulations affected learning rates for avoided pain, and to reveal the potential role of other neuromodulators in pain-avoidance learning.
Materials and methods
Participants
Ninety-one healthy students (18–26 years old; 71% female; all right handed) took part in the study. Participants reported no history of psychiatric, neurological, or pain disorders, and no current pain. Participants were instructed to abstain from using alcohol or recreational drugs 24 hr prior to testing, and to not eat or drink (except for water) 2 hr prior to testing. The study was approved by the medical ethics committee of the Leiden University Medical Center (P15.116), and all participants provided written informed consent. Participants received a fixed amount of €60 plus a variable bonus of maximally €5 related to their performance on an additional task.
Six participants were excluded from all analyses because of thermode failure, and two additional participants because of poor task performance (see ‘Pain-avoidance learning task’). In addition, nine participants were excluded from the fMRI, but not the behavioral, analyses because of excessive head movement (>3 mm in any direction). Thus, the final behavioral analyses included 83 participants (placebo group: N = 28, mean age = 21.2, 74% female; levodopa group: N = 26, mean age = 20.8, 73% female; naltrexone group: N = 29, mean age = 20.8, 68% female). The final fMRI analyses included 74 participants (placebo group: N = 24, mean age = 21.3, 71% female; levodopa group: N = 24, mean age = 20.9, 71% female; naltrexone group: N = 26, mean age = 20.8, 69% female). Sample size was based on previous studies that detected effects of dopamine (levodopa) or opioid (naloxone) manipulations on behavioral and/or fMRI measures using between-subject designs (Pessiglione et al., 2006; Eippert et al., 2008; Guitart-Masip et al., 2012; Oei et al., 2012; Beierholm et al., 2013; Bunzeck et al., 2014; Guitart-Masip et al., 2014; Wittmann and D’Esposito, 2015). These previous studies used sample sizes ranging from 13 to 30 participants per treatment group. We aimed at a sample size at the higher end of that range (25–30 participants per treatment group).
General procedure
Two to fourteen days prior to the fMRI session, we assessed participants’ eligibility using a general health questionnaire and an fMRI safety screening form. During this screening session, participants also practiced the pain-avoidance learning task.
Each eligible participant took part in one fMRI session. On the day of the fMRI session, participants received a single oral dose of either 100 mg levodopa, 50 mg naltrexone, or placebo, according to a double-blind, randomized, between-subject design. Levodopa was combined with 25 mg carbidopa—a decarboxylase inhibitor that does not cross the blood–brain barrier—to inhibit the conversion of levodopa to dopamine in the periphery. Approximately 30 min after drug administration, participants were positioned in the MRI scanner, after which we acquired a high-resolution structural scan. Approximately 53 min after drug administration, participants completed a 5-min pain-rating task during which they received a series of (unavoidable) heat stimuli of varying temperatures and rated their experienced pain following each stimulus (Figure 1—figure supplement 2). This task was included to test for drug effects on subjective pain responses, and to select a painful yet tolerable temperature for each participant in the pain-avoidance learning task. Sixty minutes after drug administration, roughly corresponding with peak plasma concentrations of levodopa and naltrexone, participants started the pain-avoidance learning task (described below), which lasted approximately 45 min. Following the pain-avoidance learning task, participants performed an 8-min probabilistic reward-learning task (not reported here).
We measured participants’ subjective state at the beginning (before drug administration) and end (2 hr after drug administration) of the test session, by means of visual analog scales measuring alertness, calmness, and contentment (Bond and Lader, 1974 Bond and Lader, 1974; Figure 1—figure supplement 1). Both subjective state measures were collected outside the scanner.
Pain-avoidance learning task
This instrumental pain-avoidance learning task contained 144 trials, divided in 4 runs of 36 trials. On each trial, participants made a choice between two options (a diamond and a circle). Choosing each option was associated with a specific probability of receiving a painful heat stimulation. The probabilities of receiving pain when choosing each option drifted across trials according to two independent random walks (Figure 1A). We used three different pairs of random walks (each pair crossed at least once); each pair was administered to approximately one-third of the participants in each treatment group.
Each trial started with the presentation of the two choice options, randomly displayed at the left and right side of the screen for 1800 ms (Figure 1B). During this period, participants had to select one option by pressing a left or right button of the response unit, using their right index or middle finger, respectively. If participants did not respond in time (1.2% of trials), the computer randomly selected an option for them. The chosen option was highlighted for 200 ms, followed by an anticipation period of 3, 5, or 7 s during which a white asterisk (*) was presented in the center of the screen. Then the outcome—a painful heat stimulus applied to participant’s leg for 1.9 s (see ‘Thermal stimulation’) or no stimulus—was presented. Outcome onset was accompanied by a change of the central asterisk to a colored plus sign (+). The plus sign was red or green during the first 200 ms of each pain and no-pain outcome, respectively, after which it turned white for the remainder of the outcome period. The color change was meant to prevent outcome uncertainty during the initial phase of the outcome period. Each trial ended with an intertrial interval of 6, 8, or 10 s during which an asterisk was presented. Except for the outcome probabilities, participants were fully informed about the task structure and procedure.
During the fMRI session, one participant switched choices more frequently following the absence of pain than following pain, and one other participant did deliberately not make a choice on 17% of the trials, due to the use of irrelevant strategies. We excluded these participants from further analysis.
Thermal stimulation
Heat stimuli (ramp rate = 40°C/s; 1 s at target temperature) were applied to the inner side of participants’ left lower leg using a Contact Heat-Evoked Potential Stimulator (CHEPS; 27 mm diameter Peltier thermode; Medoc Ltd., Israel). After the initial pain-rating task, 16% of the participants (four in the placebo, four in the levodopa, and seven in the naltrexone group) indicated that they would not tolerate repeated stimulation at the highest temperature they had received so far (50°C). For those participants, we used a temperature of 49°C in the pain-avoidance learning task. For the remaining participants, we used a temperature of 50°C. Between stimulations the thermode maintained a baseline temperature of 32°C. The total duration of each stimulation was 1850 ms (425 ms ramp-up and ramp-down periods, 1 s at target temperature) for 49°C stimuli and 1900 ms (450 ms ramp-up and ramp-down periods, 1 s at target temperature) for 50°C stimuli. After each scan run, we moved the thermode to a new site on the participant’s leg. To reduce the impact of potential site-specific habituation (Jepma et al., 2014), we administered one initial heat stimulus before starting the first trial on a new skin site.
Behavioral analyses
We tested whether the total number of received pain stimuli, the proportion of pain trials followed by a switch to the other choice option, and the proportion of no-pain trials followed by a switch to the other choice option differed between the three treatment groups, using one-way analyses of variance. In addition, for each treatment group, we used logistic regression to analyze the probability of switching choices as a function of outcome (pain vs. no-pain) over the six previous trials.
Computational modeling
We fitted two reinforcement-learning (Q learning) models to participants’ choice data: one model with a single learning rate (Model 1), and one model with separate learning rates for pain and no-pain outcomes (Model 2). On each trial t, both models update the expected probability of pain for the selected stimulus s, , in response to the experienced outcome O (pain or no pain), according to:
where reflects the prediction error on each trial. We coded O as −1 and 0 for pain and no-pain outcomes, respectively, and initialized the Q values of both stimuli to −0.5. The value of the unchosen stimulus is not updated. Learning-rate parameter controls how much Q values are updated in response to each new outcome, such that higher values of result in faster updating. The only difference between our two models is that Model 1 uses a single learning rate for all outcomes, whereas Model 2 uses separate learning rates for pain and no-pain outcomes: and , respectively.
Both models were combined with a softmax decision function, which computes the probability of choosing stimulus s on trial t () as:
Inverse-temperature parameter controls the sensitivity of choice probabilities to differences in Q values. If is 0, both stimuli are equally likely to be chosen, irrespective of their expected pain probabilities. As increases, the probability that the model chooses the stimulus with the lower expected pain probability increases. Thus, Model 1 has two free parameters ( and ) and Model 2 has three free parameters ( , , and ).
Parameter estimation
We estimated the model parameters with a hierarchical Bayesian approach, using the hBayesDM package (Ahn et al., 2017). The hierarchical Bayesian approach assumes that every participant has a different set of model parameters, which are drawn from group-level prior distributions (Gelman, 2014). The parameters governing the group-level prior distributions (hyperparameters) are also assigned prior distributions (hyperpriors). We estimated separate group-level parameters for the placebo, levodopa, and naltrexone groups. To test for treatment effects, we compared the posterior distributions of the group-level means (i.e., the hyperparameters governing the means of the group-level distributions) for each drug group vs. the placebo group.
Prior distributions
We used weakly informative priors, as implemented in the hBayesDM package. The group-level distributions for all individual parameters were assumed to be normal distributions. As hyperpriors for the mean and SD hyperparameters of these group-level distributions we used, respectively, a normal distribution with mean 0 and SD 1 and a positive half-Cauchy distribution with location 0 and scale 5 (Ahn et al., 2017). We transformed the parameters’ unconstrained values to a [0,1] range using the inverse probit transformation (Wetzels et al., 2010; Ahn et al., 2014). In addition, we transformed to a [0,20] range by multiplying its inverse-probit transformed values by 20 (Ahn et al., 2017).
MCMC sampling
The hBayesDM package performs hierarchical Bayesian parameter estimation with an MCMC sampling algorithm called Hamiltonian Monte Carlo (HMC) implemented in Stan and its R instantiation RStan. We ran 4 independent MCMC chains with different starting values, which each generated 5000 posterior samples. We discarded the first 1000 iterations of each chain as burn-in. In addition, we only used every fifth iteration to remove autocorrelation. Consequently, we obtained 3200 representative samples (800 per chain) per parameter per model fit. All chains converged (Rhat values <1.1).
Model comparison
We compared the fit of our two models using the WAIC (Watanabe, 2010). The WAIC provides an estimate of a model’s out-of-sample predictive accuracy, adjusted for the number of free parameters, in a fully Bayesian way. Lower WAIC values indicate better out-of-sample predictive accuracy. WAIC is reported on the deviance scale (Gelman, 2014) hence a difference in WAIC value of 2–6 is considered positive evidence, a difference of 6–10 strong evidence, and a difference >10 very strong evidence for one model over another (Kass and Raftery, 1995).
Computation of expected pain probability for use in fMRI analyses
We computed trial-specific expected pain probabilities—to be used as parametric modulator regressors in the fMRI analyses—by applying our winning model to each participant’s sequence of choices and outcomes. In line with previous studies (Pine et al., 2010; Voon et al., 2010; Seymour et al., 2012; Hauser et al., 2015; Kroemer et al., 2019), we used the exact same model to generate fMRI regressors for all participants by instantiating the model with the mean learning-rate parameters (the individual-level posterior medians, averaged across all participants).
fMRI acquisition and analysis
Imaging acquisition
We acquired fMRI data on a 3T Philips Achieva MRI system (Best, The Netherlands) at the Leiden University Medical Center, using a standard whole-head coil. Stimulus presentation and data acquisition were controlled using E-Prime software (Psychology Software Tools). Visual stimuli were presented via a mirror attached to the head coil, and participants responded with their right hand via an MRI-compatible response unit. The pain-avoidance learning task was divided across four scan runs. Each run lasted 581 s (264 TRs), after discarding the first 5 TRs which served as dummy scans. Functional images were acquired with a T2*-weighted whole-brain echo-planar imaging sequence (TR = 2.2 s; TE = 30 ms, flip angle = 80°, 38 transverse slices oriented parallel to the anterior commissure-posterior commissure line, voxel size = 2.75 × 2.75 × 2.75 mm + 10% interslice gap). In addition, we acquired a high-resolution T1-weighted scan (TR = 9.8 ms; TE = 4.6 ms, flip angle = 8°, 140 slices, 1.17 × 1.17 × 1.2 mm, FOV = 224 × 177 × 168), at the beginning of the scan session.
Preprocessing
Prior to preprocessing, global outlier time points (i.e. ‘spikes’ in the BOLD signal) were identified by computing both the mean and the SD (across voxels) of values for each image for all slices. Mahalanobis distances for the matrix of slice-wise mean and SD values (concatenated) × functional volumes (time) were computed, and any values with a significant χ2 value (corrected for multiple comparisons based on the more stringent of either FDR or Bonferroni methods) were considered outliers. On average 3.7% of images were outliers (SD = 1.9). The output of this procedure was later used as a covariate of noninterest in the first-level models.
Functional images were slice-acquisition-timing and motion corrected using SPM8 (Wellcome Trust Centre for Neuroimaging, London, UK). Structural T1-weighted images were coregistered to the first functional image for each subject using an iterative procedure of automated registration using mutual information coregistration in SPM8 and manual adjustment of the automated algorithm’s starting point until the automated procedure provided satisfactory alignment. Structural images were normalized to MNI space using SPM8, interpolated to 2 × 2 × 2 mm voxels, and smoothed using a 6-mm full-width at half maximum Gaussian kernel.
Analysis of outcome-specific prediction-error signals
For the first-level analysis, we create a general linear model for each participant, concatenated over the four pain-avoidance learning blocks, in SPM8. We modeled periods of decision time (cue onset until response, mean response time = 758 ms), outcome anticipation (3–7 s), onsets of pain outcomes (1 s), and onsets of no-pain outcomes (1 s), using boxcar regressors convolved with the canonical hemodynamic response function. As in our previous study (Roy et al., 2014), we only modeled the first second of the outcome periods as this is when prediction errors are triggered. We added the model-derived expected pain probability as a parametric modulator on the outcome-anticipation and outcome-onset regressors. To control for potential effects of outcome-anticipation duration, we also included anticipation duration as a first parametric modulator on the outcome-onset regressors (using serial orthogonalization, such that any shared variance between expected pain probability and anticipation duration is assigned to the anticipation-duration effect). Other regressors of noninterest (nuisance variables) were (1) ‘dummy’ regressors coding for each run (intercept for each but the last run); (2) linear drift across time within each run; (3) the 6 estimated head movement parameters (x, y, z, roll, pitch, and yaw), their mean-zeroed squares, their derivatives, and squared derivatives for each run (total 24 columns per run); (4) indicator vectors for outlier time points identified based on their multivariate distance from the other images in the sample (see above); (5) indicator vectors for the first two images in each run. Low-frequency noise was removed by employing a high-pass filter of 180 s.
To examine outcome-specific prediction-error signals, we created two contrast maps. First, to identify activation that tracks surprise more for received than avoided pain (Figure 3A), we used the following contrast: ‘negative correlation with expected pain probability at pain onset’ > ‘positive correlation with expected pain probability at no-pain onset’. Second, to identify activation tracking absolute prediction error (Figure 3B), we used the following contrast: ‘negative correlation with expected pain probability at pain onset’ > ‘negative correlation with expected pain probability at no-pain onset’. Note that this contrast is identical to: ‘positive correlation with expected pain probability at no-pain onset’ > ‘positive correlation with expected pain probability at pain onset’. It is also identical to a contrast with weights [1 1] for the ‘negative correlation with expected pain probability at pain onset’ and positive correlation with expected pain probability at no-pain onset’ regressors.
We performed a second-level analysis on each of these two contrasts using robust regression, including two second-level regressors coding for levodopa vs. placebo (weights [−1 1 0] for the treatment groups [P L N]) and naltrexone vs. placebo (weights [−1 0 1] for the treatment groups [P L N]). Maps were thresholded at FDR q < 0.05 corrected for multiple comparisons across the whole brain.
Finally, to identify regions encoding outcome-specific prediction errors which cannot be explained by a general sensitivity to expected pain probability (also see ‘Results’) we examined the conjunction between the two second-level contrast maps described above, each thresholded at FDR q < 0.05 (Figure 3C).
Acknowledgements
We thank Stephanie Bauduin, Laila Franke, Nikki Nibbering, Iliana Samara, and Iris Spruit for help with data collection.
Appendix 1
Parameter-recovery analysis
Procedure
We simulated 100 datasets, each consisting of choices from 27 synthetic participants on the pain-avoidance learning task. The number of trials and pain probabilities was the same as for the real participants (each of our three random-walk pairs was assigned to 9 synthetic participants in each dataset). We set the means of the group-level distributions for , , and to the posterior medians obtained from our fits to either the placebo group or the levodopa/naltrexone group, in 50 simulations each (as the posterior medians for the two drug groups were highly similar, we used their average values). We set the SDs of the group-level distributions to 0.1 for and , and to 1 for , in all simulations. In each simulation, the individual-level parameters were randomly drawn from their group-level distributions.
We fitted our winning model (Model 2) to each of the simulated datasets using the same hierarchical Bayesian procedure as used for the real data, and compared the recovered to the simulated (true) group-level mean parameters. In addition, we computed pairwise difference distributions (placebo − drug) for the recovered group-level mean parameters, resulting in 50 difference distributions per parameter. For each parameter, we then computed the proportion of difference distributions whose 95% HDI did not include zero (i.e., the probability that a group difference was detected in the simulated datasets).
Finally, we tested if our model-free performance measures (number of times pain was received and frequencies of switching after received and avoided pain) differed between the datasets that were simulated with our placebo and drug parameters.
Results
The recovered posterior medians of , , and were clustered around their simulated (true) values, for both the placebo and drug simulations (Appendix 1—figure 1). Recovered values of and were negatively correlated (r = −0.80, p < 0.001 and r = −0.51, p < 0.001 for the placebo and drug simulations, respectively), reflecting the typical tradeoff between learning rate and inverse temperature (Cools et al., 2011). In addition, there was a positive correlation between the recovered values of and for the placebo simulations (r = 0.38, P = 0.006)—which was driven by the data point with the highest recovered value (the correlation was not significant when excluding this data point)—but not for the drug simulations (r = −0.02, p = 0.9). The recovered values of and were uncorrelated (p’s > 0.75 for both the placebo and drug simulations).
Appendix 1—figure 1. Recovered posterior medians of , , and for fits to datasets that were simulated using the posterior medians from the placebo (black) and drug (purple) groups.
The recovered values of do not differ between the two groups, but recovered is reliably higher and recovered reliably lower in the drug group, mirroring the parameter estimates obtained from fits to the empirical data.
Despite the correlation between and , recovered was significantly higher for fits to datasets simulated with the drug, as compared to placebo, parameters (none of the 50 difference distributions’ 95% HDI included zero). In addition, recovered was significantly higher for fits to datasets simulated with the placebo parameters (none of the 50 difference distributions’ 95% HDI included zero). The recovered values of did not differ reliably between the drug and placebo simulations (80% of the difference distributions’ 95% HDI included zero). These findings imply that our modeling procedure can accurately dissociate the two patterns of parameter values we found in our placebo and drug groups.
Our model-independent performance measures—number of pain stimuli received, and frequencies of switching after received and avoided pain—did not differ between the datasets that were simulated with the placebo and drug parameters (Appendix 1—table 1), mirroring the absence of treatment effects on these measures in the empirical data. Thus, our modeling procedure can distinguish between the two parameter patterns found in our placebo and drug groups, even though these produce the same model-independent performance measures.
Appendix 1—table 1. Model-free performance measures in simulated datasets that were generated using the group-level mean parameters from our placebo and drug groups.
Placebo simulations | Drug simulations | Averagep value | |
---|---|---|---|
Number of pain stimuli | 54.9 | 55.1 | 0.63 |
Switching after pain | 45.0% | 44.1% | 0.44 |
Switching after no pain | 4.8% | 3.7% | 0.41 |
Notes: Performance measures are averaged across all synthetic participants from each of 50 simulated datasets (27 synthetic participants per dataset). We performed 50 t-tests—each comparing the scores from the synthetic participants from one placebo vs. one drug simulation—and report their average p values.
In sum, the parameter-recovery results suggest that the differences between the placebo and drug groups in both and did not merely reflect a tradeoff between these two parameters or an artifact of the parameter-optimization procedure. Instead, these results suggest that the drugs had two computational effects—an increased learning rate for avoided pain, and an increased level of decision noise—whose combination caused no significant effects on basic, model-independent performance measures.
Appendix 2
General (outcome-nonspecific) aversive and appetitive prediction-error signals
Prediction-error-related activation is often examined by regressing fMRI activity at outcome onset on model-derived prediction errors. However, as prediction errors are defined as the outcome minus the expected outcome, a problem with this approach is that the resulting brain activity may predominantly track the outcome (in our task: pain vs. no pain) or the expected outcome (in our task: the expected pain probability), which are intrinsically correlated with the prediction error.
To address this issue, and identify brain activity that truly integrates actual and expected outcomes into a prediction-error signal, a set of conditions has recently been specified (Rutledge et al., 2010; Roy et al., 2014). These conditions, or axioms, for general aversive prediction-error signals in our task are: (1) activation at outcome onset should be higher for received than avoided pain, unless pain is fully expected; (2) when pain is received, activation should be higher when pain was less expected (i.e., negative correlation with expected pain probability); and (3) when pain is avoided, activation should also be higher when pain was less expected, that is, when avoidance was more expected (Appendix 2—figure 1A, left panels). To identify regions encoding a general aversive prediction-error signal, we tested for activation that fulfilled each of these three axioms, using a whole-brain conjunction analysis. In addition, to search for regions encoding the opposite (i.e., appetitive-like) prediction-error signal, we also tested for activation that fulfilled each of the reverse axioms. Note that we did not detect activation encoding appetitive-like prediction errors in our previous study (Roy et al., 2014).
Appendix 2—figure 1. Axiomatic tests of brain activation encoding general aversive and appetitive prediction errors (N =74).
(A) Activation associated with the three axioms for aversive prediction errors in our task. Yellow regions showed the effects illustrated in the left panels, and blue regions showed the reverse effects (i.e., the axioms for appetitive prediction errors). All maps were thresholded at q < 0.05, false discovery rate (FDR) corrected for multiple comparisons across the whole brain, with higher voxel thresholds superimposed for display. (B) Conjunction results. Regions activated for each of the above three contrasts, all thresholded at q < 0.05 FDR corrected. Yellow and blue regions showed positive and negative responses for each contrast, respectively, thus encoded general aversive and appetitive prediction errors.
A fourth axiom, that applies to both aversive and appetitive prediction errors, is that activation for received and avoided pain should be equivalent if the outcome is fully predicted (i.e., when the prediction error is zero). As outcomes could never be fully predicted in our task, we could not test this axiom.
Axiom 1
A large part of the brain fulfilled the first axiom for aversive prediction errors (stronger response to received than avoided pain), including typical pain-processing regions such as the dorsal ACC, (pre)motor cortex, anterior and posterior insula, and thalamus, as well as occipital (visual) cortex (Appendix 2—figure 1A, upper panel). We found the opposite effect (stronger response to avoided than received pain) in regions of the ventromedial prefrontal cortex (vmPFC), dorsolateral prefrontal cortex (dlPFC), somatosensory cortex, posterior ACC, and later occipital cortex (LOC).
Axiom 2
A test of the second axiom for aversive prediction errors (stronger responses to more unexpected pain) revealed several activation clusters, including regions in the vmPFC, dorsal ACC, insula, amygdala, and a midbrain area covering part of the PAG (Appendix 2—figure 1A, middle panel). In addition, several other regions, including the right dlPFC and bilateral somatosensory cortex, showed the opposite effect (stronger responses to more expected pain).
Axiom 3
The third axiom for aversive prediction errors (stronger responses to more expected pain avoidance) was fulfilled by a few regions in the vmPFC and rostral ACC (rACC), as well as part of the PAG. We also found the opposite effect (stronger responses to more unexpected pain avoidance) in several regions, including the dorsal ACC, sensorimotor cortex, thalamus, putamen, and insula.
Conjunction
A conjunction analysis of the three contrasts reported above (Appendix 2—figure 1B) revealed two brain regions that satisfied all three axioms for aversive prediction errors: A midbrain region including part of the PAG (16 voxels) and an area in the rostral ACC (24 voxels). Importantly, we also identified activation that showed a negative effect for all three axioms—thus encoding appetitive-like prediction errors—in bilateral somatosensory cortex (433 and 65 voxels in the left and right hemisphere, respectively), left frontopolar cortex (47 voxels), right dlPFC (middle frontal gyrus; 27 voxels), and right LOC (253 voxels).
Together, these results replicate our previous finding that the PAG encodes general aversive prediction errors (Roy et al., 2014), and suggest a role for the rostral ACC in encoding aversive prediction errors as well. Furthermore, the identification of an additional neural circuit encoding appetitive-like prediction errors provides an important extension to our previous results, possibly owing to the larger number of participants and hence higher power in the present study.
Funding Statement
The funders had no role in study design, data collection, and interpretation, or the decision to submit the work for publication.
Contributor Information
Marieke Jepma, Email: mariekejepma@gmail.com.
Catherine Hartley, New York University, United States.
Christian Büchel, University Medical Center Hamburg-Eppendorf, Germany.
Funding Information
This paper was supported by the following grant:
Nederlandse Organisatie voor Wetenschappelijk Onderzoek Veni to Marieke Jepma.
Additional information
Competing interests
No competing interests declared.
No competing interests declared.
Author contributions
Conceptualization, Data curation, Formal analysis, Funding acquisition, Investigation, Methodology, Project administration, Supervision, Validation, Visualization, Writing - original draft.
Conceptualization, Writing – review and editing.
Investigation, Writing – review and editing.
Resources, Writing – review and editing.
Resources, Writing – review and editing.
Ethics
All participants provided written informed consent study. The study was approved by the medical ethics committee of the Leiden University Medical Center (P15.116).
Additional files
Data availability
Single-trial behavioral data and model code are available from the OSF database: https://osf.io/rqc6g/. Unthresholded t maps for all reported fMRI contrasts are available on Neurovault: https://identifiers.org/neurovault.collection:6016.
The following datasets were generated:
Jepma M. 2021. Two systems for pain-avoidance learning. Open Science Framework. rqc6g
Jepma M. 2021. Two brain systems for pain-avoidance learning. NeuroVault. neurovault.collection:6016
References
- Ahn WY, Vasilev G, Lee SH, Busemeyer JR, Kruschke JK, Bechara A, Vassileva J. Decision-making in stimulant and opiate addicts in protracted abstinence: evidence from computational modeling with pure users. Frontiers in Psychology. 2014;5:849. doi: 10.3389/fpsyg.2014.00849. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Ahn WY, Haines N, Zhang L. Revealing Neurocomputational Mechanisms of Reinforcement Learning and Decision-Making With the hBayesDM Package. Computational Psychiatry (Cambridge, Mass.) 2017;1:24–57. doi: 10.1162/CPSY_a_00002. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Bartra O, McGuire JT, Kable JW. The valuation system: a coordinate-based meta-analysis of BOLD fMRI experiments examining neural correlates of subjective value. NeuroImage. 2013;76:412–427. doi: 10.1016/j.neuroimage.2013.02.063. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Beeler JA, Daw N, Frazier CRM, Zhuang X. Tonic dopamine modulates exploitation of reward learning. Frontiers in Behavioral Neuroscience. 2010;4:170. doi: 10.3389/fnbeh.2010.00170. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Beierholm U, Guitart-Masip M, Economides M, Chowdhury R, Düzel E, Dolan R, Dayan P. Dopamine modulates reward-related vigor. Neuropsychopharmacology. 2013;38:1495–1503. doi: 10.1038/npp.2013.48. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Berridge KC, Kringelbach ML. Pleasure systems in the brain. Neuron. 2015;86:646–664. doi: 10.1016/j.neuron.2015.02.018. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Black KJ, Piccirillo ML, Koller JM, Hseih T, Wang L, Mintun MA. Levodopa effects on [ (11)C]raclopride binding in the resting human brain. F1000Research. 2015;4:23. doi: 10.12688/f1000research.5672.1. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Bond A, Lader M. The use of analogue scales in rating subjective feelings. British Journal of Medical Psychology. 1974;47:211–218. doi: 10.1111/j.2044-8341.1974.tb02285.x. [DOI] [Google Scholar]
- Boureau YL, Dayan P. Opponency revisited: competition and cooperation between dopamine and serotonin. Neuropsychopharmacology. 2011;36:74–97. doi: 10.1038/npp.2010.151. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Breitenstein C, Korsukewitz C, Flöel A, Kretzschmar T, Diederich K, Knecht S. Tonic dopaminergic stimulation impairs associative learning in healthy subjects. Neuropsychopharmacology. 2006;31:2552–2564. doi: 10.1038/sj.npp.1301167. [DOI] [PubMed] [Google Scholar]
- Brocka M, Helbing C, Vincenz D, Scherf T, Montag D, Goldschmidt J, Angenstein F, Lippert M. Contributions of dopaminergic and non-dopaminergic neurons to VTA-stimulation induced neurovascular responses in brain reward circuits. NeuroImage. 2018;177:88–97. doi: 10.1016/j.neuroimage.2018.04.059. [DOI] [PubMed] [Google Scholar]
- Brooks JCW, Faull OK, Pattinson KTS, Jenkinson M. Physiological noise in brainstem FMRI. Frontiers in Human Neuroscience. 2013;7:623. doi: 10.3389/fnhum.2013.00623. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Bunzeck N, Guitart-Masip M, Dolan RJ, Duzel E. Pharmacological dissociation of novelty responses in the human brain. Cerebral Cortex (New York, N.Y. 2014;24:1351–1360. doi: 10.1093/cercor/bhs420. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Bush G, Luu P, Posner MI. Cognitive and emotional influences in anterior cingulate cortex. Trends in Cognitive Sciences. 2000;4:215–222. doi: 10.1016/s1364-6613(00)01483-2. [DOI] [PubMed] [Google Scholar]
- Chakroun K, Mathar D, Wiehler A, Ganzer F, Peters J. Dopaminergic modulation of the exploration/exploitation trade-off in human decision-making. eLife. 2020;9:e51260. doi: 10.7554/eLife.51260. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Chamberlain SR, Müller U, Blackwell AD, Clark L, Robbins TW, Sahakian BJ. Neurochemical modulation of response inhibition and probabilistic learning in humans. Science (New York, N.Y.) 2006;311:861–863. doi: 10.1126/science.1121218. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Chew B, Hauser TU, Papoutsi M, Magerkurth J, Dolan RJ, Rutledge RB. Endogenous fluctuations in the dopaminergic midbrain drive behavioral choice variability. PNAS. 2019;116:18732–18737. doi: 10.1073/pnas.1900872116. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Cole S, McNally GP. Opioid receptors mediate direct predictive fear learning: evidence from one-trial blocking. Learning & Memory (Cold Spring Harbor, N.Y.) 2007;14:229–235. doi: 10.1101/lm.489507. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Cole S, McNally GP. Complementary roles for amygdala and periaqueductal gray in temporal-difference fear learning. Learning & Memory (Cold Spring Harbor, N.Y.) 2009;16:1–7. doi: 10.1101/lm.1120509. [DOI] [PubMed] [Google Scholar]
- Cools R, Robbins TW. Chemistry of the adaptive mind. Philosophical Transactions. Series A, Mathematical, Physical, and Engineering Sciences. 2004;362:2871–2888. doi: 10.1098/rsta.2004.1468. [DOI] [PubMed] [Google Scholar]
- Cools R, Robinson OJ, Sahakian B. Acute tryptophan depletion in healthy volunteers enhances punishment prediction but does not affect reward prediction. Neuropsychopharmacology. 2008;33:2291–2299. doi: 10.1038/sj.npp.1301598. [DOI] [PubMed] [Google Scholar]
- Cools R, D’Esposito M. Inverted-U-shaped dopamine actions on human working memory and cognitive control. Biological Psychiatry. 2011;69:e113–e125. doi: 10.1016/j.biopsych.2011.03.028. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Cools R, Nakamura K, Daw ND. Serotonin and dopamine: unifying affective, activational, and decision functions. Neuropsychopharmacology : Official Publication of the American College of Neuropsychopharmacology. 2011;36:98–113. doi: 10.1038/npp.2010.121. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Crockett MJ, Clark L, Roiser JP, Robinson OJ, Cools R, Chase HW, Ouden HD, Apergis-Schoute A, Campbell-Meiklejohn D, Campbell-Meikeljohn D, Seymour B, Sahakian BJ, Rogers RD, Robbins TW. Converging evidence for central 5-HT effects in acute tryptophan depletion. Molecular Psychiatry. 2012;17:121–123. doi: 10.1038/mp.2011.106. [DOI] [PubMed] [Google Scholar]
- Daw ND, Kakade S, Dayan P. Opponent interactions between serotonin and dopamine. Neural Networks. 2002;15:603–616. doi: 10.1016/s0893-6080(02)00052-7. [DOI] [PubMed] [Google Scholar]
- Dayan P, Huys QJM. Serotonin in affective control. Annual Review of Neuroscience. 2009;32:95–126. doi: 10.1146/annurev.neuro.051508.135607. [DOI] [PubMed] [Google Scholar]
- Deakin JFW. Roles of serotonergic systems in escape, avoidance and other behaviours. Theory in Psychopharmacology. 1983;2:149–193. [Google Scholar]
- Deakin JF, Graeff FG. 5-HT and mechanisms of defence. Journal of Psychopharmacology (Oxford, England) 1991;5:305–315. doi: 10.1177/026988119100500414. [DOI] [PubMed] [Google Scholar]
- Deichmann R, Gottfried JA, Hutton C, Turner R. Optimized EPI for fMRI studies of the orbitofrontal cortex. NeuroImage. 2003;19:430–441. doi: 10.1016/s1053-8119(03)00073-9. [DOI] [PubMed] [Google Scholar]
- den Ouden HEM, Daw ND, Fernandez G, Elshout JA, Rijpkema M, Hoogman M, Franke B, Cools R. Dissociable effects of dopamine and serotonin on reversal learning. Neuron. 2013;80:1090–1100. doi: 10.1016/j.neuron.2013.08.030. [DOI] [PubMed] [Google Scholar]
- Dinsmoor JA. Stimuli inevitably generated by behavior that avoids electric shock are inherently reinforcing. Journal of the Experimental Analysis of Behavior. 2001;75:311–333. doi: 10.1901/jeab.2001.75-311. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Dombrowski PA, Maia TV, Boschen SL, Bortolanza M, Wendler E, Schwarting RKW, Brandão ML, Winn P, Blaha CD, Da Cunha C. Evidence that conditioned avoidance responses are reinforced by positive prediction errors signaled by tonic striatal dopamine. Behavioural Brain Research. 2013;241:112–119. doi: 10.1016/j.bbr.2012.06.031. [DOI] [PubMed] [Google Scholar]
- Eippert F, Bingel U, Schoell E, Yacubian J, Büchel C. Blockade of endogenous opioid neurotransmission enhances acquisition of conditioned fear in humans. The Journal of Neuroscience. 2008;28:5465–5472. doi: 10.1523/JNEUROSCI.5336-07.2008. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Eippert F, Bingel U, Schoell ED, Yacubian J, Klinger R, Lorenz J, Büchel C. Activation of the opioidergic descending pain control system underlies placebo analgesia. Neuron. 2009;63:533–543. doi: 10.1016/j.neuron.2009.07.014. [DOI] [PubMed] [Google Scholar]
- Eldar E, Hauser TU, Dayan P, Dolan RJ. Striatal structure and function predict individual biases in learning to avoid pain. PNAS. 2016;113:4812–4817. doi: 10.1073/pnas.1519829113. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Fernando ABP, Urcelay GP, Mar AC, Dickinson A, Robbins TW. Comparison of the conditioned reinforcing properties of a safety signal and appetitive stimulus: effects of d-amphetamine and anxiolytics. Psychopharmacology. 2013;227:195–208. doi: 10.1007/s00213-012-2952-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Floel A, Garraux G, Xu B, Breitenstein C, Knecht S, Herscovitch P, Cohen LG. Levodopa increases memory encoding and dopamine release in the striatum in the elderly. Neurobiology of Aging. 2008;29:267–279. doi: 10.1016/j.neurobiolaging.2006.10.009. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Frank MJ, Seeberger LC, O’reilly RC. By carrot or by stick: cognitive reinforcement learning in parkinsonism. Science (New York, N.Y.) 2004;306:1940–1943. doi: 10.1126/science.1102941. [DOI] [PubMed] [Google Scholar]
- Frank MJ. Dynamic dopamine modulation in the basal ganglia: a neurocomputational account of cognitive deficits in medicated and nonmedicated Parkinsonism. Journal of Cognitive Neuroscience. 2005;17:51–72. doi: 10.1162/0898929052880093. [DOI] [PubMed] [Google Scholar]
- Frank MJ, Fossella JA. Neurogenetics and pharmacology of learning, motivation, and cognition. Neuropsychopharmacology : Official Publication of the American College of Neuropsychopharmacology. 2011;36:133–152. doi: 10.1038/npp.2010.96. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Gallagher DA, O’Sullivan SS, Evans AH, Lees AJ, Schrag A. Pathological gambling in Parkinson’s disease: risk factors and differences from dopamine dysregulation An analysis of published case series. Movement Disorders. 2007;22:1757–1763. doi: 10.1002/mds.21611. [DOI] [PubMed] [Google Scholar]
- Gelman A. Bayesian data analysis. Third edition. Boca Raton: CRC Press; 2014. [Google Scholar]
- Gershman SJ, Tzovaras BG. Dopaminergic genes are associated with both directed and random exploration. Neuropsychologia. 2018;120:97–104. doi: 10.1016/j.neuropsychologia.2018.10.009. [DOI] [PubMed] [Google Scholar]
- Grevert P, Goldstein A. Effects of naloxone on experimentally induced ischemic pain and on mood in human subjects. PNAS. 1977;74:1291–1294. doi: 10.1073/pnas.74.3.1291. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Guitart-Masip M, Chowdhury R, Sharot T, Dayan P, Duzel E, Dolan RJ. Action controls dopaminergic enhancement of reward representations. PNAS. 2012;109:7511–7516. doi: 10.1073/pnas.1202229109. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Guitart-Masip Marc, Economides M, Huys QJM, Frank MJ, Chowdhury R, Duzel E, Dayan P, Dolan RJ. Differential, but not opponent, effects of L -DOPA and citalopram on action learning with reward and punishment. Psychopharmacology. 2014;231:955–966. doi: 10.1007/s00213-013-3313-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Hauser TU, Iannaccone R, Walitza S, Brandeis D, Brem S. Cognitive flexibility in adolescence: neural and behavioral mechanisms of reward prediction error processing in adaptive decision making during development. NeuroImage. 2015;104:347–354. doi: 10.1016/j.neuroimage.2014.09.018. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Hindi Attar C, Finckh B, Büchel C. The influence of serotonin on fear learning. PLOS ONE. 2012;7:e42397. doi: 10.1371/journal.pone.0042397. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Huys QJM, Eshel N, O’Nions E, Sheridan L, Dayan P, Roiser JP. Bonsai trees in your head: how the pavlovian system sculpts goal-directed choices by pruning decision trees. PLOS Computational Biology. 2012;8:e1002410. doi: 10.1371/journal.pcbi.1002410. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Jepma M, Jones M, Wager TD. The dynamics of pain: evidence for simultaneous site-specific habituation and site-nonspecific sensitization in thermal pain. The Journal of Pain. 2014;15:734–746. doi: 10.1016/j.jpain.2014.02.010. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Jongkees BJ, Colzato LS. Spontaneous eye blink rate as predictor of dopamine-related cognitive function-A review. Neuroscience and Biobehavioral Reviews. 2016;71:58–82. doi: 10.1016/j.neubiorev.2016.08.020. [DOI] [PubMed] [Google Scholar]
- Josselyn SA, Falls WA, Gewirtz JC, Pistell P, Davis M. The nucleus accumbens is not critically involved in mediating the effects of a safety signal on behavior. Neuropsychopharmacology : Official Publication of the American College of Neuropsychopharmacology. 2005;30:17–26. doi: 10.1038/sj.npp.1300530. [DOI] [PubMed] [Google Scholar]
- Kass RE, Raftery AE. Bayes Factors. Journal of the American Statistical Association. 1995;90:773–795. doi: 10.1080/01621459.1995.10476572. [DOI] [Google Scholar]
- Kayser AS, Mitchell JM, Weinstein D, Frank MJ. Dopamine, locus of control, and the exploration-exploitation tradeoff. Neuropsychopharmacology : Official Publication of the American College of Neuropsychopharmacology. 2015;40:454–462. doi: 10.1038/npp.2014.193. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kim JH, Richardson R. The effect of the mu-opioid receptor antagonist naloxone on extinction of conditioned fear in the developing rat. Learning & Memory (Cold Spring Harbor, N.Y.) 2009;16:161–166. doi: 10.1101/lm.1282309. [DOI] [PubMed] [Google Scholar]
- Knutson B, Gibbs SEB. Linking nucleus accumbens dopamine and blood oxygenation. Psychopharmacology. 2007;191:813–822. doi: 10.1007/s00213-006-0686-7. [DOI] [PubMed] [Google Scholar]
- Kroemer NB, Lee Y, Pooseh S, Eppinger B, Goschke T, Smolka MN. L-DOPA reduces model-free control of behavior by attenuating the transfer of value to action. NeuroImage. 2019;186:113–125. doi: 10.1016/j.neuroimage.2018.10.075. [DOI] [PubMed] [Google Scholar]
- Lee MC, Wagner HN, Tanada S, Frost JJ, Bice AN, Dannals RF. Duration of occupancy of opiate receptors by naltrexone. Journal of Nuclear Medicine. 1988;29:1207–1211. [PubMed] [Google Scholar]
- Lefebvre G, Lebreton M, Meyniel F, Bourgeois-Gironde S, Palminteri S. Behavioural and neural characterization of optimistic reinforcement learning. Nature Human Behaviour. 2017;1:e67. doi: 10.1038/s41562-017-0067. [DOI] [Google Scholar]
- Levine JD, Gordon NC, Fields HL. The mechanism of placebo analgesia. Lancet (London, England) 1978;2:654–657. doi: 10.1016/s0140-6736(78)92762-9. [DOI] [PubMed] [Google Scholar]
- Linnman C, Moulton EA, Barmettler G, Becerra L, Borsook D. Neuroimaging of the periaqueductal gray: state of the field. NeuroImage. 2012;60:505–522. doi: 10.1016/j.neuroimage.2011.11.095. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Maia TV, Frank MJ. From reinforcement learning models to psychiatric and neurological disorders. Nature Neuroscience. 2011;14:154–162. doi: 10.1038/nn.2723. [DOI] [PMC free article] [PubMed] [Google Scholar]
- McNally GP, Westbrook RF. Opioid receptors regulate the extinction of Pavlovian fear conditioning. Behavioral Neuroscience. 2003;117:1292–1301. doi: 10.1037/0735-7044.117.6.1292. [DOI] [PubMed] [Google Scholar]
- McNally GP, Pigg M, Weidemann G. Opioid receptors in the midbrain periaqueductal gray regulate extinction of pavlovian fear conditioning. The Journal of Neuroscience. 2004;24:6912–6919. doi: 10.1523/JNEUROSCI.1828-04.2004. [DOI] [PMC free article] [PubMed] [Google Scholar]
- McNally GP, Lee BW, Chiem JY, Choi EA. The midbrain periaqueductal gray and fear extinction: opioid receptor subtype and roles of cyclic AMP, protein kinase A, and mitogen-activated protein kinase. Behavioral Neuroscience. 2005;119:1023–1033. doi: 10.1037/0735-7044.119.4.1023. [DOI] [PubMed] [Google Scholar]
- McNally GP, Cole S. Opioid receptors in the midbrain periaqueductal gray regulate prediction errors during pavlovian fear conditioning. Behavioral Neuroscience. 2006;120:313–323. doi: 10.1037/0735-7044.120.2.313. [DOI] [PubMed] [Google Scholar]
- McNally GP, Johansen JP, Blair HT. Placing prediction into the fear circuit. Trends in Neurosciences. 2011;34:283–292. doi: 10.1016/j.tins.2011.03.005. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Moustafa AA, Herzallah MM, Gluck MA. Dissociating the cognitive effects of levodopa versus dopamine agonists in a neurocomputational model of learning in Parkinson’s disease. Neuro-Degenerative Diseases. 2013;11:102–111. doi: 10.1159/000341999. [DOI] [PubMed] [Google Scholar]
- Moutoussis M, Bentall RP, Williams J, Dayan P. A temporal difference account of avoidance learning. Network (Bristol, England) 2008;19:137–160. doi: 10.1080/09548980802192784. [DOI] [PubMed] [Google Scholar]
- Mowrer OH. Two-factor learning theory: summary and comment. Psychological Review. 1951;58:350–354. doi: 10.1037/h0058956. [DOI] [PubMed] [Google Scholar]
- Mowrer OH. Two-factor learning theory reconsidered, with special reference to secondary reinforcement and the concept of habit. Psychological Review. 1956;63:114–128. doi: 10.1037/h0040613. [DOI] [PubMed] [Google Scholar]
- Nummenmaa L, Tuominen L. Opioid system and human emotions. British Journal of Pharmacology. 2018;175:2737–2749. doi: 10.1111/bph.13812. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Oei NYL, Rombouts SA, Soeter RP, van Gerven JM, Both S. Dopamine modulates reward system activity during subconscious processing of sexual stimuli. Neuropsychopharmacology. 2012;37:1729–1737. doi: 10.1038/npp.2012.19. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Ojemann JG, Akbudak E, Snyder AZ, McKinstry RC, Raichle ME, Conturo TE. Anatomic localization and quantitative analysis of gradient refocused echo-planar fMRI susceptibility artifacts. NeuroImage. 1997;6:156–167. doi: 10.1006/nimg.1997.0289. [DOI] [PubMed] [Google Scholar]
- Oleson EB, Gentry RN, Chioma VC, Cheer JF. Subsecond dopamine release in the nucleus accumbens predicts conditioned punishment and its successful avoidance. The Journal of Neuroscience. 2012;32:14804–14808. doi: 10.1523/JNEUROSCI.3087-12.2012. [DOI] [PMC free article] [PubMed] [Google Scholar]
- O’Doherty JP, Dayan P, Friston K, Critchley H, Dolan RJ. Temporal difference models and reward-related learning in the human brain. Neuron. 2003;38:329–337. doi: 10.1016/s0896-6273(03)00169-7. [DOI] [PubMed] [Google Scholar]
- Palminteri S, Clair AH, Mallet L, Pessiglione M. Similar improvement of reward and punishment learning by serotonin reuptake inhibitors in obsessive-compulsive disorder. Biological Psychiatry. 2012;72:244–250. doi: 10.1016/j.biopsych.2011.12.028. [DOI] [PubMed] [Google Scholar]
- Palminteri S, Lefebvre G, Kilford EJ, Blakemore SJ. Confirmation bias in human reinforcement learning: Evidence from counterfactual feedback processing. PLOS Computational Biology. 2017;13:e1005684. doi: 10.1371/journal.pcbi.1005684. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Palminteri S, Pessiglione M. Opponent brain systems for reward and punishment learning: causal evidence from drug and lesion studies in humans. Decision Neuroscience: An Integrative Perspective. J. C. Dreher and L. 2017;1:e3. doi: 10.1016/B978-0-12-805308-9.00023-3. [DOI] [Google Scholar]
- Palomero-Gallagher N, Vogt BA, Schleicher A, Mayberg HS, Zilles K. Receptor architecture of human cingulate cortex: evaluation of the four-region neurobiological model. Human Brain Mapping. 2009;30:2336–2355. doi: 10.1002/hbm.20667. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Parsons RG, Gafford GM, Helmstetter FJ. Regulation of extinction-related plasticity by opioid receptors in the ventrolateral periaqueductal gray matter. Frontiers in Behavioral Neuroscience. 2010;4:44. doi: 10.3389/fnbeh.2010.00044. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Pessiglione M, Seymour B, Flandin G, Dolan RJ, Frith CD. Dopamine-dependent prediction errors underpin reward-seeking behaviour in humans. Nature. 2006;442:1042–1045. doi: 10.1038/nature05051. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Phillips RG, LeDoux JE. Differential contribution of amygdala and hippocampus to cued and contextual fear conditioning. Behavioral Neuroscience. 1992;106:274–285. doi: 10.1037//0735-7044.106.2.274. [DOI] [PubMed] [Google Scholar]
- Pine A, Shiner T, Seymour B, Dolan RJ. Dopamine, time, and impulsivity in humans. The Journal of Neuroscience. 2010;30:8888–8896. doi: 10.1523/JNEUROSCI.6028-09.2010. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Ploghaus A, Tracey I, Clare S, Gati JS, Rawlins JN, Matthews PM. Learning about pain: the neural substrate of the prediction error for aversive events. PNAS. 2000;97:9281–9286. doi: 10.1073/pnas.160266497. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Poletti M, Bonuccelli U. Acute and chronic cognitive effects of levodopa and dopamine agonists on patients with Parkinson’s disease: a review. Therapeutic Advances in Psychopharmacology. 2013;3:101–113. doi: 10.1177/2045125312470130. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Preston KL, Bigelow GE. Differential naltrexone antagonism of hydromorphone and pentazocine effects in human volunteers. The Journal of Pharmacology and Experimental Therapeutics. 1993;264:813–823. [PubMed] [Google Scholar]
- Ptak R, Schnider A, Fellrath J. The Dorsal Frontoparietal Network: A Core System for Emulated Action. Trends in Cognitive Sciences. 2017;21:589–599. doi: 10.1016/j.tics.2017.05.002. [DOI] [PubMed] [Google Scholar]
- Raczka KA, Mechias ML, Gartmann N, Reif A, Deckert J, Pessiglione M, Kalisch R. Empirical support for an involvement of the mesostriatal dopamine system in human fear extinction. Translational Psychiatry. 2011;1:e12. doi: 10.1038/tp.2011.10. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Rescorla RA, Wagner AR. A theory of Pavlovian conditioning: Variations in the effectiveness of reinforcement and nonreinforcement Classical conditioning II. Current Research and Theory. 1972;1:64–99. [Google Scholar]
- Rich EL, Wallis JD. Decoding subjective decisions from orbitofrontal cortex. Nature Neuroscience. 2016;19:973–980. doi: 10.1038/nn.4320. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Robinson OJ, Cools R, Sahakian BJ. Tryptophan depletion disinhibits punishment but not reward prediction: implications for resilience. Psychopharmacology. 2012;219:599–605. doi: 10.1007/s00213-011-2410-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Roy M, Shohamy D, Wager TD. Ventromedial prefrontal-subcortical systems and the generation of affective meaning. Trends in Cognitive Sciences. 2012;16:147–156. doi: 10.1016/j.tics.2012.01.005. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Roy M, Shohamy D, Daw N, Jepma M, Wimmer GE, Wager TD. Representation of aversive prediction errors in the human periaqueductal gray. Nature Neuroscience. 2014;17:1607–1612. doi: 10.1038/nn.3832. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Rutledge RB, Dean M, Caplin A, Glimcher PW. Testing the reward prediction error hypothesis with an axiomatic model. The Journal of Neuroscience. 2010;30:13525–13536. doi: 10.1523/JNEUROSCI.1747-10.2010. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Salinas-Hernández XI, Vogel P, Betz S, Kalisch R, Sigurdsson T, Duvarci S. Dopamine neurons drive fear extinction learning by signaling the omission of expected aversive outcomes. eLife. 2018;7:e38818. doi: 10.7554/eLife.38818. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Schuh KJ, Walsh SL, Stitzer ML. Onset, magnitude and duration of opioid blockade produced by buprenorphine and naltrexone in humans. Psychopharmacology. 1999;145:162–174. doi: 10.1007/s002130051045. [DOI] [PubMed] [Google Scholar]
- Schultz W, Dayan P, Montague PR. A neural substrate of prediction and reward. Science (New York, N.Y.) 1997;275:1593–1599. doi: 10.1126/science.275.5306.1593. [DOI] [PubMed] [Google Scholar]
- Seligman ME, Maier SF. Failure to escape traumatic shock. Journal of Experimental Psychology. 1967;74:1–9. doi: 10.1037/h0024514. [DOI] [PubMed] [Google Scholar]
- Seymour B, O’Doherty JP, Dayan P, Koltzenburg M, Jones AK, Dolan RJ, Friston KJ, Frackowiak RS. Temporal difference models describe higher-order learning in humans. Nature. 2004;429:664–667. doi: 10.1038/nature02581. [DOI] [PubMed] [Google Scholar]
- Seymour B, O’Doherty JP, Koltzenburg M, Wiech K, Frackowiak R, Friston K, Dolan R. Opponent appetitive-aversive neural processes underlie predictive learning of pain relief. Nature Neuroscience. 2005;8:1234–1240. doi: 10.1038/nn1527. [DOI] [PubMed] [Google Scholar]
- Seymour B, Daw ND, Roiser JP, Dayan P, Dolan R. Serotonin selectively modulates reward value in human decision-making. The Journal of Neuroscience. 2012;32:5833–5842. doi: 10.1523/JNEUROSCI.0053-12.2012. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Sharot T, Garrett N. Forming Beliefs: Why Valence Matters. Trends in Cognitive Sciences. 2016;20:25–33. doi: 10.1016/j.tics.2015.11.002. [DOI] [PubMed] [Google Scholar]
- Sirucek L, Price RC, Gandhi W, Hoeppli M-E, Fahey E, Qu A, Becker S, Schweinhardt P. Endogenous opioids contribute to the feeling of pain relief in humans. Pain. 2021;162:2821–2831. doi: 10.1097/j.pain.0000000000002285. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Soubrié P. Reconciling the role of central serotonin neurons in human and animal behavior. Behavioral and Brain Sciences. 1986;9:319–335. doi: 10.1017/S0140525X00022871. [DOI] [Google Scholar]
- St Onge JR, Floresco SB. Dopaminergic modulation of risk-based decision making. Neuropsychopharmacology : Official Publication of the American College of Neuropsychopharmacology. 2009;34:681–697. doi: 10.1038/npp.2008.121. [DOI] [PubMed] [Google Scholar]
- Stelly CE, Haug GC, Fonzi KM, Garcia MA, Tritley SC, Magnon AP, Ramos MAP, Wanat MJ. Pattern of dopamine signaling during aversive events predicts active avoidance learning. PNAS. 2019;116:13641–13650. doi: 10.1073/pnas.1904249116. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Sutton RS, Barto AG. Reinforcement Learning: An Introduction. IEEE Transactions on Neural Networks. 1998;9:1054. doi: 10.1109/TNN.1998.712192. [DOI] [Google Scholar]
- Vellani V, de Vries LP, Gaule A, Sharot T. A selective effect of dopamine on information-seeking. eLife. 2020;9:e59152. doi: 10.7554/eLife.59152. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Voon V, Hassan K, Zurowski M, Duff-Canning S, de Souza M, Fox S, Lang AE, Miyasaki J. Prospective prevalence of pathologic gambling and medication association in Parkinson disease. Neurology. 2006;66:1750–1752. doi: 10.1212/01.wnl.0000218206.20920.4d. [DOI] [PubMed] [Google Scholar]
- Voon V, Pessiglione M, Brezing C, Gallea C, Fernandez HH, Dolan RJ, Hallett M. Mechanisms underlying dopamine-mediated reward bias in compulsive behaviors. Neuron. 2010;65:135–142. doi: 10.1016/j.neuron.2009.12.027. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Watanabe S. Asymptotic equivalence of Bayes cross validation and widely applicable information criterion in singular learning theory. Journal of Machine Learning Research. 2010;11:3571–3594. [Google Scholar]
- Weerts EM, McCaul ME, Kuwabara H, Yang X, Xu X, Dannals RF, Frost JJ, Wong DF, Wand GS. Influence of OPRM1 Asn40Asp variant (A118G) on [11C]carfentanil binding potential: preliminary findings in human subjects. The International Journal of Neuropsychopharmacology. 2013;16:47–53. doi: 10.1017/S146114571200017X. [DOI] [PubMed] [Google Scholar]
- Wenzel JM, Oleson EB, Gove WN, Cole AB, Gyawali U, Dantrassy HM, Bluett RJ, Dryanovski DI, Stuber GD, Deisseroth K, Mathur BN, Patel S, Lupica CR, Cheer JF. Phasic Dopamine Signals in the Nucleus Accumbens that Cause Active Avoidance Require Endocannabinoid Mobilization in the Midbrain. Current Biology. 2018;28:1392–1404. doi: 10.1016/j.cub.2018.03.037. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Werner MU, Pereira MP, Andersen LPH, Dahl JB. Endogenous opioid antagonism in physiological experimental pain models: a systematic review. PLOS ONE. 2015;10:e0125887. doi: 10.1371/journal.pone.0125887. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Wetzels R, Vandekerckhove J, Tuerlinckx F, Wagenmakers E-J. Bayesian parameter estimation in the Expectancy Valence model of the Iowa gambling task. Journal of Mathematical Psychology. 2010;54:14–27. doi: 10.1016/j.jmp.2008.12.001. [DOI] [Google Scholar]
- Wietzikoski EC, Boschen SL, Miyoshi E, Bortolanza M, Dos Santos LM, Frank M, Brandão ML, Winn P, Da Cunha C. Roles of D1-like dopamine receptors in the nucleus accumbens and dorsolateral striatum in conditioned avoidance responses. Psychopharmacology. 2012;219:159–169. doi: 10.1007/s00213-011-2384-3. [DOI] [PubMed] [Google Scholar]
- Wise T, Michely J, Dayan P, Dolan RJ. A computational account of threat-related attentional bias. PLOS Computational Biology. 2019;15:e1007341. doi: 10.1371/journal.pcbi.1007341. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Wittmann BC, D’Esposito M. Levodopa administration modulates striatal processing of punishment-associated items in healthy participants. Psychopharmacology. 2015;232:135–144. doi: 10.1007/s00213-014-3646-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Yarkoni T. The generalizability crisis. The Behavioral and Brain Sciences. 2020;45:1–37. doi: 10.1017/S0140525X20001685. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Zhang S, Mano H, Lee M, Yoshida W, Kawato M, Robbins TW, Seymour B. The control of tonic pain by active relief learning. eLife. 2018;7:e31949. doi: 10.7554/eLife.31949. [DOI] [PMC free article] [PubMed] [Google Scholar]