Skip to main content

This is a preprint.

It has not yet been peer reviewed by a journal.

The National Library of Medicine is running a pilot to include preprints that result from research funded by NIH in PMC and PubMed.

bioRxiv logoLink to bioRxiv
[Preprint]. 2023 Aug 21:2023.08.17.553745. [Version 2] doi: 10.1101/2023.08.17.553745

Dorsal raphe neurons signal integrated value during multi-attribute decision-making

Yang-Yang Feng 1,2, Ethan S Bromberg-Martin 1, Ilya E Monosov 1,2,3,4,5,*
PMCID: PMC10473596  PMID: 37662243

Abstract

The dorsal raphe nucleus (DRN) is implicated in psychiatric disorders that feature impaired sensitivity to reward amount, impulsivity when facing reward delays, and risk-seeking when grappling with reward uncertainty. However, whether and how DRN neurons signal reward amount, reward delay, and reward uncertainty during multi-attribute value-based decision-making, where subjects consider all these attributes to make a choice, is unclear. We recorded DRN neurons as monkeys chose between offers whose attributes, namely expected reward amount, reward delay, and reward uncertainty, varied independently. Many DRN neurons signaled offer attributes. Remarkably, these neurons commonly integrated offer attributes in a manner that reflected monkeys’ overall preferences for amount, delay, and uncertainty. After decision-making, in response to post-decision feedback, these same neurons signaled signed reward prediction errors, suggesting a broader role in tracking value across task epochs and behavioral contexts. Our data illustrate how DRN participates in integrated value computations, guiding theories of DRN in decision-making and psychiatric disease.

Introduction

The dorsal raphe nucleus (DRN), the main source of serotonin to the forebrain1, is strongly implicated in various psychiatric disorders associated with deficits in value-based decision-making211. Many of these deficits, including impaired sensitivity to reward amount12,13, impulsivity when facing reward delays1423, and risk-seeking when grappling with reward uncertainty12,2426 can be experimentally recapitulated by manipulating

serotonergic transmission or DRN activity. However, while DRN neurons have been shown to be sensitive to a diverse range of behavioral paradigms and task variables in rodents10,11,2732, much less is known about DRN’s functional roles in primates3336. Further, little is known about DRN’s role in complex cognitive tasks like multi-attribute value-based decision-making, where subjects must integrate reward, delay, and uncertainty to judge an offer’s overall subjective value and make a choice.

Despite this gap in knowledge, there are several theories about the functional role of DRN that have key implications for clinical intervention. One important set of theories suggests that DRN neurons broadcast ‘beneficialness’11,37 or ‘the availability of time and resources’10, a signal that integrates multiple distinct features or attributes to reflect the overall subjective value of offers or motivational outcomes and guide the neural computations that underlie emotion, mood, and value-guided behaviors, including multi-attribute decision-making. If so, DRN neurons should signal the subjective value of multi-attribute offers, for example, by integrating over reward amount, delay, and uncertainty. Another important set of theories suggests that DRN is involved in continuously tracking value-related statistics on multiple timescales, in the service of motivating behavior6,27,29,31,33,35,38. However, much is still unknown about whether and how DRN neurons signal such motivational variables, including surprise28,39, signed prediction errors2, or ‘global reward state’ (long-term average reward)2,18,29,40,41, particularly in the context of primate decision-making.

To test how DRN neurons process reward, delay, and uncertainty during decision-making, we recorded single DRN neurons from two monkeys as they chose between offers whose attributes, namely expected reward amount, reward delay, and reward uncertainty, varied independently. We found that many DRN neurons signaled offer attributes. Remarkably, these neurons commonly integrated offer attributes in a manner that reflected monkeys’ overall preferences for amount, delay, and uncertainty. After decision-making, in response to post-decision feedback, these same neurons signaled signed reward prediction errors, suggesting a broader role in tracking value across task epochs and behavioral contexts. Our data illustrate how DRN participates in integrated value computations, guiding theories of DRN in decision-making and psychiatric disease.

Results

Monkeys integrate reward amount, delay, and uncertainty to make multi-attribute decisions

We recorded single DRN neurons from two monkeys as they chose between two multi-attribute offers that each varied independently in expected reward amount (E[R]), reward delay (T[R]), and reward uncertainty (Unc[R]) (Figure 1A). Each offer, if chosen, provided a juice reward that was randomly drawn from a set of four possible outcomes, depicted as four bars of juice (Figure 1B, left). Thus, different offers provided different probability distributions of rewards, which could have different levels of E[R] (mean height of the bars) and Unc[R] (variability of the height of the bars, operationalized here as standard deviation). T[R], the delay between the time of choice and time of reward delivery, also varied across offers and was indicated by a ‘visual clock’ that provided explicit, online timing of progress towards reward delivery (Figure 1B, right).

Figure 1. Monkeys exhibit preferences for expected reward, delay, and uncertainty during multi-attribute decision-making.

Figure 1.

(A) Two offers (Offer 1 and Offer 2) are presented sequentially and then the monkey has 5 seconds to free view and make a choice by fixating one offer for 0.5 seconds. After a delay, the outcome is revealed and the reward is delivered. (B) Left: Vertical bars represent the probability distribution over rewards, indicating the expected reward amount (E[R]) and reward uncertainty (Unc[R]). Right: Horizontal bar acts as a clock, representing reward delay (T[R]). To allow monkeys to prepare to consume the upcoming reward, when the reward delay has elapsed, the stimulus changes to reveal the outcome, and then the reward is delivered 1.1 s later (Methods). The total time between choice and end of the trial was always 8s. (C) Fitted weights describing each monkey’s subjective preferences for E[R], T[R], and Unc[R]. Error bars denote s.e.m. Both monkeys placed significant weights on each reward attribute (** p<0.01, *** p<0.001).

To quantify the effect of reward attributes on monkeys’ decision-making, we employed a standard approach of fitting their choice behavior using a generalized linear model (GLM) where the log-odds of choosing between the two offers is equal to the difference in their overall subjective values, and where the overall subjective value of each offer is equal to a linear combination of offer attributes multiplied by their respective weights on value (Methods). Examining the resulting weights revealed that both monkeys understood the task and its offers. Consistent with previous work4244, monkeys chose offers more when they had larger E[R] and shorter T[R] (Figure 1C). The monkeys also had significant preferences for Unc[R], though in different manners, again consistent with individual variability in risk preferences seen in our42,45 and others’46,47 previous work.

Thus, monkeys demonstrated preferences for expected reward amount, reward delay, and reward uncertainty as they traded off these attributes to make multi-attribute decisions. We next asked whether and how DRN neurons signal these attributes during multi-attribute decision-making.

DRN neurons signal reward amount, delay, and uncertainty during multi-attribute decision-making

We recorded 219 DRN neurons as the two monkeys performed the task (n=117 in Monkey Z, and n=102, in Monkey P). The recording locations were verified using MRI imaging of the recording electrode within DRN (Figure 2A) and histology (Figure 2B). We reconstructed the neurons’ positions within DRN using previously established methods (Figure 2C). We also verified that our recordings took place in a subregion of DRN that is dense in serotonergic cell bodies4850, by examining our MRI- and histology-based reconstruction along with macaque DRN tissue that was stained for serotonin transporter (SERT) using previously established methods51 (Figure 2D).

Figure 2. Recording single neurons in primate dorsal raphe nucleus.

Figure 2.

(A) MRI taken with electrode in recording location, verifying its position in DRN. Shown are coronal views slightly tilted to align with the electrode track. The electrode is visible as a black ‘shadow’ on the MRI. (B) Nissl staining showing electrode track and marking lesions (Red arrows), delineating approximate range of recordings. The trochlear nucleus is visible just ventral to the recording track. (C) Anatomical reconstruction of recorded neurons, based on previously published methods238. Empty circles represent recorded neurons that were not task-responsive. Filled circles are task-responsive neurons. Top: Neurons from +3 to +4.5 interaural are shown on a plane corresponding to approximately +3.5 interaural. Bottom: Neurons from +2 to +2.5 interaural are shown on a plane corresponding to approximately +2.5 interaural. (D) SERT staining of tissue slices cut along the transverse axis of the brainstem, approximately corresponding to reconstructions shown in (C). Aq, cerebral aqueduct; MLF, medial longitudinal fasciculus; PAG, periaqueductal gray.

An analysis of variance across all epochs in our task, including trial start, choice, and post-decision epochs (Methods) revealed that 107/219 DRN neurons displayed task-related modulation. Neurons’ intrinsic electrophysiological properties and anatomical positions were not related to whether they were task-responsive (Figure 2C, Supplementary Figure 1, Supplementary Tables 1,2). We therefore focused our analyses on the 107 task-responsive DRN neurons.

We inspected the offer responses of DRN neurons and found that neurons commonly scaled their activity based on E[R], T[R], and Unc[R]. An example neuron’s responses to these offer attributes are shown in Figure 3A. Following offer presentation, this neuron displayed relatively phasic activations that scaled with E[R] (left), T[R] (middle), and Unc[R] (right). The mean cross-validated normalized activity (Methods) across task-responsive neurons demonstrated similar phasic dynamics that robustly discriminated offers based on their E[R], T[R], and Unc[R] (Figure 3B). To identify attribute-signaling neurons, we employed a standard approach of fitting neurons’ offer responses using a GLM where normalized firing rate is modeled as a linear weighted combination of offer attributes (Methods). Thus, a significant fitted weight for an attribute indicates that a neuron’s signals were related to that attribute. This approach identified significant proportions of task-responsive DRN neurons that signaled E[R], T[R], and Unc[R] (E[R]: 24/107, p=5.1E-10; T[R]: 22/107, p=1.5E-8; Unc[R]: 11/107, p=0.018; one-tailed binomial tests) (Figure 3C). Thus, all three of these offer attributes were signaled by DRN neurons.

Figure 3. DRN neurons signal expected reward, delay, and uncertainty during multi-attribute decision-making.

Figure 3.

(A) Rasters and SDFs showing example neurons’ responses to Offer 1 on trials where it had Large vs Small E[R] (left, red), Short vs Long T[R] (center, blue), or Uncertain vs Certain Unc[R] (right, green). (B) Mean cross-validated normalized activity for task-responsive DRN neurons on trials where Offer 1 E[R] (left, red), T[R] (center, blue), or Unc[R] (right, green) was preferred or non-preferred by the neuron. (C) Difference in the activity shown in B. Error bars denote s.e.m. Average discrimination during the offer response analysis window (150–575 ms after Offer 1 presentation, black bar) was significant for E[R], T[R], and Unc[R] signals (two-tailed signed-rank test, *** p<0.001). (D) Proportion of task-responsive DRN neurons with significant fitted weights for each attribute, fitting normalized activity in response to Offer 1 (one-tailed binomial test, ** p<0.01, *** p<0.001). Error bars indicate 68% confidence interval (Clopper–Pearson method).

DRN neurons integrate reward amount, delay, and uncertainty during multi-attribute decision-making

If DRN neurons signal the overall value of each offer, then they should signal attributes in accordance with each monkey’s subjective preferences, integrating these attributes by signaling their values jointly in the same direction as these preferences.

We therefore asked whether and how DRN neurons’ E[R], T[R], and Unc[R] signals reflected the subjective value monkeys assigned to each attribute based on their preferences. To test this, we quantified each task-responsive neuron’s signal for each attribute using an attribute value index, with indices above 0 indicating higher firing in response to offers with higher value based on that attribute (Methods). Strikingly, we found that all three attributes were predominantly signaled in the direction of subjective value (Figure 4). E[R]-signaling,T[R]-signaling, and Unc[R]-signaling neurons had significantly positive E[R], T[R], and Unc[R] value indices, respectively (E[R]: 0.23, p=8.1E-5; T[R]: 0.14, p=0.00098; Unc[R]: 0.16, p=0.0098; two-tailed signed-rank tests). Even considering all task-responsive neurons, the mean E[R] and T[R] value indices were significantly positive, and Unc[R] had a similar trend (E[R]: 0.068, p=6.4E-7; T[R]: 0.042, p=0.0047; Unc[R]: 0.019, p=0.10; two-tailed signed-rank tests). Moreover, for each attribute, there were significant proportions of neurons with significantly positive attribute value indices (E[R]: n=22/107, p=2.8E-14; T[R]: n=17/107, p=1.6E-9; Unc[R]: n=9/107, p=0.0015; two-tailed binomial tests), but not with significantly negative indices (E[R]: n=2/107, p=1; T[R]: n=5/107, p=0.20; Unc[R]: n=2/107, p=1; two-tailed binomial tests). Thus, task-responsive DRN neurons tended to signal attributes in accordance with monkeys’ preferences.

Figure 4. DRN neurons positively signal the value of expected reward, delay, and uncertainty.

Figure 4.

Histograms showing E[R] (left, red), T[R] (center, blue), Unc[R] (right, green) value indices (see Methods) of task-responsive neurons. Signals are biased in the direction of monkey preferences (dashed lines show mean; median significantly different from 0 by signed-rank test). The proportions of neurons with significant signaling of each attribute in the direction of monkey preferences are significant, while the proportions signaling the opposite are not (two-tailed binomial test). One outlier with an extreme E[R] value index was omitted from the left plot for presentation purposes (E[R] value index = 1.0, sig.). Colors indicate that neither (black), the x-coordinate (red), the y-coordinate (blue), or both (magenta) are significant (p < 0.05).

Next, we asked whether these attribute value signals are integrated. Specifically, we wondered whether DRN neurons’ E[R] signals, which convey information about an offer’s value based on its E[R], also conveyed information about an offer’s value based on its T[R] and Unc[R]. Consistent with integrated signaling of attributes, E[R] value indices were strongly correlated with T[R] and Unc[R] value indices, across task-responsive neurons (Figure 5A). Moreover, E[R]-signaling neurons strongly signaled T[R] and Unc[R] in the same direction as E[R] and with phasic dynamics that resembled how they signaled E[R] (Figure 5B). We also observed more neurons than expected by chance that had significant E[R] and T[R] value indices, or E[R] and Unc[R] value indices, in the same direction (E[R] and T[R]: n=10/107, p=3.3E-16; E[R] and Unc[R]: n=6/107, p=6.2E-09; two-tailed binomial tests, null proportion = 0.05*0.05/2). Thus, DRN neurons tended to integrate attributes, signaling them jointly and in directions consistent with monkeys’ preferences.

Figure 5. DRN expected reward signals integrate the value of delay or uncertainty.

Figure 5.

(A) Scatter plots showing E[R] vs T[R] (left) or E[R] vs Unc[R] (right) value signals across task-responsive neurons. E[R] signals are strongly positively correlated with T[R] and Unc[R] (Pearson and Spearman correlations, two-tailed permutation tests). Gray lines show Type 2 regression. One outlier with an extreme E[R] value index was omitted from the scatters for presentation purposes (E[R] value index = 1.0, sig.; T[R] value index = 0.14, n.s.; Unc[R] value index = 0.30, sig.). Colors indicate that neither (black), the x-coordinate (red), the y-coordinate (blue), or both (magenta) are significant (p < 0.05). (B) Top: Average normalized activity for E[R]-signaling neurons on trials where Offer 1 T[R] (left) or Unc[R] (right) was valued in the direction of the neuron’s E[R] value signal (Higher) or the opposite (Lower) (see Methods). Bottom: Difference in the activity shown in B. Error bars denote s.e.m. Average discrimination during the offer response analysis window (150–575 ms after Offer 1 presentation, black bar) was significant (two-tailed signed-rank test, ** p<0.001, *** p<0.001).

DRN neurons integrate cognitive and physical rewards

So far, we have demonstrated that DRN neurons can signal the value of offers, integrating the expected amount, delay, and uncertainty of juice, a physical reward. These data so far are consistent with the notion that DRN neurons can signal the overall value of an offer. However, if DRN neurons signal overall value, integrating all motivational variables, then they should also signal the value of non-physical or cognitive rewards.

To test this, we took advantage of recent findings showing that monkeys, like humans, often place high value on obtaining advance information to resolve uncertainty about upcoming reward outcomes42,45,5256. Importantly, they value this information even when they cannot use it to influence the physical reward outcome, suggesting that they value resolving uncertainty even when this has no objective, instrumental value for obtaining physical rewards. To measure DRN neurons’ responses to this kind of uncertainty-resolving information, we included two types of offers: informative (‘Info’) and non-informative (‘Noinfo’) (Figure 6A). After monkeys chose an Info offer, a cue would appear early in advance of reward delivery that highlighted the outcome bar, thus resolving reward uncertainty (Figure 6A; Info, ‘Cue’). Conversely, after monkeys chose a Noinfo offer, a visually matched control cue would appear that indicated a random bar, thus leaving reward uncertainty unresolved until just before reward delivery, when the upcoming outcome was revealed (Figure 6A; Noinfo, ‘Reveal’). Consistent with previous work showing that the value of information scales strongly with reward uncertainty42,45,57,58, both monkeys significantly scaled up the value of information with increasing Unc[R] (Figure 6B).

Figure 6. DRN expected reward signals integrate the value of information to reduce uncertainty.

Figure 6.

The task also measured the value of information to reduce uncertainty about upcoming rewards. Trials could either be informative (Info, orange), for which reward uncertainty would be resolved soon after choice, or non-informative (Noinfo, purple), for which reward uncertainty would not be resolved until immediately before reward delivery (see Methods). One outlier with an extreme E[R] value index was omitted from the scatter for presentation purposes (E[R] value index = 1.0, sig.; InfoxUnc[R] value index = 0.10, n.s.). (B) Bar plot showing fitted InfoxUnc[R] interaction weight (see Methods). Error bars denote s.e.m. Monkeys strongly scaled the value of information with uncertainty. (C) Same as Figure 5A, but for E[R] vs InfoxUnc[R] value signals. E[R] and InfoxUnc[R] are strongly positively correlated (Pearson and Spearman correlations, two-tailed permutation tests). Gray line shows Type 2 regression. Colors indicate that neither (black), the x-coordinate (red), the y-coordinate (blue), or both (magenta) are significant (p < 0.05). (D) Same as Figure 5B but showing differences in normalized activity between Info and Noinfo, splitting trials based on Offer 1 Unc[R].

Having replicated previous findings that monkeys integrate the values of juice reward and uncertainty-resolving information to make their decisions, we next asked whether DRN neurons integrate the values of these cognitive and physical rewards in a similar manner. To quantify neurons’ signals for the value of information to reduce uncertainty, we computed an InfoxUnc[R] value index based on the fitted interaction of Info and Unc[R], analogous to the attribute value indices for physical reward (see Methods). Consistent with integration of cognitive and physical rewards, InfoxUnc[R] and E[R] value indices were correlated across task-responsive neurons (Figure 6C). Moreover, on average, E[R]-signaling neurons signaled InfoxUnc[R] in the same direction as E[R] and with phasic dynamics that resembled how they signaled E[R] (Figure 6D). Thus, DRN neurons can integrate the values of cognitive and physical rewards.

Finally, to estimate how strongly neurons’ offer responses code an offer’s overall value, integrating across its many attributes, we computed a value coding index for each neuron using a recently-employed approach42 (see Methods). As suggested by the strong pairwise patterns of integration detailed above, we saw that DRN neurons tended to have very high value coding indices, often very close to the maximum index, consistent with coding of an offer’s overall value, integrating across its many attributes (Supplementary Figure 2).

DRN neurons track changes in value after decision-making by signaling reward prediction errors

Finally, DRN is thought to continuously track reward states on multiple timescales, potentially containing neurons sensitive to reward predictions or changes in reward predictions2,18,2729,31,33,35,38,40,41. This raises the possibility that the DRN neurons we found to signal value predictions based on offers may continue to track changes in value predictions based on post-decision feedback. We thus asked whether and how DRN neurons process post-decision feedback.

We first characterized DRN neurons’ responses to post-decision feedback, which was provided by the first post-decision event on each trial to indicate the amount of reward to be delivered (Cue on Info trials; Reveal on Noinfo trials; Figure 6A). To do this, we took advantage of several features of our task. Our task precisely manipulated reward predictions and outcomes by offering a diverse set of explicitly communicated reward distributions to the monkeys. This let us precisely test how responses to post-decision feedback were related to (a) the animal’s initial prediction about the offer’s reward value (i.e. E[R] of the chosen offer), before viewing the feedback, (b) the animal’s updated prediction of the offer’s reward value, after viewing the feedback and finding out the actual outcome to be delivered, and (c) the difference between the updated and initial predictions, known as reward prediction error (RPE), which is thought to guide learning and motivation5962. In addition, our task let us test how consistently neurons encode these variables. Specifically, we can test whether neurons encode these variables consistently for feedback with different sensory features and timing6366 (Cue on Info trials vs Reveal on Noinfo trials), consistently across their full range of values (for example, do neurons respond when the actual outcome is better than predicted, worse than predicted, or both)67,68, and in a signed or unsigned manner28,66,69.

We found that many DRN feedback responses strongly resembled a positively signed RPE signal. For example, consider the neuron in Figure 7A. This neuron was excited when feedback indicated that the reward outcome would be better than predicted (positive RPE), inhibited when feedback indicated that the reward outcome would be worse than predicted (negative RPE), and had little change in activity when feedback indicated that the reward outcome would be the same as predicted (zero RPE). The neuron did this consistently across both task events that provided post-decision feedback, the Cue for Info offers and the Reveal for Noinfo offers (Figure 7A).

Figure 7. DRN neurons that signal expected reward during decision-making signal reward prediction errors after decision-making.

Figure 7.

(A) Rasters and SDFs showing an example neuron’s responses to Info Cue (left, orange) or Noinfo Reveal (right, purple) on trials where it elicited an RPE that was positive (RPE +, solid, dark), 0 (RPE 0, solid, light) or negative (RPE −, dashed, dark). (B) Same as A but showing average normalized firing across E[R]-signaling neurons. Info and Noinfo RPE indices (see Methods) were significant (two-tailed signed-rank test, *** p<0.001). Activity was analyzed in the analysis windows shown in the black bars (250 to 750 ms post-cue, Info; 250 to 1000 ms post-reveal, Noinfo). (C) Histogram showing RPE indices (see Methods) of task-responsive neurons. Signals are predominantly positive (dashed lines show mean; median significantly different from 0 by signed-rank test). The proportion of neurons with significant positive RPE indices is large and strongly significant, while the opposite is small and weakly significant (two-tailed binomial tests). (D) Scatter plot showing RPE indices vs E[R] value signals. RPE indices and E[R] value signals were strongly positively correlated (Pearson and Spearman correlations, two-tailed permutation tests). Gray line shows Type 2 regression. One outlier with an extreme E[R] value index was omitted from the scatter for presentation purposes (E[R] value index = 1.0, sig.; RPE index = 0.84, sig.) Colors indicate that neither (black), the x-coordinate (red), the y-coordinate (blue), or both (magenta) are significant (p < 0.05).

To test whether DRN feedback responses had the key properties needed to function as a positively signed RPE signal, we quantified how each neuron’s responses discriminated between positive, zero, and negative RPEs (based on ROC analysis; Methods). First, we found that DRN feedback responses consistently and bidirectionally signaled both positive and negative RPEs. That is, they were more active when reward feedback was better than predicted and were less active when reward feedback was worse than predicted. To test this, we computed separate RPE indices for how neurons discriminated positive RPE from zero RPE (+RPE index) and zero RPE from negative RPE (−RPE index; Methods). As expected for a positively signed RPE signal, these two indices were strongly positively correlated (Supplementary Figure 3A). Second, we found that DRN feedback responses consistently signaled RPEs for both distinct feedback-related events in the task. To test this, we computed separate RPE indices for responses to the Info offer Cue (Info RPE index) and the Noinfo offer Reveal (Noinfo RPE index; Methods). As expected for a positively signed RPE signal, these two indices were strongly positively correlated (Supplementary Figure 3B). Third, we found that DRN feedback responses commonly responded in opposite directions to the initial reward prediction (the offer’s expected value) and the updated reward prediction (the actual reward outcome that would be delivered), consistent with signaling the difference between them. To test this, we fit each neuron’s activity using a GLM with separate terms for predicted and delivered reward. As expected for a positively signed RPE signal, neurons predominantly signaled delivered amounts positively and predicted amounts negatively (Supplementary Figure 4A). In sum, this activity profile is consistent with a positively signed RPE signal.

For each neuron, we summarized its RPE signal using a single RPE index computed as the mean of its + and - RPE indices for Info and Noinfo trials. The mean RPE index across all task-responsive neurons was strongly and significantly positive (+0.57, p<1E-16; two-tailed signed-rank test), consistent with a positively signed RPE signal like that of the example neuron. A large proportion of neurons had significant RPE indices (n=69/107, 64%), which were overwhelmingly positive. The proportion of neurons with significantly positive RPE indices was much greater than chance (n=62/107, p<1E-16; two-tailed binomial test). Interestingly, there was also a significant, albeit small, proportion of neurons with significantly negative RPE indices (n=7/107, p=0.018; two-tailed binomial test), in line with prior observations of some DRN neurons with negative reward discrimination27,3436,70. There was also a subpopulation of neurons with feedback responses that were less influenced by reward expectation (Supplementary Figure 4A-C), again in line with prior reports29,33,34.

We then examined the relationship between offer value signals during decision-making and RPE signals after decision-making. E[R] value indices and RPE indices were strongly correlated across task-responsive neurons (Figure 7D). Similarly, signed value coding indices and RPE indices were also strongly correlated across strongly attribute-responsive neurons (Supplementary Figure 5). Moreover, E[R]-signaling neurons signaled RPE in the same direction as Offer E[R] and with similar phasic dynamics (Figure 7B), and we observed more neurons than expected by chance that had significant RPE and E[R] value indices in the same direction (n=18/107, p<1E-16; two-tailed binomial test). E[R]-signaling neurons were also significantly more likely to respond to post-decision feedback by both positively signaling delivered amounts and negatively signaling predicted amounts, as expected for a positively signed RPE signal (Supplementary Figure 4C,D).

In summary, we find that DRN neurons that signal the value of offers during decision-making also commonly signal positively signed, bidirectional RPEs in response to feedback after decision-making. Thus, DRN signaling of offer attributes during decision-making may reflect a more generalized role in tracking changes in value predictions across task epochs and behavioral contexts.

Discussion

We found that DRN neurons signal expected reward amount, reward delay, and reward uncertainty during decision-making, in a manner reflecting the integrated value that guides monkeys’ choices. Additionally, we demonstrate that DRN neurons signal RPEs in response to post-decision feedback, suggesting a broader role in tracking value across task epochs and contexts.

Value is a complex construct thought to reflect preferences over many distinct attributes and to regulate value-based decision-making and other value-guided behaviors7173. Our results provide the first evidence that DRN reflects the integrated value of offers during decision-making based on multiple attributes of juice reward, including expected amount, delay, and uncertainty, as well as the value of uncertainty-resolving information. There are many other factors that influence value and which other brain areas have been shown to process during value-based decision-making43,7487, such as effort or punishment. While we did not assess all of these factors, our findings that DRN signals integrated value based on multiple attributes of physical and cognitive rewards, in a systematic manner across the neural population, raise the possibility that they can integrate an even wider range of the many motivational variables that govern decisions.

We found that DRN neurons can track changes in value predictions by signaling positively signed, bidirectional RPEs. This was somewhat unexpected because some previous results reported that DRN neurons may primarily signal the value of outcomes themselves29,33,34, or respond positively to both positive and negative RPEs, as if signaling surprise28. However, other results have been consistent with positively signed RPE, with DRN neurons responding more strongly to unexpected vs expected rewards29,36 and cues predicting more vs less reward28,29,34,36. It is possible that different task conditions and behavioral strategies engender different response patterns in DRN neurons. In addition, differences in recording locations could target DRN subregions with distinct functions88,89. Indeed, recent studies suggest that axons originating from more caudal subregions of DRN are predominantly inhibited by rewards90, whereas neurons in more rostral subregions, potentially more analogous to our recording locations, are predominantly excited91. There is likely also heterogeneity within subregions of DRN; for example, we also observed a subset of neurons whose responses to post-decision feedback primarily reflected the reward amount to be delivered, with relatively weak modulation by the predicted amount (Supplementary Figure 4A-C). Finally, it is also possible that some aspects of DRN function differ across species, given the known differences in DRN nuclear organization and projection patterns between rodents and primates92.

DRN receives input from several areas that could support value and RPE computations, including regions of prefrontal cortex (PFC), lateral hypothalamus, ventral pallidum (VP), and the lateral habenula (LHb)9295. PFC areas involved in emotion and cognition project strongly to DRN, including the anterior cingulate cortex (ACC) and dorsolateral PFC (DLPFC)96,97. ACC is known to signal multiple motivational variables77,98102, and we recently showed that VP signals partially integrated offer attributes in the same task used here42. Also, both ACC and VP have been implicated in processing motivational predictions and outcomes99,103. DRN could utilize these inputs from ACC and VP in flexible ways depending on changes in internal needs such as hunger or thirst, potentially modulated by lateral hypothalamus input104, or differing task demands or contingencies, potentially signaled by DLPFC105. Among DRN inputs, LHb is notable because it is well-known to potently signal negatively signed RPEs106,107. Moreover, we recently showed that it also negatively signals integrated value in this same multi-attribute decision-making task paradigm42. LHb has reciprocal connections with DRN and an established role in controlling other neuromodulators108. It also receives input from many DRN-projecting areas, including the medial PFC, VP, and lateral hypothalamus109. Future studies should investigate the relationship between integrated value and RPE signals and the dynamics of their co-emergence in the DRN-LHb circuit.

DRN is the largest component of the serotonin (5-HT) system, sending projections throughout the brain110. In addition to its 5-HT neurons, DRN contains other cell types111,112, which project to cortical and subcortical targets and regulate DRN 5-HT neurons113115. While we recorded in areas enriched with 5-HT neurons, where some estimate >60% of neurons release 5-HT50, we cannot confirm cell types in our data and it is unlikely that we exclusively recorded 5-HT neurons. DRN also contains glutamate-113,116,117, GABA-113,118, and dopamine-releasing70,119121 subpopulations that are involved in motivational processing, potentially through projections to downstream areas or local influence over DRN 5-HT neurons115,118,122125. Future studies should examine cell-type heterogeneity in primates during decision-making and other complex behaviors. Nonetheless, dissecting how value is processed within DRN is crucial for understanding the complex role of the 5-HT system in coordinating value-guided computations throughout the brain.

An influential theory of 5-HT function posits an opponent role to dopamine (DA), with phasic modulations signaling punishment prediction errors2,126. In contrast, we find that many DRN neurons positively signal offer value and RPEs, which strikingly resembles the responses of midbrain DA neurons in the ventral tegmental area (VTA) and substantia nigra pars compacta (SNc)6062. This is broadly consistent with previous observations of DRN neurons that code reward-related variables positively – some reporting a mix of positive and negative coding neurons27,3436,70 and others reporting predominantly positive29,30,117,127,128 – and with recent findings that DRN may contribute to VTA RPE signals through monosynaptic projections onto DA neurons129. While these results do not rule out the possibility that certain subpopulations of neurons in subregions of DRN91 or other 5-HT nuclei130 may indeed play an opponent role to DA, they suggest a new, potentially cooperative relationship between the 5-HT and DA systems131.

The cortex receives relatively light DA input compared to striatum132134 and within the frontal lobe, DA projections follow an increasing rostro-caudal gradient, with motor cortex receiving the most DA input and PFC receiving less135139. Thus, other neuromodulators may also be required to supply value-related signals to PFC to support its many roles in value-related processes140. 5-HT projections, while widespread, display a pattern of regional distribution that differs from and may complement that of DA. 5-HT projections are more balanced between cortex and striatum132,133 and within the frontal lobe, 5-HT projects more strongly to PFC than DA and only lightly to motor areas136,137,141. Thus, the DRN 5-HT system could broadcast value-related information to regulate networks at least partially distinctly from DA. The functional significance of these observations requires further investigation.

Our findings that DRN signals integrated value and RPE are broadly in line with existing theories of DRN and 5-HT in signaling reward variables such as ‘beneficialness’11,37, ‘the availability of time and resources’10, and ‘global reward state’ (long-term average reward rate)18,40,41. While these value-related constructs likely cannot fully explain the functions of the entire DRN and 5-HT system, a DRN signal tracking integrated value could explain how the diverse set of DRN projections89,91,110,111,142 and complex repertoire of 5-HT receptors143148 coordinately regulate multiple computational and behavioral processes4,8,10,11,38,126,149153. Firstly, DRN and the 5-HT system are theorized to regulate value-based decision-making35,20. Signaling the integrated value of offers or opportunities as they appear in the environment could support dynamic, online decision-making, potentially through DRN projections to orbitofrontal cortex (OFC)73,87,154,155 or LHb42,156,157, whose value-related activity is causally involved in decision-making, or DLPFC, which has been implicated in decision processes more generally80,155,158160. Secondly, an early theory of 5-HT proposed a role in mediating behavioral inhibition161. While this is often considered in the context of avoiding punishments4,126,162, obtaining maximal rewards can also require patience, persistence, and deliberation10,163168, which an integrated value signal could help to regulate10,169. Indeed, optogenetic stimulation of DRN 5-HT neurons170,171 or their projections to OFC and a region of rodent medial PFC thought to be homologous to ACC172 promotes waiting for rewards173,whereas chemical inhibition of DRN 5-HT neurons174 or blockade of 5-HT1A receptors175 reduces waiting for rewards. Thirdly, the 5-HT system is theorized to regulate learning and cognitive flexibility28,150,176192. Integrated value signals could indicate that time and resources are available for safe and fruitful learning10,28,176,183,190,193 ,by promoting exploration and information-seeking through DRN projections to ACC and ventrolateral PFC99,194,195 or by accelerating the rate of learning through 5-HT-mediated neuroplasticity across the brain196. Additionally, DRN’s RPE signal could update value or policy estimates, directly through 5-HT-mediated long-term depression197,198 or indirectly by influencing DA-mediated long-term potentiation197,198 through DRN projections to VTA DA neurons129,199.

Major depressive disorder (MDD) is perhaps the psychiatric illness most often linked to DRN and 5-HT system dysfunction200,201, and drugs targeting 5-HT transmission remain the main pharmaceutical therapeutics202, including conventional drugs such as selective serotonin reuptake inhibitors (SSRIs) and emerging ones such as psilocybin, a serotonergic psychedelic203205. However, computational frameworks for understanding MDD are limited, in part due to its poorly defined neural correlates and complex, heterogenous clinical presentations206208. MDD is characterized by wide-ranging symptoms, mainly low mood and anhedonia, but also other deficits such as psychomotor slowing, sleep abnormalities, and impaired cognitive function209. An integrated value signal that modulates multiple computational and behavioral processes could also contribute to the wide-ranging symptomatology of MDD and link it to DRN and the 5-HT system. A signal tracking the environment’s changes in integrated value could modulate mood, which has been hypothesized to represent ‘environmental momentum’210213. Similarly, lower estimates of value could contribute to anhedonia, defined as loss of interest or pleasure from objects or activities one previously found rewarding207,214, and negativity biases215. In turn, mood and value could regulate various behavioral parameters, including action vigor207,216, sleep207,217, and decision-making206,207,218. Moreover, while theories of serotonergic antidepressant mechanisms have been proposed at the cellular and psychological levels, it is difficult to unify them without a computational understanding of DRN and the 5-HT system. The therapeutic efficacy of both SSRIs and psychedelics has been linked to their effects on the activity of DRN neurons, 5-HT receptor availability, and neuroplasticity219224, which may then manifest as adaptive psychological changes, such as reduced negativity biases and improved cognitive flexibility219,225229. The efficacy of these drugs in treating other psychiatric illnesses such as anxiety or obsessive-compulsive disorder may also depend on similar mechanisms230237. Thus, an integrated value signal that is processed within DRN could link these neurobiological and psychological effects transdiagnostically. In summary, in addition to guiding theories of the 5-HT system in adaptive behavior, our findings about the neural computations of DRN neurons will also inform our understanding of the neural and computational basis of psychiatric disease, potentially leading to improved therapeutic avenues.

Methods

Experimental model and subject details

Two adult male monkeys (Macaca mulatta; Monkey Z and P; ages: 7–10 years old) were used for experiments. All of these procedures conform to the Guide for the Care and Use of Laboratory Animals and were approved by the Institutional Animal Care and Use Committee at Washington University.

Data acquisition

A plastic head holder and plastic recording chamber were fixed to the skull under general anesthesia and sterile surgical conditions. The chambers were tilted laterally (28° in Monkey Z; −34° in Monkey P) and aimed at DRN. After the monkeys recovered from surgery, they participated in behavioral and neurophysiological experiments.

While the monkeys participated in the behavioral procedure, we recorded single neurons in DRN. Electrode trajectories were determined with a 1 mm-spacing grid system and with the aid of MR images (3T) obtained along the direction of the recording chamber. This MRI-based estimation of neuron recording locations was aided by custom-built software238. In addition, in order to further verify the location of recording sites, after a subset of experiments the electrode was temporarily fixed in place at the recording site and the electrode tip’s location in the target area was verified by MRI (Figure 2A).

Electrophysiological recordings were performed using multi-contact electrodes (Plexon V-probes, 32 or 64 channels, 50 μm spacing) inserted through a stainless-steel guide tube and advanced by an oil-driven micromanipulator (MO-97A, Narishige). Signal acquisition (including amplification and filtering) was performed using an OmniPlex 40 kHz recording system (Plexon). Spike sorting was performed offline using publicly available software (Kilosort2) to extract clusters from the recordings, followed by manual curation to identify sets of clusters that corresponded to single neurons, the spans of time when they were well isolated, and whether they were in the regions of interest. DRN was identified based on neurons’ electrophysiological characteristics, firing patterns, and locations relative to anatomical landmarks estimated from MRIs and recordings at multiple grid locations. Specifically, we relied on visuo-auditory responses in the superior colliculus, distinct electrophysiological characteristics in the periaqueductal gray, oculomotor responses in the trochlear nucleus, and axonal characteristics in the medial longitudinal fasciculus. Neuronal and behavioral analyses were conducted offline in Matlab (Mathworks, Natick, MA).

Eye position was obtained with an infrared video camera (Eyelink, SR Research). Behavioral events and visual stimuli were controlled by Matlab (Mathworks, Natick, MA) with Psychophysics Toolbox extensions. Juice, used as reward, was delivered with a solenoid delivery reward system (CRIST Instruments).

Behavioral task

Each trial began with the appearance of a white fixation point at the center of the screen which the animal was required to fixate with its gaze. The task proceeded once the animal maintained fixation continuously for 0.25 s and at least 0.5 s had passed since fixation point onset. If the animal did not fixate within 5 s of fixation point appearance, or broke fixation before 0.5 s had passed since fixation point onset, the trial was counted as an error, moved immediately to the inter-trial interval (ITI), and repeated until it was performed successfully. After the fixation requirement was successfully completed, the fixation point disappeared and there was no longer any gaze requirement to advance the task. Simultaneously with fixation point disappearance, the first offer appeared on the screen. After 1 s, the second offer appeared on the screen. After a further 0.5 s, the choice period began. An offer was counted as chosen once the animal held its gaze on it continuously during the choice period for a fixed duration (0.4 s in Monkey Z; 0.5 s in Monkey P). For Monkey Z, if the monkey did not choose an offer within 5 s of the start of the choice period, the computer randomly selected an offer and the trial proceeded as if the animal had chosen it. For Monkey P, if the monkey did not choose an offer within 5 s of the start of the choice period, the current trial would end and a repeat trial would begin after the normal ITI, on which the same offers would be shown. For both monkeys, failure to choose within 5s occurred rarely, and computer-chosen trials and repeat trials as well as error trials were excluded from analysis unless otherwise noted. The size of each offer stimulus was approximately 6 × 5 degrees of visual angle. The locations of the two offers on the screen were randomly selected on each trial from a set of three possible locations (center, 10 degrees left of center, 10 degrees right of center).

Each offer featured a reward distribution that was visually depicted with 4 bars, with the height of each bar proportional to the volume of its juice reward, up to a maximum size Rmax (1.2 mL) corresponding to the maximum bar height. If an offer was chosen, one of its 4 bars was selected uniformly at random as the outcome, and hence the juice volume indicated by that bar was delivered to the animal at the time of outcome delivery. Juice was delivered through a metal spout placed directly in the mouth, to ensure that anticipatory actions such as licking were not required to obtain the reward and did not influence the amount of reward that was received.

We presented offers with multiple types of reward distributions. In brief, when randomly generating each offer on each trial, the offer first had its expected reward selected uniformly at random from a pre-specified set of possible expected rewards. To reduce the number of trials with trivial decisions where one offer was much better than the other, Offer 2’s expected reward was constrained to be within a pre-specified range of Offer 1’s expected reward. Then, the offer had its reward distribution randomly drawn from one of the following types: Safe distributions with 100% probability of delivering a specific reward amount; 25/50/25 distributions with 25, 50, and 25% chances of delivering small, medium, and large reward amounts, respectively, where the medium amount was equal to the mean of the small and large amounts; and 50/50 distributions with 50% chances of delivering small or large amounts. Then, after the offer’s distribution type was selected, its reward range was randomly drawn from a pre-specified set of ranges (50% or 70% of Rmax).

Each offer also could be informative (Info) or non-informative (Noinfo), indicated by its bars having distinct visual textures (we used complex visual textures sourced from images of scenes; for clarity of presentation, these are depicted in the figures as simple colors, i.e. orange or purple). If an informative offer was chosen, a visual cue was later presented that highlighted the single bar that was selected to be the outcome on that trial. If a non-informative offer was chosen, the same visual cue was presented at the same time but highlighting a single random bar. Thus, choosing an informative offer gave early access to information about the trial’s upcoming reward outcome. For monkey P, to reduce choice complexity, informativeness was always matched between Offer 1 and 2.

After the choice was made, the unchosen offer disappeared, and then a sequence of timed events occurred. First, at time Tcue, a visual “cue” appeared on the chosen offer in the form of a black box highlighting one of its bars. Then, at time Treveal, a “reveal” occurred in which three of the bars disappeared and only one bar remained, and then a further 1.1 s the reward was delivered (thus, Treveal fully determined the reward delay, with T[R] = Treveal + 1.1 s). The remaining bar always indicated the trial’s outcome, and hence, always provided full information about the trial’s upcoming outcome. This was to ensure that both informative and non-informative offers did eventually provide complete information about the trial’s outcome before it was delivered,to ensure that animals always had adequate time to physically prepare to consume the reward. Finally, the offer disappeared 8 s after the choice was made, and then a 1.2 s ITI occurred before the next trial began.

Each offer’s event timing, including Tcue and Treveal, was visually depicted as a large “clock” in the form of a horizontal bar placed at the bottom of the offer. This has been described in full detail previously42. The horizontal extent of the clock represented the full-time duration between choice and the end of the trial (8 s). The clock was divided into three sequential segments, representing the segments of time (A) from the choice to the cue (Tcue), (B) from the cue to the reveal (Tadvance = Treveal - Tcue), and (C) from the reveal to the end of the trial (8 s - Treveal). The time of reward delivery was not explicitly indicated but was always exactly 1.1 s after the reveal. The three segments of the clock each had distinct visual textures, which were also distinct for informative vs. non-informative offers (again, for clarity of presentation, these are depicted in figures as simple colors). In addition, two thick black vertical lines indicated the time of the cue and the time of the reveal. Finally, to help animals learn how the clock corresponded to real world time, and to help them anticipate the time of each event during each trial, we animated the clock. Specifically, after the choice was made, a “hand of the clock” appeared at the left edge of the clock, in the form of a thick black rectangle whose position on the clock indicated the current time during the trial. The hand moved from left to right at a constant speed as the trial proceeded. When the hand touched the first black vertical line indicating Tcue, an animation played lasting 0.6 s in which the vertical line moved upward to the center of the cued bar and then expanded to become the cue. When the hand touched the second black vertical line indicating Treveal, an animation played lasting 0.6 s in which the vertical line moved upward to the center of the chosen offer and then became an expanding rectangular region with the background color of the screen, “erasing” three of the bars thus leaving only one bar remaining. Finally, when the hand touched the right edge of the clock, the trial ended.

We presented offers with different cue and reveal (and thus, reward delivery) timing. For monkey Z, reveal timing was drawn jointly with cue timing. Specifically, the pair of event times, (Tcue, Treveal), was drawn uniformly at random from the set of all possible such pairs meeting the following requirements: Tcue was between 0.4–5.8 s; Treveal was between 1.2–6.7 s; Tcue came at least 0.8 s before Treveal; Tcue and Treveal were both multiples of 0.1 s. For monkey P, Treveal was drawn uniformly at random from the set (2.0, 2.9, 5.6,6.5 s), while Tcue was fixed at 0.5 s to maximize the value of information42 and reduce choice complexity.

Data analysis

All statistical tests were two-tailed unless otherwise noted. All statistical significance corresponds to p < 0.05 or a 95% confidence interval excluding 0 unless otherwise noted. Neural firing rates in response to each offer were analyzed in a time window 150–575 ms after offer onset. Firing rates for each neuron were converted to normalized firing rates by z-scoring. Specifically, for each neuron, we computed a vector of firing rates representing its full activity throughout the task by binning its data from all times during all trials in non-overlapping 500 ms bins. We then normalized the neuron’s firing rates in our analyses by subtracting the mean of that vector and dividing by the standard deviation of that vector. For plotting purposes, activity was smoothed using a Gaussian kernel (σ = 50 ms).

Task-responsiveness

Significant task-responsiveness was defined as variance in neural activity across all task events. To determine whether recorded neurons had significant task event-related modulations, we computed P values by comparing activities across different trial or types within the time windows of: −800 to 0 ms from Cue or Reveal onset and 0 to 900 ms from Offer 1, Offer 2, Cue, Reveal, or Reward delivery onset (Kruskal–Wallis test;P < 0.05). The windows were chosen such that they included key neuronal modulations and were wide enough to avoid bias towards a particular response pattern (e.g., phasic versus tonic). For Post-Offer 1, Post-Offer 2, Pre-Cue, and Pre-Reveal windows, 1-way comparisons were made for E[R], Unc[R], T[R], and Info, split into quantiles with 5, 5, 4, or 2 (Info vs Noinfo) levels, respectively. For Post-Cue and Post-Reveal windows, the same 1-way comparisons were made, but with the addition of delivered reward amount minus chosen offer E[R], split into quintiles. For the Post-Reward delivery window, a 1-way comparison was made only for delivered reward amount, split into quintiles. The P values were then combined using Fisher’s combined probability test239,240. This nonparametric method has fewer assumptions than parametric methods.

Offer attribute models for behavior and neuronal activity

We fit each individual’s binary choice data using a generalized linear model (GLM) designed to model a standard decision-making formulation in which values are computed for each of the two offers, and then the resulting choice probability is a logistic function of the difference in the values:

log(p(Choose Offer2)/p(Choose Offer 1))=V(Offer 2)V(Offer 1)

And the value of each offer i is a linear weighted combination of the offer’s vector of n attributes <xi,1,xi,2,xi,3,...,xi,n>:

V(Offeri)=βpreference,1*xi,1+βpreference ,2*xi,2++βpreference,n*xi,n

Thus, the resulting model was a GLM for binomial data with a logistic link function, with the equation:

log(p(Choose Offer 2)/p(Choose Offer 1))=βpreference,0+βpreference,1(x2,1x1,1)+βpreference,2(x2,2x1,2)++βpreference,n(x2,nx1,n)

…where βpreference,0 is the subject’s choice bias.

For analysis of neuronal data, we used an analogous GLM to fit neuronal firing rates in response to Offer 1, for normal data with an identity link function (i.e. equivalent to ordinary linear regression):

Rate(Offer)=βsignal,0+βsignal,1*x1+βsignal ,2*x2++βsignal,n*xn+ε

…where β0 is the neuron’s mean or baseline response to the offer and ε is the error term.

There were also several aspects of the model fitting procedure that held across models. First, to permit more interpretable comparisons between attributes, unless otherwise noted, all attributes that were not binary were standardized by z-scoring. Second, in addition to the terms described below related to the values of the offers, all models also included terms for response biases, including the location in space of each offer and the sequential order in which the offers were presented. For both behavioral and neural data, this included binary attributes representing whether or not the offer was on the left side of the screen and whether or not the offer was on the right side of the screen.

For monkey Z’s data, there were many possible variables to consider, since the data included manipulations of both Tcue and Treveal, in addition to all of the other task variables described above. Therefore, we focused on a core set of attributes that most strongly governed the subjective value of offers across monkeys, adapted from the set derived using a model selection procedure in our previous study42: E[R], T[R], Unc[R], E[R] x T[R], E[R] x SD[r], Info, Info x E[R], Info x T[R], Info x Unc[r], Info x Tadvance, Tadvance x Unc[R], Info x Tadvance x Unc[R]. For monkey P’s data, the main effect of Info and interactions involving Tadvance were excluded, since Info was always matched between offers and Tadvance was fully determined by T[R]. We then defined a model with these attributes for our analysis of both monkey behavior and neuronal activity in this task.

For each neuron, we computed its attribute value indices simply as the fitted β weight for each attribute, signflipped based on the direction of the recorded monkey’s preference for each attribute such that positive attribute value signals indicate higher firing for offers with higher value based on that attribute.

Analysis of reward prediction error coding

To quantify each neuron’s coding of conventional juice reward prediction errors, we computed a reward prediction error index (RPE Index) using the following approach. First, we defined the RPE on each correctly performed trial as the difference between the delivered reward amount and expected reward amount (i.e. the chosen offer’s E[R]). Next, we selected the subsets of trials which fell into each of four categories, based on the 2×3 combinations of the chosen offer’s informativeness (Info or Noinfo) and the resulting RPE (+,0,−), corresponding to RPEs > +0.1 mL juice, < +0.1 mL juice and > −0.1 mL juice, or < −0.1 mL juice). We then analyzed single trial firing rates on these trials after two task events where RPEs commonly occurred in our task: a post-cue time window 250–750 ms after cue onset for Info trials, and a post-reveal time window 250– 1050 ms after reveal onset for Noinfo trials. We then used this activity to compute separate measures of the strength of the positive and negative components of the RPE responses separately for Info vs Noinfo trials. Specifically, we first computed 4 separate indices as ROC(X,Y), the ROC area under the receiver operating characteristic curve241 between the set of single trial normalized activities in condition X vs. condition Y, such that the ROC area is > 0.5 if condition Y generally has higher activity than condition X, and < 0.5 if condition Y generally has less activity than condition X, and 0.5 if the distributions are the same. Significance was assessed by Wilcoxon the rank-sum test. For each index, X and Y were defined as:

For Info RPE +,X and Y were Info trials where RPE was 0 or +, respectively.
For Info RPE−, X and Y were Info trials where RPE was − or 0, respectively.
For Noinfo RPE+, X and Y were Noinfo trials where RPE was 0 or +,respectively.
For Noinfo RPE−, X and Y were Noinfo trials where RPE was − or 0, respectively.

To assess the relationship between RPEs elicited by Info Cues vs Noinfo Reveals, we computed Info and Noinfo RPE indices by averaging Info RPE+ and Info RPE− or Noinfo RPE+ and Noinfo RPE− indices, respectively. P-values for these averaged indices were computed from the p-values of the two constituent measures using Fisher’s combination test239,240. Finally, the total RPE Index was computed as the mean of the Info and Noinfo RPE indices and its p-value was again computed from the p-values of these two constituent measures using Fisher’s combination test. Since RPE is defined as the difference between the delivered reward amount and the predicted reward amount (i.e. the chosen offer’s E[R]), RPEs are positively correlated with the delivered reward and negatively correlated with the predicted reward amount. In order to verify that apparent RPE signaling did not simply reflect signaling of either the delivered reward amount or predicted reward amount, we also computed delivered reward amount and predicted reward amount indices. Specifically, we fit each neuron’s normalized activity to two GLMs for normal data with an identity link function (i.e. equivalent to ordinary linear regression). The models were fit to activity in the post-cue window on Info trials or the post-reveal window on Noinfo trials. Specifically, the models were expressed as:

Info model: Activity(Cue)=β0+βdelivered,Info*Rdelivered+βpredicted,Info *Rpredicted+ε
Noinfo model: Activity(Reveal)=β0+βdelivered,Noinfo*Rdelivered +βpredicted,Noinfo *Rpredicted+ε

…where β0 is the neuron’s mean or baseline response to the Cue or Reveal, Rdelivered and Rpredicted are the delivered and predicted reward amounts, and ε is the error term.

The delivered reward amount and predicted reward amount indices were then computed by averaging the corresponding fitted β weights across the Info and Noinfo models, and p-values for these averaged indices were computed from the p-values of the two constituent measures using Fisher’s combination test:

Delivered reward index=Mean(βdelivered,Info, βdelivered,Noinfo)
Predicted reward index=Mean(βpredicted,Info, ,βpredicted,Noinfo)

Analysis of peri-stimulus activity

To analyze cross-validated timecourses of Offer 1 E[R], T[R], and Unc[R] responses across the population of task-responsive neurons (Figure 3B), we used the following procedure. For each neuron, its data were split into two cross-validation folds, 1 and 2, which included odd-numbered and even-numbered trials, respectively. The neuronal offer attribute model was fit to each fold’s data to obtain βsignal weights. For each attribute, we split trials into High vs Low based on Offer 1 (Larger vs Smaller E[R], median split; Shorter vs Longer T[R], median split; Uncertain vs Certain Unc[R]), separately for each fold. We computed the attribute’s effect on the Offer 1 response as the difference in mean normalized firing rate during the Offer 1 analysis window on High vs Low trials. Then, we computed the total cross-validated effect by taking the mean of the effects in the two folds, after multiplying each fold’s effect by the sign of the βsignal,attribute weight derived from the other fold’s effect. This provided a cross-validated measure of the total strength of the attribute effect in units of normalized activity, such that a more positive effect indicates a stronger interaction, regardless of whether the underlying effect on firing rate had a positive or negative sign. Specifically, if the neuron had a true effect that could be measured with a consistent sign in both folds, then its total effect would be positive (regardless of whether the sign of the underlying effect reflected a positive or negative influence on firing rate). Conversely, under the null hypothesis that the neuron had no true effect, the total effect would have an expected value of 0, since each fold would be equally likely to be multiplied by either a negative or positive sign. We then tested whether the median total cross-validated effect across the population of neurons was different from 0 (signed-rank test). We also plotted the full timecourse of the mean normalized activity for High vs Low conditions and total cross-validated effect by computing it using the full spike density functions from each cross-validation fold, signflipped as in the analysis above (so that the full timecourse of each neuron’s activity in each fold was normalized using a consistent sign).

To analyze time courses of Offer 1 E[R], T[R], Unc[R], and Info x Unc[R] responses across the population of E[R]-signaling neurons, we used the following procedure. We selected E[R]-signaling neurons based on significance of βsignal,E[R] from the neuronal offer attribute model. For E[R], T[R], and Unc[R] (Figure 5B), we again split trials into two categories based on Offer 1, but this time into preferred vs unpreferred, based on animals’ behavior. For E[R] and T[R], these were again median splits of Larger vs Smaller E[R] or Shorter vs Longer T[R], respectively. For Unc[R], these were Uncertain vs Certain Unc[R] for monkey Z and Certain vs Uncertain Unc[R] for monkey P. If a neuron had a negative βsignal,E[R], its preferred vs unpreferred trials were exchanged. These categories of trials then represented what a value-signaling neuron should signal as High vs Low value for each attribute. For Info x Unc[R] (Figure 6D), we first split trials into 2×2 categories (Info, Noinfo) x (Uncertain, Certain) based on Offer 1. For each neuron, we computed the difference in Offer 1 response on Info vs Noinfo trials, separately for Uncertain vs Certain trials to obtain Uncertain Info vs Certain Info effects, respectively. If a neuron had a negative βsignal,E[R], its Uncertain vs Certain effects were exchanged, to give what a value-signaling neuron should signal as High Info value vs Low Info value.

Value coding index

We used a value coding index we previously defined42 to quantify the extent to which the offer responses of each neuron aligned with our behaviorally-derived estimate of how animals weighted those attributes to compute the offer’s overall subjective value. To do this, we first fit each neuron’s Offer 1 responses using the attribute model described above, which allows the neuron to have separate weights for each of the selected attributes described above that influenced the animal’s choices. We then fit responses with a new model called the “value model”, which only had a single attribute, ‘Value’, equal to the estimated value of the offer derived from the behavioral model (using the procedure described above). Thus, the value model represents the hypothesis that the neuron’s attribute-related activity is fully explained by encoding of a single variable, the offer’s overall subjective value; or equivalently, that the neuron’s weighting of the offer’s many attributes to govern its response, is perfectly correlated with the animal’s weighting of the offer’s many attributes to compute subjective value. Finally, we quantified what fraction of the above-chance variance in the neuron’s activity that was explained by the attribute model could also be explained by the value model. To do this, we used a measure based on R-squared (R2) to quantify the fraction of the response variance that each model explained. Specifically, to correct for the fact that the attribute model had more parameters than the value model, we computed the shuffle-corrected R-squared (R2corr), defined as the R2 computed on the neuron’s real data minus the mean of the R2 computed on 100 shuffled datasets in which the neuronal offer responses were shuffled across trials. Thus, R2corr quantified the fraction of variance of neuron’s offer responses that the model explained, above and beyond the chance level expected under the null hypothesis that attributes and values had no influence on the neuron’s activity. Finally, we defined the value coding index with the equation:

 ValueCodingIndex=c((Rcorr 2 for value model )/(Rcorr 2 for attribute model ))

…where the function c(x) = max(0,min(1,x)) clamps the index between 0 and 1 (for rare cases where the index was slightly < 0 or > 1 such as due to small variations in the shuffling used for the shuffle correction). For some analyses, we used a signed value coding index, which was simply the value coding index multiplied by the sign of the fitted ‘Value’ weight in the value model. Importantly, because the value coding index is expressed as a ratio relative to the corrected variance explained by the attribute model, it is only meaningful and reliably estimated for neurons for which a meaningful fraction of the variance was explained by the attribute model. Therefore, we only computed this index for neurons that were strongly attribute-responsive, which we defined as neurons for which the R2corr for the attribute model was greater than or equal to 0.03 and were fit with a significant effect for at least one attribute using the attribute model.

Intrinsic electrophysiological properties

Spike width was calculated as the duration between the peak and the trough of the waveform. Baseline measures of firing rate and irregularity were computed based on activity in the last 1000 ms of the ITI. We used an irregularity metric (IR) that was previously used for DRN neurons33,34,242, described here in brief. First, interspike intervals (ISIs) were computed. If spike(i − 1), spike(i), and spike(i + 1) occurred in this order, the duration between spike(i − 1) and spike(i) corresponds to ISIi; the duration between spike(i) and spike(i + 1) corresponds to ISIi + 1. Second, the difference between adjacent ISIs was computed as |log(ISIi/ISIi + 1)|. Finally, we computed IR as the median of all such differences during the baseline window. Thus, small IR values indicate regular firing and large IR values indicate irregular firing. This measure has an advantage over traditional measures of irregularity, such as the coefficient of variation of the interspike intervals, which require a constant firing rate during the measurement period. However, for comparison, we also computed the coefficient of variation (CV) of ISIs.

Supplementary Material

1

Acknowledgements.

We thank Dr. Suzanne Haber and colleagues for performing serotonin transporter (SERT) staining on macaque tissue and providing us with the data. We thank Ms. Kim Kocher for excellent animal care and handling. We thank Dr. Larry Snyder for assistance with imaging. This work was supported by the National Institute of Mental Health under Award Numbers F30MH130103 to Y-YF and R01MH128344, R01MH110594, R01MH116937, and Conte Center on the Neurocircuitry of OCD MH10643 to IEM, by the Army Research Office 78259-NS-MUR to IEM, and by the McKnight Foundation award to IEM.

References

  • 1.Ishimura K., Takeuchi Y., Fujiwara K., Tominaga M., Yoshioka H., and Sawada T. (1988).Quantitative analysis of the distribution of serotonin-immunoreactive cell bodies in the mouse brain. Neuroscience Letters 91, 265–270. 10.1016/0304-3940(88)90691-X. [DOI] [PubMed] [Google Scholar]
  • 2.Daw N.D., Kakade S., and Dayan P. (2002). Opponent interactions between serotonin and dopamine.Neural Networks 15, 603–616. 10.1016/S0893-6080(02)00052-7. [DOI] [PubMed] [Google Scholar]
  • 3.Dayan P., and Huys Q.J.M. (2009). Serotonin in Affective Control. Annual Review of Neuroscience 32,95–126. 10.1146/annurev.neuro.051508.135607. [DOI] [PubMed] [Google Scholar]
  • 4.Cools R., Nakamura K., and Daw N.D. (2011). Serotonin and Dopamine: Unifying Affective, Activational, and Decision Functions. Neuropsychopharmacology 36, 98–113. 10.1038/npp.2010.121. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Rogers R.D. (2011). The Roles of Dopamine and Serotonin in Decision Making: Evidence from Pharmacological Experiments in Humans. Neuropsychopharmacology 36, 114–132. 10.1038/npp.2010.165. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Nakamura K. (2013). The role of the dorsal raphé nucleus in reward-seeking behavior. Frontiers in Integrative Neuroscience 7. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Cowen P.J., and Browning M. (2015). What has serotonin to do with depression? World Psychiatry 14, 158–160. 10.1002/wps.20229. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Wong-Lin K., Wang D.-H., Moustafa A.A., Cohen J.Y., and Nakamura K. (2017). Toward a multiscale modeling framework for understanding serotonergic function. Journal of Psychopharmacology 31, 1121–1136. 10.1177/0269881117699612. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Amlung M., Marsden E., Holshausen K., Morris V., Patel H., Vedelago L., Naish K.R., Reed D.D., and McCabe R.E. (2019). Delay Discounting as a Transdiagnostic Process in Psychiatric Disorders: A Meta-analysis. JAMA psychiatry 76, 1176–1186. 10.1001/jamapsychiatry.2019.2102. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Doya K., Miyazaki K.W., and Miyazaki K. (2021). Serotonergic modulation of cognitive computations. Current Opinion in Behavioral Sciences 38, 116–123. 10.1016/j.cobeha.2021.02.003. [DOI] [Google Scholar]
  • 11.Liu Z., Lin R., and Luo M. (2020). Reward Contributions to Serotonergic Functions. Annual Review of Neuroscience 43, 141–162. 10.1146/annurev-neuro-093019-112252. [DOI] [PubMed] [Google Scholar]
  • 12.Rogers R.D., Everitt B.J., Baldacchino A., Blackshaw A.J., Swainson R., Wynne K., Baker N.B., Hunter J., Carthy T., Booker E., et al. (1999). Dissociable Deficits in the Decision-Making Cognition of Chronic Amphetamine Abusers, Opiate Abusers, Patients with Focal Damage to Prefrontal Cortex, and Tryptophan-Depleted Normal Volunteers: Evidence for Monoaminergic Mechanisms. Neuropsychopharmacology 20, 322–339. 10.1016/S0893-133X(98)00091-8. [DOI] [PubMed] [Google Scholar]
  • 13.Seymour B., Daw N.D., Roiser J.P., Dayan P., and Dolan R. (2012). Serotonin Selectively Modulates Reward Value in Human Decision-Making. The Journal of Neuroscience 32, 5833–5842. 10.1523/JNEUROSCI.0053-12.2012. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Wogar M.A., Bradshaw C.M., and Szabadi E. (1993). Effect of lesions of the ascending 5-hydroxytryptaminergic pathways on choice between delayed reinforcers. Psychopharmacology 111,239–243. 10.1007/BF02245530. [DOI] [PubMed] [Google Scholar]
  • 15.Bizot J.-C., Le Bihan C., Puech A.J., Hamon M., and Thiébot M.-H. (1999). Serotonin and tolerance to delay of reward in rats. Psychopharmacology 146, 400–412. 10.1007/PL00005485. [DOI] [PubMed] [Google Scholar]
  • 16.Mobini S., Chiang T.-J., Al-Ruwaitea A.S.A., Ho M.-Y., Bradshaw C.M., and Szabadi E. (2000).Effect of central 5-hydroxytryptamine depletion on inter-temporal choice: a quantitative analysis. Psychopharmacology 149, 313–318. 10.1007/s002130000385. [DOI] [PubMed] [Google Scholar]
  • 17.Mobini S., Chiang T.-J., Ho M.-Y., Bradshaw C.M., and Szabadi E. (2000). Effects of central 5-hydroxytryptamine depletion on sensitivity to delayed and probabilistic reinforcement. Psychopharmacology 152, 390–397. 10.1007/s002130000542. [DOI] [PubMed] [Google Scholar]
  • 18.Denk F., Walton M.E., Jennings K.A., Sharp T., Rushworth M.F.S., and Bannerman D.M. (2005).Differential involvement of serotonin and dopamine systems in cost-benefit decisions about delay or effort. Psychopharmacology 179, 587–596. 10.1007/s00213-004-2059-4. [DOI] [PubMed] [Google Scholar]
  • 19.Winstanley C.A., Theobald D.E.H., Dalley J.W., and Robbins T.W. (2005). Interactions between Serotonin and Dopamine in the Control of Impulsive Choice in Rats: Therapeutic Implications for Impulse Control Disorders. Neuropsychopharmacology 30, 669–682. 10.1038/sj.npp.1300610. [DOI] [PubMed] [Google Scholar]
  • 20.Schweighofer N., Bertin M., Shishida K., Okamoto Y., Tanaka S.C., Yamawaki S., and Doya K. (2008). Low-Serotonin Levels Increase Delayed Reward Discounting in Humans. Journal of Neuroscience 28, 4528–4532. 10.1523/JNEUROSCI.4982-07.2008. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Crockett M.J., Clark L., Lieberman M.D., Tabibnia G., and Robbins T.W. (2010). Impulsive choice and altruistic punishment are correlated and increase in tandem with serotonin depletion. Emotion (Washington, D.C.) 10, 855–862. 10.1037/a0019861. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Xu S., Das G., Hueske E., and Tonegawa S. (2017). Dorsal Raphe Serotonergic Neurons Control Intertemporal Choice under Trade-off. Current Biology 27, 3111–3119.e3113. 10.1016/j.cub.2017.09.008. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Mori M., Tsutsui-Kimura I., Mimura M., and Tanaka K.F. (2018). 5-HT3 antagonists decrease discounting rate without affecting sensitivity to reward magnitude in the delay discounting task in mice.Psychopharmacology 235, 2619–2629. 10.1007/s00213-018-4954-0. [DOI] [PubMed] [Google Scholar]
  • 24.Long A.B., Kuhn C.M., and Platt M.L. (2009). Serotonin shapes risky decision making in monkeys.Social Cognitive and Affective Neuroscience 4, 346–356. 10.1093/scan/nsp020. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Murphy S.E., Longhitano C., Ayres R.E., Cowen P.J., Harmer C.J., and Rogers R.D. (2009). The Role of Serotonin in Nonnormative Risky Choice: The Effects of Tryptophan Supplements on the “Reflection Effect” in Healthy Adult Volunteers. Journal of Cognitive Neuroscience 21, 1709–1719. 10.1162/jocn.2009.21122. [DOI] [PubMed] [Google Scholar]
  • 26.Macoveanu J., Rowe J.B., Hornboll B., Elliott R., Paulson O.B., Knudsen G.M., and Siebner H.R. (2013). Serotonin 2A receptors contribute to the regulation of risk-averse decisions. NeuroImage 83, 35–44. 10.1016/j.neuroimage.2013.06.063. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Ranade S.P., and Mainen Z.F. (2009). Transient Firing of Dorsal Raphe Neurons Encodes Diverse and Specific Sensory, Motor, and Reward Events. Journal of Neurophysiology 102, 3026–3037. 10.1152/jn.00507.2009. [DOI] [PubMed] [Google Scholar]
  • 28.Matias S., Lottem E., Dugué G.P., and Mainen Z.F. (2017). Activity patterns of serotonin neurons underlying cognitive flexibility. Elife 6, e20552. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.Cohen J.Y., Amoroso M.W., and Uchida N. (2015). Serotonergic neurons signal reward and punishment on multiple timescales. Elife 4, e06346. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.Paquelet G.E., Carrion K., Lacefield C.O., Zhou P., Hen R., and Miller B.R. (2022). Single-cell activity and network properties of dorsal raphe nucleus serotonin neurons during emotionally salient behaviors. Neuron 110, 2664–2679.e2668. 10.1016/j.neuron.2022.05.015. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31.Grossman C.D., Bari B.A., and Cohen J.Y. (2021). Serotonin neurons modulate learning rate through uncertainty. Current Biology. 10.1016/j.cub.2021.12.006. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Waterhouse B.D., Devilbiss D., Seiple S., and Markowitz R. (2004). Sensorimotor-related discharge of simultaneously recorded, single neurons in the dorsal raphe nucleus of the awake, unrestrained rat. Brain Research 1000, 183–191. 10.1016/j.brainres.2003.11.030. [DOI] [PubMed] [Google Scholar]
  • 33.Bromberg-Martin E.S., Hikosaka O., and Nakamura K. (2010). Coding of task reward value in the dorsal raphe nucleus. Journal of Neuroscience 30, 6262–6272. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Nakamura K., Matsumoto M., and Hikosaka O. (2008). Reward-dependent modulation of neuronal activity in the primate dorsal raphe nucleus. Journal of Neuroscience 28, 5331–5343. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35.Inaba K., Mizuhiki T., Setogawa T., Toda K., Richmond B.J., and Shidara M. (2013). Neurons in Monkey Dorsal Raphe Nucleus Code Beginning and Progress of Step-by-Step Schedule, Reward Expectation, and Amount of Reward Outcome in the Reward Schedule Task. Journal of Neuroscience 33, 3477–3491. 10.1523/JNEUROSCI.4388-12.2013. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36.Hayashi K., Nakao K., and Nakamura K. (2015). Appetitive and Aversive Information Coding in the Primate Dorsal Raphé Nucleus. Journal of Neuroscience 35, 6195–6208. 10.1523/JNEUROSCI.2860-14.2015. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37.Luo M., Li Y., and Zhong W. (2016). Do dorsal raphe 5-HT neurons encode “beneficialness”? Neurobiology of Learning and Memory 135, 40–49. 10.1016/j.nlm.2016.08.008. [DOI] [PubMed] [Google Scholar]
  • 38.Cohen J.Y., and Grossman C.D. (2020). Chapter 30 - Dorsal raphe serotonergic neurons regulate behavior on multiple timescales. In Handbook of Behavioral Neuroscience, Müller C.P., and Cunningham K.A., eds. (Elsevier; ), pp. 521–529. [Google Scholar]
  • 39.Cazettes F., Reato D., Morais J.P., Renart A., and Mainen Z.F. (2021). Phasic Activation of Dorsal Raphe Serotonergic Neurons Increases Pupil Size. Current Biology 31, 192–197.e194. 10.1016/j.cub.2020.09.090. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40.Khalighinejad N., Manohar S., Husain M., and Rushworth M.F.S. (2022). Complementary roles of serotonergic and cholinergic systems in decisions about when to act. Current Biology 32, 1150–1162.e1157. 10.1016/j.cub.2022.01.042. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41.Wittmann M.K., Fouragnan E., Folloni D., Klein-Flügge M.C., Chau B.K.H., Khamassi M., and Rushworth M.F.S. (2020). Global reward state affects learning and activity in raphe nucleus and anterior insula in monkeys. Nature Communications 11, 3771. 10.1038/s41467-020-17343-w. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42.Bromberg-Martin E.S., Feng Y.-Y., Ogasawara T., White J.K., Zhang K., and Monosov I.E. (2022).A neural mechanism for conserved value computations integrating information and rewards. 10.1101/2022.08.14.503903. [DOI] [PMC free article] [PubMed]
  • 43.Hosokawa T., Kennerley S.W., Sloan J., and Wallis J.D. (2013). Single-Neuron Mechanisms Underlying Cost-Benefit Analysis in Frontal Cortex. Journal of Neuroscience 33, 17385–17397. 10.1523/JNEUROSCI.2221-13.2013. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 44.Stevens J.R. (2014). Evolutionary pressures on primate intertemporal choice. Proceedings of the Royal Society B: Biological Sciences 281, 20140499. 10.1098/rspb.2014.0499. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 45.Jezzini A., Bromberg-Martin E.S., Trambaiolli L.R., Haber S.N., and Monosov I.E. (2021). A prefrontal network integrates preferences for advance information about uncertain rewards and punishments. Neuron 109, 2339–2352.e2335. 10.1016/j.neuron.2021.05.013. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 46.De Petrillo F., and Rosati A.G. (2021). Variation in primate decision-making under uncertainty and the roots of human economic behaviour. Philosophical Transactions of the Royal Society B: Biological Sciences 376, 20190671. 10.1098/rstb.2019.0671. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 47.Nioche A., Rougier N.P., Deffains M., Bourgeois-Gironde S., Ballesta S., and Boraud T. (2021). The adaptive value of probability distortion and risk-seeking in macaques’ decision-making. Philosophical Transactions of the Royal Society B: Biological Sciences 376, 20190668. 10.1098/rstb.2019.0668. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 48.Felten D.L., and Sladek J.R. (1983). Monoamine distribution in primate brain V. Monoaminergic nuclei: Anatomy, pathways and local organization. Brain Research Bulletin 10, 171–284. 10.1016/0361-9230(83)90045-X. [DOI] [PubMed] [Google Scholar]
  • 49.Baker K.G., Halliday G.M., and Törk I. (1990). Cytoarchitecture of the human dorsal raphe nucleus.Journal of Comparative Neurology 301, 147–161. 10.1002/cne.903010202. [DOI] [PubMed] [Google Scholar]
  • 50.Baker K.G., Halliday G.M., Hornung J.-P., Geffen L.B., Cotton R.G.H., and Törk I. (1991).Distribution, morphology and number of monoamine-synthesizing and substance P-containing neurons in the human dorsal raphe nucleus. Neuroscience 42, 757–775. 10.1016/0306-4522(91)90043-N. [DOI] [PubMed] [Google Scholar]
  • 51.Heilbronner S.R., and Haber S.N. (2014). Frontal Cortical and Subcortical Projections Provide a Basis for Segmenting the Cingulum Bundle: Implications for Neuroimaging and Psychiatric Disorders. Journal of Neuroscience 34, 10041–10054. 10.1523/JNEUROSCI.5459-13.2014. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 52.Bromberg-Martin E.S., and Monosov I.E. (2020). Neural circuitry of information seeking. Current Opinion in Behavioral Sciences 35, 62–70. 10.1016/j.cobeha.2020.07.006. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 53.Charpentier C.J., Bromberg-Martin E.S., and Sharot T. (2018). Valuation of knowledge and ignorance in mesolimbic reward circuitry. Proceedings of the National Academy of Sciences 115, E7255–E7264. 10.1073/pnas.1800547115. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 54.Kelly C.A., and Sharot T. (2021). Individual differences in information-seeking. Nature Communications 12, 7062. 10.1038/s41467-021-27046-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 55.Bromberg-Martin E.S., and Sharot T. (2020). The Value of Beliefs. Neuron 106, 561–565. 10.1016/j.neuron.2020.05.001. [DOI] [PubMed] [Google Scholar]
  • 56.Blanchard Tommy C., Hayden Benjamin Y., and Bromberg-Martin Ethan S. (2015). Orbitofrontal Cortex Uses Distinct Codes for Different Choice Attributes in Decisions Motivated by Curiosity. Neuron 85, 602–614. 10.1016/j.neuron.2014.12.050. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 57.Kobayashi K., Ravaioli S., Baranès A., Woodford M., and Gottlieb J. (2019). Diverse motives for human curiosity. Nature Human Behaviour 3, 587–595. 10.1038/s41562-019-0589-3. [DOI] [PubMed] [Google Scholar]
  • 58.Kreps D.M., and Porteus E.L. (1978). Temporal Resolution of Uncertainty and Dynamic Choice Theory. Econometrica 46, 185–200. 10.2307/1913656. [DOI] [Google Scholar]
  • 59.Niv Y., and Schoenbaum G. (2008). Dialogues on prediction errors. Trends in Cognitive Sciences 12,265–272. 10.1016/j.tics.2008.03.006. [DOI] [PubMed] [Google Scholar]
  • 60.Berke J.D. (2018). What does dopamine mean? Nature Neuroscience 21, 787–793. 10.1038/s41593-018-0152-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 61.Schultz W. (2016). Dopamine reward prediction error coding. Dialogues in Clinical Neuroscience 18,23–32. 10.31887/DCNS.2016.18.1/wschultz. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 62.Watabe-Uchida M., Eshel N., and Uchida N. (2017). Neural Circuitry of Reward Prediction Error.Annual Review of Neuroscience 40, 373–394. 10.1146/annurev-neuro-072116-031109. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 63.Schultz W. (2002). Getting Formal with Dopamine and Reward. Neuron 36, 241–263. 10.1016/S0896-6273(02)00967-4. [DOI] [PubMed] [Google Scholar]
  • 64.Ludvig E.A., Sutton R.S., and Kehoe E.J. (2008). Stimulus Representation and the Timing of Reward-Prediction Errors in Models of the Dopamine System. Neural Computation 20, 3034–3054. 10.1162/neco.2008.11-07-654. [DOI] [PubMed] [Google Scholar]
  • 65.Oemisch M., Westendorff S., Azimi M., Hassani S.A., Ardid S., Tiesinga P., and Womelsdorf T. (2019). Feature-specific prediction errors and surprise across macaque fronto-striatal circuits. Nature Communications 10, 176. 10.1038/s41467-018-08184-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 66.Hayden B.Y., Heilbronner S.R., Pearson J.M., and Platt M.L. (2011). Surprise Signals in Anterior Cingulate Cortex: Neuronal Encoding of Unsigned Reward Prediction Errors Driving Adjustment in Behavior. Journal of Neuroscience 31, 4178–4187. 10.1523/JNEUROSCI.4652-10.2011. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 67.Matsumoto M., Matsumoto K., Abe H., and Tanaka K. (2007). Medial prefrontal cell activity signaling prediction errors of action values. Nature Neuroscience 10, 647–656. 10.1038/nn1890. [DOI] [PubMed] [Google Scholar]
  • 68.Oyama K., Hernádi I., Iijima T., and Tsutsui K.-I. (2010). Reward Prediction Error Coding in Dorsal Striatal Neurons. Journal of Neuroscience 30, 11447–11457. 10.1523/JNEUROSCI.1719-10.2010. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 69.Pearce J.M., and Hall G. (1980). A model for Pavlovian learning: Variations in the effectiveness of conditioned but not of unconditioned stimuli. Psychological Review 87, 532–552. 10.1037/0033-295X.87.6.532. [DOI] [PubMed] [Google Scholar]
  • 70.Li Y., Dalphin N., and Hyland B.I. (2013). Association with Reward Negatively Modulates Short Latency Phasic Conditioned Responses of Dorsal Raphe Nucleus Neurons in Freely Moving Rats. Journal of Neuroscience 33, 5065–5078. 10.1523/JNEUROSCI.5679-12.2013. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 71.Rangel A., Camerer C., and Montague P.R. (2008). A framework for studying the neurobiology of value-based decision making. Nature Reviews Neuroscience 9, 545–556. 10.1038/nrn2357. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 72.Levy D.J., and Glimcher P.W. (2012). The root of all value: a neural common currency for choice.Current Opinion in Neurobiology 22, 1027–1038. 10.1016/j.conb.2012.06.001. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 73.Padoa-Schioppa C. (2011). Neurobiology of Economic Choice: A Good-Based Model. Annual Review of Neuroscience 34, 333–359. 10.1146/annurev-neuro-061010-113648. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 74.Padoa-Schioppa C., and Assad J.A. (2006). Neurons in the orbitofrontal cortex encode economic value. Nature 441, 223–226. 10.1038/nature04676. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 75.Lak A., Stauffer W.R., and Schultz W. (2014). Dopamine prediction error responses integrate subjective value from different reward dimensions. Proceedings of the National Academy of Sciences 111, 2343–2348. 10.1073/pnas.1321596111. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 76.Amemori K. i., and Graybiel A.M. (2012). Localized microstimulation of primate pregenual cingulate cortex induces negative decision-making. Nature Neuroscience 15, 776–785. 10.1038/nn.3088. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 77.Monosov I.E. (2017). Anterior cingulate is a source of valence-specific information about value and uncertainty. Nature Communications 8, 134. 10.1038/s41467-017-00072-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 78.Cai X., and Padoa-Schioppa C. (2019). Neuronal evidence for good-based economic decisions under variable action costs. Nature Communications 10, 393. 10.1038/s41467-018-08209-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 79.Kimmel D.L., Elsayed G.F., Cunningham J.P., and Newsome W.T. (2020). Value and choice as separable and stable representations in orbitofrontal cortex. Nature Communications 11, 3466. 10.1038/s41467-020-17058-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 80.Cai X., and Padoa-Schioppa C. (2014). Contributions of Orbitofrontal and Lateral Prefrontal Cortices to Economic Choice and the Good-to-Action Transformation. Neuron 81, 1140–1151. 10.1016/j.neuron.2014.01.008. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 81.Pastor-Bernier A., Stasiak A., and Schultz W. (2021). Reward-specific satiety affects subjective value signals in orbitofrontal cortex during multicomponent economic choice. Proceedings of the National Academy of Sciences 118, e2022650118. 10.1073/pnas.2022650118. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 82.Kim S., Hwang J., and Lee D. (2008). Prefrontal Coding of Temporally Discounted Values during Intertemporal Choice. Neuron 59, 161–172. 10.1016/j.neuron.2008.05.010. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 83.Cai X., Kim S., and Lee D. (2011). Heterogeneous coding of temporally discounted values in the dorsal and ventral striatum during intertemporal choice. Neuron 69, 170–182. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 84.Hori Y., Mimura K., Nagai Y., Fujimoto A., Oyama K., Kikuchi E., Inoue K. i., Takada M., Suhara T., Richmond B.J., and Minamimoto T. (2021). Single caudate neurons encode temporally discounted value for formulating motivation for action. eLife 10, e61248. 10.7554/eLife.61248. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 85.Kobayashi S., and Schultz W. (2008). Influence of Reward Delays on Responses of Dopamine Neurons. Journal of Neuroscience 28, 7837–7846. 10.1523/JNEUROSCI.1600-08.2008. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 86.Raghuraman A.P., and Padoa-Schioppa C. (2014). Integration of Multiple Determinants in the Neuronal Computation of Economic Values. Journal of Neuroscience 34, 11583–11603. 10.1523/JNEUROSCI.1235-14.2014. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 87.Setogawa T., Mizuhiki T., Matsumoto N., Akizawa F., Kuboki R., Richmond B.J., and Shidara M. (2019). Neurons in the monkey orbitofrontal cortex mediate reward value computation and decision-making. Communications Biology 2, 1–9. 10.1038/s42003-019-0363-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 88.Okaty Benjamin W., Freret Morgan E., Rood Benjamin D., Brust Rachael D., Hennessy Morgan L., deBairos D., Kim Jun C., Cook Melloni N., and Dymecki Susan M. (2015). Multi-Scale Molecular Deconstruction of the Serotonin Neuron System. Neuron 88, 774–791. 10.1016/j.neuron.2015.10.007. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 89.Okaty B.W., Commons K.G., and Dymecki S.M. (2019). Embracing diversity in the 5-HT neuronal system. Nature Reviews Neuroscience 20, 397–424. 10.1038/s41583-019-0151-3. [DOI] [PubMed] [Google Scholar]
  • 90.Pinto D.F.C., Pomrenze M.B., Guo M.Y., Bentzley B.S., Eshel N., and Malenka R.C. (2023). Striatal integration of inverse dopamine and serotonin signals gates learning. 10.1101/2023.06.14.544997. [DOI]
  • 91.Ren J., Friedmann D., Xiong J., Liu C.D., Ferguson B.R., Weerakkody T., DeLoach K.E., Ran C., Pun A., Sun Y., et al. (2018). Anatomically Defined and Functionally Distinct Dorsal Raphe Serotonin Sub-systems. Cell 175, 472–487.e420. 10.1016/j.cell.2018.07.043. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 92.Jacobs B.L., and Azmitia E.C. (1992). Structure and function of the brain serotonin system.Physiological Reviews 72, 165–229. 10.1152/physrev.1992.72.1.165. [DOI] [PubMed] [Google Scholar]
  • 93.Pollak Dorocic I., Fürth D., Xuan Y., Johansson Y., Pozzi L., Silberberg G., Carlén M., and Meletis K. (2014). A Whole-Brain Atlas of Inputs to Serotonergic Neurons of the Dorsal and Median Raphe Nuclei. Neuron 83, 663–678. 10.1016/j.neuron.2014.07.002. [DOI] [PubMed] [Google Scholar]
  • 94.Ogawa S.K., Cohen J.Y., Hwang D., Uchida N., and Watabe-Uchida M. (2014). Organization of monosynaptic inputs to the serotonin and dopamine neuromodulatory systems. Cell reports 8, 1105–1118. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 95.Zhou L., Liu M.-Z., Li Q., Deng J., Mu D., and Sun Y.-G. (2017). Organization of functional long-range circuits controlling the activity of serotonergic neurons in the dorsal raphe nucleus. Cell reports 18, 3018–3032. [DOI] [PubMed] [Google Scholar]
  • 96.Arnsten A.F.T., and Goldman-Rakic P.S. (1984). Selective prefrontal cortical projections to the region of the locus coeruleus and raphe nuclei in the rhesus monkey. Brain Research 306, 9–18. 10.1016/0006-8993(84)90351-2. [DOI] [PubMed] [Google Scholar]
  • 97.Chiba T., Kayahara T., and Nakano K. (2001). Efferent projections of infralimbic and prelimbic areas of the medial prefrontal cortex in the Japanese monkey, Macaca fuscata. Brain Research 888, 83–101. 10.1016/S0006-8993(00)03013-4. [DOI] [PubMed] [Google Scholar]
  • 98.Shidara M., and Richmond B.J. (2002). Anterior Cingulate: Single Neuronal Signals Related to Degree of Reward Expectancy. Science 296, 1709–1711. 10.1126/science.1069504. [DOI] [PubMed] [Google Scholar]
  • 99.Monosov I.E., and Rushworth M.F.S. (2022). Interactions between ventrolateral prefrontal and anterior cingulate cortex during learning and behavioural change. Neuropsychopharmacology: Official Publication of the American College of Neuropsychopharmacology 47, 196–210. 10.1038/s41386-021-01079-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 100.Kennerley S.W., Dahmubed A.F., Lara A.H., and Wallis J.D. (2009). Neurons in the Frontal Lobe Encode the Value of Multiple Decision Variables. Journal of Cognitive Neuroscience 21, 1162–1178. 10.1162/jocn.2009.21100. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 101.Maisson D.J.-N., Cash-Padgett T.V., Wang M.Z., Hayden B.Y., Heilbronner S.R., and Zimmermann J. (2021). Choice-relevant information transformation along a ventrodorsal axis in the medial prefrontal cortex. Nature Communications 12, 4830. 10.1038/s41467-021-25219-w. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 102.Maisson D.J.-N., Fine J.M., Yoo S.B.M., Cash-Padgett T.V., Wang M.Z., Sleezer B.J., Zimmermann J., and Hayden B.Y. (2021). Orthogonal but linked neural codes for value. 10.1101/2021.07.27.453966. [DOI]
  • 103.Ottenheimer D.J., Bari B.A., Sutlief E., Fraser K.M., Kim T.H., Richard J.M., Cohen J.Y., and Janak P.H. (2020). A quantitative reward prediction error signal in the ventral pallidum. Nature Neuroscience 23, 1267–1276. 10.1038/s41593-020-0688-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 104.Stuber G.D., and Wise R.A. (2016). Lateral hypothalamic circuits for feeding and reward. Nature Neuroscience 19, 198–205. 10.1038/nn.4220. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 105.Ott T., and Nieder A. (2019). Dopamine and Cognitive Control in Prefrontal Cortex. Trends in Cognitive Sciences 23, 213–234. 10.1016/j.tics.2018.12.006. [DOI] [PubMed] [Google Scholar]
  • 106.Matsumoto M., and Hikosaka O. (2007). Lateral habenula as a source of negative reward signals in dopamine neurons. Nature 447, 1111–1115. [DOI] [PubMed] [Google Scholar]
  • 107.Matsumoto M., and Hikosaka O. (2009). Representation of negative motivational value in the primate lateral habenula. Nature neuroscience 12, 77. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 108.Metzger M., Bueno D., and Lima L.B. (2017). The lateral habenula and the serotonergic system.Pharmacology Biochemistry and Behavior 162, 22–28. 10.1016/j.pbb.2017.05.007. [DOI] [PubMed] [Google Scholar]
  • 109.Hu H., Cui Y., and Yang Y. (2020). Circuits and functions of the lateral habenula in health and in disease. Nature Reviews Neuroscience 21, 277–295. 10.1038/s41583-020-0292-4. [DOI] [PubMed] [Google Scholar]
  • 110.Azmitia E.C., and Gannon P.J. (1986). The primate serotonergic system: a review of human and animal studies and a report on Macaca fascicularis. Advances in Neurology 43, 407–468. [PubMed] [Google Scholar]
  • 111.Huang K.W., Ochandarena N.E., Philson A.C., Hyun M., Birnbaum J.E., Cicconet M., and Sabatini B.L. (2019). Molecular and anatomical organization of the dorsal raphe nucleus. eLife 8, e46464. 10.7554/eLife.46464. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 112.Ren J., Isakova A., Friedmann D., Zeng J., Grutzner S.M., Pun A., Zhao G.Q., Kolluru S.S., Wang R., Lin R., et al. (2019). Single-cell transcriptomes and whole-brain projections of serotonin neurons in the mouse dorsal and median raphe nuclei. eLife 8, e49424. 10.7554/eLife.49424. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 113.McDevitt R.A., Tiran-Cappello A., Shen H., Balderas I., Britt J.P., Marino R.A., Chung S.L., Richie C.T., Harvey B.K., and Bonci A. (2014). Serotonergic versus nonserotonergic dorsal raphe projection neurons: differential participation in reward circuitry. Cell reports 8, 1857–1869. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 114.Xu Z., Feng Z., Zhao M., Sun Q., Deng L., Jia X., Jiang T., Luo P., Chen W., Tudi A., et al. (2021). Whole-brain connectivity atlas of glutamatergic and GABAergic neurons in the mouse dorsal and median raphe nuclei. eLife 10, e65502. 10.7554/eLife.65502. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 115.Weissbourd B., Ren J., DeLoach Katherine E., Guenthner Casey J., Miyamichi K., and Luo L. (2014). Presynaptic Partners of Dorsal Raphe Serotonergic and GABAergic Neurons. Neuron 83, 645–662. 10.1016/j.neuron.2014.06.024. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 116.Qi J., Zhang S., Wang H.-L., Wang H., de Jesus Aceves Buendia J., Hoffman A.F., Lupica C.R., Seal R.P., and Morales M. (2014). A glutamatergic reward input from the dorsal raphe to ventral tegmental area dopamine neurons. Nature Communications 5, 5390. 10.1038/ncomms6390. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 117.Liu Z., Zhou J., Li Y., Hu F., Lu Y., Ma M., Feng Q., Zhang J. e., Wang D., and Zeng J. (2014).Dorsal raphe neurons signal reward through 5-HT and glutamate. Neuron 81, 1360–1374. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 118.Challis C., Boulden J., Veerakumar A., Espallergues J., Vassoler F.M., Pierce R.C., Beck S.G., and Berton O. (2013). Raphe GABAergic Neurons Mediate the Acquisition of Avoidance after Social Defeat. Journal of Neuroscience 33, 13978–13988. 10.1523/JNEUROSCI.2383-13.2013. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 119.Matthews G.A., Nieh E.H., Vander Weele C.M., Halbert S.A., Pradhan R.V., Yosafat A.S., Glober G.F., Izadmehr E.M., Thomas R.E., Lacy G.D., et al. (2016). Dorsal Raphe Dopamine Neurons Represent the Experience of Social Isolation. Cell 164, 617–631. 10.1016/j.cell.2015.12.040. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 120.Cho J.R., Chen X., Kahan A., Robinson J.E., Wagenaar D.A., and Gradinaru V. (2021). Dorsal Raphe Dopamine Neurons Signal Motivational Salience Dependent on Internal State, Expectation, and Behavioral Context. Journal of Neuroscience 41, 2645–2655. 10.1523/JNEUROSCI.2690-20.2021. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 121.Lin R., Liang J., Wang R., Yan T., Zhou Y., Liu Y., Feng Q., Sun F., Li Y., Li A., et al. (2020). The Raphe Dopamine System Controls the Expression of Incentive Memory. Neuron 106, 498–514.e498. 10.1016/j.neuron.2020.02.009. [DOI] [PubMed] [Google Scholar]
  • 122.Soiza-Reilly M., and Commons K.G. (2011). Glutamatergic Drive of the Dorsal Raphe Nucleus.Journal of chemical neuroanatomy 41, 247–255. 10.1016/j.jchemneu.2011.04.004. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 123.Spoida K., Masseck O.A., Deneris E.S., and Herlitze S. (2014). Gq/5-HT2c receptor signals activate a local GABAergic inhibitory feedback circuit to modulate serotonergic firing and anxiety in mice. Proceedings of the National Academy of Sciences 111, 6479–6484. 10.1073/pnas.1321576111. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 124.Cathala A., Devroye C., Drutel G., Revest J.-M., Artigas F., and Spampinato U. (2019). Serotonin2B receptors in the rat dorsal raphe nucleus exert a GABA-mediated tonic inhibitory control on serotonin neurons. Experimental Neurology 311, 57–66. 10.1016/j.expneurol.2018.09.015. [DOI] [PubMed] [Google Scholar]
  • 125.Ferré S., and Artigas F. (1993). Dopamine D2 Receptor-Mediated Regulation of Serotonin Extracellular Concentration in the Dorsal Raphe Nucleus of Freely Moving Rats. Journal of Neurochemistry 61, 772–775. 10.1111/j.1471-4159.1993.tb02187.x. [DOI] [PubMed] [Google Scholar]
  • 126.Boureau Y.-L., and Dayan P. (2011). Opponency Revisited: Competition and Cooperation Between Dopamine and Serotonin. Neuropsychopharmacology 36, 74–97. 10.1038/npp.2010.151. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 127.Miyazaki K., Miyazaki K.W., and Doya K. (2011). Activation of Dorsal Raphe Serotonin Neurons Underlies Waiting for Delayed Rewards. Journal of Neuroscience 31, 469–479. 10.1523/JNEUROSCI.3714-10.2011. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 128.Li Y., Zhong W., Wang D., Feng Q., Liu Z., Zhou J., Jia C., Hu F., Zeng J., Guo Q., et al. (2016).Serotonin neurons in the dorsal raphe nucleus encode reward signals. Nature Communications 7, 10503. 10.1038/ncomms10503. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 129.Chang A.J., Wang L., Lucantonio F., Adams M., Lemire A.L., Dudman J.T., and Cohen J.Y. (2021).Neuron-type specificity of dorsal raphe projections to ventral tegmental area. bioRxiv, 2021.2001.2006.425641. 10.1101/2021.01.06.425641. [DOI] [Google Scholar]
  • 130.Kawai H., Bouchekioua Y., Nishitani N., Niitani K., Izumi S., Morishita H., Andoh C., Nagai Y., Koda M., Hagiwara M., et al. (2022). Median raphe serotonergic neurons projecting to the interpeduncular nucleus control preference and aversion. Nature Communications 13, 7708. 10.1038/s41467-022-35346-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 131.Fischer A.G., and Ullsperger M. (2017). An Update on the Role of Serotonin and its Interplay with Dopamine for Reward. Frontiers in Human Neuroscience 11. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 132.Brown R.M., Crane A.M., and Goldman P.S. (1979). Regional distribution of monoamines in the cerebral cortex and subcortical structures of the rhesus monkey: concentrations and in vivo synthesis rates. Brain Research 168, 133–150. 10.1016/0006-8993(79)90132-X. [DOI] [PubMed] [Google Scholar]
  • 133.Goldman-Rakic P.S., and Brown R.M. (1981). Regional changes of monoamines in cerebral cortex and subcortical structures of aging rhesus monkeys. Neuroscience 6, 177–187. 10.1016/0306-4522(81)90053-1. [DOI] [PubMed] [Google Scholar]
  • 134.Moghaddam B., Berridge C.W., Goldman-Rakic P.S., Bunney B.S., and Roth R.H. (1993). In vivo assessment of basal and drug-induced dopamine release in cortical and subcortical regions of the anesthetized primate. Synapse 13, 215–222. 10.1002/syn.890130304. [DOI] [PubMed] [Google Scholar]
  • 135.Lewis D.A., Campbell M.J., Foote S.L., Goldstein M., and Morrison J.H. (1987). The distribution of tyrosine hydroxylase-immunoreactive fibers in primate neocortex is widespread but regionally specific. Journal of Neuroscience 7, 279–290. 10.1523/JNEUROSCI.07-01-00279.1987. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 136.Berger B., Trottier S., Gaspar P., Verney C., and Alvarez C. (1986). Major dopamine innervation of the cortical motor areas in the Cynomolgus monkey. A radioautographic study with comparative assessment of serotonergic afferents. Neuroscience Letters 72, 121–127. 10.1016/0304-3940(86)90066-2. [DOI] [PubMed] [Google Scholar]
  • 137.Berger B., Trottier S., Verney C., Gaspar P., and Alvarez C. (1988). Regional and laminar distribution of the dopamine and serotonin innervation in the macaque cerebral cortex: A radioautographic study. Journal of Comparative Neurology 273, 99–119. 10.1002/cne.902730109. [DOI] [PubMed] [Google Scholar]
  • 138.Williams S.M., and Goldman-Rakic P.S. (1993). Characterization of the Dopaminergic Innervation of the Primate Frontal Cortex Using a Dopamine-specific Antibody. Cerebral Cortex 3, 199–222. 10.1093/cercor/3.3.199. [DOI] [PubMed] [Google Scholar]
  • 139.Lewis D.A., Melchitzky D.S., Sesack S.R., Whitehead R.E., Auh S., and Sampson A. (2001).Dopamine transporter immunoreactivity in monkey cerebral cortex: Regional, laminar, and ultrastructural localization. Journal of Comparative Neurology 432, 119–136. 10.1002/cne.1092. [DOI] [PubMed] [Google Scholar]
  • 140.Cools R., and Arnsten A.F.T. (2022). Neuromodulation of prefrontal cortex cognitive function in primates: the powerful roles of monoamines and acetylcholine. Neuropsychopharmacology: Official Publication of the American College of Neuropsychopharmacology 47, 309–328. 10.1038/s41386-021-01100-8. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 141.Lidow M.S., Goldman-Rakic P.S., Gallager D.W., and Rakic P. (1989). Quantitative autoradiographic mapping of serotonin 5-HT1 and 5-HT2 receptors and uptake sites in the neocortex of the rhesus monkey. Journal of Comparative Neurology 280, 27–42. 10.1002/cne.902800104. [DOI] [PubMed] [Google Scholar]
  • 142.Vertes R.P., and Linley S.B. (2008). Efferent and afferent connections of the dorsal and median raphe nuclei in the rat. In Serotonin and Sleep: Molecular, Functional and Clinical Aspects, Monti J.M., Pandi-Perumal S.R., Jacobs B.L., and Nutt D.J., eds. (Birkhäuser; ), pp. 69–102. [Google Scholar]
  • 143.Dag U., Nwabudike I., Kang D., Gomes M.A., Kim J., Atanas A.A., Bueno E., Estrem C., Pugliese S., Wang Z., et al. (2023). Dissecting the functional organization of the C. elegans serotonergic system at whole-brain scale. Cell 186, 2574–2592.e2520. 10.1016/j.cell.2023.04.023. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 144.Filip M., and Bader M. (2009). Overview on 5-HT receptors and their role in physiology and pathology of the central nervous system. Pharmacological Reports 61, 761–777. 10.1016/S1734-1140(09)70132-X. [DOI] [PubMed] [Google Scholar]
  • 145.Hoyer D., Clarke D.E., Fozard J.R., Hartig P.R., Martin G.R., Mylecharane E.J., Saxena P.R., and Humphrey P.P. (1994). International Union of Pharmacology classification of receptors for 5-hydroxytryptamine (Serotonin). Pharmacological Reviews 46, 157–203. [PubMed] [Google Scholar]
  • 146.Roth B.L. (2008). The Serotonin Receptors: From Molecular Pharmacology to Human Therapeutics (Springer Science & Business Media; ). [Google Scholar]
  • 147.Beliveau V., Ganz M., Feng L., Ozenne B., Højgaard L., Fisher P.M., Svarer C., Greve D.N., and Knudsen G.M. (2017). A High-Resolution In Vivo Atlas of the Human Brain’s Serotonin System. Journal of Neuroscience 37, 120–128. 10.1523/JNEUROSCI.2830-16.2016. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 148.Salvan P., Fonseca M., Winkler A.M., Beauchamp A., Lerch J.P., and Johansen-Berg H. (2023).Serotonin regulation of behavior via large-scale neuromodulation of serotonin receptor networks. Nature Neuroscience 26, 53–63. 10.1038/s41593-022-01213-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 149.Crockett M.J., Clark L., Tabibnia G., Lieberman M.D., and Robbins T.W. (2008). Serotonin Modulates Behavioral Reactions to Unfairness. Science 320, 1739–1739. 10.1126/science.1155577. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 150.Grossman C.D., and Cohen J.Y. (2022). Neuromodulation and Neurophysiology on the Timescale of Learning and Decision-Making. Annual Review of Neuroscience 45, 317–337. 10.1146/annurev-neuro-092021-125059. [DOI] [PubMed] [Google Scholar]
  • 151.Weinberg-Wolf H., and Chang S.W.C. (2019). Differences in how macaques monitor others: Does serotonin play a central role? WIREs Cognitive Science 10, e1494. 10.1002/wcs.1494. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 152.Bhave V.M., and Nectow A.R. (2021). The dorsal raphe nucleus in the control of energy balance.Trends in Neurosciences 44, 946–960. 10.1016/j.tins.2021.09.004. [DOI] [PubMed] [Google Scholar]
  • 153.Kanen J.W., Robbins T.W., and Trofimova I.N. (2022). Harnessing temperament to elucidate the complexities of serotonin function. Current Opinion in Behavioral Sciences 45, 101108. 10.1016/j.cobeha.2022.101108. [DOI] [Google Scholar]
  • 154.Ballesta S., Shi W., Conen K.E., and Padoa-Schioppa C. (2020). Values encoded in orbitofrontal cortex are causally related to economic choices. Nature 588, 450–453. 10.1038/s41586-020-2880-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 155.Wallis J.D. (2007). Orbitofrontal Cortex and Its Contribution to Decision-Making. Annual Review of Neuroscience 30, 31–56. 10.1146/annurev.neuro.30.051606.094334. [DOI] [PubMed] [Google Scholar]
  • 156.Stopper C.M., and Floresco S.B. (2014). What’s better for me? Fundamental role for lateral habenula in promoting subjective decision biases. Nature Neuroscience 17, 33–35. 10.1038/nn.3587. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 157.Stopper C.M., Maric T.L., Montes D.R., Wiedman C.R., and Floresco S.B. (2014). Overriding phasic dopamine signals redirects action selection during risk/reward decision making. Neuron 84, 177–189. [DOI] [PubMed] [Google Scholar]
  • 158.Wallis J.D., and Miller E.K. (2003). Neuronal activity in primate dorsolateral and orbital prefrontal cortex during performance of a reward preference task. European Journal of Neuroscience 18, 2069–2081. 10.1046/j.1460-9568.2003.02922.x. [DOI] [PubMed] [Google Scholar]
  • 159.Kim J.-N., and Shadlen M.N. (1999). Neural correlates of a decision in the dorsolateral prefrontal cortex of the macaque. Nature Neuroscience 2, 176–185. 10.1038/5739. [DOI] [PubMed] [Google Scholar]
  • 160.Kennerley S.W., and Wallis J.D. (2009). Evaluating choices by single neurons in the frontal lobe: outcome value encoded across multiple decision variables. The European journal of neuroscience 29, 2061–2073. 10.1111/j.1460-9568.2009.06743.x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 161.Soubrié P. (1986). Reconciling the role of central serotonin neurons in human and animal behavior.Behavioral and Brain Sciences 9, 319–335. 10.1017/S0140525X00022871. [DOI] [Google Scholar]
  • 162.Cools R., Roberts A.C., and Robbins T.W. (2008). Serotoninergic regulation of emotional and behavioural control processes. Trends in Cognitive Sciences 12, 31–40. 10.1016/j.tics.2007.10.011. [DOI] [PubMed] [Google Scholar]
  • 163.Worbe Y., Palminteri S., Savulich G., Daw N.D., Fernandez-Egea E., Robbins T.W., and Voon V. (2016). Valence-dependent influence of serotonin depletion on model-based choice strategy. Molecular psychiatry 21, 624–629. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 164.Lottem E., Banerjee D., Vertechi P., Sarra D., oude Lohuis M., and Mainen Z.F. (2018). Activation of serotonin neurons promotes active persistence in a probabilistic foraging task. Nature communications 9, 1–12. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 165.Meyniel F., Goodwin G.M., Deakin J.W., Klinge C., MacFadyen C., Milligan H., Mullings E., Pessiglione M., and Gaillard R. (2016). A specific role for serotonin in overcoming effort cost. Elife 5, e17282. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 166.Ohmura Y., Iwami K., Chowdhury S., Sasamori H., Sugiura C., Bouchekioua Y., Nishitani N., Yamanaka A., and Yoshioka M. (2021). Disruption of model-based decision making by silencing of serotonin neurons in the dorsal raphe nucleus. Current Biology. [DOI] [PubMed] [Google Scholar]
  • 167.Costa V.D., Kakalios L.C., and Averbeck B.B. (2016). Blocking serotonin but not dopamine reuptake alters neural processing during perceptual decision making. Behavioral Neuroscience 130, 461–468. 10.1037/bne0000162. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 168.Michely J., Martin I.M., Dolan R.J., and Hauser T.U. (2023). Boosting serotonin increases information gathering by reducing subjective cognitive costs. Journal of Neuroscience. 10.1523/JNEUROSCI.1416-22.2023. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 169.Namboodiri V.M., and Hussain Shuler M.G. (2016). The hunt for the perfect discounting function and a reckoning of time perception. Current Opinion in Neurobiology 40, 135–141. 10.1016/j.conb.2016.06.019. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 170.Miyazaki K.W., Miyazaki K., Tanaka K.F., Yamanaka A., Takahashi A., Tabuchi S., and Doya K. (2014). Optogenetic activation of dorsal raphe serotonin neurons enhances patience for future rewards. Current Biology 24, 2033–2040. [DOI] [PubMed] [Google Scholar]
  • 171.Fonseca M.S., Murakami M., and Mainen Z.F. (2015). Activation of Dorsal Raphe Serotonergic Neurons Promotes Waiting but Is Not Reinforcing. Current Biology 25, 306–315. 10.1016/j.cub.2014.12.002. [DOI] [PubMed] [Google Scholar]
  • 172.Van Heukelum S., Mars R.B., Guthrie M., Buitelaar J.K., Beckmann C.F., Tiesinga P.H.E., Vogt B.A., Glennon J.C., and Havenith M.N. (2020). Where is Cingulate Cortex? A Cross-Species View. Trends in Neurosciences 43, 285–299. 10.1016/j.tins.2020.03.007. [DOI] [PubMed] [Google Scholar]
  • 173.Miyazaki K., Miyazaki K.W., Sivori G., Yamanaka A., Tanaka K.F., and Doya K. (2020).Serotonergic projections to the orbitofrontal and medial prefrontal cortices differentially modulate waiting for future rewards. Science advances 6, eabc7246. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 174.Miyazaki K.W., Miyazaki K., and Doya K. (2012). Activation of Dorsal Raphe Serotonin Neurons Is Necessary for Waiting for Delayed Rewards. Journal of Neuroscience 32, 10451–10457. 10.1523/JNEUROSCI.0915-12.2012. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 175.Hori Y., Mimura K., Nagai Y., Hori Y., Kumata K., Zhang M.-R., Suhara T., Higuchi M., and Minamimoto T. (2023). Reduced serotonergic transmission alters sensitivity to cost and reward via 5-HT1A and 5-HT1B receptors in monkeys. 10.1101/2023.02.23.529630. [DOI] [PMC free article] [PubMed]
  • 176.Clarke H.F., Dalley J.W., Crofts H.S., Robbins T.W., and Roberts A.C. (2004). Cognitive inflexibility after prefrontal serotonin depletion. Science 304, 878–880. [DOI] [PubMed] [Google Scholar]
  • 177.Clarke H.F., Walker S.C., Dalley J.W., Robbins T.W., and Roberts A.C. (2007). Cognitive inflexibility after prefrontal serotonin depletion is behaviorally and neurochemically specific. Cerebral cortex 17,18–27. [DOI] [PubMed] [Google Scholar]
  • 178.Izquierdo A., Newman T.K., Higley J.D., and Murray E.A. (2007). Genetic modulation of cognitive flexibility and socioemotional behavior in rhesus monkeys. Proceedings of the National Academy of Sciences 104, 14128–14133. 10.1073/pnas.0706583104. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 179.Vallender E.J., Lynch L., Novak M.A., and Miller G.M. (2009). Polymorphisms in the 3 UTR of the serotonin transporter are associated with cognitive flexibility in rhesus macaques. Ameri ′ an Journal of Medical Genetics Part B: Neuropsychiatric Genetics 150B, 467–475. 10.1002/ajmg.b.30835. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 180.Brigman J.L., Mathur P., Harvey-White J., Izquierdo A., Saksida L.M., Bussey T.J., Fox S., Deneris E., Murphy D.L., and Holmes A. (2010). Pharmacological or Genetic Inactivation of the Serotonin Transporter Improves Reversal Learning in Mice. Cerebral Cortex 20, 1955–1963. 10.1093/cercor/bhp266. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 181.Boulougouris V., and Robbins T.W. (2010). Enhancement of Spatial Reversal Learning by 5-HT2C Receptor Antagonism Is Neuroanatomically Specific. Journal of Neuroscience 30, 930–938. 10.1523/JNEUROSCI.4312-09.2010. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 182.Bari A., Theobald D.E., Caprioli D., Mar A.C., Aidoo-Micah A., Dalley J.W., and Robbins T.W. (2010). Serotonin Modulates Sensitivity to Reward and Negative Feedback in a Probabilistic Reversal Learning Task in Rats. Neuropsychopharmacology 35, 1290–1301. 10.1038/npp.2009.233. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 183.Roberts A.C. (2011). The Importance of Serotonin for Orbitofrontal Function. Biological Psychiatry 69, 1185–1191. 10.1016/j.biopsych.2010.12.037. [DOI] [PubMed] [Google Scholar]
  • 184.Den Ouden H.E., Daw N.D., Fernandez G., Elshout J.A., Rijpkema M., Hoogman M., Franke B., and Cools R. (2013). Dissociable effects of dopamine and serotonin on reversal learning. Neuron 80,1090–1100. [DOI] [PubMed] [Google Scholar]
  • 185.Rygula R., Clarke H.F., Cardinal R.N., Cockcroft G.J., Xia J., Dalley J.W., Robbins T.W., and Roberts A.C. (2015). Role of Central Serotonin in Anticipation of Rewarding and Punishing Outcomes: Effects of Selective Amygdala or Orbitofrontal 5-HT Depletion. Cerebral Cortex 25, 3064–3076. 10.1093/cercor/bhu102. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 186.Barlow R.L., Alsiö J., Jupp B., Rabinovich R., Shrestha S., Roberts A.C., Robbins T.W., and Dalley J.W. (2015). Markers of Serotonergic Function in the Orbitofrontal Cortex and Dorsal Raphé Nucleus Predict Individual Variation in Spatial-Discrimination Serial Reversal Learning. Neuropsychopharmacology 40, 1619–1630. 10.1038/npp.2014.335. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 187.Scholl J., Kolling N., Nelissen N., Browning M., Rushworth M.F.S., and Harmer C.J. (2017). Beyond negative valence: 2-week administration of a serotonergic antidepressant enhances both reward and effort learning signals. PLOS Biology 15, e2000756. 10.1371/journal.pbio.2000756. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 188.Iigaya K., Fonseca M.S., Murakami M., Mainen Z.F., and Dayan P. (2018). An effect of serotonergic stimulation on learning rates for rewards apparent after long intertrial intervals. Nature Communications 9, 2477. 10.1038/s41467-018-04840-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 189.Kanen J.W., Apergis-Schoute A.M., Yellowlees R., Arntz F.E., van der Flier F.E., Price A., Cardinal R.N., Christmas D.M., Clark L., Sahakian B.J., et al. (2021). Serotonin depletion impairs both Pavlovian and instrumental reversal learning in healthy humans. Molecular Psychiatry 26, 7200–7210. 10.1038/s41380-021-01240-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 190.Luo Q., Kanen J.W., Bari A., Skandali N., Langley C., Knudsen G.M., Alsiö J., Phillips B.U., Sahakian B.J., Cardinal R.N., and Robbins T.W. (2023). Common roles for serotonin in rats and humans for computations underlying flexible decision-making. 10.1101/2023.02.15.527569. [DOI] [PMC free article] [PubMed]
  • 191.Hyun J.H., Hannan P., Iwamoto H., Blakely R.D., and Kwon H.-B. (2023). Serotonin in the orbitofrontal cortex enhances cognitive flexibility. 10.1101/2023.03.09.531880. [DOI]
  • 192.Morgan A.A., Alves N.D., Stevens G.S., Yeasmin T.T., Mackay A., Power S., Sargin D., Hanna C.,Adib A.L., Ziolkowski-Blake A., et al. (2023). Medial Prefrontal Cortex Serotonin Input Regulates Cognitive Flexibility in Mice. 10.1101/2023.03.30.534775. [DOI]
  • 193.Doya K. (2002). Metalearning and neuromodulation. Neural Networks 15, 495–506. 10.1016/S0893-6080(02)00044-8. [DOI] [PubMed] [Google Scholar]
  • 194.Porrino L.J., and Goldman-Rakic P.S. (1982). Brainstem innervation of prefrontal and anterior cingulate cortex in the rhesus monkey revealed by retrograde transport of HRP. Journal of Comparative Neurology 205, 63–76. 10.1002/cne.902050107. [DOI] [PubMed] [Google Scholar]
  • 195.Wilson M.A., and Molliver M.E. (1991). The organization of serotonergic projections to cerebral cortex in primates: Retrograde transport studies. Neuroscience 44, 555–570. 10.1016/0306-4522(91)90077-2. [DOI] [PubMed] [Google Scholar]
  • 196.Kraus C., Castrén E., Kasper S., and Lanzenberger R. (2017). Serotonin and neuroplasticity – Links between molecular, functional and structural pathophysiology in depression. Neuroscience & Biobehavioral Reviews 77, 317–326. 10.1016/j.neubiorev.2017.03.007. [DOI] [PubMed] [Google Scholar]
  • 197.Hong S.Z., Mesik L., Grossman C.D., Cohen J.Y., Lee B., Severin D., Lee H.-K., Hell J.W., and Kirkwood A. (2022). Norepinephrine potentiates and serotonin depresses visual cortical responses by transforming eligibility traces. Nature Communications 13, 3202. 10.1038/s41467-022-30827-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 198.He K., Huertas M., Hong S.Z., Tie X., Hell J.W., Shouval H., and Kirkwood A. (2015). Distinct Eligibility Traces for LTP and LTD in Cortical Synapses. Neuron 88, 528–538. 10.1016/j.neuron.2015.09.037. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 199.Watabe-Uchida M., Zhu L., Ogawa S.K., Vamanrao A., and Uchida N. (2012). Whole-brain mapping of direct inputs to midbrain dopamine neurons. Neuron 74, 858–873. [DOI] [PubMed] [Google Scholar]
  • 200.Jauhar S., Arnone D., Baldwin D.S., Bloomfield M., Browning M., Cleare A.J., Corlett P., Deakin J.F.W., Erritzoe D., Fu C., et al. (2023). A leaky umbrella has little value: evidence clearly indicates the serotonin system is implicated in depression. Molecular Psychiatry, 1–4. 10.1038/s41380-023-02095-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 201.Michelsen K.A., Schmitz C., and Steinbusch H.W.M. (2007). The dorsal raphe nucleus—From silver stainings to a role in depression. Brain Research Reviews 55, 329–342. 10.1016/j.brainresrev.2007.01.002. [DOI] [PubMed] [Google Scholar]
  • 202.Cipriani A., Furukawa T.A., Salanti G., Chaimani A., Atkinson L.Z., Ogawa Y., Leucht S., Ruhe H.G., Turner E.H., Higgins J.P.T., et al. (2018). Comparative efficacy and acceptability of 21 antidepressant drugs for the acute treatment of adults with major depressive disorder: a systematic review and network meta-analysis. The Lancet 391, 1357–1366. 10.1016/S0140-6736(17)32802-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 203.Goodwin G.M., Aaronson S.T., Alvarez O., Arden P.C., Baker A., Bennett J.C., Bird C., Blom R.E., Brennan C., Brusch D., et al. (2022). Single-Dose Psilocybin for a Treatment-Resistant Episode of Major Depression. New England Journal of Medicine; 387, 1637–1648. 10.1056/NEJMoa2206443. [DOI] [PubMed] [Google Scholar]
  • 204.Gukasyan N., Davis A.K., Barrett F.S., Cosimano M.P., Sepeda N.D., Johnson M.W., and Griffiths R.R. (2022). Efficacy and safety of psilocybin-assisted treatment for major depressive disorder: Prospective 12-month follow-up. Journal of Psychopharmacology 36, 151–158. 10.1177/02698811211073759. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 205.Rotz R.v., Schindowski E.M., Jungwirth J., Schuldt A., Rieser N.M., Zahoranszky K., Seifritz E., Nowak A., Nowak P., Jäncke L., et al. (2023). Single-dose psilocybin-assisted therapy in major depressive disorder: a placebo-controlled, double-blind, randomised clinical trial. eClinicalMedicine 56. 10.1016/j.eclinm.2022.101809. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 206.Kao C.-H., Feng G.W., Hur J.K., Jarvis H., and Rutledge R.B. (2023). Computational models of subjective feelings in psychiatry. Neuroscience & Biobehavioral Reviews 145, 105008. 10.1016/j.neubiorev.2022.105008. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 207.Huys Q.J.M., Daw N.D., and Dayan P. (2015). Depression: A Decision-Theoretic Analysis. Annual Review of Neuroscience 38, 1–23. 10.1146/annurev-neuro-071714-033928. [DOI] [PubMed] [Google Scholar]
  • 208.Huys Q.J.M., Browning M., Paulus M.P., and Frank M.J. (2021). Advances in the computational understanding of mental illness. Neuropsychopharmacology 46, 3–19. 10.1038/s41386-020-0746-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 209.Association A.P. (2013). Diagnostic and Statistical Manual of Mental Disorders, Fifth Edition (American Psychiatric Publishing; ). [Google Scholar]
  • 210.Michely J., Eldar E., Martin I.M., and Dolan R.J. (2020). A mechanistic account of serotonin’s impact on mood. Nature Communications 11, 2335. 10.1038/s41467-020-16090-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 211.Eldar E., Rutledge R.B., Dolan R.J., and Niv Y. (2016). Mood as Representation of Momentum.Trends in Cognitive Sciences 20, 15–24. 10.1016/j.tics.2015.07.010. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 212.Eldar E., Roth C., Dayan P., and Dolan R.J. (2018). Decodability of Reward Learning Signals Predicts Mood Fluctuations. Current Biology 28, 1433–1439.e1437. 10.1016/j.cub.2018.03.038. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 213.Eldar E., and Niv Y. (2015). Interaction between emotional state and learning underlies mood instability. Nature Communications 6, 6149. 10.1038/ncomms7149. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 214.Der-Avakian A., and Markou A. (2012). The neurobiology of anhedonia and other reward-related deficits. Trends in Neurosciences 35, 68–77. 10.1016/j.tins.2011.11.005. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 215.Disner S.G., Beevers C.G., Haigh E.A.P., and Beck A.T. (2011). Neural mechanisms of the cognitive model of depression. Nature Reviews Neuroscience 12, 467–477. 10.1038/nrn3027. [DOI] [PubMed] [Google Scholar]
  • 216.Shadmehr R., Reppert T.R., Summerside E.M., Yoon T., and Ahmed A.A. (2019). Movement Vigor as a Reflection of Subjective Economic Utility. Trends in Neurosciences 42, 323–336. 10.1016/j.tins.2019.02.003. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 217.Perogamvros L., and Schwartz S. (2012). The roles of the reward system in sleep and dreaming. Neuroscience & Biobehavioral Reviews 36, 1934–1951. 10.1016/j.neubiorev.2012.05.010. [DOI] [PubMed] [Google Scholar]
  • 218.Mendl M., and Paul E.S. (2020). Animal affect and decision-making. Neuroscience & Biobehavioral Reviews 112, 144–163. 10.1016/j.neubiorev.2020.01.025. [DOI] [PubMed] [Google Scholar]
  • 219.Harmer C.J., Duman R.S., and Cowen P.J. (2017). How do antidepressants work? New perspectives for refining future treatment approaches. The Lancet Psychiatry 4, 409–418. 10.1016/S2215-0366(17)30015-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 220.Wang L., Zhou C., Zhu D., Wang X., Fang L., Zhong J., Mao Q., Sun L., Gong X., Xia J., et al. (2016). Serotonin-1A receptor alterations in depression: a meta-analysis of molecular imaging studies. BMC Psychiatry 16, 319. 10.1186/s12888-016-1025-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 221.Yohn C.N., Gergues M.M., and Samuels B.A. (2017). The role of 5-HT receptors in depression.Molecular Brain 10, 28. 10.1186/s13041-017-0306-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 222.Piñeyro G., and Blier P. (1999). Autoregulation of Serotonin Neurons: Role in Antidepressant Drug Action. Pharmacological Reviews 51, 533–591. [PubMed] [Google Scholar]
  • 223.De Gregorio D., Inserra A., Enns J.P., Markopoulos A., Pileggi M., El Rahimy Y., Lopez-Canul M., Comai S., and Gobbi G. (2022). Repeated lysergic acid diethylamide (LSD) reverses stress-induced anxiety-like behavior, cortical synaptogenesis deficits and serotonergic neurotransmission decline. Neuropsychopharmacology 47, 1188–1198. 10.1038/s41386-022-01301-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 224.Kwan A.C., Olson D.E., Preller K.H., and Roth B.L. (2022). The neural basis of psychedelic action.Nature Neuroscience 25, 1407–1419. 10.1038/s41593-022-01177-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 225.Roiser J.P., Elliott R., and Sahakian B.J. (2012). Cognitive Mechanisms of Treatment in Depression.Neuropsychopharmacology 37, 117–136. 10.1038/npp.2011.183. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 226.Skandali N., Rowe J.B., Voon V., Deakin J.B., Cardinal R.N., Cormack F., Passamonti L., Bevan-Jones W.R., Regenthal R., Chamberlain S.R., et al. (2018). Dissociable effects of acute SSRI (escitalopram) on executive, learning and emotional functions in healthy humans. Neuropsychopharmacology 43, 2645–2651. 10.1038/s41386-018-0229-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 227.Doss M.K., Považan M., Rosenberg M.D., Sepeda N.D., Davis A.K., Finan P.H., Smith G.S., Pekar J.J., Barker P.B., Griffiths R.R., and Barrett F.S. (2021). Psilocybin therapy increases cognitive and neural flexibility in patients with major depressive disorder. Translational Psychiatry 11, 1–10. 10.1038/s41398-021-01706-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 228.Magaraggia I., Kuiperes Z., and Schreiber R. (2021). Improving cognitive functioning in major depressive disorder with psychedelics: A dimensional approach. Neurobiology of Learning and Memory 183, 107467. 10.1016/j.nlm.2021.107467. [DOI] [PubMed] [Google Scholar]
  • 229.Pasquereau B., Drui G., Saga Y., Richard A., Millot M., Météreau E., Sgambato V., Tobler P.N., and Tremblay L. (2021). Selective serotonin reuptake inhibitor treatment retunes emotional valence in primate ventral striatum. Neuropsychopharmacology, 1–10. 10.1038/s41386-021-00991-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 230.Nutt D. (2004). Anxiety and depression: individual entities or two sides of the same coin? International Journal of Psychiatry in Clinical Practice 8, 19–24. 10.1080/13651500410005513. [DOI] [PubMed] [Google Scholar]
  • 231.Gorka S.M., Young C.B., Klumpp H., Kennedy A.E., Francis J., Ajilore O., Langenecker S.A., Shankman S.A., Craske M.G., Stein M.B., and Phan K.L. (2019). Emotion-based brain mechanisms and predictors for SSRI and CBT treatment of anxiety and depression: a randomized trial. Neuropsychopharmacology 44, 1639–1648. 10.1038/s41386-019-0407-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 232.Goddard A.W., Shekhar A., Whiteman A.F., and McDougle C.J. (2008). Serotoninergic mechanisms in the treatment of obsessive–compulsive disorder. Drug Discovery Today 13, 325–332. 10.1016/j.drudis.2007.12.009. [DOI] [PubMed] [Google Scholar]
  • 233.Goodwin G.M. (2015). The overlap between anxiety, depression, and obsessive-compulsive disorder. Dialogues in Clinical Neuroscience 17, 249–260. 10.31887/DCNS.2015.17.3/ggoodwin. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 234.Oikonomidis L., Santangelo A.M., Shiba Y., Clarke F.H., Robbins T.W., and Roberts A.C. (2017). A dimensional approach to modeling symptoms of neuropsychiatric disorders in the marmoset monkey. Developmental Neurobiology 77, 328–353. 10.1002/dneu.22446. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 235.Roberts C., Sahakian B.J., and Robbins T.W. (2020). Psychological mechanisms and functions of 5-HT and SSRIs in potential therapeutic change: Lessons from the serotonergic modulation of action selection, learning, affect, and social cognition. Neuroscience & Biobehavioral Reviews 119, 138–167. 10.1016/j.neubiorev.2020.09.001. [DOI] [PubMed] [Google Scholar]
  • 236.Chamberlain S.R., Solly J.E., Hook R.W., Vaghi M.M., and Robbins T.W. (2021). Cognitive Inflexibility in OCD and Related Disorders. In The Neurobiology and Treatment of OCD: Accelerating Progress, Fineberg N.A., and Robbins T.W., eds. (Springer International Publishing; ), pp. 125–145. [DOI] [PubMed] [Google Scholar]
  • 237.Apergis-Schoute A.M., van der Flier F.E., Ip S.H.Y., Kanen J.W., Vaghi M.M., Fineberg N.A., Sahakian B.J., Cardinal R.N., and Robbins T.W. (2023). Perseveration and shifting in obsessive–compulsive disorder as a function of uncertainty, punishment, and serotonergic medication. Biological Psychiatry Global Open Science. 10.1016/j.bpsgos.2023.06.004. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 238.Daye P.M., Monosov I.E., Hikosaka O., Leopold D.A., and Optican L.M. (2013). pyElectrode: An open-source tool using structural MRI for electrode positioning and neuron mapping. Journal of Neuroscience Methods 213, 123–131. 10.1016/j.jneumeth.2012.12.012. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 239.Fisher R.A. (1992). Statistical Methods for Research Workers. In Breakthroughs in Statistics: Methodology and Distribution, Kotz S., and Johnson N.L., eds. (Springer; ), pp. 66–70. [Google Scholar]
  • 240.Littell R.C., and Folks J.L. (1971). Asymptotic Optimality of Fisher’s Method of Combining Independent Tests. Journal of the American Statistical Association 66, 802–806. 10.1080/01621459.1971.10482347. [DOI] [Google Scholar]
  • 241.Green D.M., and Swets J.A. (1966). Signal detection theory and psychophysics (Wiley New York; ). [Google Scholar]
  • 242.Davies R.M., Gerstein G.L., and Baker S.N. (2006). Measurement of Time-Dependent Changes in the Irregularity of Neural Spiking. Journal of Neurophysiology 96, 906–918. 10.1152/jn.01030.2005. [DOI] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

1

Articles from bioRxiv are provided here courtesy of Cold Spring Harbor Laboratory Preprints

RESOURCES