Abstract
Primates must adapt to changing environments by optimizing their behavior to make beneficial choices. At the core of adaptive behavior is the orbitofrontal cortex (OFC) of the brain, which updates choice value through direct experience or knowledge-based inference. Here, we identify distinct neural circuitry underlying these two separate abilities. We designed two behavioral tasks in which two male macaque monkeys updated the values of certain items, either by directly experiencing changes in stimulus-reward associations, or by inferring the value of unexperienced items based on the task’s rules. Chemogenetic silencing of bilateral OFC combined with mathematical model-fitting analysis revealed that monkey OFC is involved in updating item value based on both experience and inference. In vivo imaging of chemogenetic receptors by positron emission tomography allowed us to map projections from the OFC to the rostromedial caudate nucleus (rmCD) and the medial part of the mediodorsal thalamus (MDm). Chemogenetic silencing of the OFC-rmCD pathway impaired experience-based value updating, while silencing the OFC-MDm pathway impaired inference-based value updating. Our results thus demonstrate dissociable contributions of distinct OFC projections to different behavioral strategies, and provide new insights into the neural basis of value-based adaptive decision-making in primates.
Subject terms: Neural circuits, Decision
The OFC is at the core of value-based adaptive decision-making. Here, the authors demonstrate dissociable contributions of primate OFC-subcortical pathways to different behavioral strategies by combining chemogenetics with a model-fitting approach.
Introduction
To survive in a constantly changing world, animals naturally adapt quickly to new environments and adjust their behavior to maximize the benefits. This involves making decisions that will lead to maximum subjective benefit, based on the changing relationships between specific events and outcomes, and thus requires knowing the current worth of each option when making a choice. Typically, an item’s worth is learned through direct experience, a process often explained through the concept of classical reinforcement learning1. However, animals with highly developed brains, particularly primates, have also evolved the ability to infer the value of unexperienced events/items from their knowledge of the world. This ability is described by the rule or theory of shifting relationships. For example, if a monkey is eating a banana and notices that it is ripe, it may be able to infer that other nearby bananas are also ripe, based on its knowledge of how fruits ripen. Optimal decision-making relies on a balance between experience- and inference-based behavioral strategies.
The primate orbitofrontal cortex (OFC) is thought to contribute to such adaptive behavior by leveraging both direct experience and knowledge of the current context. The OFC has long been thought to play an essential role in encoding the subjective values of alternative events/items that guide subsequent decision-making, and in learning/updating these values by integrating past experiences as consequences of our choices2–6. At the same time, the OFC has also been shown to be necessary for inferring value based on mental simulation of outcomes, even in the absence of direct experience, by generalizing knowledge of the current situation or environment7–9. A recent report suggests that the OFC regulates the balance between these two valuation strategies, rather than simply initiating one or the other10. Thus, there is ongoing debate about the core function of the OFC in adaptive behavior.
The complexity of OFC function might arise from its interactions with other brain regions through direct anatomical connections11. For example, subcortical structures, such as the rostromedial part of the caudate nucleus (rmCD) and the medial part of the mediodorsal thalamus (MDm), receive direct projections from the OFC12–14. Lesions to these areas have been shown to produce deficits that are similar, but not identical, to those produced by OFC lesions15–20 with a tendency that rmCD and MDm are particularly involved in value-updating based on experience and inference, respectively, suggesting that they have overlapping yet distinct roles in adaptive behavior. These findings have raised the possibility that two pathways originating from OFC, namely the OFC-rmCD and OFC-MDm pathways, are needed for different value-updating strategies. To investigate this possibility, it is essential to independently manipulate the prefronto-subcortical circuits, which is technically challenging, especially in behaving nonhuman primates.
To investigate the causal roles of the OFC and its originating pathways in these two types of valuation, we use a chemogenetic tool called designer receptors exclusively activated by designer drugs (DREADDs)21. This tool allows neurons to be silenced by activating an inhibitory DREADD (hM4Di) following systemic administration of a DREADD agonist. Additionally, local agonist infusion that activates hM4Di expressed at axon terminals can suppress synaptic transmission22,23. By combining these techniques with positron emission tomography (PET) as an in vivo imaging tool for hM4Di-positive projection sites, we have previously developed an imaging-guided chemogenetic synaptic-silencing method that is dramatically more efficient and accurate, especially when applied to nonhuman primates24. Leveraging this technique and a model-fitting analysis in a reinforcement learning framework, the present study addresses the contributions of these two OFC-subcortical pathways to different value-updating strategies. Our results suggest that experience- and inference-based strategies for updating stimulus-reward associations rely on the OFC-rmCD and OFC-MDm pathways, respectively.
Results
Experience- and inference-based behavioral adaptation in multi-reward value-updating tasks
To address the question of how the OFC and its projections to subcortical structures contribute to behavioral adaptation through experience- and inference-based valuation, we devised two behavioral tasks for macaque monkeys: NOVEL and FAMILIAR tasks, respectively (Fig. 1a). In both tasks, the monkeys were required to choose either of two presented visual stimuli (out of a set of five abstract images), each of which was associated with 1, 2, 3, 4, or 5 drops of juice. The order of associations was reversed within a session (Fig. 1a). To maximize their reward, monkeys had to learn the values of the visual stimuli and then update them following subsequent changes in stimulus-reward associations. In the NOVEL task, which was aimed at assessing experience-based updating, a new set of stimuli was introduced in each session, thus requiring the monkeys to learn new stimulus-reward associations, as well as the association reversal that was imposed mid-session (90 trials after the beginning of each session). After several months of training, two monkeys (Mk#1 and Mk#2) were able to learn new stimulus-reward associations within 80–90 trials and adapted to their reversal within 30–50 trials (Fig. 1b, right). In the FAMILIAR task, which was aimed at assessing inference-based updating, a fixed set of five visual stimuli was used throughout the experiments, and the reversals were imposed several times after performance reached a predetermined criterion (see Methods). Following several months of training on this task, the monkeys were able to adapt to the reversal within 3–5 trials (Fig. 1c, right). They even showed optimal choice for “unexperienced” stimulus-reward associations after experiencing the other associations following the reversal (for details, see the last paragraph of the Results section), indicating that they solved this task based on inference, that is, their prior knowledge of the limited patterns of stimulus-reward associations.
To examine whether the monkeys solved these tasks based on experience- or inference-based strategies, we conducted simulations using two types of reinforcement learning models (see Methods for details); one assuming that the values were updated through direct experience based on standard model-free reinforcement learning that was driven by reward prediction errors (“EXP” model), and one assuming that the monkeys had a priori knowledge of the two possible stimulus-reward association patterns (1,2,3,4,5 or 5,4,3,2,1 drops) through daily training, and that prediction errors drove the transition between these two already learned value sets, allowing them to infer the values of any unexperienced stimuli (“INF” model). As expected, this analysis revealed that behavior during the NOVEL task was better explained by the EXP model (Fig. 1d), while that during the FAMILIAR task was better explained by the INF model (Fig. 1e). These results suggest that monkeys “solved” the two tasks with different strategies — experience-based for the NOVEL task and inference-based for the FAMILIAR task.
OFC silencing impairs both experience- and inference-based strategies
Next, we chemogenetically inactivated the OFC to determine whether it contributes to experience- and/or inference-based valuation strategies. First, we bilaterally introduced the inhibitory DREADD hM4Di into the OFC (Brodmann’s area 11/13, hereafter referred to as “OFC11/13”) of each monkey via injections of an adeno-associated virus (AAV) vector (AAV2-CMV-hM4Di and AAV2.1-CaMKII-hM4Di-IRES-AcGFP for Mk#1 and Mk#2, respectively) (Fig. 2a). Several weeks after the injections, we non-invasively visualized hM4Di expression using PET imaging with DREADD-PET tracers. In both monkeys, we consistently observed increased PET signal in the bilateral OFC (Fig. 2a; Supplementary Fig. 1a), which was confirmed by post-mortem immunohistochemistry to reflect hM4Di expression. We then examined the effect that silencing OFC11/13 had on behavior; task performances were compared after systemically administering either a vehicle control or the DREADD agonist deschloroclozapine (DCZ) (Fig. 2b). While silencing OFC11/13 did not alter performance on the NOVEL task before reversal of the stimulus-reward contingencies (acquisition phase), it did impair performance following the reversal, which was particularly true in the early phase (100 post-reversal trials, PoE in Fig. 2c, e). Similarly, OFC11/13 silencing also impaired performance on the FAMILIAR task after reversals (Fig. 2d, g). To quantify the silencing effect in both tasks, we fitted two behavioral models to the behavioral data following DCZ and vehicle administration. Model-fitting analysis revealed that the impaired performance after OFC11/13 silencing could be attributed to a decrease in the learning rate (α) after the reversals (Fig. 2f, h, see Methods for details), but not to the change in the extent of exploration (the inverse temperature, β; Supplementary Fig. 2a, d), suggesting that silencing OFC11/13 led to deficits in value-updating when using either strategy.
We conducted four additional experiments to confirm that the observed behavioral changes were due to the loss of normal OFC11/13 function that is involved in adapting to the shift in the task context, and which requires updating the values of the external stimuli. First, to confirm that healthy OFC11/13 function is required at the time of the reversal, but not when learning the stimulus-reward associations or during any other process in the acquisition phase, we administered DCZ after the acquisition phase (just before reversal, see Methods for details). This manipulation produced deficits similar to those observed when DCZ was administered before the beginning of each session (Supplementary Fig. 3a, b), suggesting that the effects of OFC11/13 silencing were limited to the reversal of stimulus-reward associations. Second, we confirmed that DCZ alone did not significantly affect behavioral performance before the introduction of hM4Di (Mk#2, Supplementary Fig. 3c, d), indicating that the effects observed after DCZ administration were due to DREADD activation. Third, we confirmed that OFC11/13 silencing did not significantly impact simple reversal learning (Supplementary Fig. 4), consistent with a previous lesion study25. Fourth, to test whether the OFC11/13 is also essential for situations in which knowledge-based value-updating is required and the cognitive load is high, but the item values are binary (i.e., reward or no reward), we examined the effects of OFC11/13 silencing using an analog of the Wisconsin Card Sorting Test26 (Mk#2, Supplementary Fig. 5b). In this case, we found that silencing OFC11/13 had no effect on task performance (Supplementary Fig. 5c), suggesting that it is required specifically when updating external items with complex, non-binary values. Importantly, chemogenetic silencing of OFC11/13 remained effective at the end of all experiments, including these control experiments and the pathway-selective manipulations (see below) (Supplementary Fig. 5a), as demonstrated by the significant effects of DCZ administration on performance during the final devaluation task (Supplementary Fig. 5d, e), which is one of the most common tasks requiring normal OFC function7,27. Taken together, these results suggest that OFC11/13 is essential for adapting behavioral responses that are specifically contingent upon being able to update multiple values of multiple external stimuli.
The OFC11/13-rmCD and OFC11/13-MDm pathways are necessary for experience- and inference-based value-updating, respectively
Having demonstrated that OFC11/13 is essential for both experience- and inference-based behavioral adaptation, the next question is whether these different strategies are governed by separate neural pathways. To answer this, we conducted a chemogenetic pathway-selective manipulation wherein information flow can be temporarily inactivated by local agonist infusion into axonal terminals expressing hM4Di. It is generally challenging to precisely localize and target axonal projection sites in vivo in monkeys, which have relatively large and complexly shaped brains. We overcame this obstacle by using an imaging-guided chemogenetic technique24 in which PET allows localization of the projection sites for hM4Di-positive OFC11/13 neurons. Aside from OFC11/13, subtraction PET images (post-AAV injection minus pre-AAV injection) showed increased PET signals in the striatum and the thalamus, specifically in the rmCD, MDm, and the medial part of the putamen (Fig. 3a, b; Supplementary Fig. 1b–d). These regions colocalized with GFP-positive axon terminals under immunohistological examination (Supplementary Fig. 1b–d). In contrast, the hM4Di signal was not clearly observed via PET or histology in other brain regions that are known to also receive projections from OFC11/13, such as the amygdala, and was not comparable to what we observed in these three regions (Supplementary Fig. 1e). Thus, our next experiments focused on the projections from OFC11/13 to the three DREADD-positive terminal regions.
To reversibly block neuronal transmission from OFC11/13, we infused DCZ into one of the three terminal regions under the guidance of MR, PET, and CT imaging (Fig. 3a, b). Compared with the control vehicle infusion, local DCZ infusion into the rmCD impaired performance on the NOVEL task just after reversals, similar to the systemic DCZ injections (Fig. 3c, e, left; effect size of treatment, 0.50 vs. 0.48 for systemic injections and local infusions into the rmCD). However, this was not the case for the FAMILIAR task (Fig. 3c, e, right). Conversely, DCZ infusion into the MDm did not impair performance on the NOVEL task (Fig. 3d, f, left), but did impair performance on the FAMILIAR task (Fig. 3d, f, right; effect size of treatment, 0.46 vs. 0.29 for systemic injections and local infusions into the MDm). The effect of local DCZ infusion appeared to be consistent throughout the session, with no systematic differences in task performance between the 1st and 2nd halves (Supplementary Fig. 6). Although we also injected DCZ into the medial part of the putamen, where we found weak PET and histological signals (Supplementary Fig. 1d), this did not affect performance on the NOVEL task (Supplementary Fig. 7a, b). In addition, control DCZ infusions into the hM4Di-negative brain region (anterior thalamus; the structure between the rmCD and the MDm) had no significant impact on behavior in the NOVEL task (Supplementary Fig. 7c, d), thus supporting the claim that our data described above are attributable to the chemogenetic silencing of hM4Di-positive OFC11/13-derived axon terminals. We therefore focused further tests on the OFC11/13 to rmCD and OFC11/13 to MDm pathways. The order of the experiments, including systemic injections and local infusions into each area, for each monkey is summarized in Supplementary Fig. 8.
Model fitting analysis using two reinforcement learning models revealed that silencing the OFC11/13-rmCD pathway significantly reduced the learning rate during the NOVEL task (EXP model; Fig. 3g, left), but not during the FAMILIAR task (INF model; Fig. 3g, right). Conversely, silencing the OFC11/13-MDm pathway had no impact on the learning rate during the NOVEL task (Exp model; Fig. 3h, left), but significantly reduced it during the FAMILIAR task (INF model; Fig. 3h, right). Similar to OFC11/13 silencing, the inverse temperature was not affected by silencing either pathway (Supplementary Fig. 2b, c, e, f).
Damage to the OFC in humans has been associated with increased impulsivity28. In monkeys, OFC inactivation or lesioning has resulted in faster reaction times on experimental tasks29, which is generally interpreted as a sign of impulsivity or lack of control. To assess whether silencing OFC11/13 and its projections affected impulsivity in our task context, and if so, whether this behavioral change is related to the observed impairment in performance, we examined reaction times after each type of silencing (Supplementary Fig. 9). OFC11/13 silencing resulted in shorter reaction times for both monkeys during all task phases, but we did not observe any direct relationship between reaction time and performance (Supplementary Fig. 9a). In contrast, silencing the OFC11/13-rmCD and OFC11/13-MDm pathways induced complex and contradicting results; silencing the OFC11/13-rmCD pathway increased reaction time (Supplementary Fig. 9b), whereas silencing the OFC11/13-MDm pathway did not influence reaction time (Supplementary Fig. 9c). Although these effects on reaction time differed, we can conclude that the difficulty in updating item values was unrelated to increased impulsivity.
Taken together, the outcomes after selectively silencing the two OFC11/13 projections indicated that both are involved in updating stimulus values; the OFC11/13-rmCD pathway is needed when updating based on direct experience of stimulus-reward associations, whereas the OFC11/13-MDm pathway is needed when updating based on inference from previously learned knowledge.
The OFC11/13-rmCD and OFC11/13-MDm pathways differentially contribute to the sensitivity to past outcomes
Our results suggest that the two pathways contribute distinctly to different behavioral strategies. To further corroborate this dissociation, we performed a learning-model agnostic analysis in which we decomposed trials into specific events related to each behavioral strategy. First, we focused on experience-based updating during the NOVEL task. If the ability to update values based on past experience is impaired, the behavior following unpredicted positive experiences (i.e., obtaining a good result after choosing the previously unchosen option) should favor repeating the same choice, and vice versa following negative experiences (Fig. 4a). Consistent with the model-fitting analysis, silencing OFC11/13 or the OFC11/13-rmCD pathway significantly impaired performance following both positive and negative experiences (Fig. 4c, d), suggesting that their contribution to updating stimulus-reward associations is based on both positive and negative experiences. In contrast, silencing the OFC11/13-MDm pathway only impaired performance following negative experiences (Fig. 4e), likely reflecting minor deficits at the immediate post-reversal period (Fig. 3d). Similarly, an asymmetric deficit was induced by silencing the OFC11/13-MDm pathway during the FAMILIAR task (Supplementary Fig. 10).
Next, we focused on inference-based updating during the FAMILIAR task. We found that the monkeys could adapt to the new stimulus-reward association after a single experience of a reversal (1st trial, Fig. 4b, left), even when the subsequent trial did not include options that appeared in the previous trial (inference trial, Fig. 4b, right). Monkeys exhibited a greater percentage of optimal choices in the inference trial than in the 1st trial (inference vs. first trials; Mk#1: 72.2% vs. 19.6%; Mk#2: 59.3% vs. 17.7%) under the baseline control conditions, indicating that they inferred the stimulus-reward associations without needing direct experience. This inference-based behavioral adaptation was significantly impaired after silencing OFC11/13 or the OFC11/13-MDm pathway (Fig. 4f, h). In contrast, silencing the OFC11/13-rmCD pathway had no effect on the inference trials (Fig. 4g). Taken together, these results support the conclusion that the OFC11/13-rmCD pathway is essential for updating value via positive and negative experiences, whereas the OFC11/13-MDm pathway is selectively involved in behavioral changes following negative experiences—a capacity that is critical for rapid behavioral adaptation using an inference-based strategy based on prior knowledge of the situation.
Discussion
Here, by combining chemogenetic silencing of individual neural pathways with a model-fitting approach, we demonstrate that the two natural strategies for updating subjective value of external stimuli rely on two distinct neural pathways, each originating in the OFC11/13 and projecting to a different subcortical brain region. Silencing the OFC11/13-rmCD pathway impaired performance when monkeys updated option values through direct experience of both positive and negative changes of stimulus-reward associations, whereas silencing the OFC11/13-MDm pathway impaired performance when monkeys updated the values based on inference that was guided by the negative experience that resulted from a change in the task context. This dissociable contribution of neural pathways provides new insights into the neural basis of value-based adaptive decision-making in primates.
In this study, we devised two behavioral tasks that allowed us to dissociate how stimulus-reward associations are updated: the NOVEL task for experience-based updating and the FAMILIAR task for inference-based updating. Indeed, our model-fitting analysis suggested that the monkeys solved these tasks using, primarily, the respective value-updating strategies proposed (Fig. 1). However, we had extensively trained the monkeys on the NOVEL task before the experimental sessions, which likely led to the development of a general understanding task structure, including the potential for reversal. Improved performance in the post-reversal phase, relative to the acquisition phase, suggests a cognitive ability that goes beyond purely making new experience-dependent associations from scratch. Therefore, we cannot exclude the possibility that some degree of inference was implemented during the NOVEL task, and deficits in both tasks induced by OFC silencing could have resulted from no longer being able to infer the changes in stimulus-reward associations following the reversals. However, the differential effects observed after silencing the OFC-rmCD pathway in the NOVEL task (specifically during the post-reversal phase), but not in the FAMILIAR task, led us to a more straightforward interpretation: the critical involvement of the OFC-rmCD pathway in the NOVEL task is specific to task context. Specifically, it appears to be essential for updating learned values based on experience following novel cue-reward associations. While this interpretation is straightforward, it does not rule out the possibility that some type of inference, such as awareness of potential reversal, could be a common element in both tasks.
The OFC has been thought to play an essential role in learning/updating stimulus-reward associations by integrating past experiences as consequences of our choices2–6, although recent research has highlighted a specialized role in inference-based, or model-based, learning30,31. Recent reports suggest that rodent OFC, or the ventral PFC in humans, is not simply engaged in either experience- or inference-based value-updating strategies, but rather in regulating the switch between them according to current contexts10,32. Our present results are consistent with and extend this idea; we can now assign distinct roles for at least two different OFC projections to subcortical structures (the rmCD and MDm). Two regions downstream from the PFC—the striatum and MD thalamus—have been shown to be involved in value-updating. Many studies have suggested the involvement of the striatum in both experience- and inference-based updating, with a gradation within the striatum15–17,33–35. Importantly, a pathway-selective manipulation study in rodents demonstrated that selective ablation of the OFC-accumbens pathway induced impaired performance following a reversal that was similar to what we show here, with a slight difference in the sensitivity to positive outcomes36. The MD thalamus has also been suggested to be involved in inference-based updating18,37, or in adapting to shifts in task context, not simply in changes to stimulus-outcome contingencies38. The critical contribution of the interactions between the PFC and the MD thalamus for adaptation to shifts in task context has also been demonstrated in rodents39,40. These previous reports imply that OFC sends information necessary for implementing each type of value update, e.g., reward prediction and prediction error signal41–43 or task context44, to rmCD and MDm, which should be identified in future studies.
The OFC is widely thought to play a central role in updating and maintaining information about possible outcomes6,11. Both lesion and recording studies have shown that the OFC is critical for updating associations between stimuli and outcomes, but not for acquiring such associations20,45, which was also the case in our study (Fig. 2). Previous research has also suggested that the OFC is required for reversal learning in various species46–48. However, recent studies have challenged this view by showing that selective lesion/inactivation of the OFC in macaque monkeys had no effect in a simple reversal learning paradigm25,49, a result consistent with our observations (Supplementary Fig. 4). Recently, another study suggested that the OFC is essential for updating the desirability (i.e., quality or quantity) of reward-associated stimuli, but not their availability (i.e., the probability of receiving rewards)20. Our current results, in which OFC silencing severely impaired performance on a behavioral task that requires monkeys to associate multiple stimuli with multiple reward amounts and to update their associations, support this view. Notably, OFC silencing did not impact performance on the Wisconsin Card Sorting Task (WCST, Supplementary Fig. 5), which is commonly used to assess behavioral flexibility. Although previous lesion studies in monkeys have reported deficits in WCST performance following aspirating lesions26, subsequent studies have suggested that these deficits may have been due to damaged fibers that pass near the OFC25. Because outcomes on the WCST are typically binary, either receiving a reward or not, our results suggest that primate OFC, at least the area that we focused on in this study (Brodmann’s area 11/13), is specialized for representing stimulus-reward associations that are based on the desirability of possible outcomes.
Importantly, the two pathways we focused on form part of a broader circuit for value-updating. For example, the amygdala is known to be involved in value-based behavior through interactions with the OFC. It has been shown that lesion of the monkey amygdala affects value representation in OFC neurons50,51, and that crossed lesions of the OFC and amygdala interrupt reinforcer devaluation52,53, suggesting that their interactions are important for inference-based value updating. Lesions of the amygdala were also reported to result in inference-based enhancements due to reductions in the amount of evidence suggesting a reversal has occurred, which causes frequent switching after a negative outcome54,55. However, we did not examine the OFC-amygdala pathway because positive PET signals were not detected in the amygdala (Supplementary Fig. 1), which is consistent with previous anatomical studies indicating relatively weak direct projections from OFC11/13 to this area56,57. Even without any direct OFC-to-amygdala connection, we must consider the fact that these brain regions may interact through other routes, such as direct projections from the amygdala to the OFC58 or indirect connections via midbrain dopamine neurons59,60. Future studies should aim to investigate how the OFC and the amygdala interact and contribute to value updating.
In the current study, we were able to dissect the functions of the OFC-subcortical circuits by using the imaging-guided pathway-selective synaptic silencing technique that we recently developed24. In that study, we also demonstrated a functional double dissociation in the pathways from the dorsolateral prefrontal cortex (dlPFC) to the CD and MD thalamus. Given the parallel forms of cortico-basal ganglia-thalamo-cortical circuits61, these consistent observations lead us to predict that other prefronto-striatal and prefronto-thalamus pathways, e.g., pathways from the ventrolateral PFC to the caudate and MD thalamus, may also have specialized roles aligned with the cognitive functions associated with their prefrontal origin, as recently investigated in the rodents62. On the other hand, convergence of projections from different prefrontal areas to specific striatal zones has also been reported63. Indeed, the PET data from the current study and our previous research24, indicate that a part of the rmCD receives projections from both the dorsolateral and orbital prefrontal cortex, seemingly corresponding to a previously identified “convergence zone”63. These observations suggest that this area may be involved in the integration of information from different cortico-basal ganglia-thalamo-cortical circuits. Future studies should aim to better understand the overall architecture of these networks by identifying where and how they communicate with each other and how the information processed in each network is integrated.
The current study had some limitations. First, our data were obtained from only two monkeys, fewer than the norm for behavioral ablation studies. Nevertheless, the strength of our conclusions about the functional roles of the OFC and its projections to the two subcortical regions is supported by statistically significant and consistent findings across multiple experiments in these two monkeys (for a few exceptions, please see each figure legend). Our chemogenetic approach allows for a rigorous within-subject design—in contrast to ablation techniques that generally require larger sample sizes (typically three or more subjects) due to individual variability—to study the effect of region/pathway manipulations in the same individuals. Second, although our results demonstrated that selective silencing of either pathway induced robust effects comparable to those observed by OFC11/13 silencing per se, especially in silencing the OFC11/13-rmCD pathway in the NOVEL task, this does not exclude the possibility that other regions/pathways also play important roles in generating these behaviors. Furthermore, we cannot determine whether we were able to inhibit an entire subset of projections from the OFC11/13—as our methods rely on the diffusion of agonist solution and proportion of neurons that expressed DREADDs—only that it was sufficient to drive selective behavioral deficits.
In summary, leveraging the technical advantage of imaging-guided pathway-selective chemogenetic silencing, we have demonstrated the dissociable contributions of two OFC-subcortical pathways to different value-updating strategies. The identification of causal relationships between a specific neural pathway and a cognitive function, as demonstrated in this study, can complement what we have learned from human studies, but which cannot be directly tested in humans for ethical reasons. Thus, in addition to providing new insights into the neural basis of value-based adaptive decision-making in nonhuman primates, our findings have potential implications for understanding certain psychiatric disorders, such as obsessive-compulsive disorder (OCD). Our findings are particularly relevant to OCD because abnormal functional connectivity between the OFC and the caudate nucleus is frequently reported in patients with OCD64.
Methods
Subjects
Two experimentally naïve male Japanese monkeys (Macaca fuscata) participated in the experiments (Mk#1: 7.2 kg; Mk#2: 6.8 kg; both aged 4 years at the beginning of experiments). The monkeys were kept in individual primate cages in an air-conditioned room. A standard diet, supplementary fruits/vegetables, and a tablet of vitamin C (200 mg) were provided daily. All experimental procedures involving the monkeys were carried out in accordance with the Guide for the Care and Use of Nonhuman primates in Neuroscience Research (The Japan Neuroscience Society; https://www.jnss.org/en/animal_primates) and were approved by the Animal Ethics Committee of the National Institutes for Quantum Science and Technology.
Viral vector production
Mk#1 was co-injected with two AAV vectors, one expressing hM4Di and the other expressing GFP (AAV2-CMV-hM4Di and AAV2-CMV-AcGFP; 2.3 × 1013 and 4.6 × 1012 particles/mL, respectively). Mk#2 was injected with an AAV vector expressing both hM4Di and GFP (AAV2.1-CaMKII-hM4Di-IRES-AcGFP, 1.0 × 1013 particles/mL). AAV vectors were produced by a helper-free triple transfection procedure, and were purified by affinity chromatography (GE Healthcare, Chicago, USA). Viral titer was determined by quantitative PCR using Taq-Man technology (Life Technologies, Waltham, USA)65.
Surgical procedures and viral vector injections
Surgeries were performed under aseptic conditions in a fully equipped operating suite. We monitored body temperature, heart rate, SpO2, and tidal CO2 throughout all surgical procedures. Monkeys were immobilized by intramuscular (i.m.) injection of ketamine (5–10 mg/kg) and xylazine (0.2–0.5 mg/kg) and intubated with an endotracheal tube. Anesthesia was maintained with isoflurane (1–3%, to effect). Before surgery, magnetic resonance (MR) imaging (7 tesla 400 mm/SS system, NIRS/KOBELCO/Brucker) and X-ray computed tomography (CT) scans (Accuitomo170, J. MORITA CO., Kyoto, Japan) were performed under anesthesia (continuous intravenous infusion of propofol 0.2–0.6 mg/kg/min). Overlay MR and CT images were created using PMOD® image-analysis software (PMOD Technologies Ltd, Zurich, Switzerland) to estimate stereotaxic coordinates of target brain structures. After surgery, prophylactic antibiotics and analgesics (cefmetazole, 25–50 mg/kg; ketoprofen, 1–2 mg/kg) were administered.
The bilateral OFCs (BA11 & BA13) of each monkey were injected with the AAV vectors (Fig. 1a). The injections were performed under direct vision using the same types of surgical procedures as in a previous study66. Briefly, after retracting skin, galea, and muscle, the frontal cortex was exposed by removing a bone flap and reflecting the dura mater. Then, handheld injections were made under visual guidance through an operating microscope (Leica M220, Leica Microsystems GmbH, Wetzlar, Germany), with care taken to place the beveled tip of a microsyringe (Model 1701RN, Hamilton) containing the viral vector at an angle oblique to the brain surface. The needle (26 Gauge, PT2) was inserted into the intended area of injection by one person and a second person pressed the plunger to expel approximately 1 µL per penetration. For Mk#1, total volumes of 54 µL and 50 µL were injected via 54 and 50 tracks in the left and right hemispheres, respectively. For Mk#2, total volumes of 53 µL and 49 µL were injected via 50 and 47 tracks in the left and right hemispheres, respectively.
PET imaging
PET imaging was conducted as previously reported24. Briefly, PET scans were conducted before injection of vectors and at 45 days after injection for both monkeys. PET scans were performed using a microPET Focus 220 scanner (Siemens Medical Solutions USA, Malvern, USA). Monkeys were immobilized by ketamine (5–10 mg/kg) and xylazine (0.2–0.5 mg/kg) and then maintained under anesthetized condition with isoflurane (1%–3%) during all PET procedures. Transmission scans were performed for approximately 20 min with a Ge-68 source. Emission scans were acquired in 3D list mode with an energy window of 350–750 keV after intravenous bolus injection of [11C]clozapine (for Mk#1; 375.5–394.7 MBq) or [11C]DCZ (for MK#2; 324.9–382.3 MBq). Emission data acquisition lasted 90 min. To estimate the specific binding of [11C]DCZ in Mk#2, regional binding potential relative to nondisplaceable radioligand (BPND) was calculated by PMOD® with an original multilinear reference tissue model (MRTMo). To visualize the expression of DREADDs, contrast (subtraction) of images taken before and 45 days after vector injection were created using PMOD for SUV (standardized uptake value) for Mk#1 and BPND for Mk#2 by investigating whether differential PET signals were observed at the target sites.
Drug administration
DCZ (HY-42110; MedChemExpress) was dissolved in dimethyl sulfoxide (DMSO, FUJIFILM Wako Pure Chemical Co.), aliquoted and stored at −30 °C. For systemic intramuscular injection, this stock solution was first diluted in saline to a final volume of 1 mL (2.5% DMSO in saline), and thus achieving a dose of 100 μg/kg. DCZ solution was injected 15 min before the beginning of the experiments, unless otherwise noted. Fresh solution was prepared on each day of usage. We performed at most one injection experiment per week for each treatment (vehicle and DCZ), repeating the NOVEL and FAMILIAR tasks seven and six times, respectively.
For microinfusion, DCZ was first dissolved in DMSO and then diluted in PBS to a final concentration of 100 nM. We prepared fresh solutions on the day of usage. We used two stainless steel infusion cannulae (outer diameter 300 μm; Muromachi-Kikai) inserted into each target region: rmCD and MDm, and ventral putamen for additional experiments (Fig. S7). Each cannula was connected to a 10-μL microsyringe (#7105KH; Hamilton) via polyethylene tubing. These cannulae were advanced via guide tube by means of an oil-drive micromanipulator. DCZ solution or PBS was injected at a rate of 0.25 μL/min by auto-injector (Legato210; KD Scientific) for a total volume of 3 μL for each hemisphere. The injection volumes were determined based on a previous study reporting that injections of 3 μL and 1.5 μL resulted in a diameter of aqueous spread in the monkey brain of approximately 5–6 mm and 3–4 mm, respectively67. We chose sufficient volumes to cover the hM4Di-positive terminal sites, which had diameters of 5–7 mm and 3–4 mm for the rmCD and MDm, respectively, as measured by increased PET signals. Because the MDm is located close to the midline, we placed the canulae laterally near the MDm so that the injected solution would diffuse into the entirety of the MDm (Fig. 3b). CT image was obtained to visualize the infusion cannulae in relation to the chambers and skull following each infusion. The CT image was overlaid on MR and PET images obtained beforehand using PMOD to verify that the infusion sites (tips of the infusion cannulae) were located in the target (presumed hM4Di-positive terminal regions identified as increased PET signals). The behavioral session began approximately 30 min after the end of the infusion and lasted approximately one hour in both tasks, thus the sessions ended 90 min after the end of the infusions. We performed at most one infusion in one are per week for one area, and repeated five times for each area. The order of the experiments for each monkey, including systemic injections and local infusions into each area, is summarized in Supplementary Fig. 8.
Behavioral tasks
The monkeys were tested with two versions of modified reversal learning tasks in which they were required to choose either of two visual stimuli (out of a set of five) presented on a computer screen. Behavioral testing was conducted in a sound-attenuated room. The monkeys sat on a monkey chair from which they could reach out one with hand to touch an LCD display placed in front of them. The behavioral task was controlled by a computer using commercially available software (Inquisit, Millisecond). A monkey initiated a trial by touching a sensor mounted on the chair, which caused a small white circle to appear in the center of the display. After a delay of 0.5 s, the circle disappeared and two stimuli of the five possible stimuli were presented simultaneously on the left and right side of the display. If the monkey touched either stimulus, it could receive a reward from the spout placed in front of its mouth. If the monkey released the touch lever before the presentation of visual stimuli, the trial was aborted and repeated after a 3–4 s inter-trial interval. Each stimulus was associated with 1, 2, 3, 4, or 5 drops of juice.
In the NOVEL task, a new set of visual stimuli was introduced each session, which required the monkeys to learn a new set of stimulus-reward associations. A daily session consisted of 90 acquisition-phase trials, followed by the reversal of the stimulus-reward associations, and then 210 post-reversal trials. The reversal was conducted such that a stimulus previously associated with 1 drop of juice became associated with 5 drops of juice, and one associated with 2 drops became associated with 4 drops, and vice versa. The combination of visual stimuli seen on each trial was pre-determined pseudorandomly so that each combination appeared once every 10 trials in a round-robin fashion.
In the FAMILIAR task, a fixed sets of five visual stimuli were used throughout the experiments. This ensured that the monkeys became familiar with all possible stimulus-reward associations for the two sets before and after reversals. If the optimal choice rate (i.e., the proportion of trials in which the option associated with the greater reward was chosen) in 30 consecutive trials passed 76%, the associations were reversed. Daily sessions comprised 300 and 400 trials for Mk#1 and Mk#2, respectively. Both monkeys were trained on the NOVEL task and then the FAMILIAR task. As a control, the monkeys were tested on a simple reversal learning task (Supplementary Fig. 4) in which only two novel stimuli were introduced in each session. The stimuli were associated with 5 drops of juice reward or no reward.
One monkey (Mk#2) was also tested on a reinforcer devaluation task and on a modified version of the Wisconsin Card Sorting task, as previously described25,26. Briefly, in the reinforcer devaluation task, the monkey was required to choose one of two objects that were placed above two holes located in a wooden plate. For one set of objects, food 1 (peanut) was delivered when the object was selected, whereas for the other set of objects, food 2 (raisin) was delivered. The associations between the objects and reward type were fixed throughout training. The monkey was tested on 4 consecutive days in a week: review of object-reward associations (Day 1), baseline choice test (Day 2), review of object-reward associations (Day 3), and choice test following selective satiation (devaluation of a food) with vehicle or DCZ administration 15 min before devaluation (Day 4). For devaluation, the monkey was given up to 30 min to consume as much of either food as it wanted. The devaluation procedure was deemed to be complete when the monkey refrained from retrieving food from the food chamber for 5 min. The monkey’s ability to adaptively shift away from choosing objects associated with the devalued food was calculated as the “proportion shifted” as below,
1 |
where F1 and F2 represent choices associated with the two food types (peanut and raisin) in each week in which that food type was devalued, and D and N respectively represent the data for devaluation (Day 4) and baseline (Day 2). A version of the Wisconsin Card Sorting Test, which was modified for monkeys to comprise only two rules (color and shape), was used as previously reported26. Briefly, in each trial, the monkey initiated a trial by touching a sensor mounted on the chair, which caused a small white circle to appear in the center of the display. After a delay of 0.5 s, the circle disappeared and a sample stimulus appeared in the center of the display, followed after 1 s by horizontal (left, center, right) presentation of three test stimuli below the sample stimulus, one matching the sample stimulus in color, another matching in shape, and the third not matching in either color or shape. The sample and test stimuli were randomly chosen from a set of colored shapes (colors = red, yellow, green, and blue; shapes = triangle, circle, square, and cross). The monkeys had to touch the test stimulus that matched the sample in color or in shape within 2000 ms. The monkey was given a reward (drops of juice) after a correct target selection, and the sample and test items were extinguished upon reward delivery. If the animal chose an incorrect option, reward was not given and a white square around the chosen stimulus appeared for 1 s as a visual feedback for error. When the monkey reached 85% correct performance in 20 consecutive trials or 90% correct in 10 consecutive trials, the relevant rule (matching by color or matching by shape) was changed without notice to the monkey. A minimum of 20 trials was provided in each block even if the monkey reached the criterion earlier. Each daily session consisted of 400 trials and the first rule of the day alternated between days.
Statistics
Throughout the manuscript, we compare optimal choice rate and learning rate between conditions. For analysis of optimal choice rate, we divided the session into three phases. For the NOVEL task, the phases were as follows: pre-reversal (Pre; the 90 trials before the reversal), early phase of the post-reversal (PoE; the first 100 trials after the reversal), and late phase of the post-reversal (PoL; the next 110 trials after reversal). For the FAMILIAR task, they were Pre (10 trials before the reversal), PoE (the first 5 trials after reversal), PoL (the next 5 trials after reversal). Then, we first conducted a three-way ANOVA (subject × phase × treatment) to determine whether there was a significant difference in treatment (Vehicle vs. DCZ) or any interaction between phase and treatment, indicating deficits in a specific phase(s). If there were, we subsequently conducted two-way ANOVAs (subject × treatment) for each phase to determine in which phase(s) any significant differences in treatment occurred. For analysis of learning rate, comparisons were analyzed with two-way repeated-measures ANOVA (subject × treatment) to find the effect of each treatment and any individual differences. The analyses were conducted using GraphPad Prism 9. For the NOVEL task, the optimal choice rate was averaged across 10 trials of pseudo-random stimulus-reward combinations (see “Behavioral task”) for each session. In the FAMILIAR task, the optimal choice rate around the time of reversal was averaged across each reversal for each session. Reaction time was defined time between releasing the bar and touching the object on the screen, and data were averaged across 10 trials as described above.
For a learning-model that incorporates an agnostic analysis of the ability to update values based on past experience (Fig. 4), we extracted the trials in the PoE phase following unpredicted positive and negative experiences as follows. An “unpredicted positive” experience was defined as a single choice of higher valued options (4 or 5 drops) following a non-choice of that higher valued option in a preceding trial, and vice-versa (lower valued options, 1 or 2 drops) for negative experiences. Then, the averaged optimal choice rate in the next trials that included that option following these unexpected outcomes was calculated. A concrete example for a positive-experience would be as follows: the monkey is presented with “5-drops-of-juice” and “2-drops-of-juice” options, but does not choose the 5-drops option. Subsequently, it is presented with the choice between 3-drops and 5-drops stimuli and chooses the 5-drops option (this is the unpredicted positive experience). A future trial in which the monkey has to choose between the 1-drop and 5-drops options would be the one extracted for analysis. Similarly, for a negative experience: the monkey is presented with a choice between 1-drop and 2-drops stimuli, and does not choose the 1-drop option. Subsequently, when presented with 1-drop and 4-drop stimuli, it chooses the 1-drop option (this is the unexpected negative experience). A future trial in which the monkey has to choose between 1-drop and 3-drop stimuli would be extracted for analysis.
Histology and immunostaining
For histological inspection, monkeys were deeply anesthetized with an overdose of sodium pentobarbital (80 mg/kg, i.v.) and transcardially perfused with saline at 4 °C, followed by 4% paraformaldehyde in 0.1 M phosphate buffered saline (PBS), pH 7.4. The brain was removed from the skull, postfixed in the same fresh fixative overnight, saturated with 30% sucrose in phosphate buffer (PB) at 4 °C, and then cut serially into 50-μm-thick sections with a freezing microtome. For visualization of immunoreactive GFP signals (co-expressed with hM4Di), a series of every 6th section was immersed in 1% skim milk for 1 h at room temperature and incubated overnight at 4 °C with rabbit anti-GFP monoclonal antibody (1:500, G10362, Thermo Fisher Scientific) in PBS containing 0.1% Triton X-100 and 1% normal donkey serum (S30-100ML, Sigma-Aldrich) for 2 days at 4 °C. The sections were then incubated in the same fresh medium containing biotinylated donkey anti-rabbit IgG antibody (1:1,000; Jackson ImmunoResearch, West Grove, PA, USA) for 2 h at room temperature, followed by avidin-biotin-peroxidase complex (ABC Elite, Vector Laboratories, Burlingame, CA, USA) for 2 h at room temperature. For visualizing the antigen, the sections were reacted in 0.05 M Tris-HCl buffer (pH 7.6) containing 0.04% diaminobenzidine (DAB), 0.04% NiCl2, and 0.003% H2O2. The sections were mounted on gelatin-coated glass slides, air-dried, and cover-slipped. A portion of the other sections was Nissl-stained with 1% Cresyl violet. Images of sections were digitally captured using an optical microscope equipped with a high-grade charge-coupled device (CCD) camera (Biorevo, Keyence, Osaka, Japan).
Model-fitting analysis
We constructed two mathematical models termed “EXP” and “INF” to investigate whether monkey behavior could be explained by experience- or inference-based strategies.
In the EXP model, the probability that the learner (i.e., monkey) chooses stimulus Si when stimuli Si and Sj are presented (i,j = 1,⋯,5; i ≠ j) is given by
2 |
where β is a parameter controlling the exploration-exploitation trade-off (inverse temperature), and Vi is the subjective value of stimulus Si, which is interpreted as the amount of reward expected by the learner after choosing Si. Vi is updated after each trial based on the Rescorla–Wagner rule, which applies the stochastic gradient-descent algorithm to minimize the squared error between the expected reward for the chosen stimulus and the experienced reward. The update rule after choosing stimulus Si is given by
3 |
where α is the learning rate and robs is the amount of the reward provided in the current trial. All Vi (i = 1,⋯,5) were assumed to be zero at the beginning of each session, and the learning rate and inverse temperature were fitted to the behavioral data in each session via maximum likelihood estimation.
In the INF model, we assume that the learner knows that the stimulus-reward association takes one of the two possible patterns. Taking this assumption, we prepared two sets of Vi:
4 |
for pattern A, and
5 |
for pattern B. In the INF model, the learner’s choice is assumed to follow a mixture of the two softmax distributions corresponding to the two possible stimulus-reward association patterns. The probability that stimulus Si is chosen when stimuli Si and Sj are presented is given by
6 |
where β is the inverse temperature and w is a weight parameter adaptively learned based on the experienced rewards. If w = 1, the learner’s choice is assumed to follow the softmax distribution corresponding to stimulus-reward association pattern A, and if w = 0, the choice is assumed to follow the softmax distribution corresponding to stimulus-reward association pattern B. Thus, this weight parameter can be interpreted as a parameter representing the confidence that the current stimulus-reward association is pattern A out of the two possible patterns. Unlike in the EXP model, in the INF model, the inputs to the softmax function are fixed, and the parameter weighting the two softmax distributions is updated after each trial. Like the EXP model, the weight parameter is updated by the update rule derived from the stochastic gradient-descent algorithm, which minimizes the squared error between the expected reward for the chosen stimulus and the experienced reward. The update rule after choosing stimulus Si is given by
7 |
where α is the learning rate and robs is the observed reward in the current trial. w is assumed to be 0.5 at the beginning of each session and is assumed to take a value in the range of [0.0 1.0] at each update by replacing its value with 0.0/1.0 if it is smaller/larger than 0.0/1.0. As in the EXP model, the learning rate and inverse temperature were fitted to the behavioral data in each session via maximum likelihood estimation.
Because we were interested in the updating behavior during the post-reversal period, we modified the two models described above for the current study. In the modified models, the trials in each session were divided into two groups: 1) the trials in the post-reversal period, and 2) others. The post-reversal period consisted of all trials after the reversal in the NOVEL task, and the five trials after each reversal in the FAMILIAR task. While the learning rate in the basic models is fixed in each session, in the modified models we assumed that it took a different value each of the two trial groups. The learning rate for the post-reversal period is considered to better reflect the updating behavior, and a total of three parameters (two learning rates and inverse temperature) were fitted to the given behavioral data. Unless stated otherwise, the learning rates for the post-reversal periods derived from the modified models were reported in this paper.
To compare the ability of the EXP and INF models to explain the behavioral data, we performed Bayesian model comparison. Following the procedure in previous studies68,69, the Bayesian information criterion (BIC) was computed for each model and each session, then the Bayesian posterior probability for each model was computed based on those BIC values.
Reporting summary
Further information on research design is available in the Nature Portfolio Reporting Summary linked to this article.
Supplementary information
Source data
Acknowledgements
This study was supported by MEXT/JSPS KAKENHI Grant Numbers JP18K15353, JP21K07268, JP22H05521 (to K.O.), and, JP17H02219 (to T.H.), JP22H05157 (to K.I.), JP19H05467 (to M.T.), and JP15K21742, JP20H05955 (to T.M.), by JST PRESTO Grant Numbers JPMJPR22S3 (to K.O.), and JPMJPR2128 (to KMajima) and by AMED Grant Numbers JP23dm0307007 (to T.H.), JP21dm0107146 (to T.M.), JP20dm0307021 (to K.I.), and JP21dm0207077 (to M.T.). Two Japanese monkeys were provided by the Japan MEXT National Bio-Resource Project “Japanese Monkeys”. We thank Drs. Brian Russ and Junya Hirokawa for their valuable comments on an earlier version of the manuscript, and Jun Kamei, Ryuji Yamaguchi, Yuichi Matsuda, Yoshio Sugii, Takashi Okauchi, Rie Yoshida, Maki Fujiwara, and Mayuko Nakano for their technical assistance. We also thank Katsushi Kumata and Dr. Ming-Rong Zhang for producing the radioligand.
Author contributions
Conceptualization: T.M.; investigation: K.O., Yukiko.Hori., and Y.N.; Model construction: K.Majima; Resources: K.I. and M.T.; Surgery: M.E. and R.S.; Visualization: K.O.; Writing – review & editing: K.O., K. Majima, Y.N., Yukiko.Hori, T.H., M.E., K. Mimura., N.M., A.F., Yuki Hori, H.I., K.I., R.S., M.T., N.Y., M.H., B.R., and T.H.
Peer review
Peer review information
Nature Communications thanks the anonymous reviewer(s) for their contribution to the peer review of this work. A peer review file is available.
Data availability
The Source data are provided as a Source Data file at https://github.com/minamimoto-lab/2023-Oyama-OFC. Source data are provided with this paper.
Code availability
The Python code used for the analysis is available at https://github.com/minamimoto-lab/2023-Oyama-OFC.
Competing interests
The authors declare no competing interests.
Footnotes
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary information
The online version contains supplementary material available at 10.1038/s41467-024-50505-8.
References
- 1.Sutton, R. S. & Barto, A. G. Reinforcement Learning: An Introduction (MIT Press, 2018).
- 2.Tremblay, L. & Schultz, W. Relative reward preference in primate orbitofrontal cortex. Nature398, 704–708 (1999). 10.1038/19525 [DOI] [PubMed] [Google Scholar]
- 3.Padoa-Schioppa, C. & Assad, J. A. Neurons in the orbitofrontal cortex encode economic value. Nature441, 223–226 (2006). 10.1038/nature04676 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.Tsuchida, A., Doll, B. B. & Fellows, L. K. Beyond reversal: a critical role for human orbitofrontal cortex in flexible learning from probabilistic feedback. J. Neurosci.30, 16868–16875 (2010). 10.1523/JNEUROSCI.1958-10.2010 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.Miller, K. J., Botvinick, M. M. & Brody, C. D. Value representations in the rodent orbitofrontal cortex drive learning, not choice. eLife11, e64575 (2022). 10.7554/eLife.64575 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.Rolls, E. T. The orbitofrontal cortex and reward. Cereb. Cortex10, 284–294 (2000). 10.1093/cercor/10.3.284 [DOI] [PubMed] [Google Scholar]
- 7.Stalnaker, T. A., Cooch, N. K. & Schoenbaum, G. What the orbitofrontal cortex does not do. Nat. Neurosci.18, 620–627 (2015). 10.1038/nn.3982 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Kahnt, T. & Schoenbaum, G. Cross-species studies on orbitofrontal control of inference-based behavior. Behav. Neurosci.135, 109–119 (2021). 10.1037/bne0000401 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.Zhou, J. et al. Rat orbitofrontal ensemble activity contains multiplexed but dissociable representations of value and task structure in an odor sequence task. Curr. Biol.29, 897–907.e893 (2019). 10.1016/j.cub.2019.01.048 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Panayi, M., Khamassi, M. & Killcross, S. The rodent lateral orbitofrontal cortex as an arbitrator selecting between model-based and model-free learning systems. Behav. Neurosci.135, 226–244 (2021). 10.1037/bne0000454 [DOI] [PubMed] [Google Scholar]
- 11.Rudebeck, P. H. & Rich, E. L. Orbitofrontal cortex. Curr. Biol.28, 1083–1088 (2018). 10.1016/j.cub.2018.07.018 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.McFarland, N. R. & Haber, S. N. Thalamic relay nuclei of the basal ganglia form both reciprocal and nonreciprocal cortical connections, linking multiple frontal cortical areas. J. Neurosci.22, 8117–8132 (2002). 10.1523/JNEUROSCI.22-18-08117.2002 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Xiao, D., Zikopoulos, B. & Barbas, H. Laminar and modular organization of prefrontal projections to multiple thalamic nuclei. Neuroscience161, 1067–1081 (2009). 10.1016/j.neuroscience.2009.04.034 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Choi, E. Y., Ding, S. L. & Haber, S. N. Combinatorial inputs to the ventral striatum from the temporal cortex, frontal cortex, and amygdala: Implications for segmenting the striatum. eNeuro4, ENEURO.0392-0317.2017 10.1523/ENEURO.0392-17.2017 (2017). [DOI] [PMC free article] [PubMed]
- 15.Clarke, H. F., Robbins, T. W. & Roberts, A. C. Lesions of the medial striatum in monkeys produce perseverative impairments during reversal learning similar to those produced by lesions of the orbitofrontal cortex. J. Neurosci.28, 10972–10982 (2008). 10.1523/JNEUROSCI.1521-08.2008 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Costa, V. D., Dal Monte, O., Lucas, D. R., Murray, E. A. & Averbeck, B. B. Amygdala and ventral striatum make distinct contributions to reinforcement learning. Neuron92, 505–517 (2016). 10.1016/j.neuron.2016.09.025 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Taswell, C. A., Costa, V. D., Murray, E. A. & Averbeck, B. B. Ventral striatum’s role in learning from gains and losses. Proc. Natl. Acad. Sci. USA115, E12398–E12406 (2018). 10.1073/pnas.1809833115 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Mitchell, A. S., Browning, P. G. F. & Baxter, M. G. Neurotoxic lesions of the medial mediodorsal nucleus of the thalamus disrupt reinforcer devaluation effects in rhesus monkeys. J. Neurosci.27, 11289–11295 (2007). 10.1523/JNEUROSCI.1914-07.2007 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.Chakraborty, S., Kolling, N., Walton, M. E. & Mitchell, A. S. Critical role for the mediodorsal thalamus in permitting rapid reward-guided updating in stochastic reward environments. eLife5, e13588 (2016). 10.7554/eLife.13588 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Rudebeck, P. H., Saunders, R. C., Lundgren, D. A. & Murray, E. A. Specialized representations of value in the orbital and ventrolateral prefrontal cortex: desirability versus availability of outcomes. Neuron95, 1208–1220.e1205 (2017). 10.1016/j.neuron.2017.07.042 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Armbruster, B. N., Li, X., Pausch, M. H., Herlitze, S. & Roth, B. L. Evolving the lock to fit the key to create a family of G protein-coupled receptors potently activated by an inert ligand. Proc. Natl. Acad. Sci. USA104, 5163–5168 (2007). 10.1073/pnas.0700293104 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22.Stachniak, T. J., Ghosh, A. & Sternson, S. M. Chemogenetic synaptic silencing of neural circuits localizes a hypothalamus→midbrain pathway for feeding behavior. Neuron82, 797–808 (2014). 10.1016/j.neuron.2014.04.008 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23.Mahler, S. V. et al. Designer receptors show role for ventral pallidum input to ventral tegmental area in cocaine seeking. Nat. Neurosci.17, 577–585 (2014). 10.1038/nn.3664 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24.Oyama, K. et al. Chemogenetic dissection of the primate prefronto-subcortical pathways for working memory and decision-making. Sci. Adv.7, eabg4246 (2021). 10.1126/sciadv.abg4246 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.Rudebeck, P. H., Saunders, R. C., Prescott, A. T., Chau, L. S. & Murray, E. A. Prefrontal mechanisms of behavioral flexibility, emotion regulation and value updating. Nat. Neurosci.16, 1140–1145 (2013). 10.1038/nn.3440 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Buckley, M. J. et al. Dissociable components of rule-guided behavior depend on distinct medial and prefrontal regions. Science325, 52–58 (2009). 10.1126/science.1172377 [DOI] [PubMed] [Google Scholar]
- 27.Schoenbaum, G., Roesch, M. R., Stalnaker, T. A. & Takahashi, Y. K. A new perspective on the role of the orbitofrontal cortex in adaptive behaviour. Nat. Rev. Neurosci.10, 885–892 (2009). 10.1038/nrn2753 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28.Torregrossa, M. M., Quinn, J. J. & Taylor, J. R. Impulsivity, compulsivity, and habit: the role of orbitofrontal cortex revisited. Biol. Psychiatry63, 253–255 (2008). 10.1016/j.biopsych.2007.11.014 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Setogawa, T. et al. Neurons in the monkey orbitofrontal cortex mediate reward value computation and decision-making. Commun. Biol.2, 126 (2019). 10.1038/s42003-019-0363-0 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30.Pickens, C. L. et al. Different roles for orbitofrontal cortex and basolateral amygdala in a reinforcer devaluation task. J. Neurosci.23, 11078–11084 (2003). 10.1523/JNEUROSCI.23-35-11078.2003 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 31.Murray, E. A., Moylan, E. J., Saleem, K. S., Basile, B. M. & Turchi, J. Specialized areas for value updating and goal selection in the primate orbitofrontal cortex. eLife4, e11695 (2015). 10.7554/eLife.11695 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 32.Kim, D., Park, G. Y., JP, O. D. & Lee, S. W. Task complexity interacts with state-space uncertainty in the arbitration between model-based and model-free learning. Nat. Commun.10, 5738 (2019). 10.1038/s41467-019-13632-1 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33.Huang, Y., Yaple, Z. A. & Yu, R. Goal-oriented and habitual decisions: neural signatures of model-based and model-free learning. Neuroimage215, 116834 (2020). 10.1016/j.neuroimage.2020.116834 [DOI] [PubMed] [Google Scholar]
- 34.McDannald, M. A., Lucantonio, F., Burke, K. A., Niv, Y. & Schoenbaum, G. Ventral striatum and orbitofrontal cortex are both required for model-based, but not model-free, reinforcement learning. J. Neurosci.31, 2700–2705 (2011). 10.1523/JNEUROSCI.5499-10.2011 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35.Daw, N. D., Gershman, S. J., Seymour, B., Dayan, P. & Dolan, R. J. Model-based influences on humans’ choices and striatal prediction errors. Neuron69, 1204–1215 (2011). 10.1016/j.neuron.2011.02.027 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 36.Groman, S. M. et al. Orbitofrontal circuits control multiple reinforcement-learning processes. Neuron103, 734–746.e733 (2019). 10.1016/j.neuron.2019.05.042 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37.Chakraborty, S., Ouhaz, Z., Mason, S. & Mitchell, A. S. Macaque parvocellular mediodorsal thalamus: dissociable contributions to learning and adaptive decision-making. Eur. J. Neurosci.49, 1041–1054 (2019). 10.1111/ejn.14078 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 38.Ouhaz, Z., Perry, B. A. L., Nakamura, K. & Mitchell, A. S. Mediodorsal thalamus is critical for updating during extradimensional shifts but not reversals in the attentional set-shifting task. eNeuro9. 10.1523/ENEURO.0162-21.2022 (2022). [DOI] [PMC free article] [PubMed]
- 39.Mukherjee, A., Lam, N. H., Wimmer, R. D. & Halassa, M. M. Thalamic circuits for independent control of prefrontal signal and noise. Nature600, 100–104 (2021). 10.1038/s41586-021-04056-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40.Rikhye, R. V., Gilra, A. & Halassa, M. M. Thalamic regulation of switching between cortical representations enables cognitive flexibility. Nat. Neurosci.21, 1753–1763 (2018). 10.1038/s41593-018-0269-z [DOI] [PMC free article] [PubMed] [Google Scholar]
- 41.Schultz, W., Tremblay, L., Hollerman, J. R. & Schultz. Reward processing in primate orbitofrontal cortex and basal ganglia. Cerebral Cortex (New York, N.Y.: 1991)10, 272–284 (2000). [DOI] [PubMed]
- 42.Oyama, K. et al. Discrete coding of stimulus value, reward expectation, and reward prediction error in the dorsal striatum. J. Neurophysiol.114, 2600–2615 (2015). 10.1152/jn.00097.2015 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43.Oyama, K., Hernádi, I., Iijima, T. & Tsutsui, K.-I. Reward prediction error coding in dorsal striatal neurons. J. Neurosci.30, 11447–11457 (2010). 10.1523/JNEUROSCI.1719-10.2010 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 44.Stalnaker, T. A., Raheja, N. & Schoenbaum, G. Orbitofrontal state representations are related to choice adaptations and reward predictions. J. Neurosci.41, 1941–1951 (2021). 10.1523/JNEUROSCI.0753-20.2020 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 45.Schoenbaum, G., Nugent, S. L., Saddoris, M. P. & Setlow, B. Orbitofrontal lesions in rats impair reversal but not acquisition of go, no-go odor discriminations. NeuroReport13, 885–890 (2002). 10.1097/00001756-200205070-00030 [DOI] [PubMed] [Google Scholar]
- 46.Dias, R., Robbins, T. W. & Roberts, A. C. Dissociation in prefrontal cortex of affective and attentional shifts. Nature380, 69–72 (1996). 10.1038/380069a0 [DOI] [PubMed] [Google Scholar]
- 47.Rudebeck, P. H. & Murray, E. A. Amygdala and orbitofrontal cortex lesions differentially influence choices during object reversal learning. J. Neurosci.28, 8338–8343 (2008). 10.1523/JNEUROSCI.2272-08.2008 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 48.Rygula, R., Walker, S. C., Clarke, H. F., Robbins, T. W. & Roberts, A. C. Differential contributions of the primate ventrolateral prefrontal and orbitofrontal cortex to serial reversal learning. J. Neurosci.30, 14552–14559 (2010). 10.1523/JNEUROSCI.2631-10.2010 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 49.Rudebeck, P. H. & Murray, E. A. Dissociable effects of subtotal lesions within the macaque orbital prefrontal cortex on reward-guided behavior. J. Neurosci.31, 10569–10578 (2011). 10.1523/JNEUROSCI.0091-11.2011 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 50.Rudebeck, P. H., Mitz, A. R., Chacko, R. V. & Murray, E. A. Effects of amygdala lesions on reward-value coding in orbital and medial prefrontal cortex. Neuron80, 1519–1531 (2013). 10.1016/j.neuron.2013.09.036 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 51.Rudebeck, P. H., Ripple, J. A., Mitz, A. R., Averbeck, B. B. & Murray, E. A. Amygdala contributions to stimulus–reward encoding in the macaque medial and orbital frontal cortex during learning. J. Neurosci.37, 2186–2202 (2017). 10.1523/JNEUROSCI.0933-16.2017 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 52.Pujara, M. S., Ciesinski, N. K., Reyelts, J. F., Rhodes, S. E. V. & Murray, E. A. Selective prefrontal-amygdala circuit interactions underlie social and nonsocial valuation in rhesus macaques. J. Neurosci.42, 5593–5604 (2022). 10.1523/JNEUROSCI.0794-21.2022 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 53.Fiuzat, E. C., Rhodes, S. E. V. & Murray, E. A. The role of orbitofrontal–amygdala interactions in updating action–outcome valuations in macaques. J. Neurosci.37, 2463–2470 (2017). 10.1523/JNEUROSCI.1839-16.2017 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 54.Jang, A. I. et al. The role of frontal cortical and medial-temporal lobe brain areas in learning a Bayesian prior belief on reversals. J. Neurosci.35, 11751–11760 (2015). 10.1523/JNEUROSCI.1594-15.2015 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 55.Taswell, C. A. et al. Effects of amygdala lesions on object-based versus action-based learning in Macaques. Cereb. Cortex31, 529–546 (2020). 10.1093/cercor/bhaa241 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 56.Ghashghaei, H. T. & Barbas, H. Pathways for emotion: interactions of prefrontal and anterior temporal pathways in the amygdala of the rhesus monkey. Neuroscience115, 1261–1279 (2002). 10.1016/S0306-4522(02)00446-3 [DOI] [PubMed] [Google Scholar]
- 57.Zikopoulos, B., Höistad, M., John, Y. & Barbas, H. Posterior orbitofrontal and anterior cingulate pathways to the amygdala target inhibitory and excitatory systems with opposite functions. J. Neurosci.37, 5051–5064 (2017). 10.1523/JNEUROSCI.3940-16.2017 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 58.Ghashghaei, H. T., Hilgetag, C. C. & Barbas, H. Sequence of information processing for emotions based on the anatomic dialogue between prefrontal cortex and amygdala. NeuroImage34, 905–923 (2007). 10.1016/j.neuroimage.2006.09.046 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 59.Fudge, J. L. et al. Beyond the classic VTA: extended amygdala projections to DA-striatal paths in the primate. Neuropsychopharmacology42, 1563–1576 (2017). 10.1038/npp.2017.38 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 60.Frankle, W. G., Laruelle, M. & Haber, S. N. Prefrontal cortical projections to the midbrain in primates: evidence for a sparse connection. Neuropsychopharmacology31, 1627–1636 (2006). 10.1038/sj.npp.1300990 [DOI] [PubMed] [Google Scholar]
- 61.Alexander, G. E. & Crutcher, M. D. Functional architecture of basal ganglia circuits: neural substrates of parallel processing. Trends Neurosci.13, 266–271 (1990). 10.1016/0166-2236(90)90107-L [DOI] [PubMed] [Google Scholar]
- 62.de Kloet, S. F. et al. Bi-directional regulation of cognitive control by distinct prefrontal cortical output neurons to thalamus and striatum. Nat. Commun.12, 1994 (2021). 10.1038/s41467-021-22260-7 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 63.Averbeck, B. B., Lehman, J., Jacobson, M. & Haber, S. N. Estimates of projection overlap and zones of convergence within frontal-striatal circuits. J. Neurosci.34, 9497–9505 (2014). 10.1523/JNEUROSCI.5806-12.2014 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 64.Maia, T. V., Cooney, R. E. & Peterson, B. S. The neural bases of obsessive–compulsive disorder in children and adults. Dev. Psychopathol.20, 1251–1283 (2008). 10.1017/S0954579408000606 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 65.Kimura, K. et al. A mosaic adeno-associated virus vector as a versatile tool that exhibits high levels of transgene expression and neuron specificity in primate brain. Nat. Commun.14, 4762 (2023). 10.1038/s41467-023-40436-1 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 66.Oyama, K. et al. Chemogenetic disconnection between the orbitofrontal cortex and the rostromedial caudate nucleus disrupts motivational control of goal-directed action. J. Neurosci.42, 6267–6275 (2022). 10.1523/JNEUROSCI.0229-22.2022 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 67.Murata, Y. et al. Temporal plasticity involved in recovery from manual dexterity deficit after motor cortex lesion in macaque monkeys. J. Neurosci.35, 84–95 (2015). 10.1523/JNEUROSCI.1737-14.2015 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 68.Li, J.-A. et al. Quantum reinforcement learning during human decision-making. Nat. Hum. Behav.4, 294–307 (2020). 10.1038/s41562-019-0804-2 [DOI] [PubMed] [Google Scholar]
- 69.Stephan, K. E., Penny, W. D., Daunizeau, J., Moran, R. J. & Friston, K. J. Bayesian model selection for group studies. Neuroimage46, 1004–1017 (2009). 10.1016/j.neuroimage.2009.03.025 [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
The Source data are provided as a Source Data file at https://github.com/minamimoto-lab/2023-Oyama-OFC. Source data are provided with this paper.
The Python code used for the analysis is available at https://github.com/minamimoto-lab/2023-Oyama-OFC.