Abstract
The amygdala and orbitofrontal cortex (OFC) are often thought of as components of a neural circuit that assigns affective significance—or value—to sensory stimuli so as to anticipate future events and adjust behavioral and physiological responses. Much recent work has been aimed at understanding the distinct contributions of the amygdala and OFC to these processes, but a detailed understanding of the physiological mechanisms underlying learning about value remains lacking. To gain insight into these processes, we have focused initially on characterizing the neural signals of the primate amygdala, and more recently of the primate OFC, during appetitive and aversive reinforcement learning procedures. We have employed a classical conditioning procedure whereby monkeys form associations between visual stimuli and rewards or aversive stimuli. After learning these initial associations, we reverse the stimulus-reinforcement contingencies, and monkeys learn these new associations. We have discovered that separate populations of neurons in the amygdala represent the positive and negative value of conditioned visual stimuli. This representation of value updates rapidly upon image value reversal, as fast as monkeys learn, often within a single trial. We suggest that representations of value in the amygdala may change through multiple interrelated mechanisms: some that arise from fairly simple Hebbian processes, and others that may involve gated inputs from other brain areas, such as the OFC.
Keywords: amygdala, OFC, orbitofrontal cortex, reinforcement learning, conditioning, learning, reward, aversive, value, monkey
EMOTION, VALUATION, AND REINFORCEMENT LEARNING
In humans, the regulation of emotion is extremely flexible, adapting to different sensory cues, social situations, and cognitive operations, such as the application of rules. How does the brain mediate these different aspects of emotional processing? Most prior efforts to understand emotion at the neural level have employed rodents, often using fear-conditioning and related behavioral paradigms. In human and non-human primates, however, a more flexible control of emotion is thought to be conferred by interactions between the amygdala and prefrontal cortex (PFC).1 Indeed, in primates, as compared to non-primates, there is an extensive elaboration of the PFC and its connections with the amygdala.2 In addition, because the visual system is a dominant sensory modality in primates, there are dense connections among the amygdala, PFC, and visual system. Thus, although many aspects of the function and organization of the amygdala and interconnected structures are conserved across species, amygdala function in primates likely expands upon and differs from processing in rodents in significant ways. For these reasons, it is important to elucidate the complex neural circuitry that regulates emotion in the rhesus monkey. Their rich behavioral and cognitive repertoire makes rhesus monkeys ideal for helping to fill a critical gap between studies in rodents and humans.
One way to approach these questions is to exploit different conditioning procedures developed by experimental psychologists to investigate the neural basis of appetitive and aversive reinforcement learning. During reinforcement learning, subjects learn that particular sensory stimuli are associated with rewards and punishments. Emotional responses, such as excitement or fear, commonly occur upon exposure to sensory stimuli that have been endowed with affective value through reinforcement learning.
Associative learning, such as that which links a sensory stimulus with punishment or reward, is frequently assumed to arise gradually via changing synaptic weights; computational models of reinforcement learning have often worked from this assumption.3 The seminal Rescorla–Wagner model posits that the value representation should be updated on a trial-by-trial basis by “error signals”—that is, signals reflecting the difference between expected and received reinforcement—and recent theories (e.g., temporal difference [TD] models), have extended this model so that reinforcement learning may be described quantitatively in real time.4-6 TD models require a neural representation of value as a function of time, with error signals being computed continuously by taking the difference in the value of situations or “states” at successive time steps. However, more complex forms of conditioning that often occur in nature—such as conditioning in which the value of stimuli changes depending upon context—require representations of value that can be flexibly activated depending on contextual cues or other information. Consider the game of black-jack, whereby being dealt the same card, such as a king, can be rewarding (if it makes a total of 21) or punishing (if it makes a total greater than 21) depending upon the cards dealt earlier in the hand, which define a context that dictates whether receiving the king would be good or bad. TD and other reinforcement learning models, in their simplest form and on their own, may not be able to explain this kind of flexible control of neural activity reflecting value. Instead, such models would need to be extended to incorporate information, such as rules, so that a potentially infinite number of contexts can be interpreted appropriately; furthermore, describing these processes may require mechanisms that operate on a short timescale to gate, or regulate, neural representations of value. Finally, beyond model development, we still need to understand how the brain actually represents value in a flexible manner.
As a first step to understand how value may be represented flexibly, and building upon the theoretical framework of reinforcement learning,4,5,7 we have targeted two brain areas, the amygdala and orbitofrontal cortex (OFC). We focused on these brain areas because of their known anatomic connectivity and because of a long history of research linking them to reinforcement learning and emotional processes.8-10 The amygdala is a structurally and functionally heterogeneous collection of nuclei lying in the anterior medial portion of each temporal lobe.11,12 Sensory information is provided to the amygdala from advanced levels of sensory cortices, the olfactory system, and polysensory brain areas such as perirhinal cortex.11 This information enters the amygdala primarily, but not exclusively, in the lateral nucleus, and then flows—either directly or through multiple synapses—to the basal, accessory basal, central, and other more medial nuclei. Output from the amygdala is directed to a wide range of target structures, including PFC, sensory cortices, the hippocampus, perirhinal cortex, entorhinal cortex, the striatum, and the basal forebrain; and also to subcortical structures responsible for physiological responses related to emotion, such as autonomic responses, hormonal responses, and startle. In general, subcortical projections originate from the central nucleus, and projections to cortex and the striatum originate from the basal, accessory basal and in some cases the lateral nuclei.
Modulation of intrinsic processing in the amygdala can occur via multiple pathways. In particular, OFC projects to numerous amygdala nuclei, including the basal, accessory basal, intercalated masses, and lateral nuclei.13 Other inputs, such as dopaminergic input from the ventral tegmental area and the substantia nigra and serotonergic input from the raphe nuclei, also probably modulate intrinsic amygdala processing, though this has primarily been studied in the rodent.11 Each of these processing streams, as well as others not mentioned, may modulate amygdala processing substantially, potentially helping induce plastic changes important to learning and memory formation, as well as facilitating the expression of physiological and behavioral responses. Thus the amygdala can receive information from all sensory modalities about conditioned and unconditioned stimuli (CSs and USs), and also from structures that might transmit instructive or supervisory signals to the amygdala, such as midbrain dopamine neurons or PFC, especially the OFC.13,14
THE ROLE OF OFC AND THE AMYGDALA
Prior neurophysiological studies in rodent amygdala have suggested that the amygdala supplies neural signals involved in associative learning.15-25 Most of these studies have been conducted in the context of fear conditioning or related tasks, and they have revealed that amygdala neural activity changes during fear conditioning. Compared with investigations in rodents, only a few studies have investigated amygdala neurophysiology in primates,26-32 and results were often conflicting. Prior to our recent paper,33 there had been no systematic study of the neurophysiological properties of primate amygdala neurons during classical conditioning with a well-controlled experimental design.
The OFC, which comprises much of the ventral surface of the frontal lobe, is an area that also has frequently been implicated in the control of emotional behavior. It receives extensive innervation from the amygdala, hippocampus, striatum, and hypothalamus, as well as from many other cortical areas.34,35 The amygdala sends projections throughout OFC, but most densely to the caudal areas (e.g., area 13).36 Similarly, OFC (mainly area 13, but also areas 12o, 14 and 11) sends projections to several nuclei of the amygdala, including the basal and lateral nuclei, which partially overlaps with input from temporal cortices.13,14,37 OFC may therefore modulate the processing of sensory information in the amygdala.
The OFC, together with other parts of the PFC, the striatum and amygdala, is thought to assign values to stimuli, which in turn can contribute to emotional responses and decision making.38,39 Human patients with lesions of the OFC have abnormal emotional and social behavior (including disinhibition), are impaired on decision-making tasks, and have an inability to generate normal physiological reactions to negatively CSs.38,40,41 Lesion studies in non-human primates have been consistent with these findings; for example, monkeys with OFC lesions, or even an effective “disconnection” of OFC and amygdala, are impaired on a reinforcer-devaluation paradigm.42,43 Functional imaging studies have revealed OFC involvement in anticipating and acting upon positive and negative outcomes.44-46
Studies of OFC neurophysiology in rats and monkeys have suggested a role for the OFC in associative learning.16,47-51 Nearly all of these studies have used instrumental tasks, that is, tasks in which the subject must perform (or actively refrain from performing) an action in order to obtain reward or avoid punishment. In these types of tasks, aversive stimuli do not occur after learning occurs; therefore, one cannot conclude that neural activity recorded during task performance represents the anticipation of aversive stimuli. Conclusions about the encoding of negative value are best supported by tasks in which aversive stimuli always occur, even after learning. However, the responses of OFC neurons have not been examined using classical aversive conditioning, in which the CS is consistently predictive of punishment.
Recent primate studies have largely focused on operant forms of appetitive conditioning. Many OFC neurons develop responses to visual cues that predict reward, as well as anticipatory responses to uncued predictable rewards, which are often modulated by the type and amount of reward anticipated.52-56 These signals have often been interpreted as reflecting expectation of the value of the anticipated reward, although two new studies have called into question whether all aspects of value are integrated in the OFC.57,58 Furthermore, one recent study suggests that OFC in rodents is involved in updating value representations during procedures containing Pavlovian features, but not during procedures containing only instrumental procedures.59 Interestingly, several studies have suggested that reward signals in OFC are context dependent—that is, some OFC neurons respond differentially to a stimulus predicting a particular reward depending upon the available alternatives.52,55,56 Overall, OFC carries information about the identity and value of rewards that are available, rewards that are expected, and rewards that are actually received. Considered together with the OFC’s anatomic connectivity, this makes OFC a prime candidate for the flexible modulation of value representations for stimuli associated with reward or punishment.
In the studies described here, we have employed both appetitive and aversive classical conditioning procedures to ask two fundamental questions: (1) What information is represented by neurons in the amygdala and OFC during reinforcement learning? and (2) How are these response properties related to behavioral learning?
HOW IS VALUE REPRESENTED IN THE AMYGDALA?
Our initial studies have been aimed at describing neural signals in the amygdala during the learning and reversal of affective associations. Monkeys performed a trace-conditioning task in which novel, abstract visual stimuli were followed by USs: either a liquid reward, nothing, or an aversive air-puff directed at the face (Fig. 1A). Trace conditioning is a version of classical conditioning in which a brief temporal gap is inserted between CS offset and US onset.60,61 It is possible that differential responses to CSs associated with rewards and punishments could be attributed to the sensory properties of the CSs themselves, rather than to their reinforcement contingencies; therefore, after initial learning occurred, without warning, and at a variable trial number across experiments, we reversed the contingencies of the images initially associated with rewards and air-puffs, and the monkeys learned the new associations. The monkeys demonstrated their learning by licking a spout in anticipation of a liquid reward, or closing their eyes—a defensive behavior—in anticipation of an air-puff. Note that in our task, CSs predicted USs with 100% certainty, so aversive stimuli could not be avoided.
FIGURE 1.

The trace-conditioning task and localization of recording sites in the amygdala. (A) Sequence of events in the three trial types during trace-conditioning. Top and bottom squares: images that reverse values from positive to negative or vice-versa. Middle square: image does not reverse and was always non-reinforced. (B) Coronal MRI acquired with a two-dimensional (2D) spoiled gradient recalled acquisition (SPGR) sequence in monkey V. The susceptibility artifact from a tungsten microelectrode dorsal to the amygdala (circled) is visible. (C–F) Coronal MRI with 2D inversion recovery (IR) sequence (C, arrows point to the electrode artifact, which is less evident). Magnified images show the recording site locations (slice in F is immediately posterior to E). The arrow in D corresponds to a possible border of the lateral nucleus, which contains a fiber tract. Recording sites spanning 2 mm in the anterior–posterior dimension were collapsed onto each image slice. In many cases, this resulted in the superposition of multiple cells with different properties (the key above F gives the properties denoted by symbols: “+” denotes positive value-coding, “−” denotes negative value-coding, and “no” symbol indicates no value-coding). Recording sites from monkey P occurred in an overlapping region of the amygdala. (From Paton et al.33 Reproduced by permission.) (In color in Annals online.)
While monkeys performed the trace-conditioning task, we recorded the activity of individual amygdala neurons. We hypothesized that neurons encoding value would change their response profile to the same CSs once the US associated with them switched. Furthermore, we hypothesized that separate populations of neurons would preferentially respond to positive and negative values, respectively, and that these neurons would rapidly update their responses to CSs during learning. We defined “value” operationally, with positive and negative “CS value-coding neurons” referring to neurons whose activity was higher in response to a CS paired with rewards or a CS paired with aversive air-puffs, respectively.
We found that some amygdala neurons encoded positive value, whereas other amygdala neurons encoded negative value. Figure 2 shows an experiment in which we recorded a neuron encoding positive value. Anticipatory licking and blinking behavior (Fig. 2A,B) demonstrated that the monkey learned about the value of the images. We scored every trial according to whether the monkey licked or blinked during the last 500 ms of the trace interval. For both images, licking response rates were greater, and blinking response rates lower, when an image was positive than when the same image was negative. In this experiment, activity in the neuron under study was higher during the trace interval when images had a positive value compared to when the same images had a negative value (Fig. 2C–F), typical of a positive value-coding neuron. Figure 3 shows the results from another experiment, in which we recorded the activity of a neuron encoding negative value, predominantly during the visual stimulus interval. Neurons encoding positive and negative value were dispersed throughout our recording sites in the amygdala, which largely spanned the lateral, basal, accessory basal, and central nuclei, as estimated by reconstructing recording sites with MRI (Fig. 1B–F). These recording sites overlapped those used in prior monkey neurophysiology studies of the amygdala.26-32
FIGURE 2.

Neural activity from a single amygdalar neuron that encoded positive value during learning, in relation to behavioral learning. (A,B) Behavioral performance. Cumulative (curves) and trial-by-trial (tick marks) licking (red) and blinking (blue) responses, plotted as a function of trial number for images 1 and 2; black dots represent change points. Value reversals occurred at the vertical green lines. (C–F) Rasters and peri-stimulus time histograms (PSTHs) for the amygdala cell recorded during the same experiment. The plots are truncated at US delivery. Each dot represents one action potential, and each row of dots represents the timing of action potentials during one trial. PSTHs sum and average activity across trials, smoothing with a 10-ms moving average of activity. Blue ticks indicate fixation point onset; red ticks indicate visual stimulus onset/offset. (G,H) Spike count and cumulative spike count during the trace interval for the cell depicted in C–F, plotted as a function of trial number separately for images 1 and 2. Red dots show change points that demarcate the onset of a significant change in neural firing rate. (From Paton et al.33 Reproduced by permission.) (In color in Annals online.)
FIGURE 3.

Neural activity from an amygdala neuron encoding negative value in relation to behavioral learning. (A,B) Behavioral performance. (C–F) Rasters and peri-stimulus time histograms (PSTHs) for the neuron recorded during the same experiment as the behavior shown in A,B. (Gs,H) Change-point analysis of neural responses during the visual stimulus epoch. All labeling conventions in this figure follow those shown in Figure 2. (In color in Annals online.)
The activity of neurons encoding positive and negative values was not related to motor responses. When neural activity was aligned on licking and blinking onset, no amygdala neurons had neural responses that were related to either motor action.33 Thus value-related response properties are not well described as representing the link between a CS and a particular response. Moreover, many neurons responded to both rewards and air-puff, suggesting that these cells do not simply represent a link between CS and the sensory modality of the US it is associated with. Finally, it is difficult to account for the value-related signals as simply being related to autonomic reactivity. In general, arousing stimuli of both positive and negative valences can trigger autonomic responses.62
To understand how neural activity in the amygdala might relate to activity in other brain areas, such as the OFC, it is critical to characterize the response dynamics of amygdala neurons during learning so that these dynamics can be compared across areas. In particular, it is important to know how rapidly amygdala neurons learn in relation to changes in reinforcement contingencies, and how rapidly amygdala neurons can encode value after the onset of a CS. Therefore, to examine how individual neurons changed their response level when a CS reversed its value, we applied a change-point test63 to neural responses that reflected value during either the trace (Fig. 2) or visual stimulus (Fig. 3) time epochs. The change-point test identifies the onset of a significant change in response rate in relation to the reversal of image value (P<0.05), represented graphically by a change in slope of the cumulative record of responses.63 Inspection of the data for the neuron depicted in Figure 2 reveals that neural activity changed over the course of a number of trials after the change point for Image 1, but activity changed in a single trial for Image 2. We used the same change-point test to detect the onset of changes in behavioral responses in this experiment (Fig. 2A,B). For the experiments depicted in Figures 2 and 3, changes in neural activity and in behavioral responses indicative of learning occurred at about the same time.
Across experiments, behavioral learning of image value reversals was correlated with changes in neural activity. Figure 4A,B shows the change points of neural activity plotted against the corresponding change points for the behavioral data. Each data point compares a neural activity change point with either a blinking or licking change point. The distributions of licking and blinking change points were not significantly different from one another, so they were combined here (P>0.25, t-test). For both monkeys tested, and in both the visual stimulus and trace intervals, behavioral learning was significantly correlated with changes in neural activity (visual stimulus interval: monkey V, P=0.02, r=0.24; monkey P, P<10−5, r=0.66; trace interval: monkey V, P<10−5, r =0.63; monkey P, P<10−5, r=0.57). Moreover, there was not a significant difference between neural and behavioral change points in all cases (paired t-test, P>0.05). Finally, change points during the visual stimulus and trace intervals were not significantly different from each other (P>0.1, t-test).
FIGURE 4.

Neural activity changes as fast as behavioral learning. (A,B) Change points from neural data in the visual stimulus and trace intervals plotted against the change points found for licking or blinking responses reflecting learning. Histograms show that on average there is no difference between neural and behavioral change points. Data and regression lines for monkey V are shown in blue; data and regression lines for monkey P are shown in red. (C) Average normalized neural activity and behavioral responses plotted as a function of trial number relative to the reversal in image value. Shaded regions indicate 95% prediction intervals for best-fit Weibull functions. Behavioral and neural learning curves overlap. (D) Neural responses do not change to the non-reinforced image, as shown by applying the same analysis done for C to the trials containing non-reinforced images. Shaded regions show SEM of data points. (From Paton et al.33 Reproduced by permission.) (In color in Annals online.)
The tight correlation between the onset of changes in behavior and neural activity suggested that the time course of behavioral and neural learning was similar. We confirmed this by comparing, across neurons encoding image value, the time course of average neural responses with the time course of average behavioral responses. Figure 4C shows the normalized and then averaged neural activity and behavior from the 20 trials before and after the value reversal of each image. The data were fit with sigmoidal functions to construct “neural” and “behavioral” learning curves. The time courses of these curves were quite similar and statistically indistinguishable. Moreover, the changing activity and behavior was specific for the images that changed image value, as demonstrated by the same analysis applied to the same cells from the trials with non-reinforced images (Fig. 4D). Thus, the dynamics of behavioral learning could be accounted for if monkeys based their decisions to lick or blink on the evolving representation of value in the amygdala.
It is also worth emphasizing that both neural activity and behavior changed very rapidly after image value reversal, becoming asymptotic within 4–10 trials, on average. These data imply that amygdala neurons update their response profile on a short timescale, since changes in their activity occur within very few trials of a change in image value. Amygdala neurons may update their representation of value rapidly by virtue of receiving information about reinforcement that does not match expectations; indeed, amygdala responses to reinforcement are often stronger when rewards or aversive stimuli occur unexpectedly, such as immediately after a reversal.64 Neural signals representing reinforcement in relation to expectation could come to the amygdala from a variety of subcortical and cortical pathways. In particular, OFC is a candidate for providing rapid, flexible regulation of amygdala neural activity, consistent with a proposed role for OFC in helping account for amygdala response flexibility during reversal learning.65
In addition to rapidly updating its representation during learning of reversed CS-US contingencies, amygdala neurons provide a temporally extended representation of value that appears shortly after visual CS onset and spans until the time of reinforcement. To quantify this, we used a receiver operating characteristic (ROC) analysis66 to estimate the extent to which activity was different before and after an identified change point (Fig. 5). By convention, ROC values > 0.5 indicated activity that was greater when an image was associated with reward, and values < 0.5 indicated activity that was greater when an image was associated with air-puff. Using this approach, we characterized how neurons represent value across time during a trial by repeating the ROC analysis in consecutive overlapping time-windows of 100 ms (advanced in steps of 20 ms). Figure 5 shows how each value-coding cell represented the value of images as a function of time within trials (each row corresponds to how a single cell represented the value of a single image, as quantified by the ROC analysis). On average, for positive-coding cells, the first bin significantly greater than 0.5 occurred 80–180 ms after visual stimulus onset; all subsequent bins were also significantly greater than 0.5 (downward white arrow, P<0.05, t-test). The first significant bin for neurons encoding negative value was 120–220 ms after visual stimulus onset (upward white arrow). This representation of value over time could correspond to the sort of representation posited by models of reinforcement learning to be required for computations of prediction error signals.6 However, the temporal dynamics of amygdala neurons have not been compared quantitatively during learning with neurons encoding prediction errors, such as dopamine neurons.6
FIGURE 5.

The temporally extended representation of value in the amygdala. Value signals in the amygdala plotted as a function of time, with positive and negative value-coding neurons represented by different colors. Each row in the color map shows how value was represented by a neuron during the presentation of a single image. Positive and negative cell rows are sorted in opposite order according to the latency of the first post-visual-stimulus data point significantly different from 0.5 (P < 0.05, permutation test). The white curves depict the mean ROC values across the populations of neurons encoding positive and negative value. Time 0 corresponds to the start of the bin spanning from 0–99 ms after visual stimulus onset. White arrows indicate the first bins significantly different from 0.5 for the mean ROC values (P < 0.05, t-test). (Adapted from Fig. 3D in Paton et al.33) (In color in Annals online.)
HOW IS VALUE REPRESENTED IN PRIMATE OFC?
Given the anatomic interconnections between the amygdala and OFC, a significant challenge for neuroscientists is to understand how neural activity in the two brain areas is interrelated during appetitive and aversive reinforcement conditioning. We have therefore been interested in comparing neural signals in the OFC and amygdala during learning induced by classical conditioning. How and when might signals representing value develop during learning in each area, and what is the timing of these signals relative to each other? OFC is frequently thought of as being critical for reversal learning, but are the response dynamics of OFC neurons appropriate for regulating amygdala responses during our task?
To address these questions, we have been recording simultaneously from individual neurons in the OFC (primarily area 13) and amygdala while monkeys performed a trace-conditioning task similar to the one described above. We find that the responses of neurons in the OFC, like those in the amygdala, can be modulated by image value during one or more periods of the task.67 Value-related signals in the amygdala and OFC develop with a range of overlapping latencies after CS onset. Furthermore, both OFC and amygdala neurons can rapidly change their activity during reversal learning, with the time course appearing to be similar in a small data set.67 Both OFC and the amygdala therefore contain signals that could underlie appetitive and aversive learning during classical conditioning. On the basis of these preliminary data, amygdala and OFC may be seen to represent closely connected and interrelated components of a neural circuit that assigns affective significance to sensory stimuli, and the two areas may largely act in unison during learning.
SUMMARY AND REFLECTIONS ON THE REGULATION OF VALUE REPRESENTATIONS IN THE BRAIN
Both the amygdala and OFC have long been hypothesized to play a role in aspects of emotional learning and behavior. We are investigating the neurophysiology of the amygdala and OFC during learning induced by appetitive and aversive classical conditioning, and during the reversal of learned reinforcement contingencies. We have discovered that the amygdala contains different populations of neurons, some that respond more strongly to a CS associated with a reward, and others that respond more strongly to a CS associated with an aversive stimulus. This representation of CS “value” is rapidly updated during reversal of contingencies. In some cells, and for some CSs, it occurs within a single trial, and on average it occurs within 4–10 trials, at the same time as changes in behavior indicative of learning. In addition, the representation of value appears rapidly after CS onset. Similar signals appear to be present in OFC. We are in the process of characterizing the rates of learning in OFC neurons compared to amygdala neurons, and the relative latency and duration of value-related signaling in the two brain areas. If one brain area is driving the other, then learning rates should be faster, and latencies shorter, in that brain area. In general we are interested in determining the different roles of OFC and amygdala during learning induced by classical conditioning. Does the amygdala or OFC drive learning in the other brain area, or do the two brain areas function in a parallel fashion? Do distinct response properties in each area depend upon input from the other brain area? What mechanisms are responsible for the flexible representation of value in these brain areas?
The data presented here and in our recent paper,33 characterize the neurophysiological properties of primate amygdala neurons during appetitive and aversive classical conditioning. Prior work provided conflicting data about whether primate amygdala neurons rapidly changed their response properties upon reversal of reinforcement contingencies using instrumental tasks,28,29,68 but analogous instrumental tasks in rodents had shown that amygdala neurons were sensitive to reversals in contingencies.16,48 Our use of classical conditioning procedures (rather than avoidance tasks, where the aversive stimuli do not occur after learning), our use of novel CSs in every experiment, and our tighter requirements for visual fixation all may have contributed to our identification of a flexible representation of value in primate amygdala.
Several lines of experiments have suggested that the OFC is itself a critical structure for reversal learning. In monkeys, lesions of OFC disrupt performance on an operant task involving stimulus-reward reversal learning.42 Similar findings have been reported in human clinical populations.69 Moreover, selective lesions in primate amygdala, in contrast to OFC, do not appear to impair performance on a similar operant task.70 Overall, these results mirror the effects of OFC lesions reported in rodents on a different operant task in which animals learned to use cues to acquire rewards and to avoid aversive stimuli.71 Interestingly, rat OFC lesions decreased the proportion of amygdala neurons that reversed cue-related responses during this task. The extent of homology between the parts of rat OFC that were lesioned in these studies, and those which we are studying physiologically in monkeys, needs to be determined. Nonetheless, these studies raise the possibility that the flexible representation of value in the amygdala may depend, directly or indirectly, on OFC input. If this is the case, however, the regulation must be sufficiently rapid to account for the largely overlapping temporal dynamics of amygdala and OFC neural activity during learning. Furthermore, it is worth noting that the tasks establishing that OFC is required for reversal learning have all been operant tasks, and it remains unclear how amygdala and OFC lesions would affect performance on a classical conditioning task that includes a reversal of contingencies, such as the one we have employed.
Our experiments have revealed that the primate amygdala provides a flexible representation of the positive and negative value of visual CSs. The rate of learning exhibited by cells, however, was variable. Even the same cell could learn at different rates for different CSs (see, for example, the cell in Fig. 2). This raises the possibility that combinations of different types of mechanisms may be employed in regulating the representation of value in the amygdala. At one extreme, synaptic weights may gradually change during learning through a Hebbian mechanism; this is possible because the amygdala is anatomically well-situated to receive the required coincident input from sensory stimuli and reinforcers. At the other extreme, gated inputs might dictate the response level of amygdala neurons, perhaps explaining how single-trial reversal learning could occur. For a model of this type of flexible learning, we might look to supervised response modulation.72,73 In this scheme, a neural network is set up to perform a certain function (e.g., compute expected reinforcement); based on the result of the network’s performance, a supervising neural circuit will make direct adjustments to the response properties of the neurons in the network. By interacting with Hebbian plasticity in the synapses of the main network, this kind of supervising circuit can result in efficient learning, even when multiple outputs are required. Thus, gradually changing synaptic weights are not the only physiologically plausible mechanisms for modulating the output of amygdala neurons. OFC could constitute a part of the supervising circuit that may help control representations of value in the amygdala.
Acknowledgments
We thank S. Dashnaw and J. Hirsch for MRI support. This work was supported by the Keck Foundation, grants from the NIMH, and the Klingenstein, Sloan, James S. McDonnell, Gatsby and NARSAD Foundations, and by a Charles E. Culpeper Scholarship award from Goldman Philanthropic Partnerships to C.D.S. J.J.P. received support from NICHD and NEI institutional training grants. S.E.M. received support from an NSF graduate research fellowship.
References
- 1.Ochsner KN, Gross JJ. The cognitive control of emotion. Trends Cogn Sci. 2005;9:242–249. doi: 10.1016/j.tics.2005.03.010. [DOI] [PubMed] [Google Scholar]
- 2.Murray EA. Neuropsychology of primate reward processes. In: Squire LR, editor. New Encyclopedia of Neuroscience. Elsevier; 2008. In press. [Google Scholar]
- 3.Pennartz CM. Reinforcement learning by Hebbian synapses with adaptive thresholds. Neuroscience. 1997;81:303–319. doi: 10.1016/s0306-4522(97)00118-8. [DOI] [PubMed] [Google Scholar]
- 4.Rescorla RA, Wagner AR. A theory of Pavlovian conditioning: variations in the effectiveness of reinforcement and non-reinforcement. In: Black AH, Prokasy WF, editors. Classical Conditioning II: Current Research and Theory. Appleton Century Crofts; New York: 1972. pp. 64–99. [Google Scholar]
- 5.Sutton R, Barto A. Reinforcement Learning. MIT Press; Cambridge, MA: 1998. [Google Scholar]
- 6.Schultz W, Dayan P, Montague PR. A neural substrate of prediction and reward. Science. 1997;275:1593–1599. doi: 10.1126/science.275.5306.1593. [DOI] [PubMed] [Google Scholar]
- 7.Pearce J, Hall G. A model for Pavlovian conditioning: variations in the effectiveness of conditioned but not unconditioned stimuli. Psychol Rev. 1980;87:532–552. [PubMed] [Google Scholar]
- 8.LeDoux JE. Emotion circuits in the brain. Annu Rev Neurosci. 2000;23:155–184. doi: 10.1146/annurev.neuro.23.1.155. [DOI] [PubMed] [Google Scholar]
- 9.Everitt BJ, Cardinal RN, Parkinson JA, Robbins TW. Appetitive behavior: impact of amygdala-dependent mechanisms of emotional learning. Ann N Y Acad Sci. 2003;985:233–250. [PubMed] [Google Scholar]
- 10.Baxter M, Murray EA. The amygdala and reward. Nat Rev Neurosci. 2002;3:563–573. doi: 10.1038/nrn875. [DOI] [PubMed] [Google Scholar]
- 11.Amaral D, Price J, Pitkanen A, Carmichael S. Anatomical organization of the primate amygdaloid complex. In: Aggleton J, editor. The Amygdala: Neurobiological Aspects of Emotion, Memory, and Mental Dysfunction. Wiley–Liss; New York: 1992. pp. 1–66. [Google Scholar]
- 12.Swanson LW, Petrovich GD. What is the amygdala? Trends Neurosci. 1998;21:323–331. doi: 10.1016/s0166-2236(98)01265-x. [DOI] [PubMed] [Google Scholar]
- 13.Ghashghaei HT, Barbas H. Pathways for emotion: interactions of prefrontal and anterior temporal pathways in the amygdala of the rhesus monkey. Neuroscience. 2002;115:1261–1279. doi: 10.1016/s0306-4522(02)00446-3. [DOI] [PubMed] [Google Scholar]
- 14.Stefanacci L, Amaral DG. Some observations on cortical inputs to the macaque monkey amygdala: an anterograde tracing study. J Comp Neurol. 2002;451:301–323. doi: 10.1002/cne.10339. [DOI] [PubMed] [Google Scholar]
- 15.Repa JC, et al. Two different lateral amygdala cell populations contribute to the initiation and storage of memory. Nat Neurosci. 2001;4:724–731. doi: 10.1038/89512. [DOI] [PubMed] [Google Scholar]
- 16.Schoenbaum G, Chiba AA, Gallagher M. Neural encoding in orbitofrontal cortex and basolateral amygdala during olfactory discrimination learning. J Neurosci. 1999;19:1876–1884. doi: 10.1523/JNEUROSCI.19-05-01876.1999. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Pascoe JP, Kapp BS. Electrophysiological characteristics of amygdaloid central nucleus neurons during Pavlovian fear conditioning in the rabbit. Behav Brain Res. 1985;16:117–133. doi: 10.1016/0166-4328(85)90087-7. [DOI] [PubMed] [Google Scholar]
- 18.Quirk GJ, Repa C, LeDoux JE. Fear conditioning enhances short-latency auditory responses of lateral amygdala neurons: parallel recordings in the freely behaving rat. Neuron. 1995;15:1029–1039. doi: 10.1016/0896-6273(95)90092-6. [DOI] [PubMed] [Google Scholar]
- 19.Applegate CD, Frysinger RC, Kapp BS, Gallagher M. Multiple unit activity recorded from amygdala central nucleus during Pavlovian heart rate conditioning in rabbit. Brain Res. 1982;238:457–462. doi: 10.1016/0006-8993(82)90123-8. [DOI] [PubMed] [Google Scholar]
- 20.Goosens KA, Hobin JA, Maren S. Auditory-evoked spike firing in the lateral amygdala and Pavlovian fear conditioning: mnemonic code or fear bias? Neuron. 2003;40:1013–1022. doi: 10.1016/s0896-6273(03)00728-1. [DOI] [PubMed] [Google Scholar]
- 21.Collins DR, Pare D. Differential fear conditioning induces reciprocal changes in the sensory responses of lateral amygdala neurons to the CS(+) and CS(-) Learn Mem. 2000;7:97–103. doi: 10.1101/lm.7.2.97. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22.Maren S, Quirk GJ. Neuronal signaling of fear memory. Nat Rev Neurosci. 2004;5:844–852. doi: 10.1038/nrn1535. [DOI] [PubMed] [Google Scholar]
- 23.Schoenbaum G, Chiba AA, Gallagher M. Orbitofrontal cortex and basolateral amygdala encode expected outcomes during learning. Nat Neurosci. 1998;1:155–159. doi: 10.1038/407. [DOI] [PubMed] [Google Scholar]
- 24.Maren S. Auditory fear conditioning increases CS-elicited spike firing in lateral amygdala neurons even after extensive overtraining. Eur J Neurosci. 2000;12:4047–4054. doi: 10.1046/j.1460-9568.2000.00281.x. [DOI] [PubMed] [Google Scholar]
- 25.Rorick-Kehn LM, Steinmetz JE. Amygdalar unit activity during three learning tasks: eyeblink classical conditioning, Pavlovian fear conditioning, and signaled avoidance conditioning. Behav Neurosci. 2005;119:1254–1276. doi: 10.1037/0735-7044.119.5.1254. [DOI] [PubMed] [Google Scholar]
- 26.Sugase-Miyamoto Y, Richmond BJ. Neuronal signals in the monkey basolateral amygdala during reward schedules. J Neurosci. 2005;25:11071–11083. doi: 10.1523/JNEUROSCI.1796-05.2005. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Dehaene S, Changeux JP. Reward-dependent learning in neuronal networks for planning and decision making. Prog Brain Res. 2000;126:217–229. doi: 10.1016/S0079-6123(00)26016-0. [DOI] [PubMed] [Google Scholar]
- 28.Sanghera MK, Rolls ET, Roper-Hall A. Visual responses of neurons in the dorsolateral amygdala of the alert monkey. Exp Neurol. 1979;63:610–626. doi: 10.1016/0014-4886(79)90175-4. [DOI] [PubMed] [Google Scholar]
- 29.Nishijo H, Ono T, Nishino H. Single neuron responses in amygdala of alert monkey during complex sensory stimulation with affective significance. J Neurosci. 1988;8:3570–3583. doi: 10.1523/JNEUROSCI.08-10-03570.1988. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30.Nakamura K, Mikami A, Kubota K. Activity of single neurons in the monkey amygdala during performance of a visual discrimination task. J Neurophysiol. 1992;67:1447–1663. doi: 10.1152/jn.1992.67.6.1447. [DOI] [PubMed] [Google Scholar]
- 31.Wilson FA, Rolls ET. The primate amygdala and reinforcement: a dissociation between rule-based and associatively-mediated memory revealed in neuronal activity. Neuroscience. 2005;133:1061–1072. doi: 10.1016/j.neuroscience.2005.03.022. [DOI] [PubMed] [Google Scholar]
- 32.Rolls E. Neurophysiology and functions of the primate amygdala. In: Aggleton J, editor. The Amygdala: Neurobiological Aspects of Emotion, Memory, and Mental Dysfunction. Wiley–Liss; New York: 1992. pp. 143–166. [Google Scholar]
- 33.Paton J, Belova M, Morrison S, Salzman C. The primate amygdala represents the positive and negative value of visual stimuli during learning. Nature. 2006;439:865–870. doi: 10.1038/nature04490. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34.Haber SN, Kunishio K, Mizobuchi M, Lynd-Balta E. The orbital and medial prefrontal circuit through the primate basal ganglia. J Neurosci. 1995;15:4851–4867. doi: 10.1523/JNEUROSCI.15-07-04851.1995. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35.Cavada C, Company T, Tejedor J, et al. The anatomical connections of the macaque monkey orbitofrontal cortex: a review. Cereb Cortex. 2000;10:220–242. doi: 10.1093/cercor/10.3.220. [DOI] [PubMed] [Google Scholar]
- 36.Carmichael ST, Price JL. Limbic connections of the orbital and medial prefrontal cortex in macaque monkeys. J Comp Neurol. 1995;363:615–641. doi: 10.1002/cne.903630408. [DOI] [PubMed] [Google Scholar]
- 37.Stefanacci L, Amaral DG. Topographic organization of cortical inputs to the lateral nucleus of the macaque monkey amygdala: a retrograde tracing study. J Comp Neurol. 2000;421:52–79. doi: 10.1002/(sici)1096-9861(20000522)421:1<52::aid-cne4>3.0.co;2-o. [DOI] [PubMed] [Google Scholar]
- 38.Bechara A, Damasio H, Damasio AR. Emotion, decision making and the orbitofrontal cortex. Cereb Cortex. 2000;10:295–307. doi: 10.1093/cercor/10.3.295. [DOI] [PubMed] [Google Scholar]
- 39.Bechara A, Damasio H, Damasio AR. Role of the amygdala in decision-making. Ann N Y Acad Sci. 2003;985:356–369. doi: 10.1111/j.1749-6632.2003.tb07094.x. [DOI] [PubMed] [Google Scholar]
- 40.Bechara A, Tranel D, Damasio H, Damasio AR. Failure to respond autonomically to anticipated future outcomes following damage to prefrontal cortex. Cereb Cortex. 1996;6:215–225. doi: 10.1093/cercor/6.2.215. [DOI] [PubMed] [Google Scholar]
- 41.Chow TW. Personality in frontal lobe disorders. Curr Psychiatry Rep. 2000;2:446–451. doi: 10.1007/s11920-000-0031-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 42.Izquierdo A, Suda RK, Murray EA. Bilateral orbital prefrontal cortex lesions in rhesus monkeys disrupt choices guided by both reward value and reward contingency. J Neurosci. 2004;24:7540–7548. doi: 10.1523/JNEUROSCI.1921-04.2004. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43.Baxter M, Parker A, Lindner CC, et al. Control of response selection by reinforcer value requires interaction of amygdala and orbital prefrontal cortex. J Neurosci. 2000;20:4311–4319. doi: 10.1523/JNEUROSCI.20-11-04311.2000. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 44.Cox SM, Andrade A, Johnsrude IS. Learning to like: a role for human orbitofrontal cortex in conditioned reward. J Neurosci. 2005;25:2733–2740. doi: 10.1523/JNEUROSCI.3360-04.2005. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 45.Ursu S, Carter CS. Outcome representations, counterfactual comparisons and the human orbitofrontal cortex: implications for neuroimaging studies of decision-making. Cogn Brain Res. 2005;23:51–60. doi: 10.1016/j.cogbrainres.2005.01.004. [DOI] [PubMed] [Google Scholar]
- 46.O’Doherty J, Kringelbach ML, Rolls ET, et al. Abstract reward and punishment representations in the human orbitofrontal cortex. Nature Neurosci. 2001;4:95–102. doi: 10.1038/82959. [DOI] [PubMed] [Google Scholar]
- 47.Thorpe SJ, Rolls ET, Maddison S. The orbitofrontal cortex: neuronal activity in the behaving monkey. Exp Brain Res. 1983;49:93–115. doi: 10.1007/BF00235545. [DOI] [PubMed] [Google Scholar]
- 48.Schoenbaum G, Chiba A, Gallagher M. Orbitalfrontal cortex and basolateral amygdala encode expected outcomes during learning. Nature Neurosci. 1998;1:155–159. doi: 10.1038/407. [DOI] [PubMed] [Google Scholar]
- 49.Schoenbaum G, Chiba AA, Gallagher M. Neural encoding in orbitofrontal cortex and basolateral amygdala during olfactory discrimination learning. J Neurosci. 1999;19:1876–1884. doi: 10.1523/JNEUROSCI.19-05-01876.1999. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 50.Schoenbaum G, Setlow B, Saddoris MP, Gallagher M. Encoding predicted outcome and acquired value in orbitofrontal cortex during cue sampling depends upon input from basolateral amygdala. Neuron. 2003;39:855–867. doi: 10.1016/s0896-6273(03)00474-4. [DOI] [PubMed] [Google Scholar]
- 51.Rolls ET, Critchley HD, Mason R, Wakeman EA. Orbitofrontal cortex neurons: role in olfactory and visual association learning. J Neurophysiol. 1996;75:1970–1981. doi: 10.1152/jn.1996.75.5.1970. [DOI] [PubMed] [Google Scholar]
- 52.Tremblay L, Schultz W. Relative reward preference in primate orbitofrontal cortex. Nature. 1999;398:704–708. doi: 10.1038/19525. [DOI] [PubMed] [Google Scholar]
- 53.Wallis JD, Miller EK. Neuronal activity in primate dorsolateral and orbital prefrontal cortex during performance of a reward preference task. Eur J Neurosci. 2003;18:2069–2081. doi: 10.1046/j.1460-9568.2003.02922.x. [DOI] [PubMed] [Google Scholar]
- 54.Tremblay L, Schultz W. Modifications of reward expectation-related neuronal activity during learning in primate orbitofrontal cortex. J Neurophysiol. 2000;83:1877–1885. doi: 10.1152/jn.2000.83.4.1877. [DOI] [PubMed] [Google Scholar]
- 55.Roesch MR, Olson CR. Neuronal activity related to reward value and motivation in primate frontal cortex. Science. 2004;304:307–310. doi: 10.1126/science.1093223. [DOI] [PubMed] [Google Scholar]
- 56.Padoa-Schioppa C, Assad JA. Neurons in the orbitofrontal cortex encode economic value. Nature. 2006;441:223–226. doi: 10.1038/nature04676. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 57.Feierstein CE, Quirk MC, Uchida N, et al. Representation of spatial goals in rat orbitofrontal cortex. Neuron. 2006;51:495–507. doi: 10.1016/j.neuron.2006.06.032. [DOI] [PubMed] [Google Scholar]
- 58.Roesch MR, Taylor AR, Schoenbaum G. Encoding of time-discounted rewards in orbitofrontal cortex is independent of value representation. Neuron. 2006;51:509–520. doi: 10.1016/j.neuron.2006.06.027. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 59.Ostlund SB, Balleine BW. Orbitofrontal cortex mediates outcome encoding in Pavlovian but not instrumental conditioning. J Neurosci. 2007;27:4819–4825. doi: 10.1523/JNEUROSCI.5443-06.2007. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 60.Mazur JE. Learning and Behavior. Prentice Hall; Upper Saddle River, NJ: 2006. [Google Scholar]
- 61.Pavlov IP. Conditioned Reflexes. Oxford University Press; London: 1927. [Google Scholar]
- 62.Lang PJ, Bradley MM, Cuthbert BN. Emotion, attention, and the startle reflex. Psychol Rev. 1990;97:377–395. [PubMed] [Google Scholar]
- 63.Gallistel CR, Fairhurst S, Balsam P. The learning curve: implications of a quantitative analysis. Proc Natl Acad Sci USA. 2004;101:13124–13131. doi: 10.1073/pnas.0404965101. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 64.Belova MA, Paton JJ, Morrison SE, Salzman CD. Expectation modulates neural responses to pleasant and aversive stimuli in primate amygdala. Neuron. 2007;55(6):970–984. doi: 10.1016/j.neuron.2007.08.004. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 65.Saddoris MP, Gallagher M, Schoenbaum G. Rapid associative encoding in basolateral amygdala depends on connections with orbitofrontal cortex. Neuron. 2005;46:321–331. doi: 10.1016/j.neuron.2005.02.018. [DOI] [PubMed] [Google Scholar]
- 66.Green DM, Swets JA. Signal Detection Theory and Psychophysics. Wiley; New York: 1966. [Google Scholar]
- 67.Morrison SE, Salzman CD. Primate orbitofrontal cortex and amygdala encode stimulus value with a similar time course during reinforcement learning. Soc Neurosci Abstr. 2007:934.4. [Google Scholar]
- 68.Rolls E. Neurophysiology and functions of the primate amygdala, and the neural basis of emotion. In: Aggleton J, editor. The Amygdala: a Functional Analysis. Oxford University Press; New York: 2000. pp. 447–478. [Google Scholar]
- 69.Fellows LK, Farah MJ. Ventromedial frontal cortex mediates affective shifting in humans: evidence from a reversal learning paradigm. Brain. 2003;126:1830–1837. doi: 10.1093/brain/awg180. [DOI] [PubMed] [Google Scholar]
- 70.Izquierdo A, Murray EA. Selective bilateral amygdala lesions in rhesus monkeys fail to disrupt object reversal learning. J Neurosci. 2007;27:1054–1062. doi: 10.1523/JNEUROSCI.3616-06.2007. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 71.Schoenbaum G, Setlow B, Nugent SL, Saddoris MP, Gallagher M. Lesions of orbitofrontal cortex and basolateral amygdala complex disrupt acquisition of odor-guided discriminations and reversals. Learn Mem. 2003;10:129–140. doi: 10.1101/lm.55203. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 72.Swinehart CD, Abbott LF. Supervised learning through neuronal response modulation. Neural Comput. 2005;17:609–631. doi: 10.1162/0899766053019980. [DOI] [PubMed] [Google Scholar]
- 73.Swinehart CD, Bouchard K, Partensky P, Abbott LF. Control of network activity through neuronal response modulation. Neurocomputing. 2004:58–60. 327–335. [Google Scholar]
