Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2011 Apr 1.
Published in final edited form as: Curr Opin Neurobiol. 2010 Mar 19;20(2):191–198. doi: 10.1016/j.conb.2010.02.009

Heterogeneous reward signals in prefrontal cortex

Jonathan D Wallis 1,2, Steven W Kennerley 1,2,3
PMCID: PMC2862852  NIHMSID: NIHMS183334  PMID: 20303739

Summary

Neurons encode upcoming rewards throughout frontal cortex. Recent papers have helped to determine that these signals play different roles in different frontal regions. Neurons in orbitofrontal cortex (PFo) appear to be responsible for calculating the specific value of an expected reward, information that can help efficiently guide decision-making. Similar signals are also present in the cingulate sulcus (PFcs). In contrast, reward signals in lateral prefrontal cortex (PFl) are consistent with a role in using reward to guide other cognitive processes, such as the allocation of attentional resources and using value information to guide learning other relationships in the environment such as arbitrary stimulus-response mappings. A remaining issue for future work is to specify the precise roles of PFo and PFcs. These two areas show very different patterns of connectivity with other brain areas, and it is currently unclear how this effects their contribution to decision-making.

Introduction

During the 1980s, two distinct lines of research highlighted the importance of PFo for processing reward information and using such information to guide behavior. First, PFo neurons encoded whether a sensory stimulus was predictive of a rewarding outcome [1]. Second, patients with PFo damage had specific impairments in their ability to make decisions in everyday life [2]. These findings elegantly complemented one another. PFo patients would be forced to make decisions more or less at random if they lacked the signals as to the expected rewards that might result from different possible choices. Despite this initial consilience, neuropsychological and neurophysiological findings subsequently diverged. Laboratory tests confirmed that frontal patients had very specific deficits in value-based decision-making (e.g. which color do you prefer?), and not other forms of decision-making such as perceptual decision-making (e.g. which color is more blue?) [3]. Furthermore, such deficits only occur following damage to PFo, not PFl [3], consistent with the notion that PFo has a specialized role in this process. However, reward-predictive neurons, far from being specific to PFo, were subsequently found in every frontal area that neurophysiologists cared to record from, including PFl [4], PFcs [5], and motor areas such as the premotor and primary motor cortex [6].

These findings raised two important questions. First, if the presence of reward-predictive neurons does not indicate a necessary role in decision-making, what is it about the neuronal encoding in PFo that makes it so important to this process? Second, what function does reward-predictive encoding serve in other frontal areas? Rewards have many aspects. They have sensory (e.g. flavor, intensity) and emotional components (they make us happy), they satisfy motivational drives, they have a value (some rewards are more preferable than others are), they are reinforcing (we repeat behavior that produced a reward), and they can guide the allocation of cognitive and attentional resources. How do reward-predictive neurons in different frontal areas relate to these different functional properties of reward?

Reward signals underlying the calculation of value

How does one determine the value of one good relative to another? One approach is through barter: if I want three chickens for my pig then we can say that I find pigs three times more valuable than chickens. By bartering with a monkey using different types of juice, researchers showed that many PFo neurons encode the value of one juice reward relative to another [7,8]. One problem facing such a signal is that there is a huge range in the value of potential goods. Consider browsing the real estate section of the newspaper in a café, when the server interrupts to take our order. We are effortlessly able to switch from valuations in the hundreds of thousands of dollars to those for just a few dollars. Yet the PFo neurons representing this value range have a limited firing rate range (typically <50 Hz). A recent study shows that PFo neurons solve this problem by dynamically adjusting their firing rate range to the range of values on offer [9] (Fig. 1). An interesting question is whether and how top-down influences, such as the context one is in, might be able to accomplish such an adjustment. PFo neurons also encode a wide range of other variables necessary for decision-making [10], including positive and negative expected outcomes [11,12], the costs involved in acquiring an outcome [13,14], the probability that one's choice will be successful [14] and one's confidence in the choice [15].

Figure 1.

Figure 1

a) The figure illustrates the hypothetical relationship between firing rate and value if the neuron adapts its firing rate to match the value of the range under consideration (ΔV). Specifically, as the value range increases, the slope of the relationship between firing rate and value decreases. b) The raw data taken from the entire population of recorded neurons from PFo. Each line is the average firing rate of a specific neuron for a given value. Different colors indicate the different value ranges under consideration. c) Average of the neuronal responses in b). The data clearly match the predictions of the hypothetical model shown in a). As the value range under consideration increases, the slope of the relationship between value and firing rate decreases, indicative of neuronal adaptation. Reprinted with permission from [9].

A recent neuropsychological study in rats provided greater insight into the nature of the PFo value signal by using the phenomenon of transreinforcer blocking [16]. A subject only learns that a cue predicts a reward if the reward's occurrence is surprising. Thus, learning about a cue can be blocked if a second cue is present that fully predicts the delivery of a reward. However, supposing the two cues predict different rewards (e.g. apple versus orange). Now the second cue only partially blocks learning to the first cue: it blocks learning to those factors that the two rewards share (general reward information) but leaves intact learning to the factors that differentiate the two rewards (specific reward information). It is only this latter information that is sensitive to PFo lesions. But what exactly is this specific reward information? One explanation is that PFo encodes the sensory features (e.g. appearance) of the specific reward that the cue predicts [17]. A problem with this interpretation is that it seems to draw a somewhat arbitrary distinction between reward-predictive cues and the sensory properties of a reward. For example, is the sight of an apple a predictive cue of the rewarding properties of the apple (e.g. its sweet taste) or a sensory feature of the apple? An alternative explanation is that there are multiple value scales in the brain, a value system that signals the general affective properties of apples (i.e. they are positive things that will make you happy) and a value system that calculates the specific value of the apples (i.e. 2 apples are worth 5 oranges) [16]. This specific value system is the one instantiated by PFo. The notion of distinct value systems is not novel. For example, previous investigators have suggested distinct value systems related to liking something (hedonic value) and wanting something (incentive value) [18].

In addition to PFo, dysfunction of PFcs has also been associated with impairments in using reward information to guide decision-making [19,20]. Neurons in PFcs also encode many of the same factors as PFo neurons including positive and negative outcomes [21-23], probability of success [14], and reward preference [24]. We trained monkeys to make choices between different behavioral outcomes that varied in terms of either payoff (volume of juice), cost (number of lever presses necessary to earn the juice) or probability (probability that the juice would be delivered) [14]. We recorded from PFo, PFcs and PFl and found that neurons encoded value across the different decision variables in diverse ways. For example, some neurons encoded the value of just a single decision variable, others encoded the value of choices for two of the decisions but not the third, while still others encoded value across all three decision variables (Fig. 2). The prevalence of neurons encoding a single decision variable was similar in all areas, but neurons encoding two or three decision variables were most prevalent in PFcs and PFo. Thus, PFcs and PFo neurons encode a multiplexed representation, which may allow the integration of the individual components of a decision and underlie their critical contribution to decision-making. Consistent with this notion, patients with PFo damage are impaired in integrating multiple attributes of a choice outcome in order to make a decision (e.g. choosing between a small apartment in a low crime neighborhood versus a larger apartment in a higher crime neighborhood) [25].

Figure 2.

Figure 2

Spike density histograms illustrating the activity recorded from single neurons under three different types of value manipulation (probability, payoff or effort). The vertical lines indicate the onset of the pictures indicating the value of the choice (left) and the time at which the animal was able to make his choice (right). The different colored lines indicate the value of the choice under consideration. Four neurons are shown: a) encodes the value of probability choices alone, b) encodes payoff and cost but not probability, c) encodes probability and payoff but not cost and d) encodes the value of all three decision variables. Across the population, there was an approximately equal number of neurons showed a positive relationship between firing rate and value (d) as showed a negative relationship (a, b, and c).

Using value signals for other cognitive processes

Value information can potentially serve other functions beyond allowing efficient action selection. More broadly, value information allows prioritization, which is necessary whenever one has limited resources. Action selection is just one example of system with capacity constraints - we can only pursue one course of action at a time. However, cognitive processes also have capacity constraints, most notably the process of working memory. Neurons in PFl encode information in working memory [26,27]. A common finding is that performance of a working memory task improves when a subject expects a larger reward, which might arise if value information can modulate the fidelity of information stored in working memory. To examine this, we trained monkeys to maintain spatial locations in working memory [28,29]. At the beginning of each trial, a picture informed the subject how much juice they would receive for correct performance of the task. Figure 3 illustrates an example of reward modulating the spatial selectivity of a PFl neuron. We recorded from PFl, PFo and PFcs, but we only saw such modulation in PFl. Reward-dependent modulation is not limited to spatial information; reward can also modulate PFl encoding of high-level information, such as categories [30]. Furthermore, the modulation may be bidirectional: the contents of working memory may also be able to modulate the reward signal. In dieters exercising self-control regarding choices involving healthy and unhealthy snacks, there is increased activation of PFl and a concomitant decrease in areas representing value information, as though PFl is dampening the value signal [31].

Figure 3.

Figure 3

Spike density histograms illustrating how the size of an expected reward can modulate spatial information held in working memory. The graphs illustrate neuronal activity as animals remember different locations on a computer screen under the expectancy of receiving either a small or a large reward for correct performance. The grey bar indicates the presentation of the mnemonic spatial cue. To enable clear visualization, the spatial data is collapsed into four groups consisting of 6 of the 24 possible spatial locations tested. The inset indicates the mean standardized firing rate of the neuron across the 24 spatial locations. a) When the subject expected a small reward, the neuron showed moderate spatial selectivity, which consisted of an increase in firing rate when the subject was remembering locations in the top right of the screen. When the subject expected a large reward for correct performance, spatial selectivity dramatically increased with a high firing rate for locations in the top right of the screen and a low firing rate for locations in the bottom left. b) A neuron that showed little spatial selectivity when the subject expected a small reward, but a dramatic increase in spatial selectivity when the subject expected a large reward.

Value information can also serve simply as feedback. For example, consider a child learning an association between a stimulus (e.g. the configuration of a door handle and hinges) and a response (e.g. push or pull). The precise value of the reward is not necessarily relevant and can be very variable (e.g. food in a refrigerator, a toy in a cupboard, an adult saying, “Good job”). Instead, the reward serves as feedback indicating to the child that they selected the correct response. This differs from the actions themselves having different values (e.g. a child choosing between opening two cupboards, one full of toys, the other containing clothes).

Neurophysiological evidence shows that PFl neurons encode the success or failure of a selected response [32], consistent with a feedback signal. PFl neurons integrate such signals across multiple trials [33], thereby potentially providing a mechanism by which sequences of actions can be learned [34]. In addition, PFl feedback signals are sustained and influence both future behavior as well as the neuronal encoding of that behavior [35]. Taken together, these results suggest that PFl neurons may play an important role in linking behavior with temporally separated feedback, in order to determine correct responses. Although neurons in PFl encode signals relating to the history of previously performed actions, signals relating to the values of different actions are more common in PFcs. For example, in a task where subjects must learn what response to make to a stimulus in order to obtain a specific outcome, PFl neurons encode the association between the stimulus and response rather than information about the outcome [5]. In contrast, neurons in PFcs encoded which response led to which outcome [5]. Similarly, in a task where a monkey chose between two different responses each associated with one of three different juices, PFl neurons only encoded the responses whereas PFcs neurons encoded the responses and the outcomes [24]. Furthermore, the firing rate of PFcs neurons to the outcomes correlated with the subject's preferences between the different juices, consistent with a value signal [24]. In sum, the activity of PFl neurons is consistent with using value information as feedback to determine the correct response to make given a particular sensory environment. In contrast, PFcs neurons are similar to PFo neurons in that they appear to encode value information directly.

Functional differences between PFo and PFcs

If PFo and PFcs are both capable of encoding value signals and damage to either area impairs decision-making, a clear question is what the two areas are doing differently from one another. A comparison of PFcs and PFo anatomy suggests potential answers to this question. PFcs and PFo both connect with areas responsible for processing rewards, such as the amygdala and hypothalamus [36,37], and receive dopaminergic inputs [38] that also carry reward signals [39,40]. PFcs strongly connects with cingulate motor areas and limbic structures but has few direct connections with sensory cortex [41-43]. In contrast, PFo only weakly connects with motor areas, but heavily connects with all sensory areas, including visual, olfactory and gustatory cortices [41,42,44]. Consistent with this anatomy, PFcs neurons tend to encode rewards and the action that led to the reward in monkeys [5,45] and humans [46], whereas PFo neurons encode the value of sensory stimuli with little encoding of motor responses [7,8,14,29,47]. PFcs may have an additional role in monitoring the outcome of the selected action to guide behavioral adjustment, as PFcs activity is sensitive to errors or deviations in reward expectancy [48], reflects the history of reward associated with different actions [19,49], tracks volatility in reward rate of different actions [50], and encodes “fictive” rewards, counterfactual rewards that might have been obtained had the subject selected an alternate course of action [51]. Thus, there may be a division of labor between PFo and PFcs with regard to choice evaluation, with PFo calculating the value of possible outcomes, while PFcs calculates the value of the action producing the outcome and monitors the success of behavioral outcomes over time to guide adaptive behavior.

Recent neuropsychological studies support this dissociation. Lesions of PFo impaired probabilistic matching tasks when the discriminanda were visual pictures, but not when the discriminanda were motor responses, while the opposite pattern occurred following lesions of PFcs [19,52]. In rats, lesions to the corresponding PFcs region disrupts choices when rats are faced with options differing in their payoff and the effort necessary to obtain that payoff [53], but such lesions do not effect other types of cost-benefit decisions (e.g. delay-based decisions), that do not require the evaluation of an action [54]. In contrast, rats with PFo lesions show impaired delay-based decision-making, but intact effort-based decision-making [54].

Interpretational issues

Adaptive decision-making depends on a number of different computations, including representing the value of one's current internal (motivational) state, the value of possible (chosen and unchosen) alternatives, and the generation of a prediction error signal when outcomes deviate from expectations [55]. Importantly, although these value representations reflect different computations, they can be highly correlated with one another [50,56]. Moreover, given choices can be made on the order of a few hundreds milliseconds, these processes are likely to occur in close temporal proximity and thus be difficult to differentiate. Finally, it is difficult to infer functional specialization by comparing results across studies that use different tasks, subjects and statistical analyses. To address these issues several studies have recorded from different populations of neurons in the same subjects, demonstrating the degree to which different frontal areas are recruited for different value computations [6,12,14,21,57], as well as demonstrating a hierarchy in the order in which reward is encoded, first in PFo and then later in PFl and PFcs [29,47].

A final issue is that it is not clear how these ideas relate to the functional organization of the human frontal cortex. The human frontal area that is most consistently activated across a broad range of value-based decision-making studies is the ventral portion of the medial surface of the frontal lobe (PFvm) [31,58-61]. This area most likely corresponds to either area 32 or area 10 (Fig 4). Primate neurophysiologists have done little recording from these areas due to their relative inaccessibility (although see [62]).

Figure 4.

Figure 4

Cytoarchitectonic maps of the ventral and medial surface of the human and macaque frontal cortex. Granular cortex (containing a layer 4) appear in blue, agranular areas in green and allocortical areas in yellow. Abbreviations: a, agranular; AON, anterior olfactory ‘nucleus’; I, insula; Pir, piriform cortex; l, lateral; m, medial; o, orbital; r, rostral; c, caudal; i, inferior; p, posterior; s, sulcal; v, ventral. (Note that the letter ‘a’ has two meanings: in Ia it means agranular; in 13a it distinguishes that area from area 13b.) Architectonics are taken from [63,64]. Reprinted with permission from [65].

Conclusion

Over the last decade, it has become clear that reward signals are ubiquitous in the frontal lobe. This should not be surprising given that a central goal of behavior is to obtain reward. It is also clear that reward information is useful for many different functions, and so reward signals in different frontal regions may be serving very different roles. A thesis compatible with the most recent research is that neurons in PFcs and PFo are directly involved in calculating value, with PFo calculating the value of possible outcomes, while PFcs calculates the value of the action to guide future behavior. In contrast, PFl may use value information to modulate the contents of working memory and as a feedback signal to learn arbitrary relationships such as those that underlie stimulus-response mappings.

Acknowledgments

The preparation of this manuscript was supported by NIDA grant R01DA19028 and NINDS grant P01NS040813 to J.D.W. and NIMH training grant F32MH081521 to S.W.K.

Footnotes

Publisher's Disclaimer: This is a PDF file of an unedited manuscript that has been accepted for publication. As a service to our customers we are providing this early version of the manuscript. The manuscript will undergo copyediting, typesetting, and review of the resulting proof before it is published in its final citable form. Please note that during the production process errors may be discovered which could affect the content, and all legal disclaimers that apply to the journal pertain.

The authors have no competing interests.

References

  • 1.Thorpe SJ, Rolls ET, Maddison S. The orbitofrontal cortex: neuronal activity in the behaving monkey. Exp Brain Res. 1983;49:93–115. doi: 10.1007/BF00235545. [DOI] [PubMed] [Google Scholar]
  • 2.Eslinger PJ, Damasio AR. Severe disturbance of higher cognition after bilateral frontal lobe ablation: patient EVR. Neurology. 1985;35:1731–1741. doi: 10.1212/wnl.35.12.1731. [DOI] [PubMed] [Google Scholar]
  • 3.Fellows LK, Farah MJ. The role of ventromedial prefrontal cortex in decision making: Judgment under uncertainty or judgment per se? Cereb Cortex. 2007;17:2669–2674. doi: 10.1093/cercor/bhl176. [DOI] [PubMed] [Google Scholar]
  • 4.Watanabe M. Reward expectancy in primate prefrontal neurons. Nature. 1996;382:629–632. doi: 10.1038/382629a0. [DOI] [PubMed] [Google Scholar]
  • 5.Matsumoto K, Suzuki W, Tanaka K. Neuronal correlates of goal-based motor selection in the prefrontal cortex. Science. 2003;301:229–232. doi: 10.1126/science.1084204. [DOI] [PubMed] [Google Scholar]
  • 6.Roesch MR, Olson CR. Impact of expected reward on neuronal activity in prefrontal cortex, frontal and supplementary eye fields and premotor cortex. J Neurophysiol. 2003;90:1766–1789. doi: 10.1152/jn.00019.2003. [DOI] [PubMed] [Google Scholar]
  • 7.Padoa-Schioppa C, Assad JA. Neurons in the orbitofrontal cortex encode economic value. Nature. 2006;441:223–226. doi: 10.1038/nature04676. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Padoa-Schioppa C, Assad JA. The representation of economic value in the orbitofrontal cortex is invariant for changes of menu. Nat Neurosci. 2008;11:95–102. doi: 10.1038/nn2020. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9 •.Padoa-Schioppa C. Range-adapting representation of economic value in the orbitofrontal cortex. J Neurosci. 2009;29:14004–14014. doi: 10.1523/JNEUROSCI.3751-09.2009. [Neurons in PFo adapt the dynamic range of their firing rate depending on the maximum value offered in a given behavioral context. This mechanism can explain how we are able to make decisions involving rewards whose values may differ by many orders of magnitude.] [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Wallis JD. Orbitofrontal cortex and its contribution to decision-making. Annu Rev Neurosci. 2007;30:31–56. doi: 10.1146/annurev.neuro.30.051606.094334. [DOI] [PubMed] [Google Scholar]
  • 11.Morrison SE, Salzman CD. The convergence of information about rewarding and aversive stimuli in single neurons. J Neurosci. 2009;29:11471–11483. doi: 10.1523/JNEUROSCI.1815-09.2009. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Roesch MR, Olson CR. Neuronal activity related to reward value and motivation in primate frontal cortex. Science. 2004;304:307–310. doi: 10.1126/science.1093223. [DOI] [PubMed] [Google Scholar]
  • 13.Roesch MR, Olson CR. Neuronal activity in primate orbitofrontal cortex reflects the value of time. J Neurophysiol. 2005;94:2457–2471. doi: 10.1152/jn.00373.2005. [DOI] [PubMed] [Google Scholar]
  • 14 •.Kennerley SW, Dahmubed AF, Lara AH, Wallis JD. Neurons in the frontal lobe encode the value of multiple decision variables. J Cogn Neurosci. 2009;21:1162–1178. doi: 10.1162/jocn.2009.21100. [The authors recorded from multiple frontal areas while subjects made choices where the value was manipulated via either the payoff, probability or cost of the decision. PFcs and PFo neurons differed from PFl neurons in that they encoded value for multiple decision variables enabling them to calculate an abstract value signal.] [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15 ••.Kepecs A, Uchida N, Zariwala HA, Mainen ZF. Neural correlates, computation and behavioural impact of decision confidence. Nature. 2008;455:227–231. doi: 10.1038/nature07200. [Using a psychophysical categorization task, the authors show that PFo neurons encode how confident the animal is in a perceptual decision, a factor that would be important in determining the likelihood of a successful outcome.] [DOI] [PubMed] [Google Scholar]
  • 16 ••.Burke KA, Franz TM, Miller DN, Schoenbaum G. The role of the orbitofrontal cortex in the pursuit of happiness and more specific rewards. Nature. 2008;454:340–344. doi: 10.1038/nature06993. [Using a sophisticated behavioral paradigm this study shows that learning about the general affective properties of a reward can be dissociated from learning about its specific value. Lesions of PFo impair the latter not the former.] [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Mainen ZF, Kepecs A. Neural representation of behavioral outcomes in the orbitofrontal cortex. Curr Opin Neurobiol. 2009;19:84–91. doi: 10.1016/j.conb.2009.03.010. [DOI] [PubMed] [Google Scholar]
  • 18.Berridge KC, Robinson TE. What is the role of dopamine in reward: hedonic impact, reward learning, or incentive salience? Brain Res Brain Res Rev. 1998;28:309–369. doi: 10.1016/s0165-0173(98)00019-8. [DOI] [PubMed] [Google Scholar]
  • 19.Kennerley SW, Walton ME, Behrens TE, Buckley MJ, Rushworth MF. Optimal decision making and the anterior cingulate cortex. Nat Neurosci. 2006;9:940–947. doi: 10.1038/nn1724. [DOI] [PubMed] [Google Scholar]
  • 20.Amiez C, Joseph JP, Procyk E. Reward encoding in the monkey anterior cingulate cortex. Cereb Cortex. 2006;16:1040–1055. doi: 10.1093/cercor/bhj046. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Seo H, Lee D. Behavioral and neural changes after gains and losses of conditioned reinforcers. J Neurosci. 2009;29:3627–3641. doi: 10.1523/JNEUROSCI.4726-08.2009. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22 ••.Quilodran R, Rothe M, Procyk E. Behavioral shifts and action valuation in the anterior cingulate cortex. Neuron. 2008;57:314–325. doi: 10.1016/j.neuron.2007.11.031. [Neurons in PFcs encode both positive and negative outcomes associated with different actions. In addition, they show stronger responses when the animal first discovers a rewarded response (exploration), but not when the animal repeats the response to receive further rewards (exploitation).] [DOI] [PubMed] [Google Scholar]
  • 23.Sallet J, Quilodran R, Rothe M, Vezoli J, Joseph JP, Procyk E. Expectations, gains, and losses in the anterior cingulate cortex. Cogn Affect Behav Neurosci. 2007;7:327–336. doi: 10.3758/cabn.7.4.327. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24 •.Luk CH, Wallis JD. Dynamic encoding of responses and outcomes by neurons in medial prefrontal cortex. J Neurosci. 2009;29:7526–7539. doi: 10.1523/JNEUROSCI.0386-09.2009. [If choice alternatives are presented sequentially, PFcs encode the first potential action and it's value followed by the value of the second alternative relative to the first. Thus, PFcs appear to encode all of the information necessary to make optimal choices.] [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Fellows LK. Deciding how to decide: ventromedial frontal lobe damage affects information acquisition in multi-attribute decision making. Brain. 2006;129:944–952. doi: 10.1093/brain/awl017. [DOI] [PubMed] [Google Scholar]
  • 26.Funahashi S, Bruce CJ, Goldman-Rakic PS. Mnemonic coding of visual space in the monkey's dorsolateral prefrontal cortex. J Neurophysiol. 1989;61:331–349. doi: 10.1152/jn.1989.61.2.331. [DOI] [PubMed] [Google Scholar]
  • 27.Rao SC, Rainer G, Miller EK. Integration of what and where in the primate prefrontal cortex. Science. 1997;276:821–824. doi: 10.1126/science.276.5313.821. [DOI] [PubMed] [Google Scholar]
  • 28.Kennerley SW, Wallis JD. Reward-dependent modulation of working memory in lateral prefrontal cortex. J Neurosci. 2009;29:3259–3270. doi: 10.1523/JNEUROSCI.5353-08.2009. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.Kennerley SW, Wallis JD. Encoding of reward and space during a working memory task in the orbitofrontal cortex and anterior cingulate sulcus. J Neurophysiol. 2009 doi: 10.1152/jn.00273.2009. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.Pan X, Sawa K, Tsuda I, Tsukada M, Sakagami M. Reward prediction based on stimulus categorization in primate lateral prefrontal cortex. Nat Neurosci. 2008;11:703–712. doi: 10.1038/nn.2128. [DOI] [PubMed] [Google Scholar]
  • 31 ••.Hare TA, Camerer CF, Rangel A. Self-control in decision-making involves modulation of the vmPFC valuation system. Science. 2009;324:646–648. doi: 10.1126/science.1168450. [This study highlights the importance of PFl in the top-down control of reward signals. One mechanism by which we are able to exert self-control over our behavior may involve PFl dampening value signals.] [DOI] [PubMed] [Google Scholar]
  • 32.Watanabe M. The appropriateness of behavioral responses coded in post-trial activity of primate prefrontal units. Neurosci Lett. 1989;101:113–117. doi: 10.1016/0304-3940(89)90450-3. [DOI] [PubMed] [Google Scholar]
  • 33 •.Seo H, Barraclough DJ, Lee D. Dynamic signals related to choices and outcomes in the dorsolateral prefrontal cortex. Cereb Cortex. 2007;17(Suppl 1):i110–117. doi: 10.1093/cercor/bhm064. [PFl neurons keep track of the rewards associated with multiple actions, which could provide a mechanism by which a a particular outcome can be linked to its causative action.] [DOI] [PubMed] [Google Scholar]
  • 34.Averbeck BB, Lee D. Prefrontal neural correlates of memory for sequences. J Neurosci. 2007;27:2204–2211. doi: 10.1523/JNEUROSCI.4483-06.2007. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35 •.Histed MH, Pasupathy A, Miller EK. Learning substrates in the primate prefrontal cortex and striatum: sustained activity related to successful actions. Neuron. 2009;63:244–253. doi: 10.1016/j.neuron.2009.06.019. [PFl neurons maintain information about whether an action produced a reward between trials, and this information influences both behavior and neuronal activity on the next trial. This mechanism allows rewards to affect future behavior thereby enabling adaptive learning to occur.] [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36.Ongur D, An X, Price JL. Prefrontal cortical projections to the hypothalamus in macaque monkeys. J Comp Neurol. 1998;401:480–505. [PubMed] [Google Scholar]
  • 37.Carmichael ST, Price JL. Limbic connections of the orbital and medial prefrontal cortex in macaque monkeys. J Comp Neurol. 1995;363:615–641. doi: 10.1002/cne.903630408. [DOI] [PubMed] [Google Scholar]
  • 38.Williams SM, Goldman-Rakic PS. Characterization of the dopaminergic innervation of the primate frontal cortex using a dopamine-specific antibody. Cereb Cortex. 1993;3:199–222. doi: 10.1093/cercor/3.3.199. [DOI] [PubMed] [Google Scholar]
  • 39.Bayer HM, Glimcher PW. Midbrain dopamine neurons encode a quantitative reward prediction error signal. Neuron. 2005;47:129–141. doi: 10.1016/j.neuron.2005.05.020. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40.Schultz W. Predictive reward signal of dopamine neurons. J Neurophysiol. 1998;80:1–27. doi: 10.1152/jn.1998.80.1.1. [DOI] [PubMed] [Google Scholar]
  • 41.Carmichael ST, Price JL. Sensory and premotor connections of the orbital and medial prefrontal cortex of macaque monkeys. Journal of Comparative Neurology. 1995;363:642–664. doi: 10.1002/cne.903630409. [DOI] [PubMed] [Google Scholar]
  • 42.Croxson PL, Johansen-Berg H, Behrens TE, Robson MD, Pinsk MA, Gross CG, Richter W, Richter MC, Kastner S, Rushworth MF. Quantitative investigation of connections of the prefrontal cortex in the human and macaque using probabilistic diffusion tractography. J Neurosci. 2005;25:8854–8866. doi: 10.1523/JNEUROSCI.1311-05.2005. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43.Dum RP, Strick PL. Cingulate motor areas. In: Vogt BA, Gabriel M, editors. Neurobiology of cingulate cortex and limbic thalamus: a comprehensive handbook. Birkhaeuser; 1993. pp. 415–441. [Google Scholar]
  • 44.Cavada C, Company T, Tejedor J, Cruz-Rizzolo RJ, Reinoso-Suarez F. The anatomical connections of the macaque monkey orbitofrontal cortex. A review. Cereb Cortex. 2000;10:220–242. doi: 10.1093/cercor/10.3.220. [DOI] [PubMed] [Google Scholar]
  • 45.Shima K, Tanji J. Role for cingulate motor area cells in voluntary movement selection based on reward. Science. 1998;282:1335–1338. doi: 10.1126/science.282.5392.1335. [DOI] [PubMed] [Google Scholar]
  • 46.Williams ZM, Bush G, Rauch SL, Cosgrove GR, Eskandar EN. Human anterior cingulate neurons and the integration of monetary reward with motor responses. Nat Neurosci. 2004;7:1370–1375. doi: 10.1038/nn1354. [DOI] [PubMed] [Google Scholar]
  • 47.Wallis JD, Miller EK. Neuronal activity in primate dorsolateral and orbital prefrontal cortex during performance of a reward preference task. Eur J Neurosci. 2003;18:2069–2081. doi: 10.1046/j.1460-9568.2003.02922.x. [DOI] [PubMed] [Google Scholar]
  • 48.Amiez C, Joseph JP, Procyk E. Anterior cingulate error-related activity is modulated by predicted reward. Eur J Neurosci. 2005;21:3447–3452. doi: 10.1111/j.1460-9568.2005.04170.x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 49.Seo H, Lee D. Temporal filtering of reward signals in the dorsal anterior cingulate cortex during a mixed-strategy game. J Neurosci. 2007;27:8366–8377. doi: 10.1523/JNEUROSCI.2369-07.2007. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 50.Behrens TE, Woolrich MW, Walton ME, Rushworth MF. Learning the value of information in an uncertain world. Nat Neurosci. 2007;10:1214–1221. doi: 10.1038/nn1954. [DOI] [PubMed] [Google Scholar]
  • 51 ••.Hayden BY, Pearson JM, Platt ML. Fictive reward signals in the anterior cingulate cortex. Science. 2009;324:948–950. doi: 10.1126/science.1168488. [In addition to encoding the reward associated with a chosen action, PFcs neurons also encode rewards that would have been obtained had the animal chosen an alternate course of action. Such signals could improve learning and decision-making, and may also underlie counterfactual emotions such as regret and guilt.] [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 52 •.Rudebeck PH, Behrens TE, Kennerley SW, Baxter MG, Buckley MJ, Walton ME, Rushworth MF. Frontal cortex subregions play distinct roles in choices between actions and stimuli. J Neurosci. 2008;28:13775–13785. doi: 10.1523/JNEUROSCI.3541-08.2008. [This study reports a double dissociation between the effects of PFo and PFcs lesions, with the former impairing learning about stimulus-value contingencies and the latter learning about action-value contingencies.] [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 53.Walton ME, Bannerman DM, Rushworth MF. The role of rat medial frontal cortex in effort-based decision making. J Neurosci. 2002;22:10996–11003. doi: 10.1523/JNEUROSCI.22-24-10996.2002. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 54.Rudebeck PH, Walton ME, Smyth AN, Bannerman DM, Rushworth MF. Separate neural pathways process different decision costs. Nat Neurosci. 2006;9:1161–1168. doi: 10.1038/nn1756. [DOI] [PubMed] [Google Scholar]
  • 55.Rangel A, Camerer C, Montague PR. A framework for studying the neurobiology of value-based decision making. Nat Rev Neurosci. 2008;9:545–556. doi: 10.1038/nrn2357. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 56.Hare TA, O'Doherty J, Camerer CF, Schultz W, Rangel A. Dissociating the role of the orbitofrontal cortex and the striatum in the computation of goal values and prediction errors. J Neurosci. 2008;28:5623–5630. doi: 10.1523/JNEUROSCI.1309-08.2008. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 57.Matsumoto M, Matsumoto K, Abe H, Tanaka K. Medial prefrontal cell activity signaling prediction errors of action values. Nat Neurosci. 2007;10:647–656. doi: 10.1038/nn1890. [DOI] [PubMed] [Google Scholar]
  • 58.Kable JW, Glimcher PW. The neural correlates of subjective value during intertemporal choice. Nat Neurosci. 2007;10:1625–1633. doi: 10.1038/nn2007. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 59.Ballard K, Knutson B. Dissociable neural representations of future reward magnitude and delay during temporal discounting. Neuroimage. 2009;45:143–150. doi: 10.1016/j.neuroimage.2008.11.004. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 60.Glascher J, Hampton AN, O'Doherty JP. Determining a Role for Ventromedial Prefrontal Cortex in Encoding Action-Based Value Signals During Reward-Related Decision Making. Cereb Cortex. 2008 doi: 10.1093/cercor/bhn098. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 61.FitzGerald TH, Seymour B, Dolan RJ. The role of human orbitofrontal cortex in value comparison for incommensurable objects. J Neurosci. 2009;29:8388–8395. doi: 10.1523/JNEUROSCI.0717-09.2009. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 62.Tsujimoto S, Genovesio A, Wise SP. Evaluating self-generated decisions in frontal pole cortex of monkeys. Nat Neurosci. doi: 10.1038/nn.2453. in press. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 63.Carmichael ST, Price JL. Architectonic subdivision of the orbital and medial prefrontal cortex in the macaque monkey. J Comp Neurol. 1994;346:366–402. doi: 10.1002/cne.903460305. [DOI] [PubMed] [Google Scholar]
  • 64.Ongur D, Ferry AT, Price JL. Architectonic subdivision of the human orbital and medial prefrontal cortex. J Comp Neurol. 2003;460:425–449. doi: 10.1002/cne.10609. [DOI] [PubMed] [Google Scholar]
  • 65.Wise SP. Forward frontal fields: phylogeny and fundamental function. Trends Neurosci. 2008;31:599–608. doi: 10.1016/j.tins.2008.08.008. [DOI] [PMC free article] [PubMed] [Google Scholar]

RESOURCES