Abstract
The application of a neuroeconomic approach to the study of reward-related processes has provided significant insights in our understanding of human learning and decision-making. Much of this research has primarily focused on the contributions of the cortico-striatal circuitry, involved in trial and error reward learning. As a result, less consideration has been allotted to the potential influence of different neural mechanisms such as the hippocampus, or to more common ways in human society in which information is acquired and utilized to reach a decision, such as through explicit instruction rather than trial and error learning. This review examines the basic and applied value of examining the individual contributions of multiple learning and memory neural systems and their interactions during human decision-making in normal individuals and neuropsychiatric populations. Specifically, the anatomical and functional connectivity across multiple memory systems are highlighted to suggest that probing the role of the hippocampus and its interactions with the cortico-striatal circuitry via the application of model-based neuroeconomic approaches may provide novel insights into several neuropsychiatric populations who suffer from damage to one of these structures, and as a consequence have deficits in learning, memory, or decision-making.
Keywords: Striatum, hippocampus, prediction error, dopamine, reinforcement learning, memory systems
Introduction
There is an emerging literature investigating the neural basis of decision-making that encompasses the field of neuroeconomics (1). Across a variety of techniques and interdisciplinary efforts, neuroeconomics research has emphasized two key points: a) choice behavior is influenced by learning systems involved in trial and error learning (such as instrumental conditioning), and b) cortico-striatal neural systems are involved in the computation of basic principles of conditioning and reinforcement learning that underlie simple decision-making. This cross-species account of decision-making represents a significant contribution to the field, and informs various influential models of decision-making (2–4), but it does not fully reflect the complexity of learning in humans, where the existence of multiple memory systems has been well documented (5, 6). Given the intricacy of the environment, it would be peculiar if one neural system was primarily responsible for learning signals that contribute to decision-making. In fact, choices made in everyday human society are not solely driven by experience through trial and error learning, as we also learn and allow our decisions to be influenced by communication (e.g., talking to others), knowledge of rules (e.g., when to cross the street) and facts (e.g., this person is untrustworthy). Thus, consideration of how distinct neural circuits interact during reward-related learning to influence choice behavior can enhance the contributions of neuroeconomics research and contain potential clinical applications.
In this review, we focus on the distinct and common contributions of the basal ganglia (BG; involved in learning through trial and error), and the medial temporal lobes (MTL), particularly the hippocampus (involved in declarative learning of rules and facts) during learning and how they translate to decision-making in both healthy and clinical populations. We first discuss the concept for and evidence supporting multiple memory systems, followed by subsections describing relevant learning and decision-making research involving the BG and MTL, along with potential interactions between these structures. Finally, we evaluate how neuroeconomics research may be expanded upon in the context of multiple memory systems and synthesized to help inform neuropsychiatric disorders that involve compromised BG and MTL structures.
Multiple memory systems
The values we assign to different stimuli and prior experiences tend to influence how we make decisions. For instance, an individual may choose to shop at the local family-owned grocery store versus a new larger grocery store in town because of prior positive experiences at the local establishment (Figure 1). Within this example, one explanation of how the choice is made is based on a greater value assigned to the local shop because of reinforcement learning. An impressive collection of studies has highlighted the role of cortico-striatal circuits along with modulation of dopamine neurons as integral for this type of learning (7–9). Not surprisingly, individuals with pathologies that involve BG dysfunction, such as Parkinson’s disease (PD), have difficulties learning by trial and error and updating learned contingencies (10–15), which may manifest as perseverative choices in some contexts (16). Defined by a loss of the majority of dopaminergic neurons in the midbrain, PD consists of several motor impairments (e.g., akinesia) as well as cognitive deficits (17, 18). Importantly, patients afflicted with PD can learn through different neural mechanisms, such as MTL structures involved in declarative learning, which can also help guide choices (19) and highlights PD as a candidate disorder for investigating multiple memory systems.
Figure 1.
Multiple memory systems during human decision-making. In everyday life, humans are required to make hundreds of decisions, ranging from the complex, e.g., what career path to follow, to the simple, such as where to buy groceries. One may rely on past experiences (positive or negative) in order to make future decisions, as illustrated in the figure by positive past experiences at a local Mom & Pop Grocery. Repeated good experiences at one store may lead to continually shopping at the establishment, a form of learning and decision-making thought to involve the basal ganglia (BG), particularly the input unit of the BG – the striatum. However, decisions can be influenced by distinct neural systems underlying multiple learning and memory systems in addition to the striatum, including the hippocampus, part of the medial temporal lobe (MTL) which may store declarative information about one’s environment. For example, speaking with a friend about a new larger grocery store in town or reading a favorable review in the newspaper may lead one to make the decision to try out the new store. In this manner multiple memory systems may interact in everyday decision making, via either direct connections between these regions (MTL to striatum in BG) or via midbrain dopaminergic influences stemming from the ventral tegmental area and substantia nigra (VTA/SN), hypothesized to facilitate learning and memory processes in the MTL and BG. Abbreviations: MTL, medial temporal lobe; BG, basal ganglia; VTA, ventral tegmental area; SN, substantia nigra.
Over the past few decades, an increasingly supported theory suggests that memory is not a unitary process, but rather consists of multiple processes which engage distinct neural substrates (5, 6, 20). A major dissociation is between declarative and nondeclarative memory. Declarative memory consists of knowledge of facts and events; this type of memory may be rapidly formed and later readily recalled, is flexible in nature and relies on the integrity of the MTL (21). In contrast, nondeclarative memory encompasses several different types of learning including habit learning, procedural, conditioning, and priming (20). Such learning is thought to involve regions of the BG, along with cortical structures and the amygdala, and is characterized by a slower time course of acquisition and less flexibility with respect to updating, as in the case of habits (21, 22).
Support for this theory comes from a vast array of research ranging from non-human animal investigations to studies with clinical populations (19). In a classic paradigm, a rodent is trained to navigate a maze to retrieve a food reward (23, 24). A clear dissociation between multiple memory systems is apparent by poor performance in a spatial learning version of a radial maze (the “win-shift” task) in rats with lesions to the fimbria-fornix – the output of the MTL – whereas rats with lesions in the BG are unimpaired at remembering the previously rewarded locations. In contrast, MTL lesioned rats show intact performance in a stimulus-response version of the maze (the “win-stay” task) where approach behavior is dictated by a conditioned cue, and is impaired in rats with lesions in the dorsal striatum, an input structure in the BG (23–25). This dissociation between the MTL and BG with respect to declarative and nondeclarative-like learning tasks respectively is mirrored across different discrimination related paradigms, such as delayed-nonmatching to sample (26) or concurrent discrimination (27) which suggests that lesions to the MTL adversely affect a monkey’s ability to successfully complete discrimination tasks (27, 28), but not skill learning (28). Additionally, patients suffering from MTL amnesia have extremely compromised declarative memory, while nondeclarative types of learning (e.g., skill learning) remain intact (29–31). Likewise, patients afflicted with PD have difficulty with some forms of trial and error learning, while declarative knowledge persists (30, 32).
There are several examples where this dissociation is challenged, however, primarily based on disagreements on what constitutes declarative versus nondeclarative memory (33), in part as evidenced by conflicting results in patient populations (34–39). For instance, successful probabilistic trial and error learning has been observed in patients with PD (39), leading to the argument that deficits observed may be more nuanced than originally hypothesized, dependent on the type of task (e.g., skill learning versus perceptual) and the locus of impairment (e.g., cortico-striatal circuits versus distributed neocortical association regions) (35, 38). The emergence of neuroeconomics research promises to shed light on this debate by introducing new paradigms and quantitative model-based approaches. A popular method budding from cognitive and computational neuroscience research (40, 41), that is rooted in classical learning theories (42), is the application and characterization of model-based learning signals (9). Widely employed in the interdisciplinary study of neuroeconomics, these model-derived signals can be used to quantify how a cognitive process is implemented in a specific brain region and can be expanded across different memory systems to explore how such systems interact in the context of goal-directed behaviors.
The basal ganglia & reward learning
Learning what choice provides the best outcome can be an arduous process often accomplished by trial and error. In an experimental setting, an animal will learn to predict positive outcomes after repeated successful pairings of a reward with a conditioned stimulus, such as a light or a lever, leading to increased approach behavior towards the stimulus. Early in the decision-making process, therefore, the animal learns to integrate the action-outcome association (e.g., lever-pressing leads to liquid reward) with the value attributed to such outcome (e.g., high value for liquids when thirsty), in a manner that allows constant updating of action-contingencies (e.g., when lever does not yield predicted outcome, a prediction error is coded, leading to updating of current contingencies). These types of studies illustrate how basic learning principles of conditioning can shape current decision-making and future behavior. Across species and through different methodologies, research has converged on the role of the cortico-striatal circuitry in the computation of basic learning principles of conditioning and reinforcement underlying simple decision-making (9, 43–46).
The BG complex is comprised of multiple cortico-striatal loops which exert distinct functions (8). The striatum, in particular, has been linked to various types of learning involving procedural skills (47), habits (30, 48, 49), and reward learning (43, 44, 50). Functional dissociations between subdivisions of the striatum have also been proposed, with the dorsomedial striatum involved in processing action-contingency (43, 51), goal-directed learning (52), and instrumental conditioning (46), and the dorsolateral striatum linked with habit and motor learning (49). Activity in the ventral striatum has been primarily associated with reward-related processing, demonstrating sensitivity to changes in subjective value (9, 53) and prediction-based learning (46). Not surprisingly, these striatum subdivisions are also distinguished with respect to cortical connectivity, with the dorsomedial striatum connected to an associative network with the prefrontal and parietal association cortices, the dorsolateral striatum connected to sensorimotor regions and the ventral striatum connected with ventral portions of the prefrontal cortex (e.g., orbitofrontal cortex) (44, 49).
One of the most thought-provoking findings with respect to BG circuitry and affective learning concerns the role of midbrain dopaminergic neurons in coding prediction error signals (54). It was observed that dopaminergic neurons in non-human primates respond to unexpected rewards (e.g., drop of juice) but no longer respond to the juice once its delivery becomes predictable. Instead, the neurons fire to the earliest predictor of a reward (e.g., tone predicting juice). When a predicted reward fails to occur, dopamine neurons show a depression in neuronal firing, thought to be a prediction error signal used to update contingencies during affective learning. Using functional magnetic resonance imaging (fMRI), prediction error learning signals have correlated with blood oxygen level dependent (BOLD) responses in the human midbrain (55), but primarily in dopaminergic targets such as the striatum using a variety of rewards (e.g., juice, money) and paradigms (e.g., classical and instrumental conditioning) where predictions and contingencies are typically acquired and updated through trial and error (46, 56, 57).
The human striatum has been shown to be involved in various facets of reward-related processing that encompass affective learning and inform future decision-making (9, 44) – from anticipating or predicting a potential reward (58) to coding the differential response between positive and negative outcomes (59), particularly when a contingency between an action and an outcome is perceived (51). Thus, it is not surprising that BOLD responses in the human striatum correlate with prediction error learning signals during conditioning paradigms (46, 60), with stronger correlations suggesting greater learning rates (61) and exploitative, rather than explorative, decision-making (62). More recently, such model-based approaches have begun to be applied to clinical populations with BG dysfunction, namely PD patients. Consistent with the pathology of the disorder, prediction error signals have been found to be preserved in more ventral regions of the striatum, but impaired in dorsolateral regions relative to healthy controls (13) suggesting a potential explanation for why PD patients are impaired in probabilistic trial and error paradigms (30, 32, 39, 63).
The medication status of PD patients during scientific studies (e.g., on or off dopaminergic medicine) has also become a formidable way of investigating the potential role of a dopaminergic learning signal and its influences on decision-making. The mere enhancement or reduction of dopaminergic function by pharmacological treatment (L-DOPA or haloperidol respectively) in healthy participants modulates the magnitude of a prediction error learning signal, suggesting that increases in dopaminergic function lead to choices being perceived as more rewarding (64). Similar manipulations in PD patients hint at specific impairments in learning from positive outcomes when in an unmedicated state (but intact learning from negative outcomes), and normal learning from positive (but not negative) outcomes when treated with dopaminergic agonists (11, 13, 16). Interestingly, in some individual cases of PD, administration of dopaminergic agonists targeting the dorsal striatum can lead to enhanced impulse control impairments such as increased gambling behavior (65–67), posited to be due to an overflow of dopamine in the ventral striatum (68–70). Neuroeconomic approaches of modeling dopaminergic learning signals raise interesting possibilities for future clinical research involving BG dysfunction (13, 16) which allow for extensions beyond the classification of a particular impairment (e.g. learning deficits in feedback-based tasks in PD) to identification of the mechanism underlying such impairment (e.g., learning deficits due to prediction error impairments and medication status). An exciting future direction is the incorporation of multiple learning signals arising from distinct structures, such as the MTL, that underlie different forms of learning signals which participate in goal-directed behavior.
Medial temporal lobe, declarative learning & reward-related contributions
When deciding to try out the new grocery store in town, an individual may rely not only on past experiences with the local shop, but also declarative information obtained regarding the new store. One may read a positive review about the wide array of food available at the new store or be informed by a friend about the great selection of produce (Figure 1). This type of declarative information, thought to be dependent on the integrity of the MTL, is a major influence on our decisions. The MTL consists of many substructures, including the hippocampus, entorhinal cortex, the parahippocampus and perirhinal cortices (71). Collectively, these substructures support multiple types of declarative learning including memory for facts and events (21), novelty detection (72, 73), mediating flexible transfer of knowledge (31, 74, 75), as well as spatial learning and navigation (76, 77). Studies in non-human animals and patient populations with damage to the MTL have demonstrated the importance of this structure in forming new declarative memories and flexibly applying knowledge in new scenarios (6, 14, 31, 78).
More recently, the hippocampus has been highlighted in reward-related studies (79–86). In comparison to the substantial existing evidence of the involvement of distinct MTL structures in various aspects of declarative memory, however, there is still less known about the role of the hippocampus in reward-related learning and decision-making. One important consideration with regards to the hippocampus and reward processing is its connectivity with cortico-striatal circuits (44, 83). Specifically, the hippocampus is anatomically connected with the striatum (87, 88) and it has been suggested that cortical mechanisms may mediate the interaction between them (89). The potential functional significance of these interactions is illustrated in specific learning contexts, such as spatial learning (81). In such context, distinct but parallel contributions to successful path navigation have been proposed for the hippocampus, involved in prospection or the representation of future paths, the ventral striatum, involved in the valuation of future rewards, and the dorsal striatum, tracking action-related task components, that together aid decision-making (77).
The hippocampus also receives projections from midbrain dopaminergic structures (90, 91) and an alternative, perhaps complementary view, is that dopaminergic modulation provides a learning signal for both the hippocampus and the striatum (75, 89). Whether these dopaminergic signals exert similar or distinct influences upon both these structures is an intriguing question for neuroeconomics currently. A popular hypothesis posits that a distinct function of dopaminergic input into the hippocampus facilitates the entry of information into long-term memory by novelty detection (73). Consistent with this idea, dopamine has been thought to play a key role in adaptive memory and determining what information is learned and subsequently remembered (e.g., those associated with a high reward value (79, 83, 86)).
This connectivity between the hippocampus and midbrain dopaminergic centers is also involved in mediating the association between different environmental contexts and reward (92), and the ability to generalize prior learned associations into future decisions (74). Research on neuropsychiatric populations has specified a double dissociation between patients with PD and those with hippocampal atrophy during initial learning (impaired in PD) and the subsequent ability to transfer (impaired in hippocampal patients) during an associative learning task (31), highlighting the role these regions play in learning and flexibly using information. Interestingly, hippocampal damage does not typically correlate with deficits in performance in some learning tasks associated with dopamine and striatal function, such as trial and error feedback-based learning (14, 93). Moreover, during instrumental conditioning tasks with delayed reinforcement, rats with lesions of the hippocampus actually display better learning than control animals (94), perhaps because such lesions hinder the formation of context-outcome associations, focusing instead on action-outcome contingencies dependent on cortico-striatal systems.
A potential similar function of dopaminergic modulation in both the hippocampus and striatum may be the encoding of prediction error learning signals during specific learning contexts. In associative learning tasks, learning-related spiking activity is correlated with a reward-predictive value signal in monkey hippocampal neurons (82), while in human probabilistic learning paradigms, activity in the hippocampus has been suggested to track differences between positive and negative outcomes (95) potentially encoding a feedback-based model-derived reward prediction error signal (96) in parallel with cortico-striatal systems. A challenge for neuroeconomics currently is understanding the functional and temporal significance of this learning signal as the human hippocampus has also been associated with coding for the uncertainty of cue-outcome associations (97), a novelty response (98, 99), and a mismatch signal (100, 101). The few reports that exist thus far advocating an association between the hippocampus and outcome prediction (95, 96, 102) do not necessarily claim that the hippocampal activity exclusively correlates with a scalar prediction error, since plausible alternative models, such as a mismatch signal or an unsigned prediction error, were not tested. Rather, these initial reports point to neuroeconomic approaches as a potential tactic for advancing these questions for future consideration. It is likely that the nature of a learning signal coded in the hippocampus depends on the learning context (e.g., probabilistic task with declarative and nondeclarative elements), and the anatomical location and functional connectivity of loci of activation (e.g., CA1 subfield of the hippocampus has been linked with tracking associative mismatches (100)). Nevertheless, understanding the contributions of the dopaminergic learning signal encoded in the hippocampus may present new directions in decision-making research and translational applications for neuropsychiatric disorders by highlighting how humans flexibly use information in novel contexts to guide future choices.
Interactive memory systems during decision-making
It is theorized that the inadequacy of a single learning and memory system’s ability to deal with a complex situation may have led to the evolution of multiple memory systems (5). To illustrate this in the context of decision-making and neuroeconomics more specifically, one can consider “the trust game”, a classic paradigm in behavioral economics used to investigate risky decision-making and social interactions (103). Briefly, in its simplest form the game involves a one shot interaction between two players, where the first player, the proposer, is given an initial sum of money (e.g., $20) and decides whether to keep it and end the game or invest some or all of that money with the second player, the responder. Any amount invested gets tripled (e.g., $60), and the responder now decides whether to share (or not) some of the investment back to the proposer (e.g., each player gets $30). It is a risky proposition, but a potentially profitable one for the proposer who has to show “trust” in their partner. Adaptations of this game to neuroimaging settings involving repeated interactions evoke trial and error learning and suggest that reputation building is dependent on learning signals being processed in the striatum (104–106), with a shift in the magnitude of the BOLD signal in the striatum to the earliest prediction of a trustworthy interaction based on the development of reputation (106), akin to the shift observed in reinforcement learning models of prediction error signals (54, 57).
The concept of multiple memory systems can be applied to these complex learning and decision-making situations where other types of information can influence decisions. For instance, a proposer may have either good or bad priors with respect to the responder, such as knowledge about their ethical and business practices. In this situation, the additional information affects the belief about probability of reciprocation and decisions to trust (104, 105, 107). Furthermore, learning signals in the striatum involved in updating predictions based on current outcomes are diminished when prior declarative information is present, as if a proposer places less weight on the current experience or outcome due to the value assigned to the prior information (105).
One hypothesis is that declarative knowledge may bias decision-making in this case, potentially overriding exploration of alternatives favored by trial and error learning mechanisms, and suggesting dynamic, perhaps competitive, interactions between the MTL and BG that can affect future choices. This is consistent with evidence from probabilistic learning studies which propose that these memory systems may compete during some learning scenarios (108, 109). An alternative hypothesis is that MTL and BG may at times cooperate during learning situations, thus allowing compensation for a faulty system in a state where the brain is diseased (110, 111) or when such an interaction is behaviorally relevant (112).
A third idea is that the nature of the interactions between MTL and BG involves parallel processing of information, with the interactions being competitive or cooperative depending on task demands (25, 113, 114). This debate across the theory of multiple memory systems regarding the nature of the interactions between MTL and BG can profit from neuroeconomics research and the setting conjectured by experimental paradigms such as the trust game. For example, in the trust game the dynamic interactions between the MTL and BG could be perceived as competitive, however reinforcement learning signals have been observed during probabilistic learning tasks in both structures (95, 96), conceivably suggesting parallel engagement or even a cooperative interaction between these regions. Introducing instructed knowledge to a probabilistic learning paradigm diminishes the reinforcement learning signals in these areas, concurrent with an increase in activity in the dorsolateral prefrontal cortex (95), posited to mediate the interaction between the MTL and BG (89). Thus, parallel processing within these distinct neural circuits in the MTL and BG, which are influenced by dopaminergic input and shaped by cortical modulation, may underlie learning and decision-making processes that occur during complex social interactions such as the trust game.
Further considerations
Several models posited to underlie decision-making processes have been highlighted in the literature and merit brief discussion (3, 4, 9, 115). One dual system account of decision-making, for example, distinguishes between a cognitively-based, controlled system (deliberative system) and an automatic, emotional system (impulsive system), acutely important during inter-temporal choices (116–119). Deliberative processes have been posited to involve prefrontal and parietal cortices, while automatic processes have been linked with subcortical regions such as the striatum ((4, 119); although see (116)). Another influential model highlights goal-directed and habitual learning as two key systems involved in guiding choice behavior (115, 120), with goal-directed learning involving the anterior caudate nucleus and medial prefrontal regions, and habitual learning associated with the posterior putamen (48, 49, 115). Current work is also exploring a distinction between reinforcement learning error signals related to reward prediction, referred to as model-free, and expectations regarding current states given previous choices, referred to as model-based (121, 122).
Integrating these models in a cohesive account of how distinct neural systems involved in learning, such as the striatum (typically highlighted across models) and the hippocampus (observed less often), can inform decision-making is an important future endeavor. One recent proposal posits four key decision-making systems in the brain: goal-directed (dorsolateral PFC and dorsomedial striatum), habitual (dorsolateral striatum), episodic (MTL), and Pavlovian (123). These distinct systems help guide decisions independently, but may also interact to steer choices depending on the context of the decision.
As this review primarily focused on the potential contributions of learning signals associated with the BG and MTL to decision-making, less discussion was attributed to other important regions that contribute to learning and decision-making, namely the amygdala and the prefrontal cortex (124, 125). The amygdala is a critical structure for aversive learning (126) and emotional memory (127) that has implications for decision-making in particular contexts, such as when decisions involve ambiguity (128) or are framed as gains or losses (129). Interestingly, the amygdala is another potential link between interactions of the MTL and BG, as it has been shown to modulate activity in these structures during specific learning tasks (e.g., win-shift and win-stay respectively; (130)). Thus, a clear future direction is how learning signals originating from the amygdala can influence the MTL and BG during decision-making processes.
In contrast, the involvement of the PFC in decision-making processes has been well documented (4, 9, 115). Both dorsolateral and ventromedial prefrontal cortical regions, in particular, have been associated with the representation of goal values (9) and integrating reward-related information along with a striatal and dopaminergic network (131), consistent with the previously highlighted theories of cortical modulation of MTL and BG interactions (89). Future research using computational models of learning signals will prove useful in further understanding the relationship between PFC regions and the striatum during action-outcome learning (95) as potentially informed by cortico-hippocampal connections that mediate rule governance (132).
Conclusion
Understanding how neural structures underlying distinct learning and memory systems interact has profound potential to inform neuropsychiatric disorders, especially in populations where decision-making is adversely affected by the disease or disease treatment (e.g., pathological gambling and PD (65–67)). The advent of fMRI has allowed for extension of elegant animal models and patient studies in support of multiple learning and memory systems – although a debate remains with respect to how underlying neural structures, namely the BG and MTL, interact to facilitate learning and decision-making. The emergence of neuroeconomic approaches, such as the application of signals derived from models of cognitive processes that are used to more carefully identify neural mechanisms, promises to shed light on this debate and extend this knowledge beyond basic principles of learning that inform decision-making to other important forms of learning common to humans (e.g., instructed and observational learning (95, 133)). The evolution of neuroeconomics based models, paradigms, and initial efforts of their application in special neuropsychiatric populations bodes well for future advances in understanding the connectivity and interactions of neural mechanisms involved in learning, memory, and decision-making to promote more efficient diagnosis and treatment options.
Acknowledgments
This work was supported by a National Institute of Mental Health grant (MH-084041) and a National Science Foundation Grant (0718153).
References
- 1.Glimcher PW, Camerer C, Fehr E, Poldrack RA. Neuroeconomics Decision Making and the Brain. London, UK: Academic Press; 2009. [Google Scholar]
- 2.Dayan P, Niv Y. Reinforcement learning: the good, the bad and the ugly. Curr Opin Neurobiol. 2008;18:185–196. doi: 10.1016/j.conb.2008.08.003. [DOI] [PubMed] [Google Scholar]
- 3.Frank MJ, Cohen MX, Sanfey AG. Multiple systems in decision making: A neurocomputational perspective. Current Directions in Psychological Science. 2009;18:73–77. [Google Scholar]
- 4.Sanfey AG, Loewenstein G, McClure SM, Cohen JD. Neuroeconomics: cross-currents in research on decision-making. Trends Cogn Sci. 2006;10:108–116. doi: 10.1016/j.tics.2006.01.009. [DOI] [PubMed] [Google Scholar]
- 5.Sherry DF, Schacter DL. The evolution of multiple memory systems. Psychological Review. 1987;94:439–454. [Google Scholar]
- 6.Squire LR, Zola SM. Structure and function of declarative and nondeclarative memory systems. Proc Natl Acad Sci U S A. 1996;93:13515–13522. doi: 10.1073/pnas.93.24.13515. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Balleine BW, Delgado MR, Hikosaka O. The role of the dorsal striatum in reward and decision-making. J Neurosci. 2007;27:8161–8165. doi: 10.1523/JNEUROSCI.1554-07.2007. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Middleton FA, Strick PL. Basal ganglia output and cognition: evidence from anatomical, behavioral, and clinical studies. Brain Cogn. 2000;42:183–200. doi: 10.1006/brcg.1999.1099. [DOI] [PubMed] [Google Scholar]
- 9.Rangel A, Camerer C, Montague PR. A framework for studying the neurobiology of value-based decision making. Nat Rev Neurosci. 2008;9:545–556. doi: 10.1038/nrn2357. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Cools R, Barker RA, Sahakian BJ, Robbins TW. Mechanisms of cognitive set flexibility in Parkinson’s disease. Brain. 2001;124:2503–2512. doi: 10.1093/brain/124.12.2503. [DOI] [PubMed] [Google Scholar]
- 11.Frank MJ, Seeberger LC, O’Reilly RC. By carrot or by stick: cognitive reinforcement learning in parkinsonism. Science. 2004;306:1940–1943. doi: 10.1126/science.1102941. [DOI] [PubMed] [Google Scholar]
- 12.Peterson DA, Elliott C, Song DD, Makeig S, Sejnowski TJ, Poizner H. Probabilistic reversal learning is impaired in Parkinson’s disease. Neuroscience. 2009;163:1092–1101. doi: 10.1016/j.neuroscience.2009.07.033. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Schonberg T, O’Doherty JP, Joel D, Inzelberg R, Segev Y, Daw ND. Selective impairment of prediction error signaling in human dorsolateral but not ventral striatum in Parkinson’s disease patients: evidence from a model-based fMRI study. Neuroimage. 2010;49:772–781. doi: 10.1016/j.neuroimage.2009.08.011. [DOI] [PubMed] [Google Scholar]
- 14.Shohamy D, Myers CE, Hopkins RO, Sage J, Gluck MA. Distinct hippocampal and basal ganglia contributions to probabilistic learning and reversal. J Cogn Neurosci. 2009;21:1821–1833. doi: 10.1162/jocn.2009.21138. [DOI] [PubMed] [Google Scholar]
- 15.Shohamy D, Myers CE, Onlaor S, Gluck MA. Role of the basal ganglia in category learning: how do patients with Parkinson’s disease learn? Behav Neurosci. 2004;118:676–686. doi: 10.1037/0735-7044.118.4.676. [DOI] [PubMed] [Google Scholar]
- 16.Rutledge RB, Lazzaro SC, Lau B, Myers CE, Gluck MA, Glimcher PW. Dopaminergic drugs modulate learning rates and perseveration in Parkinson’s patients in a dynamic foraging task. J Neurosci. 2009;29:15104–15114. doi: 10.1523/JNEUROSCI.3524-09.2009. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Lang AE, Lozano AM. Parkinson’s disease. First of two parts. N Engl J Med. 1998;339:1044–1053. doi: 10.1056/NEJM199810083391506. [DOI] [PubMed] [Google Scholar]
- 18.Lang AE, Lozano AM. Parkinson’s disease. Second of two parts. N Engl J Med. 1998;339:1130–1143. doi: 10.1056/NEJM199810153391607. [DOI] [PubMed] [Google Scholar]
- 19.Packard MG, Knowlton BJ. Learning and memory functions of the Basal Ganglia. Annu Rev Neurosci. 2002;25:563–593. doi: 10.1146/annurev.neuro.25.112701.142937. [DOI] [PubMed] [Google Scholar]
- 20.Squire LR. Declarative and nondeclarative memory: Multiple brain systems supporting learning and memory. Journal of Cognitive Neuroscience. 1992;4:232–243. doi: 10.1162/jocn.1992.4.3.232. [DOI] [PubMed] [Google Scholar]
- 21.Squire LR. Memory systems of the brain: a brief history and current perspective. Neurobiol Learn Mem. 2004;82:171–177. doi: 10.1016/j.nlm.2004.06.005. [DOI] [PubMed] [Google Scholar]
- 22.Reber PJ, Knowlton BJ, Squire LR. Dissociable properties of memory systems: differences in the flexibility of declarative and nondeclarative knowledge. Behav Neurosci. 1996;110:861–871. doi: 10.1037//0735-7044.110.5.861. [DOI] [PubMed] [Google Scholar]
- 23.Packard MG, Hirsh R, White NM. Differential effects of fornix and caudate nucleus lesions on two radial maze tasks: evidence for multiple memory systems. J Neurosci. 1989;9:1465–1472. doi: 10.1523/JNEUROSCI.09-05-01465.1989. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24.Packard MG, McGaugh JL. Double dissociation of fornix and caudate nucleus lesions on acquisition of two water maze tasks: further evidence for multiple memory systems. Behav Neurosci. 1992;106:439–446. doi: 10.1037//0735-7044.106.3.439. [DOI] [PubMed] [Google Scholar]
- 25.White NM, McDonald RJ. Multiple parallel memory systems in the brain of the rat. Neurobiol Learn Mem. 2002;77:125–184. doi: 10.1006/nlme.2001.4008. [DOI] [PubMed] [Google Scholar]
- 26.Alvarez P, Zola-Morgan S, Squire LR. Damage limited to the hippocampal region produces long-lasting memory impairment in monkeys. May 1995. The Journal of Neuroscience. 1995;15:3796–3807. doi: 10.1523/JNEUROSCI.15-05-03796.1995. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Zola-Morgan S, Squire LR. Medial temporal lesions in monkeys impair memory on a variety of tasks sensitive to human amnesia. Behav Neurosci. 1985;99:22–34. doi: 10.1037//0735-7044.99.1.22. [DOI] [PubMed] [Google Scholar]
- 28.Zola-Morgan S, Squire LR. Preserved learning in monkeys with medial temporal lesions: sparing of motor and cognitive skills. J Neurosci. 1984;4:1072–1085. doi: 10.1523/JNEUROSCI.04-04-01072.1984. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Corkin S. Lasting Consequences of Bilateral Medial Temporal Lobectomy - Clincal Course and Experimental Findings in HM. Seminars in Neurology. 1984;4:249–259. [Google Scholar]
- 30.Knowlton BJ, Mangels JA, Squire LR. A neostriatal habit learning system in humans. Science. 1996;273:1399–1402. doi: 10.1126/science.273.5280.1399. [DOI] [PubMed] [Google Scholar]
- 31.Myers CE, Shohamy D, Gluck MA, Grossman S, Kluger A, Ferris S, et al. Dissociating hippocampal versus basal ganglia contributions to learning and transfer. J Cogn Neurosci. 2003;15:185–193. doi: 10.1162/089892903321208123. [DOI] [PubMed] [Google Scholar]
- 32.Shohamy D, Myers CE, Grossman S, Sage J, Gluck MA, Poldrack RA. Cortico-striatal contributions to feedback-based learning: converging data from neuroimaging and neuropsychology. Brain. 2004;127:851–859. doi: 10.1093/brain/awh100. [DOI] [PubMed] [Google Scholar]
- 33.Henke K. A model for memory systems based on processing modes rather than consciousness. Nat Rev Neurosci. 2010;11:523–532. doi: 10.1038/nrn2850. [DOI] [PubMed] [Google Scholar]
- 34.Bondi MW, Kaszniak AW. Implicit and explicit memory in Alzheimer’s disease and Parkinson’s disease. J Clin Exp Neuropsychol. 1991;13:339–358. doi: 10.1080/01688639108401048. [DOI] [PubMed] [Google Scholar]
- 35.Heindel WC, Salmon DP, Shults CW, Walicke PA, Butters N. Neuropsychological evidence for multiple implicit memory systems: a comparison of Alzheimer’s, Huntington’s, and Parkinson’s disease patients. J Neurosci. 1989;9:582–587. doi: 10.1523/JNEUROSCI.09-02-00582.1989. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 36.Meeter M, Radics G, Myers CE, Gluck MA, Hopkins RO. Probabilistic categorization: how do normal participants and amnesic patients do it? Neurosci Biobehav Rev. 2008;32:237–248. doi: 10.1016/j.neubiorev.2007.11.001. [DOI] [PubMed] [Google Scholar]
- 37.Owen AM, Beksinska M, James M, Leigh PN, Summers BA, Marsden CD, et al. Visuospatial memory deficits at different stages of Parkinson’s disease. Neuropsychologia. 1993;31:627–644. doi: 10.1016/0028-3932(93)90135-m. [DOI] [PubMed] [Google Scholar]
- 38.Reber PJ, Squire LR. Relaxing decision criteria does not improve recognition memory in amnesic patients. Mem Cognit. 1999;27:501–511. doi: 10.3758/bf03211544. [DOI] [PubMed] [Google Scholar]
- 39.Wilkinson L, Lagnado DA, Quallo M, Jahanshahi M. The effect of feedback on non-motor probabilistic classification learning in Parkinson’s disease. Neuropsychologia. 2008;46:2683–2695. doi: 10.1016/j.neuropsychologia.2008.05.008. [DOI] [PubMed] [Google Scholar]
- 40.O’Doherty JP, Hampton A, Kim H. Model-based fMRI and its application to reward learning and decision making. Ann N Y Acad Sci. 2007;1104:35–53. doi: 10.1196/annals.1390.022. [DOI] [PubMed] [Google Scholar]
- 41.Sutton RS, Barto AG. Reinforcement Learning. Cambridge, M.A: MIT Press; 1998. [Google Scholar]
- 42.Rescorla RA, Wagner AR. A theory of Pavlovian conditioning: variations in the effectiveness of reinforcement and nonreinforcement. In: Black AH, Prokasy WF, editors. Classical Conditioning II: Current Research and Theory. New York: Appleton Century Crofts; 1972. pp. 64–99. [Google Scholar]
- 43.Delgado MR. Reward-related responses in the human striatum. Ann N Y Acad Sci. 2007;1104:70–88. doi: 10.1196/annals.1390.002. [DOI] [PubMed] [Google Scholar]
- 44.Haber SN, Knutson B. The reward circuit: linking primate anatomy and human imaging. Neuropsychopharmacology. 2010;35:4–26. doi: 10.1038/npp.2009.129. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 45.Montague PR, Berns GS. Neural economics and the biological substrates of valuation. Neuron. 2002;36:265–284. doi: 10.1016/s0896-6273(02)00974-1. [DOI] [PubMed] [Google Scholar]
- 46.O’Doherty J, Dayan P, Schultz J, Deichmann R, Friston K, Dolan RJ. Dissociable roles of ventral and dorsal striatum in instrumental conditioning. Science. 2004;304:452–454. doi: 10.1126/science.1094285. [DOI] [PubMed] [Google Scholar]
- 47.Poldrack RA, Prabhakaran V, Seger CA, Gabrieli JD. Striatal activation during acquisition of a cognitive skill. Neuropsychology. 1999;13:564–574. doi: 10.1037//0894-4105.13.4.564. [DOI] [PubMed] [Google Scholar]
- 48.Tricomi E, Balleine BW, O’Doherty JP. A specific role for posterior dorsolateral striatum in human habit learning. Eur J Neurosci. 2009;29:2225–2232. doi: 10.1111/j.1460-9568.2009.06796.x. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 49.Yin HH, Knowlton BJ. The role of the basal ganglia in habit formation. Nat Rev Neurosci. 2006;7:464–476. doi: 10.1038/nrn1919. [DOI] [PubMed] [Google Scholar]
- 50.O’Doherty JP. Reward representations and reward-related learning in the human brain: insights from neuroimaging. Curr Opin Neurobiol. 2004;14:769–776. doi: 10.1016/j.conb.2004.10.016. [DOI] [PubMed] [Google Scholar]
- 51.Tricomi EM, Delgado MR, Fiez JA. Modulation of caudate activity by action contingency. Neuron. 2004;41:281–292. doi: 10.1016/s0896-6273(03)00848-1. [DOI] [PubMed] [Google Scholar]
- 52.Voorn P, Vanderschuren LJ, Groenewegen HJ, Robbins TW, Pennartz CM. Putting a spin on the dorsal-ventral divide of the striatum. Trends Neurosci. 2004;27:468–474. doi: 10.1016/j.tins.2004.06.006. [DOI] [PubMed] [Google Scholar]
- 53.Knutson B, Taylor J, Kaufman M, Peterson R, Glover G. Distributed neural representation of expected value. J Neurosci. 2005;25:4806–4812. doi: 10.1523/JNEUROSCI.0642-05.2005. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 54.Schultz W, Dayan P, Montague PR. A neural substrate of prediction and reward. Science. 1997;275:1593–1599. doi: 10.1126/science.275.5306.1593. [DOI] [PubMed] [Google Scholar]
- 55.D’Ardenne K, McClure SM, Nystrom LE, Cohen JD. BOLD responses reflecting dopaminergic signals in the human ventral tegmental area. Science. 2008;319:1264–1267. doi: 10.1126/science.1150605. [DOI] [PubMed] [Google Scholar]
- 56.McClure SM, Berns GS, Montague PR. Temporal prediction errors in a passive learning task activate human striatum. Neuron. 2003;38:339–346. doi: 10.1016/s0896-6273(03)00154-5. [DOI] [PubMed] [Google Scholar]
- 57.O’Doherty JP, Dayan P, Friston K, Critchley H, Dolan RJ. Temporal difference models and reward-related learning in the human brain. Neuron. 2003;38:329–337. doi: 10.1016/s0896-6273(03)00169-7. [DOI] [PubMed] [Google Scholar]
- 58.Knutson B, Adams CM, Fong GW, Hommer D. Anticipation of increasing monetary reward selectively recruits nucleus accumbens. The Journal of Neuroscience. 2001;21:RC159. doi: 10.1523/JNEUROSCI.21-16-j0002.2001. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 59.Delgado MR, Nystrom LE, Fissell C, Noll DC, Fiez JA. Tracking the hemodynamic responses to reward and punishment in the striatum. J Neurophysiol. 2000;84:3072–3077. doi: 10.1152/jn.2000.84.6.3072. [DOI] [PubMed] [Google Scholar]
- 60.Haruno M, Kawato M. Different neural correlates of reward expectation and reward expectation error in the putamen and caudate nucleus during stimulus-action-reward association learning. J Neurophysiol. 2006;95:948–959. doi: 10.1152/jn.00382.2005. [DOI] [PubMed] [Google Scholar]
- 61.Schonberg T, Daw ND, Joel D, O’Doherty JP. Reinforcement learning signals in the human striatum distinguish learners from nonlearners during reward-based decision making. J Neurosci. 2007;27:12860–12867. doi: 10.1523/JNEUROSCI.2496-07.2007. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 62.Daw ND, O’Doherty JP, Dayan P, Seymour B, Dolan RJ. Cortical substrates for exploratory decisions in humans. Nature. 2006;441:876–879. doi: 10.1038/nature04766. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 63.Knowlton BJ, Squire LR, Gluck MA. Probabilistic classification learning in amnesia. Learn Mem. 1994;1:106–120. [PubMed] [Google Scholar]
- 64.Pessiglione M, Seymour B, Flandin G, Dolan RJ, Frith CD. Dopamine-dependent prediction errors underpin reward-seeking behaviour in humans. Nature. 2006;442:1042–1045. doi: 10.1038/nature05051. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 65.Dodd ML, Klos KJ, Bower JH, Geda YE, Josephs KA, Ahlskog JE. Pathological gambling caused by drugs used to treat Parkinson disease. Arch Neurol. 2005;62:1377–1381. doi: 10.1001/archneur.62.9.noc50009. [DOI] [PubMed] [Google Scholar]
- 66.Molina JA, Sainz-Artiga MJ, Fraile A, Jimenez-Jimenez FJ, Villanueva C, Orti-Pareja M, et al. Pathologic gambling in Parkinson’s disease: a behavioral manifestation of pharmacologic treatment? Mov Disord. 2000;15:869–872. doi: 10.1002/1531-8257(200009)15:5<869::aid-mds1016>3.0.co;2-i. [DOI] [PubMed] [Google Scholar]
- 67.Voon V, Fox SH. Medication-related impulse control and repetitive behaviors in Parkinson disease. Arch Neurol. 2007;64:1089–1096. doi: 10.1001/archneur.64.8.1089. [DOI] [PubMed] [Google Scholar]
- 68.Cools R, Barker RA, Sahakian BJ, Robbins TW. Enhanced or impaired cognitive function in Parkinson’s disease as a function of dopaminergic medication and task demands. Cereb Cortex. 2001;11:1136–1143. doi: 10.1093/cercor/11.12.1136. [DOI] [PubMed] [Google Scholar]
- 69.Gotham AM, Brown RG, Marsden CD. ‘Frontal’ cognitive function in patients with Parkinson’s disease ‘on’ and ‘off’ levodopa. Brain. 1988;111(Pt 2):299–321. doi: 10.1093/brain/111.2.299. [DOI] [PubMed] [Google Scholar]
- 70.Swainson R, Rogers RD, Sahakian BJ, Summers BA, Polkey CE, Robbins TW. Probabilistic learning and reversal deficits in patients with Parkinson’s disease or frontal or temporal lobe lesions: possible adverse effects of dopaminergic medication. Neuropsychologia. 2000;38:596–612. doi: 10.1016/s0028-3932(99)00103-7. [DOI] [PubMed] [Google Scholar]
- 71.Squire LR, Stark CE, Clark RE. The medial temporal lobe. Annu Rev Neurosci. 2004;27:279–306. doi: 10.1146/annurev.neuro.27.070203.144130. [DOI] [PubMed] [Google Scholar]
- 72.Knight R. Contribution of human hippocampal region to novelty detection. Nature. 1996;383:256–259. doi: 10.1038/383256a0. [DOI] [PubMed] [Google Scholar]
- 73.Lisman JE, Grace AA. The hippocampal-VTA loop: controlling the entry of information into long-term memory. Neuron. 2005;46:703–713. doi: 10.1016/j.neuron.2005.05.002. [DOI] [PubMed] [Google Scholar]
- 74.Shohamy D, Wagner AD. Integrating memories in the human brain: hippocampal-midbrain encoding of overlapping events. Neuron. 2008;60:378–389. doi: 10.1016/j.neuron.2008.09.023. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 75.Wimmer GE, Shohamy D. The striatum and beyond: Hippocampal contributions to decision making. In: Delgado MR, Phelps EA, Robbins TW, editors. Attention and Performance XXII. Oxford, UK: Oxford University Press; 2011. [Google Scholar]
- 76.Burgess N, Maguire EA, O’Keefe J. The human hippocampus and spatial and episodic memory. Neuron. 2002;35:625–641. doi: 10.1016/s0896-6273(02)00830-9. [DOI] [PubMed] [Google Scholar]
- 77.van der Meer MA, Johnson A, Schmitzer-Torbert NC, Redish AD. Triple dissociation of information processing in dorsal striatum, ventral striatum, and hippocampus on a learned spatial decision task. Neuron. 2010;67:25–32. doi: 10.1016/j.neuron.2010.06.023. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 78.Corkin S. What’s new with the amnesic patient H.M.? Nat Rev Neurosci. 2002;3:153–160. doi: 10.1038/nrn726. [DOI] [PubMed] [Google Scholar]
- 79.Adcock RA, Thangavel A, Whitfield-Gabrieli S, Knutson B, Gabrieli JD. Reward-motivated learning: mesolimbic activation precedes memory formation. Neuron. 2006;50:507–517. doi: 10.1016/j.neuron.2006.03.036. [DOI] [PubMed] [Google Scholar]
- 80.Johnson A, Redish AD. Hippocampal replay contributes to within session learning in a temporal difference reinforcement learning model. Neural Netw. 2005;18:1163–1171. doi: 10.1016/j.neunet.2005.08.009. [DOI] [PubMed] [Google Scholar]
- 81.Johnson A, van der Meer MAA, Redish AD. Integrating hippocampus and striatum in decision-making. Current Opinion in Neurobiology. 2007;17:692–697. doi: 10.1016/j.conb.2008.01.003. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 82.Okatan M. Correlates of reward-predictive value in learning-related hippocampal neural activity. Hippocampus. 2009;19:487–506. doi: 10.1002/hipo.20535. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 83.Shohamy D, Adcock RA. Dopamine and adaptive memory. Trends Cogn Sci. 2010;14:464–472. doi: 10.1016/j.tics.2010.08.002. [DOI] [PubMed] [Google Scholar]
- 84.Wirth S, Avsar E, Chiu CC, Sharma V, Smith AC, Brown E, et al. Trial outcome and associative learning signals in the monkey hippocampus. Neuron. 2009;61:930–940. doi: 10.1016/j.neuron.2009.01.012. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 85.Wittmann BC, Bunzeck N, Dolan RJ, Duzel E. Anticipation of novelty recruits reward system and hippocampus while promoting recollection. Neuroimage. 2007;38:194–202. doi: 10.1016/j.neuroimage.2007.06.038. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 86.Wittmann BC, Schott BH, Guderian S, Frey JU, Heinze HJ, Duzel E. Reward-related FMRI activation of dopaminergic midbrain is associated with enhanced hippocampus-dependent long-term memory formation. Neuron. 2005;45:459–467. doi: 10.1016/j.neuron.2005.01.010. [DOI] [PubMed] [Google Scholar]
- 87.Brog JS, Salyapongse A, Deutch AY, Zahm DS. The patterns of afferent innervation of the core and shell in the “accumbens” part of the rat ventral striatum: immunohistochemical detection of retrogradely transported fluoro-gold. J Comp Neurol. 1993;338:255–278. doi: 10.1002/cne.903380209. [DOI] [PubMed] [Google Scholar]
- 88.Groenewegen HJ, Vermeulen-Van der Zee E, te Kortschot A, Witter MP. Organization of the projections from the subiculum to the ventral striatum in the rat. A study using anterograde transport of Phaseolus vulgaris leucoagglutinin. Neuroscience. 1987;23:103–120. doi: 10.1016/0306-4522(87)90275-2. [DOI] [PubMed] [Google Scholar]
- 89.Poldrack RA, Rodriguez P. How do memory systems interact? Evidence from human classification learning. Neurobiol Learn Mem. 2004;82:324–332. doi: 10.1016/j.nlm.2004.05.003. [DOI] [PubMed] [Google Scholar]
- 90.Bolam JP, Hanley JJ, Booth PA, Bevan MD. Synaptic organisation of the basal ganglia. J Anat. 2000;196(Pt 4):527–542. doi: 10.1046/j.1469-7580.2000.19640527.x. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 91.Swanson LW. The projections of the ventral tegmental area and adjacent regions: a combined fluorescent retrograde tracer and immunofluorescence study in the rat. Brain Res Bull. 1982;9:321–353. doi: 10.1016/0361-9230(82)90145-9. [DOI] [PubMed] [Google Scholar]
- 92.Luo AH, Tahsili-Fahadan P, Wise RA, Lupica CR, Aston-Jones G. Linking context with reward: a functional circuit from hippocampal CA3 to ventral tegmental area. Science. 2011;333:353–357. doi: 10.1126/science.1204622. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 93.Filoteo JV, Maddox WT, Davis JD. Quantitative modeling of category learning in amnesic patients. J Int Neuropsychol Soc. 2001;7:1–19. doi: 10.1017/s1355617701711010. [DOI] [PubMed] [Google Scholar]
- 94.Cheung TH, Cardinal RN. Hippocampal lesions facilitate instrumental learning with delayed reinforcement but induce impulsive choice in rats. BMC Neurosci. 2005;6:36. doi: 10.1186/1471-2202-6-36. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 95.Li J, Delgado MR, Phelps EA. How instructed knowledge modulates the neural systems of reward learning. Proc Natl Acad Sci U S A. 2011;108:55–60. doi: 10.1073/pnas.1014938108. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 96.Dickerson KC, Li J, Delgado MR. Parallel contributions of distinct human memory systems during probabilistic learning. Neuroimage. 2011;55:266–276. doi: 10.1016/j.neuroimage.2010.10.080. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 97.Vanni-Mercier G, Mauguiere F, Isnard J, Dreher JC. The hippocampus codes the uncertainty of cue-outcome associations: an intracranial electrophysiological study in humans. J Neurosci. 2009;29:5287–5294. doi: 10.1523/JNEUROSCI.5298-08.2009. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 98.Fyhn M, Molden S, Hollup S, Moser MB, Moser E. Hippocampal neurons responding to first-time dislocation of a target object. Neuron. 2002;35:555–566. doi: 10.1016/s0896-6273(02)00784-5. [DOI] [PubMed] [Google Scholar]
- 99.Kumaran D, Maguire EA. Novelty signals: a window into hippocampal information processing. Trends Cogn Sci. 2009;13:47–54. doi: 10.1016/j.tics.2008.11.004. [DOI] [PubMed] [Google Scholar]
- 100.Chen J, Olsen RK, Preston AR, Glover GH, Wagner AD. Associative retrieval processes in the human medial temporal lobe: Hippocampal retrieval success and CA1 mismatch detection. Learn Mem. 2011;18:523–528. doi: 10.1101/lm.2135211. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 101.Kumaran D, Maguire EA. An unexpected sequence of events: mismatch detection in the human hippocampus. PLoS Biol. 2006;4:e424. doi: 10.1371/journal.pbio.0040424. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 102.Foerde K, Shohamy D. Feedback timing modulates brain systems for learning in humans. J Neurosci. 2011;31:13157–13167. doi: 10.1523/JNEUROSCI.2701-11.2011. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 103.Berg J, Dickhaut J, McCabe K. Trust, Reciprocity, and Social History. Games and Economic Behavior. 1995;10:122–142. [Google Scholar]
- 104.Baumgartner T, Heinrichs M, Vonlanthen A, Fischbacher U, Fehr E. Oxytocin shapes the neural circuitry of trust and trust adaptation in humans. Neuron. 2008;58:639–650. doi: 10.1016/j.neuron.2008.04.009. [DOI] [PubMed] [Google Scholar]
- 105.Delgado MR, Frank RH, Phelps EA. Perceptions of moral character modulate the neural systems of reward during the trust game. Nat Neurosci. 2005;8:1611–1618. doi: 10.1038/nn1575. [DOI] [PubMed] [Google Scholar]
- 106.King-Casas B, Tomlin D, Anen C, Camerer CF, Quartz SR, Montague PR. Getting to know you: reputation and trust in a two-person economic exchange. Science. 2005;308:78–83. doi: 10.1126/science.1108062. [DOI] [PubMed] [Google Scholar]
- 107.Chang LJ, Doll BB, van’t Wout M, Frank MJ, Sanfey AG. Seeing is believing: Trustworthiness as a dynamic belief. Cognitive Psychology. 2010;61:87–105. doi: 10.1016/j.cogpsych.2010.03.001. [DOI] [PubMed] [Google Scholar]
- 108.Lee AS, Duman RS, Pittenger C. A double dissociation revealing bidirectional competition between striatum and hippocampus during learning. Proc Natl Acad Sci U S A. 2008;105:17163–17168. doi: 10.1073/pnas.0807749105. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 109.Poldrack RA, Clark J, Pare-Blagoev EJ, Shohamy D, Creso Moyano J, Myers C, et al. Interactive memory systems in the human brain. Nature. 2001;414:546–550. doi: 10.1038/35107080. [DOI] [PubMed] [Google Scholar]
- 110.Dagher A, Owen AM, Boecker H, Brooks DJ. The role of the striatum and hippocampus in planning: a PET activation study in Parkinson’s disease. Brain. 2001;124:1020–1032. doi: 10.1093/brain/124.5.1020. [DOI] [PubMed] [Google Scholar]
- 111.Voermans NC, Petersson KM, Daudey L, Weber B, Van Spaendonck KP, Kremer HP, et al. Interaction between the human hippocampus and the caudate nucleus during route recognition. Neuron. 2004;43:427–435. doi: 10.1016/j.neuron.2004.07.009. [DOI] [PubMed] [Google Scholar]
- 112.Sadeh T, Shohamy D, Levy DR, Reggev N, Maril A. Cooperation between the hippocampus and the striatum during episodic encoding. J Cogn Neurosci. 2011;23:1597–1608. doi: 10.1162/jocn.2010.21549. [DOI] [PubMed] [Google Scholar]
- 113.Doeller CF, King JA, Burgess N. Parallel striatal and hippocampal systems for landmarks and boundaries in spatial memory. Proc Natl Acad Sci U S A. 2008;105:5915–5920. doi: 10.1073/pnas.0801489105. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 114.Tricomi E, Fiez JA. Feedback signals in the caudate reflect goal achievement on a declarative memory task. Neuroimage. 2008;41:1154–1167. doi: 10.1016/j.neuroimage.2008.02.066. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 115.Balleine BW, O’Doherty JP. Human and rodent homologies in action control: corticostriatal determinants of goal-directed and habitual action. Neuropsychopharmacology. 2010;35:48–69. doi: 10.1038/npp.2009.131. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 116.Kable JW, Glimcher PW. The neural correlates of subjective value during intertemporal choice. Nat Neurosci. 2007;10:1625–1633. doi: 10.1038/nn2007. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 117.Kahneman D, Tversky A. Prospect theory: an analysis of decision under risk. Econometrica. 1979;47:263–292. [Google Scholar]
- 118.McClure SM, Ericson KM, Laibson DI, Loewenstein G, Cohen JD. Time discounting for primary rewards. J Neurosci. 2007;27:5796–5804. doi: 10.1523/JNEUROSCI.4246-06.2007. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 119.McClure SM, Laibson DI, Loewenstein G, Cohen JD. Separate neural systems value immediate and delayed monetary rewards. Science. 2004;306:503–507. doi: 10.1126/science.1100907. [DOI] [PubMed] [Google Scholar]
- 120.Balleine BW, Dickinson A. The effect of lesions of the insular cortex on instrumental conditioning: evidence for a role in incentive memory. J Neurosci. 2000;20:8954–8964. doi: 10.1523/JNEUROSCI.20-23-08954.2000. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 121.Daw ND, Gershman SJ, Seymour B, Dayan P, Dolan RJ. Model-based influences on humans’ choices and striatal prediction errors. Neuron. 2011;69:1204–1215. doi: 10.1016/j.neuron.2011.02.027. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 122.Glascher J, Daw N, Dayan P, O’Doherty JP. States versus rewards: dissociable neural prediction error signals underlying model-based and model-free reinforcement learning. Neuron. 2010;66:585–595. doi: 10.1016/j.neuron.2010.04.016. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 123.Dayan P. Better Than Conscious? Decision Making, the Human Mind, and Implications for Institutions. Cambridge: MIT Press; 2008. The Role of Value Systems in Decision Making. In: Engel C, Singer W, editors; pp. 51–70. [Google Scholar]
- 124.Kringelbach ML. The human orbitofrontal cortex: linking reward to hedonic experience. Nat Rev Neurosci. 2005;6:691–702. doi: 10.1038/nrn1747. [DOI] [PubMed] [Google Scholar]
- 125.Seymour B, Dolan R. Emotion, decision making, and the amygdala. Neuron. 2008;58:662–671. doi: 10.1016/j.neuron.2008.05.020. [DOI] [PubMed] [Google Scholar]
- 126.Phelps EA, LeDoux JE. Contributions of the amygdala to emotion processing: from animal models to human behavior. Neuron. 2005;48:175–187. doi: 10.1016/j.neuron.2005.09.025. [DOI] [PubMed] [Google Scholar]
- 127.McGaugh JL. The amygdala modulates the consolidation of memories of emotionally arousing experiences. Annu Rev Neurosci. 2004;27:1–28. doi: 10.1146/annurev.neuro.27.070203.144157. [DOI] [PubMed] [Google Scholar]
- 128.Hsu M, Bhatt M, Adolphs R, Tranel D, Camerer CF. Neural systems responding to degrees of uncertainty in human decision-making. Science. 2005;310:1680–1683. doi: 10.1126/science.1115327. [DOI] [PubMed] [Google Scholar]
- 129.De Martino B, Kumaran O, Seymour B, Dolan RJ. Frames, Biases, and Rational Decision-Making in the Human Brain. Science. 2006;313:684–687. doi: 10.1126/science.1128356. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 130.Packard MG, Teather LA. Amygdala modulation of multiple memory systems: Hippocampus and caudate-putamen. Neurobiology of Learning and Memory. 1998;69:163–203. doi: 10.1006/nlme.1997.3815. [DOI] [PubMed] [Google Scholar]
- 131.Ballard IC, Murty VP, Carter RM, MacInnes JJ, Huettel SA, Adcock RA. Dorsolateral prefrontal cortex drives mesolimbic dopaminergic regions to initiate motivated behavior. J Neurosci. 2011;31:10340–10346. doi: 10.1523/JNEUROSCI.0895-11.2011. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 132.Doll BB, Jacobs WJ, Sanfey AG, Frank MJ. Instructional control of reinforcement learning: a behavioral and neurocomputational investigation. Brain Res. 2009;1299:74–94. doi: 10.1016/j.brainres.2009.07.007. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 133.Burke CJ, Tobler PN, Baddeley M, Schultz W. Neural mechanisms of observational learning. Proc Natl Acad Sci U S A. 2010;107:14431–14436. doi: 10.1073/pnas.1003111107. [DOI] [PMC free article] [PubMed] [Google Scholar]

