Skip to main content
Learning & Memory logoLink to Learning & Memory
. 2020 Oct;27(10):418–422. doi: 10.1101/lm.051714.120

Dorsolateral striatum engagement during reversal learning

Hadley C Bergstrom 1, Abby G Lieberman 1, Carolyn Graybeal 1, Anna M Lipkin 1, Andrew Holmes 1
PMCID: PMC7497112  PMID: 32934094

Abstract

Most experimental preparations demonstrate a role for dorsolateral striatum (DLS) in stimulus-response, but not outcome-based, learning. Here, we assessed DLS involvement in a touchscreen-based reversal task requiring mice to update choice following a change in stimulus-reward contingencies. In vivo single-unit recordings in the DLS showed reversal produced a population-level shift from excited to inhibited neuronal activity prior to choices being made. The larger the shift, the faster mice reversed. Furthermore, optogenetic photosilencing DLS neurons during choice increased early reversal errors. These findings suggest dynamic DLS engagement may facilitate reversal, possibly by signaling a change in contingencies to other striatal and cortical regions.


The dorsolateral striatum (DLS) is key neural locus for stimulus-response learning and habit formation, and DLS dysfunction is implicated in the pathophysiology of addictions (White 1996; Everitt and Robbins 2005; Yin and Knowlton 2006; Graybiel 2008; Balleine and O'Doherty 2010). In rodent instrumental learning paradigms, extensive training and certain schedules of reinforcement (random interval) favor the generation of DLS-dependent habitual behaviors (Adams 1982; Dickinson 1985) and these behaviors are expedited by alcohol and other drugs of abuse (Nelson and Killcross 2006; Nordquist et al. 2007; DePoy et al. 2013; Gremel and Lovinger 2017).

The involvement of DLS later in training is posited to reflect its role in constructing slow-to-acquire, stimulus-driven motor sequences (Balleine et al. 2009; Dezfouli and Balleine 2012). This contribution contrasts with that of the neighboring dorsomedial striatum (DMS), which is involved in cognitive flexibility and other processes that entail the updating of behavior when outcome values change (Packard and Knowlton 2002; Balleine et al. 2007; Ragozzino 2007; Hart et al. 2018). Hence, current models propose that outcome-sensitive learning, mediated by the DMS and prefrontal cortical regions, dominates performance early in training, and is supplanted by DLS-mediated stimulus-bound performance with further training (Dickinson and Balleine 1995).

Supporting this model, response-related neuronal activity in the caudate (homologue of DMS) emerges before activity in the putamen (homologue of DLS) of nonhuman primates learning a visual learning task (Williams and Eskandar 2006) and the development of habit in human subjects coincides with an increase in BOLD signal in the dorsolateral posterior putamen (Tricomi et al. 2009). Moreover, drug-seeking is disrupted by inactivation of the DLS, but not DMS, following prolonged, but not limited, training in rodents (Zapata et al. 2010; Corbit et al. 2012). However, DLS single-unit activity is evident from the beginning of motor, maze-based and instrumental training (Jog et al. 1999; Barnes et al. 2005; DeCoteau et al. 2007; Tang et al. 2007; Kimchi et al. 2009; Vandaele et al. 2019). Furthermore, lesions or optogenetic inactivation of the DLS in rodents can expedite early acquisition of spatial and stimulus discriminations when choice-outcome associations are still being formed (Bradfield and Balleine 2013; Bergstrom et al. 2018) and pharmacological inactivation of the anterior putamen impair visual reversal learning in marmosets (Jackson et al. 2019).

These prior findings suggest the DLS may exert a greater influence over outcome-based performance than currently appreciated. This led us, in the current study, to test for a potential role for the DLS in outcome-based reversal learning in which mice are required to update choice based on a new stimulus-reward contingency. To that end, we performed in vivo recordings and optogenetic manipulations of the DLS in a touchscreen-based reversal paradigm, a task previously shown to recruit dorsal and ventral striatum (Brigman et al. 2013; Bergstrom et al. 2018; Piantadosi et al. 2019; Radke et al. 2019).

We trained male C57BL/6J mice to learn to accurately discriminate (>85% correct choice) between two visual stimuli presented on a touch-sensitive screen to obtain a food pellet, and then reversed the stimulus-reward contingency and retrained mice to criterion (Fig. 1A,B). In parallel, single-unit DLS activity was recorded via chronically implanted microelectrode arrays (Fitzgerald et al. 2014; Gunduz-Cinar et al. 2019; Halladay et al. 2020) during three test sessions: (1) the final (late) discrimination (LD), (2) the first (early) reversal (ER), and (3) the final (late) reversal (LR) session.

Figure 1.

Figure 1.

DLS neuronal recordings in a pairwise discrimination and reversal touchscreen task. (A) Mice had microarray electrode arrays unliterally implanted in the DLS (dots depict estimated location of arrays) and were trained to discriminate between two visual stimuli (“fan” and “marble”) presented on a touch-sensitive screen to obtain a food pellet. The stimulus-reward contingency was then reversed, and mice retrained to criterion. DLS neuronal activity was recorded on three sessions: late discrimination (LD), early reversal (ER), and late reversal (LR). (B) Discrimination and reversal criteria (>85% correct choice) were attained in ∼12–14 test sessions. (C) Percent correct choice was high at LD and LR, and low at ER (repeated ANOVA stage effect: F(2,24) = 129.56, P < 0.01, followed by Newman-Keuls post hoc tests). Errors were low at LD and LR, and high ER, whether measured as total errors (F(2,24) = 26.41, P < 0.01) (D), perseverative errors (an error following an error) (F(2,24) = 26.33, P < 0.01) (E) or nonperseverative errors (an error following a correct) (F(2,24) = 15.43, P < 0.01) (F). For corresponding choice and reward-collection latencies, see Supplemental Figure S1. n = 13 mice. Data are means ± SEM. (*) P < 0.05.

As expected from the results of previous studies in our laboratory, percent correct choice was high at LD and LR, but low at ER, while total errors showed the inverse pattern (Fig. 1C,D). Perseverative errors (an error following an error) were highest at ER; nonperseverative errors (an error following a correct choice) were also elevated (Fig. 1E,F). Choice and reward-collection and latencies were lowest at ER (Supplemental Fig. S1). Overall, these behavioral profiles resemble those previously reported in our laboratory (Izquierdo et al. 2006; Graybeal et al. 2014).

Next, we aligned the in vivo recording data obtained from 203 (n = 64–72 per session) DLS units (average firing rate = 5.95 ± 0.44 Hz) to the time a correct choice or error was made during each recording session (without attempting to segregate neurons into medium spiny neurons versus fast-spiking interneurons). We classified units into those exhibiting phasically increased (“excited”) or decreased (“inhibited”) activity in the 2-sec epoch either prior to or immediately after a choice was made (Fig. 2A). Then, to test whether the recruitment of these units changed in association with performance, we compared the proportion of prechoice and postchoice excited and inhibited activity across the three recording sessions.

Figure 2.

Figure 2.

DLS neuronal activity changes on reversal. (A) Raster and perievent histogram examples of DLS neurons exhibiting increased (“excited,” red shading) or decreased (“inhibited,” blue shading) activity either prior to or immediately after a choice was made. (B) On the LD recording session, the majority of (pre or post) choice related DLS neurons were excited. This pattern was maintained for postchoice neurons on the ER and LR recording sessions (χ2 comparison ER versus LD: P > 0.05; LR versus LD: P > 0.05). Conversely, there was a significant shift increase in the proportion of inhibited prechoice neurons at ER, which was maintained at LR (ER versus LD: χ2 = 21.95, P < 0.01; LR versus ER: χ2 = 10.32, P < 0.01), and a corresponding decrease in excited neurons (ER versus LD: χ2 = 5.02, P < 0.05; LR versus ER: χ2 = 9.47, P < 0.01). (C) The proportion of DLS neurons exhibiting prechoice inhibited activity at ER predicted fewer sessions to reach reversal criterion. (D) Segregation of slow and fast learners based on a median split of sessions to reversal criterion. Slow learners made more errors during reversal (t(9) = 2.38, P < 0.05) (E) but showed similar latencies to choose (t(9) = 0.69, P > 0.05) and collect reward (t(9) = 1.13, P > 0.05) (F), as compared to fast learners. (G) Slow learners had a significantly small proportion of prechoice inhibited DLS neurons than fast leaners at ER (t(9) = 3.49, P < 0.01), but not other stages. Data are means ± SEM. Data are means ± SEM. (*) P < 0.05.

Most choice-related neurons were excited during LD, irrespective of whether their activity was evident pre- or postchoice (Fig. 2B). Strikingly, however, at ER there was a marked shift from excited to inhibited activity, and specifically so during the prechoice (not postchoice) period. Notably, this preponderance of inhibited activity was largely maintained, though somewhat attenuated, at LR even though behavioral performance at this stage was now comparable to that at LD, and markedly different from that at ER. Thus, the shift to prechoice inhibited activity at ER is not simply due to the preponderance of erroneous responding but, rather, reflects the new task contingencies. In fact, the maintenance of inhibited activity at LR suggests that DLS units signal a “record” of the change in contingencies throughout reversal even as performance resolves to high-choice accuracy.

The finding that DLS unit activity is sensitive to a change in stimulus-reward contingencies with reversal suggests the DLS may play a dynamic role in modifying behavior in response to the contingency reversal. To explore this possibility in more detail, we asked whether DLS activity at ER correlated with measures of reversal performance. This revealed a significant inverse relationship between the proportion of DLS units exhibiting prechoice inhibited activity at ER and the number of sessions to reach reversal criterion (Pearson correlation coefficient r = −0.61, Fisher's test: P < 0.05); i.e., the more DLS units that were phasically inhibited prior to choice at early reversal, the faster mice eventually relearned the new task contingencies (Fig. 2C). Importantly, further analysis showed this relationship was specific, in that neither prechoice excited activity (r = 0.06, P > 0.05) nor postchoice inhibited (r = 0.30, P > 0.05) or excited (r = −0.41, P > 0.05) activity correlated with sessions to reverse (Supplemental Fig. S2). The degree of inhibited activity at ER was also unrelated to the number of discrimination sessions to criterion (r = −0.34, P > 0.05), suggesting the correlation with sessions to reverse does not reflect a relationship with learning more generally. In this context, we have previously shown that measures of discrimination and reversal learning load on separate principal components (Izquierdo et al. 2006; Graybeal et al. 2014).

To further investigate the relationship between DLS neuronal activity at ER and reversal performance, we split mice into slow (n = 6) and fast (n = 5) learners based on a median split (=13 sessions) of the average number of sessions to attain reversal criterion (Fig. 2D). Slow learners made significantly more errors throughout reversal than fast learners, but did not differ in latency to make a choice or collect the reward, consistent with a cognitive but not motivational deficit in this group (Fig. 2E,F). Of note, the two groups were similar in their rates of discrimination learning (overall = 12.0 ± 2.1 SEM, slow = 10.0 ± 1.3, fast = 13.5 ± 6.6 sessions to criterion, t-test: P < 0.05), again indicating the differences in reversal were specific to that problem. We then compared DLS unit activity at ER between the two groups. There was a significantly smaller proportion of prechoice inhibited DLS units in the slow, relative to the fast, learner group (t(9) = 3.49, P < 0.01) (Fig. 2G). These data are again consistent with a positive relationship between inhibited activity in DLS neurons at early reversal and the efficiency of subsequent reversal learning.

The results of our recording data led us to perform a causal test for the contribution of the DLS to reversal performance. To do so, we used in vivo optogenetics to silence DLS neurons as mice performed the task (Fig. 3A). Stereotaxic surgery was used to bilaterally transfect DLS neurons with adeno-associated virus containing the inhibitory opsin, archaerhodopsin (ArchT, rAAV8/CAG-ArchT-GFP) or a control construct (rAAV8/CAG-GFP), and chronically implant optic-fibers directed at the DLS (Bergstrom et al. 2018; Sengupta and Holmes 2019). Green light (561 nm, 7 mW) was shone on the DLS during each trial throughout reversal, beginning at the initiation of the trial through to reward collection (correct trials) or 3-sec postchoice (error trials) (light on duration averaged 7–10 sec). In accordance with our previous study of DLS-photosilencing effects on discrimination (Bergstrom et al. 2018), performance was analyzed by binning sessions into mutually exclusive early, mid and late phases. This was done by equally subdividing the number of sessions to criterion for each mouse into the first, second, and third groups of sessions (e.g., nine sessions to criterion = three early, three mid, three late sessions). For session totals indivisible by three, the additional session(s) accrued to the late, then mid phase. The cumulative values in each phase were calculated and either expressed as a total value (errors) or average by phase (%correct and latencies).

Figure 3.

Figure 3.

Optogenetic silencing of DLS disrupts early reversal. (A) DLS photosilencing during touchscreen reversal learning. (B) DLS-silencing did not affect the number of sessions to attain reversal criterion (unpaired Student's t-test: t(13) = 0.94, P > 0.05) or (C) correct responding at any stage of reversal (ANOVA group-effect: F(1,13) = 1.61, P > 0.05; stage-effect: F(2,26) = 177.15, P < 0.01; interaction: F(2,26) = 0.42, P > 0.05). (D) Silencing did, however, increase the number of total errors made, specifically at ER (ANOVA group-effect: F(1,13) = 0.76, P > 0.05; stage-effect: F(2,26) = 134.83, P < 0.01; interaction: F(2,26) = 8.71, P < 0.01). (E) This effect was largely driven more perseverative errors (ANOVA group-effect: F(1,13) = 1.03, P > 0.05; stage-effect: F(2,26) = 203.78, P < 0.01; interaction: F(2,26) = 9.93, P < 0.01), (F) though nonperseverative errors were also visibly higher (ANOVA group-effect: F(1,13) = 0.32, P > 0.05; stage-effect: F(2,26) = 14.24, P < 0.01; interaction: F(2,26) = 3.39, P < 0.05). n = 7–8 per group. Data are means ± SEM. (*) P < 0.05.

We found that DLS-silencing did not affect the number of sessions to attain reversal criterion or percent correct responding during any stage of reversal (Fig. 3B,C) (choice and reward-collection latencies were also unaffected, Supplemental Fig. S3A,B). Photosilencing did, however, significantly increase the total number of errors (P < 0.05), specifically (at least statistically) the number of perseverative errors (P < 0.05), at ER but not the later reversal stages (Fig. 3D–F). Given the specificity of the effect to the early testing stage, this silencing-induced increase in errors cannot be due to a nonspecific increase in touchscreen responding. Instead, these data are consistent with an impairment in early reversal performance as a result of DLS silencing. This deficit likely failed to manifest as a decrease in overall percent correct responding because this measure was already near floor levels in GFP controls. As such, these data are consistent with a recent study demonstrating that pharmacological inhibition of the anterior putamen causes visual reversal impairments in nonhuman primates (Jackson et al. 2019), and contrast with the ability of DLS photosilencing to facilitate early discrimination (Bergstrom et al. 2018).

How do the results of these optogenetic manipulations align with our recording data? One possibility is that the shift to inhibited activity in DLS units prior to early reversal choices reflects the suppression of the old contingency to enable formation of the new. If this were the case, however, DLS silencing would be predicted to result in a facilitation of reversal performance (e.g., by removing competition between the old and new memories, or liberating other structures, such as the DMS, to support reversal learning unhindered), rather than the deficit we observed. An alternative interpretation is that the change in DLS unit activity signals that the contingencies have changed, and this signal is utilized to update behavior to reflect the new contingency. This would explain the positive relationship between size of the inhibited DLS population at early reversal and the subsequent rate of reversal, as well as the impairment in early reversal caused by DLS silencing. However, given the effect of silencing was limited to the early reversal stage, this putative contribution of the DLS is not necessary for the contingencies to be reversed in full. Instead, it may serve to catalyze the ability of other regions to fulfill this function, such as the DMS and PFC. In turn, this predicts that augmenting DLS activity, e.g., via optogenetic photostimulation, could expedite early reversal learning.

In sum, the current data provide preliminary support for a contribution of the DLS to reversal learning in mice, such that a dynamic increase in the prechoice inhibition of DLS units during a change in stimulus-reward contingencies may facilitate the ability to modify behavior accordingly. These findings add to a growing body of evidence supporting a contribution of the DLS to forms of learning that utilize outcome information to guide behavior (Dezfouli et al. 2014; Bergstrom et al. 2018; Vandaele et al. 2019).

Supplementary Material

Supplemental Material

Acknowledgments

Research supported by the NIAAA Intramural Research Program.

Footnotes

[Supplemental material is available for this article.]

References

  1. Adams CD. 1982. Variations in the sensitivity of instrumental responding to reinforcer devaluation. Q J Exp Psychol 34: 77–98. 10.1080/14640748208400878 [DOI] [Google Scholar]
  2. Balleine BW, O'Doherty JP. 2010. Human and rodent homologies in action control: corticostriatal determinants of goal-directed and habitual action. Neuropsychopharmacology 35: 48–69. 10.1038/npp.2009.131 [DOI] [PMC free article] [PubMed] [Google Scholar]
  3. Balleine BW, Delgado MR, Hikosaka O. 2007. The role of the dorsal striatum in reward and decision-making. J Neurosci 27: 8161–8165. 10.1523/JNEUROSCI.1554-07.2007 [DOI] [PMC free article] [PubMed] [Google Scholar]
  4. Balleine BW, Liljeholm M, Ostlund SB. 2009. The integrative function of the basal ganglia in instrumental conditioning. Behav Brain Res 199: 43–52. 10.1016/j.bbr.2008.10.034 [DOI] [PubMed] [Google Scholar]
  5. Barnes TD, Kubota Y, Hu D, Jin DZ, Graybiel AM. 2005. Activity of striatal neurons reflects dynamic encoding and recoding of procedural memories. Nature 437: 1158–1161. 10.1038/nature04053 [DOI] [PubMed] [Google Scholar]
  6. Bergstrom HC, Lipkin AM, Lieberman AG, Pinard CR, Gunduz-Cinar O, Brockway ET, Taylor WW, Nonaka M, Bukalo O, Wills TA, et al. 2018. Dorsolateral striatum engagement interferes with early discrimination learning. Cell Rep 23: 2264–2272. 10.1016/j.celrep.2018.04.081 [DOI] [PMC free article] [PubMed] [Google Scholar]
  7. Bradfield LA, Balleine BW. 2013. Hierarchical and binary associations compete for behavioral control during instrumental biconditional discrimination. J Exp Psychol Anim Behav Process 39: 2–13. 10.1037/a0030941 [DOI] [PMC free article] [PubMed] [Google Scholar]
  8. Brigman JL, Daut RA, Wright T, Gunduz-Cinar O, Graybeal C, Davis MI, Jiang Z, Saksida LM, Jinde S, Pease M, et al. 2013. GluN2B in corticostriatal circuits governs choice learning and choice shifting. Nat Neurosci 16: 1101–1110. 10.1038/nn.3457 [DOI] [PMC free article] [PubMed] [Google Scholar]
  9. Corbit LH, Nie H, Janak PH. 2012. Habitual alcohol seeking: time course and the contribution of subregions of the dorsal striatum. Biol Psychiatry 72: 389–395. 10.1016/j.biopsych.2012.02.024 [DOI] [PMC free article] [PubMed] [Google Scholar]
  10. DeCoteau WE, Thorn C, Gibson DJ, Courtemanche R, Mitra P, Kubota Y, Graybiel AM. 2007. Oscillations of local field potentials in the rat dorsal striatum during spontaneous and instructed behaviors. J Neurophysiol 97: 3800–3805. 10.1152/jn.00108.2007 [DOI] [PubMed] [Google Scholar]
  11. DePoy L, Daut R, Brigman JL, MacPherson K, Crowley N, Gunduz-Cinar O, Pickens CL, Cinar R, Saksida LM, Kunos G, et al. 2013. Chronic alcohol produces neuroadaptations to prime dorsal striatal learning. Proc Natl Acad Sci 110: 14783–14788. 10.1073/pnas.1308198110 [DOI] [PMC free article] [PubMed] [Google Scholar]
  12. Dezfouli A, Balleine BW. 2012. Habits, action sequences and reinforcement learning. Eur J Neurosci 35: 1036–1051. 10.1111/j.1460-9568.2012.08050.x [DOI] [PMC free article] [PubMed] [Google Scholar]
  13. Dezfouli A, Lingawi NW, Balleine BW. 2014. Habits as action sequences: hierarchical action control and changes in outcome value. Philos Trans R Soc Lond B Biol Sci 369: 20130482 10.1098/rstb.2013.0482 [DOI] [PMC free article] [PubMed] [Google Scholar]
  14. Dickinson A. 1985. Actions and habits: the development of behavioural autonomy. Philos Trans R Soc Lond B 308: 67–78. 10.1098/rstb.1985.0010 [DOI] [Google Scholar]
  15. Dickinson A, Balleine BW. 1995. Overtraining and the motivational control of instrumental action. Anim Learn Behav 22: 197–206. 10.3758/BF03199935 [DOI] [Google Scholar]
  16. Everitt BJ, Robbins TW. 2005. Neural systems of reinforcement for drug addiction: from actions to habits to compulsion. Nat Neurosci 8: 1481–1489. 10.1038/nn1579 [DOI] [PubMed] [Google Scholar]
  17. Fitzgerald PJ, Whittle N, Flynn SM, Graybeal C, Pinard CR, Gunduz-Cinar O, Kravitz AV, Singewald N, Holmes A. 2014. Prefrontal single-unit firing associated with deficient extinction in mice. Neurobiol Learn Mem 113: 69–81. 10.1016/j.nlm.2013.11.002 [DOI] [PMC free article] [PubMed] [Google Scholar]
  18. Graybeal C, Bachu M, Mozhui K, Saksida LM, Bussey TJ, Sagalyn E, Williams RW, Holmes A. 2014. Strains and stressors: an analysis of touchscreen learning in genetically diverse mouse strains. PLoS One 9: e87745 10.1371/journal.pone.0087745 [DOI] [PMC free article] [PubMed] [Google Scholar]
  19. Graybiel AM. 2008. Habits, rituals, and the evaluative brain. Annu Rev Neurosci 31: 359–387. 10.1146/annurev.neuro.29.051605.112851 [DOI] [PubMed] [Google Scholar]
  20. Gremel CM, Lovinger DM. 2017. Associative and sensorimotor cortico-basal ganglia circuit roles in effects of abused drugs. Genes Brain Behav 16: 71–85. 10.1111/gbb.12309 [DOI] [PMC free article] [PubMed] [Google Scholar]
  21. Gunduz-Cinar O, Brockway E, Lederle L, Wilcox T, Halladay LR, Ding Y, Oh H, Busch EF, Kaugars K, Flynn S, et al. 2019. Identification of a novel gene regulating amygdala-mediated fear extinction. Mol psychiatry 24: 601–612. 10.1038/s41380-017-0003-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
  22. Halladay LR, Kocharian A, Piantadosi PT, Authement ME, Lieberman AG, Spitz NA, Coden K, Glover LR, Costa VD, Alvarez VA, et al. 2020. Prefrontal regulation of punished ethanol self-administration. Biol Psychiatry 87: 967–978. 10.1016/j.biopsych.2019.10.030 [DOI] [PMC free article] [PubMed] [Google Scholar]
  23. Hart G, Bradfield LA, Fok SY, Chieng B, Balleine BW. 2018. The bilateral prefronto-striatal pathway is necessary for learning new goal-directed actions. Curr Biol 28: 2218–2229 e2217. 10.1016/j.cub.2018.05.028 [DOI] [PubMed] [Google Scholar]
  24. Izquierdo A, Wiedholz LM, Millstein RA, Yang RJ, Bussey TJ, Saksida LM, Holmes A. 2006. Genetic and dopaminergic modulation of reversal learning in a touchscreen-based operant procedure for mice. Behav Brain Res 171: 181–188. 10.1016/j.bbr.2006.03.029 [DOI] [PubMed] [Google Scholar]
  25. Jackson SAW, Horst NK, Axelsson SFA, Horiguchi N, Cockcroft GJ, Robbins TW, Roberts AC. 2019. Selective role of the putamen in serial reversal learning in the marmoset. Cereb cortex 29: 447–460. 10.1093/cercor/bhy276 [DOI] [PMC free article] [PubMed] [Google Scholar]
  26. Jog MS, Kubota Y, Connolly CI, Hillegaart V, Graybiel AM. 1999. Building neural representations of habits. Science 286: 1745–1749. 10.1126/science.286.5445.1745 [DOI] [PubMed] [Google Scholar]
  27. Kimchi EY, Torregrossa MM, Taylor JR, Laubach M. 2009. Neuronal correlates of instrumental learning in the dorsal striatum. J Neurophysiol 102: 475–489. 10.1152/jn.00262.2009 [DOI] [PMC free article] [PubMed] [Google Scholar]
  28. Nelson A, Killcross S. 2006. Amphetamine exposure enhances habit formation. J Neurosci 26: 3805–3812. 10.1523/JNEUROSCI.4305-05.2006 [DOI] [PMC free article] [PubMed] [Google Scholar]
  29. Nordquist RE, Voorn P, de Mooij-van Malsen JG, Joosten RN, Pennartz CM, Vanderschuren LJ. 2007. Augmented reinforcer value and accelerated habit formation after repeated amphetamine treatment. Eur Neuropsychopharmacol 17: 532–540. 10.1016/j.euroneuro.2006.12.005 [DOI] [PubMed] [Google Scholar]
  30. Packard MG, Knowlton BJ. 2002. Learning and memory functions of the basal ganglia. Annu Rev Neurosci 25: 563–593. 10.1146/annurev.neuro.25.112701.142937 [DOI] [PubMed] [Google Scholar]
  31. Piantadosi PT, Lieberman AG, Pickens CL, Bergstrom HC, Holmes A. 2019. A novel multichoice touchscreen paradigm for assessing cognitive flexibility in mice. Learn Mem 26: 24–30. 10.1101/lm.048264.118 [DOI] [PMC free article] [PubMed] [Google Scholar]
  32. Radke AK, Kocharian A, Covey DP, Lovinger DM, Cheer JF, Mateo Y, Holmes A. 2019. Contributions of nucleus accumbens dopamine to cognitive flexibility. Eur J Neurosci 50: 2023–2035. 10.1111/ejn.14152 [DOI] [PMC free article] [PubMed] [Google Scholar]
  33. Ragozzino ME. 2007. The contribution of the medial prefrontal cortex, orbitofrontal cortex, and dorsomedial striatum to behavioral flexibility. Ann N Y Acad Sci 1121: 355–375. 10.1196/annals.1401.013 [DOI] [PubMed] [Google Scholar]
  34. Sengupta A, Holmes A. 2019. A discrete dorsal raphe to basal amygdala 5-HT circuit calibrates aversive memory. Neuron 103: 489–505 e487. 10.1016/j.neuron.2019.05.029 [DOI] [PMC free article] [PubMed] [Google Scholar]
  35. Tang C, Pawlak AP, Prokopenko V, West MO. 2007. Changes in activity of the striatum during formation of a motor habit. Eur J Neurosci 25: 1212–1227. 10.1111/j.1460-9568.2007.05353.x [DOI] [PubMed] [Google Scholar]
  36. Tricomi E, Balleine BW, O'Doherty JP. 2009. A specific role for posterior dorsolateral striatum in human habit learning. Eur J Neurosci 29: 2225–2232. 10.1111/j.1460-9568.2009.06796.x [DOI] [PMC free article] [PubMed] [Google Scholar]
  37. Vandaele Y, Mahajan NR, Ottenheimer DJ, Richard JM, Mysore SP, Janak PH. 2019. Distinct recruitment of dorsomedial and dorsolateral striatum erodes with extended training. Elife 8: e49536 10.7554/eLife.49536 [DOI] [PMC free article] [PubMed] [Google Scholar]
  38. White NM. 1996. Addictive drugs as reinforcers: multiple partial actions on memory systems. Addiction 91: 921–949. 10.1111/j.1360-0443.1996.tb03586.x [DOI] [PubMed] [Google Scholar]
  39. Williams ZM, Eskandar EN. 2006. Selective enhancement of associative learning by microstimulation of the anterior caudate. Nat Neurosci 9: 562–568. 10.1038/nn1662 [DOI] [PubMed] [Google Scholar]
  40. Yin HH, Knowlton BJ. 2006. The role of the basal ganglia in habit formation. Nat Rev Neurosci 7: 464–476. 10.1038/nrn1919 [DOI] [PubMed] [Google Scholar]
  41. Zapata A, Minney VL, Shippenberg TS. 2010. Shift from goal-directed to habitual cocaine seeking after prolonged experience in rats. J Neurosci 30: 15457–15463. 10.1523/JNEUROSCI.4072-10.2010 [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Supplemental Material

Articles from Learning & Memory are provided here courtesy of Cold Spring Harbor Laboratory Press

RESOURCES