Skip to main content
Frontiers in Neuroscience logoLink to Frontiers in Neuroscience
. 2018 Mar 19;12:165. doi: 10.3389/fnins.2018.00165

Contributions of Lateral and Orbital Frontal Regions to Abstract Rule Acquisition and Reversal in Monkeys

Giancarlo La Camera 1,2,3,*, Sebastien Bouret 3,4, Barry J Richmond 3,*
PMCID: PMC5867347  PMID: 29615854

Abstract

The ability to learn and follow abstract rules relies on intact prefrontal regions including the lateral prefrontal cortex (LPFC) and the orbitofrontal cortex (OFC). Here, we investigate the specific roles of these brain regions in learning rules that depend critically on the formation of abstract concepts as opposed to simpler input-output associations. To this aim, we tested monkeys with bilateral removals of either LPFC or OFC on a rapidly learned task requiring the formation of the abstract concept of same vs. different. While monkeys with OFC removals were significantly slower than controls at both acquiring and reversing the concept-based rule, monkeys with LPFC removals were not impaired in acquiring the task, but were significantly slower at rule reversal. Neither group was impaired in the acquisition or reversal of a delayed visual cue-outcome association task without a concept-based rule. These results suggest that OFC is essential for the implementation of a concept-based rule, whereas LPFC seems essential for its modification once established.

Keywords: prefrontal cortex, abstract concepts, learning, behavior, animal, decision making, primates, memory

Introduction

Primate prefrontal cortex (PFC) supports executive, mnemonic and attentional functions critical for learning and invoking rule-based strategies to control behavior (Miller and Cohen, 2001; Curtis and D'Esposito, 2004; Bunge et al., 2005; Tanji et al., 2007; Fuster, 2008; Tanji and Hoshi, 2008; Wise, 2008; Buckley et al., 2009). Different prefrontal regions seem to support different aspects of rule-based behavior (Lee et al., 2007). Lateral prefrontal cortex (LPFC) has been implicated in working memory (Fuster and Alexander, 1971; Funahashi et al., 1989, 1993), attention (Lebedev et al., 2004), executive control (Huettel et al., 2004; Tanji and Hoshi, 2008), self-organized behavior (Procyk and Goldman-Rakic, 2006), rule-based behavior (Rushworth et al., 1997; Wallis et al., 2001; Bunge et al., 2003; Shima et al., 2007; Tanji et al., 2007; Moore et al., 2009, 2012), and context-dependent decisions (Wise et al., 1996; Mante et al., 2013; Rigotti et al., 2013), whereas orbitofrontal cortex (OFC) has been implicated in reversal learning (Dias et al., 1996, 1997; Izquierdo et al., 2004; Walton et al., 2010), reinforcement learning (Rolls et al., 1996; Rolls, 2000; Hampton et al., 2006; Salzman et al., 2007; Simmons and Richmond, 2008; McDannald et al., 2011), reward evaluation and comparison (Tremblay and Schultz, 1999; Schultz et al., 2000; Wallis and Miller, 2003; Salzman et al., 2007; Bouret and Richmond, 2010; Simmons et al., 2010), and the evaluation of alternative options (Padoa-Schioppa and Assad, 2006, 2008).

In monkeys, OFC seems to have a role in decisions based on expected outcome value beyond simple stimulus-response associations (Walton et al., 2010; Clark et al., 2013), whereas LPFC seems essential in updating a rule-based strategy to optimize a rewarding outcome (Dias et al., 1996; Buckley et al., 2009; Moore et al., 2012). An important determinant of reward-based learning is the nature of the predictive cues used for learning, which may be simple stimuli acting in isolation, compound stimuli, or abstract concepts. Presumably, a concept-based task such as a delayed match to sample (DMS) task (“if the 2 stimuli match, then reward”) results in different cognitive demands than a simpler cue-outcome association task (“if stimulus A, then reward”). Since LPFC is implicated in rule learning and OFC is implicated in assessing outcome value, we tested the effects of LPFC and OFC removals in learning a behavior that requires the formation of abstract concepts compared to a behavior that requires simple visual stimulus-outcome associations. Monkeys with bilateral LPFC and OFC lesions were tested in a novel version of DMS, requiring the formation of the concepts of “same” and “different,” and in two simpler rule-based tasks that required no concept formation. This comparison should expose significant differences in the roles of LPFC and OFC in behavior that depends on concept-based as opposed to sensory-cue based predictions of forthcoming contingencies.

To be able to compare the learning times across all tasks we developed a variation of DMS (dubbed “indirect DMS,” or iDMS) that was learned rapidly by the monkeys. Control monkeys quickly learned to use the association between the abstract concepts of “same” and “different” with their predicted outcome. Monkeys with bilateral OFC lesions were impaired at both acquiring and reversing the associations between the concept and the outcome, whereas monkeys with bilateral LPFC lesions acquired the task as quickly as the control group, but were impaired at reversing the association between concept and outcome. Both lesion groups learned the simpler cue-outcome associations (with and without a memory component) as quickly as controls, showing that the impairments were related to forming the abstract concept and/or applying the abstract concept to infer the rule of the task. These results suggest that OFC is essential for acquiring and updating an association between a concept and a reward, whereas LPFC seems essential for its modification once established.

Methods

Subjects and surgical procedures

Nine rhesus monkeys were used for this study, 3 unoperated controls, 3 monkeys with bilateral LPFC lesions (Figure 1A) and 3 monkeys with bilateral OFC lesions (Figure 1C; see below for details). All the experimental procedures were carried out in accordance with the ILAR Guide for Care and Use of Laboratory Animals and approved by the Animal Care and Use Committee of the National Institute of Mental Health. Monkeys received bilateral lesions of orbital or lateral PFC using a combination of suction and electrocautery. The intended lateral prefrontal lesion (Figure 1B) extended laterally from the dorsal midline to the orbital surface of the inferior convexity.

Figure 1.

Figure 1

Lesions. (A) Lateral view (top) and coronal sections (bottom) of a standard rhesus monkey brain showing the extent of the LPFC lesion (shaded regions) in all three monkeys. Areas corresponding to cytoarchitectonic regions 9, 12, and 45 are indicated on coronal sections. (B) Lateral view of the intended LPFC lesion. (C) Ventral view (top) and coronal sections (bottom) of a standard rhesus monkey brain showing the extent of the OFC lesion (shaded regions) in all three monkeys. Areas corresponding to cytoarchitectonic regions 11, 12, and 13 are indicated on coronal sections. (D) Ventral view of the intended OFC lesion. The numerals next to each coronal section in panels (A,C) indicate the distance in millimeters from the interaural plane. ps, principal sulcus; L, left hemisphere; R, right hemisphere.

The rostral limit of the lesion was the frontal pole. The caudal limit was the caudal extent of the principal sulcus. The frontal eye fields and the banks of the arcuate sulci were intentionally spared. In total, the intended lateral prefrontal lesion included areas 9, 46, 45, 12, and dorsal area 10 (Walker, 1940; Petrides and Pandya, 1994). The intended orbital prefrontal lesion (Figure 1C) extended from the fundus of the lateral orbital sulcus to the fundus of the rostral sulcus. The rostral limit of the lesion was a line joining the anterior tips of the lateral and medial orbital sulci. The caudal limit was ~ 5 mm rostral to the junction of the frontal and temporal lobes. In total, the intended orbital prefrontal lesion included areas 11, 13, 14 and the caudal part of ventral area 10 (Walker, 1940; Petrides and Pandya, 1994). The lateral and orbital prefrontal lesions shared a common boundary at the lateral orbital sulcus.

Lesion location and extent were largely as intended within each experimental group (Figures 1A–C). In the LPFC group, all three lesions extended from 23 to 43 mm rostral to the interaural line. All LPFC lesions included regions 9, 46, 45, and 10, as intended. However, most of 12o (on the orbital surface) and the caudal part of 12l (on the ventrolateral surface) were spared in all three animals. The banks of the arcuate sulcus were spared, as intended. There were no areas of unintended damage. In the OFC group, two of the three lesions (JE and SU) extended from 25 to 42 mm rostral to the interaural line; the other (BR) was placed slightly more caudally, extending from 23 to 41 mm rostral to the interaural line. Except for a narrow strip of area 14 immediately ventral to the rostral sulcus and a small part of area 13l immediately medial to the lateral orbital sulcus, OFC lesions in all three monkeys included all intended regions. There were no areas of unintended damage. More details are reported in Simmons et al. (2010).

Prior to their use in this study, the LPFC monkeys were trained in the bar release task (Figure 2A) and tested in a Wisconsin-analog task (Lerchner et al., 2007), in reward postponement tasks (Minamimoto et al., 2009; Simmons et al., 2010), and the categorization task of Minamimoto et al. (2010). The OFC monkeys were tested in configural discrimination tasks administered on touch screen similar to those used in Bussey et al. (2002, 2003), reinforcement devaluation tasks (Chudasama et al., 2007), and reward postponement tasks (Simmons et al., 2010). The time elapsed between the ablations and testing for this study depended on the monkey and ranged between 6 and 24 months. The control monkeys were naive prior to exposure to the tasks described in this study. None of the monkeys had had prior exposure to our iDMS task, our control tasks, or more traditional variants of DMS prior to their use in this study.

Figure 2.

Figure 2

Behavioral tasks. (A) Bar release task. The monkeys were required to release a bar when a red dot, located centrally on a computer screen, turned green. Failure to do so resulted in an incorrect trial that was immediately aborted. Correct trials were rewarded with a liquid reward; incorrect trials were aborted and no reward was given. Transition times between events are reported to the left (see section Methods for details). (B) iDMS task. Two visual cues, a “sample” and a “test” image, were presented in sequence and separated by a temporal delay of 1 s. The sample image was turned off after 500 ms from its onset, starting the delay period during which only the red dot was visible on the screen. At the end of the delay period, the test image appeared on the screen behind the red dot; between 500 and 1,500 ms later, the red dot turned green. A bar release at this point (between 200 and 1,000 ms after the onset of the green dot), led to either reward or a time-out depending on the rule in effect. A bar release outside the allowed interval caused the immediate abortion of the trial (see section Methods for details). Inset at the bottom shows a sample of the cue set. Sets of 50 images were used in each session. (C) Simple association task where one of two visual cues was associated to reward while the other predicted a time-out. Regardless of the outcome of the previous trial, each cue had an equal chance of being selected in the following trial. Transition times between events are reported to the left and the two visual stimuli used as cues are shown below the main plot. All other task features were as in the iDMS task of (B) (see section Methods for details). (D) Delayed association task. In this task, the visual cue disappeared after 250–500 ms from the onset of the red dot, which turned green after an additional delay period of 1 s. Everything else was as in the non-delayed association task of (C). Two different visual cues (shown below the main plot) were used.

Behavioral training

The monkeys were initially trained on a simple bar release task (Figure 2A). They sat in a monkey chair facing a computer screen displaying a white noise background and had access to a bar located on the primate chair. When the monkey touched the bar, a visual stimulus (cue) appeared in the center of the screen (13° on a side; black square in Figure 2A), followed 400 ms later by the appearance of a red dot (0.5° on a side), superimposed to the visual cue in the center of the cue. After a randomly selected time interval of 500, 750, 1,000, 1,250, or 1,500 ms, the red dot turned green. The monkey was required to release the bar 200–1,000 ms from the appearance of the green dot. This caused the green dot to turn blue, followed by the reward (one drop of juice) after 300 ms. A bar release outside of the 200–1,000 ms interval (or not occurring at all within 5 s of the onset of the green dot) caused the trial to be immediately aborted, with all stimuli disappearing from the screen until the next trial. No reward was given after an aborted trial. After each trial, whether correct or not, there was an inter-trial interval (ITI) of 1 s before the monkey could initiate a new trial.

Once the monkeys got proficient in this task (>75% correct), which occurred within 2 weeks for all the monkeys, they were tested on a variation of this task where in some specific correct trials the reward was given only after a delay (“reward postponement” task; Minamimoto et al., 2009). Specifically, three different visual cues (instead of one) were associated with three different reward postponements of 0.3 ± 0.07, 3.6 ± 0.4, and 7.2 ± 0.8 s, but the rewards were all equal in size (1 drop of juice). All monkeys (lesioned and controls) learned quickly the meaning of the visual cues (1–3 sessions, χ2-test for proportions, p < 0.01) and developed a linear dependence of error rates vs. predicted reward postponement, following the same pattern reported elsewhere (see Figure 3 of Minamimoto et al., 2009 and Figure 5 of Simmons et al., 2010). This test was conducted merely to establish sensitivity to reward postponement, which is instrumental to the main iDMS task described next.

iDMS task

In the iDMS task (Figure 2B), two visual images were presented sequentially in each trial, initially separated by a temporal delay of 1 s. The first image (“sample”) disappeared after 500 ms and only the red dot remained visible. Then the second image (“test”) appeared and remained on. After 500–1,500 ms the red dot turned green. In “match” trials, the test matched the sample; in “non-match” trials, the sample and the test cue were different. The stimuli used as sample and test images were chosen randomly from a set of 50 stimuli (see below and inset of Figure 2B for a sample). As before, the monkeys would have to release the bar on green to complete a trial. However, completing a trial would not always be the correct response. The outcome of a completed trial depended on the rule in effect: in the “reward-if-match” rule, a completed trial was met with either a reward or a “time-out,” depending on whether the two images were identical (“match” trial) or different (“non-match” trial).

In a completed trial ending with a reward, the green dot would turn blue and the test cue would disappear, followed by a reward after 300 ms, and then by the ITI. In a completed time-out trial, the green dot would turn blue and the test cue would disappear, followed by a pseudo-randomly chosen time period of 5.8–8.2 s preceding the ITI. No reward was given in a time-out trial. Because of the time-out period and lack of reward, completed time-out trials were considered erroneous and were scored as incorrect trials (see below for complete definitions).

The opposite would occur in the “time-out-if-match” rule (i.e., time-out would be the outcome of a completed match trial and reward of a completed non-match trial). As in the previous tasks, an improperly timed bar release would immediately abort the trial, causing all stimuli to be turned off immediately before entering the 1 s ITI. The next trial was always chosen to be a match or a non-match trial with 50% chance, regardless of the behavior in the current trial. The rationale for this task structure was to take advantage of the spontaneous tendency of the monkeys not to release on green when this predicts a time-out (Minamimoto et al., 2009). The monkeys were allowed to engage in the task until they stopped by themselves (they completed an average of 353 ± 103 trials in each session (mean ± SD) prior to the acquisition of the task).

In scoring the monkey's decisions in the iDMS task, we considered “correct” trials either completed trials that led to reward, or aborted trials that would result in a time-out; “incorrect” trials were either completed trials that led to a time-out, or aborted trials that would lead to a reward. Trials with pre-test bar releases (bar releases occurring before the appearance of the test image) were not informative of trial type and were not scored. Bar releases occurring erroneously after the appearance of the test image led to the abortion of the current trial and, in time-out trials, to the avoidance of the time-out. Monkeys learned to do this on purpose and for this reason we termed those trials “skipped” trials. The difference in skip rates between match and non-match trials is an alternative measure of performance, and was used as a criterion to reverse the task's rule during the experiments (see below). However, since this criterion turned out to be a less reliable measure of performance compared to the correct rates metric defined above, the criterion to task acquisition was defined in terms of percent correct.

The rule of the task was reversed when the difference in skip rates in match vs. non-match trials was significantly different in at least 4 out of 5 consecutive sessions (χ2-test for proportions, p < 0.01). In terms of percent correct, this criterion translated into 2–5 consecutive sessions with correct rates significantly above chance (χ2-test for proportions, p < 0.01; see filled circles in Figure 4A). Although less reliable than correct rates, 5 sessions with 4 significant skip rates indicate that the monkey has learned the task, and this criterion insured that all monkeys were asked to reverse the task after similar post-learning experience. Reversal of the task amounted to swapping the association of trial type with reward contingency. For example, match trials, if initially associated with reward, would become associated with time-out, and vice-versa. Rule changes only occurred at the beginning of a new session.

After 5 sessions with a significant separation of skip rates in the reversed task, the delay between sample and test was increased gradually from 1 to 21 s. Duration increments were 1 s per session; only one delay duration was used each day. A delay increment was introduced after each day of significant separation of correct rates with the current delay duration. In most cases, a significant separation of correct rates with new delay duration was reached within the same day the new delay duration was introduced. During testing with longer delays, on occasion (once every seven sessions on average) the delay was kept fixed and a new cue set, never shown before to the monkey, was used. After significant separation of correct rates (which occurred always on the first day a new cue set was introduced), the cue set was kept fixed and the delay incremented by 1 s, resuming the incremental delay schedule. Thus, once a new set had been introduced, it would be used for a subsequent number of daily sessions until a new set was introduced (or until testing stopped).

We used overall 5 sets of 50 stimuli as sample and test cues. Each visual cue was a 200 × 200 pixel resolution image. The 50 images in each set represented a large variety of subjects including landscapes, animals, vegetables, man-made objects (planes, cars, tools, etc.; see Figure 2B for a few examples). Different cue sets differed in both the object displayed within a category (e.g., different landscapes in different sets) and the category of objects (e.g., landscapes in one set and tables in another set). Many categories were used in each cue set. A match between sample and test required the stimuli to be identical (e.g., two different images containing similar, but not identical, tables were a non-match).

Control tasks

After successfully reaching the session with 21 s delay, all monkeys were tested in a number of control tasks, described below in the order in which they were executed.

Cue-outcome association task (Figure 2C)

A single visual cue was presented in each trial. There were only 2 cues in this cue set, one predicting reward and the other predicting a time-out. The cue remained on the screen throughout the trial. A bar release after green was required to complete the trial, which would result in reward or time-out depending on which of the two visual cues had been presented. The relative timings of events (except the presentation of the delay and the second image) were as in the iDMS task. The stimuli, shown in Figure 2C, had not been used before, but were of same type and resolution as the stimuli used in the iDMS task. After the correct rates for the two stimuli had been significantly different for 4 consecutive sessions (χ2-test, p < 0.05), the task was reversed, so that the visual cue previously predicting reward now predicted time-out, and vice-versa.

Delayed cue-outcome association task (Figure 2D)

The monkeys were then tested in a delayed version of the previous task. In this version, the visual cue disappeared between 250 and 500 ms after the onset of the red dot, and a 1-s delay followed before the appearance of the green dot. Everything else was as in the non-delayed cue-outcome association task. Two new stimuli were used (Figure 2D). After 4 consecutive days of different correct rates (χ2-test, p < 0.05), the cue-outcome rule was reversed.

Delayed matching-to-sample with short delay

Finally, the monkeys were tested in the iDMS task with a sample-test delay interval of 100 ms. A 100 ms delay was preferred to no delay at all to prevent the monkeys from approaching the task as a perceptual change detection task. The same cue set and abstract rule learned by the monkeys when last exposed to the iDMS task were used; percent correct difference was immediately above chance for all monkeys. The monkeys were tested for 4 sessions in this task, then the association between trial type and outcome was reversed. After 4 consecutive days of different correct rates in the two trial types (χ2-test, p < 0.05), the delay between sample and test cue was brought back to 1 s, and a new cue set was used. However, the rule in effect remained the same, and performance (difference in percent correct) was immediately above chance for all monkeys. After 4 sessions, the task was reversed again, and then again 3 more times, each time after 4 consecutive days of significant difference in percent correct (χ2-test, p < 0.05). During these additional reversals, the cue set was not changed.

Data analysis

All data were analyzed in the R statistical computing environment (R Development Core Team, 2008). Completed trials that led to reward, or aborted trials that would result in a time-out, were scored as “correct”; completed trials that led to a time-out, or aborted trials that would lead to a reward, were scored as “incorrect.” In the iDMS task of Figure 2B, trials with pre-test bar releases (bar releases occurring before the appearance of the test image) were not informative of trial type and were not scored (the monkey's behavior could not be the consequence of the predicted outcome).

Performance was quantified as overall percent correct (e.g., Miyashita, 1988; Miyashita and Chang, 1988; Miller et al., 1993; Amit et al., 1997; Yakovlev et al., 1998; Wallis and Miller, 2003). In completed trials, reaction times were defined as the time interval between the onset of the green dot and the onset of bar release. Inferences based on reaction times did not alter the conclusions based on correct-rates alone, and thus are not reported.

The first session with percent correct significantly above chance (at p < 0.01, χ2-test), followed by at least one more significant session among the next two, was taken as the onset of discriminative behavior (i.e., as evidence that the association between trial type and reward contingency had been acquired). 99% confidence intervals for percent correct were based on Wilson “score” interval for a binomial proportion (Brown et al., 2001; La Camera and Richmond, 2008).

Analysis of the iDMS task

After all monkeys had accomplished both acquisition and reversal of the iDMS task with 1 s delay, the number of sessions to criterion were analyzed with a 2-way, mixed design ANOVA, with two factors (“group” and “protocol”) and within-subjects repeated measures. Post-hoc multiple comparisons were based on a Mann-Whitney test. A Kruskal-Wallis test was used to analyze the number of sessions to criterion for reversal in the iDMS task with 100 ms delay.

Analysis of control tasks: change-point procedure

For the control tasks of Figures 2C,D we also performed a trials-to-criterion analysis. Trials-to-criterion were obtained with the change-point procedure (Gallistel et al., 2001, 2004). The procedure selected a sequence of trials that marked the putative onset of acquisition (or reversal). These trials are called “change-points” and are identified as the points where the change in slope of the cumulative record of correct responses exceeded a chosen criterion. This algorithm follows an iterative procedure: starting from the initial trial x and any point y > x in the cumulative record, the putative change-point z was the point between x and y placed at the maximal distance from the straight-line connecting x and y. The selected point was then checked for statistical significance, by comparing performance (percent correct) between trials x and z vs. performance between trials z and y2-test with P-value “p”). The putative change-point was accepted as valid if logit(p) = log10((1–p)/p) was >10 (results were robust to variations in this criterion). Given a valid change-point z, the search for the next change-point would start again, this time starting from x = z as the initial trial. Otherwise, the procedure was repeated using y+1 as the new end point, until a valid change-point (if any) was found. The three following scenarios could occur:

  1. At least one change-point was found, in which case the earliest change-point was taken as the acquisition/reversal point (Figure 3A);

  2. No change-point was found, but the overall percent correct in the session was significantly higher than chance (p < 0.01, χ2-test). In such a case, the first of 10 consecutive correct responses was taken as the valid change-point (such a point could always be found). This scenario occurred when the cumulative response curve was a straight line, i.e., the integrated correct response rate did not change over time (Figure 3B);

  3. A special case of (ii) occurs when the monkey is indifferent to trial type and always releases on green (the percent correct in this case is 50%). This could happen when monkeys were first exposed to a task, or in the first session after a task reversal. In those cases, all subsequent sessions were analyzed until a valid change-point was found (typically, this required the analysis of 1–2 additional sessions). All trials in previous sessions were added to the count of the trials performed up until the acquisition/reversal point. This circumstance occurred only infrequently and across all groups.

Figure 3.

Figure 3

Change-point procedure. Two examples of the change-point procedure used to determine the number of trials-to-criterion in the tasks of Figures 2C,D (see section Methods). In both panels, full lines show the cumulative record of correct responses, which grows by one for each correct trial (see section Methods). Gray bars show the difference in correct rates between trial types in successive blocks of 15 trials (the horizontal dotted line marks 100% difference in correct rates between trial types). The time-course of the difference in correct rates confirms the validity of the procedure but was not used to find the change-point. (A) In this example, the change-point (vertical broken line) marks the successful reversal of the cue-outcome association task of Figure 2C on the first day the experimenter had reversed the task (monkey JE). The change-point was reached at trial 180 (corresponding to 48 completed trials). (B) Cumulative record of correct responses for monkey SH during the first day of exposure to the same task as in (A). The cumulative record is a straight line implying immediate acquisition of the task, as is apparent from the time course of the difference in correct rates (gray bars). The overall percent correct (with 99% confidence interval) was 91.8 < 95.1 < 97.1. No valid change-point was found in this case, and the first of 10 consecutive correct responses was used as a criterion, according to which acquisition of the task occurred after 6 completed trials.

Since the monkeys received feedback about the rules of the task only in completed trials, to compute the number of trials to criterion we always used the number of completed trials (and not the number of total trials), which were reported as trials-to-criterion. Trials-to-criterion across the groups were analyzed with the same 2-way, mixed design, ANOVA used to analyze the number of sessions to criterion in the iDMS task.

Results

Nine rhesus monkeys, 3 normal controls, 3 with large lateral prefrontal (LPFC) cortex lesions (Figures 1A,B), and 3 with large orbitofrontal (OFC) lesions (Figures 1C,D), were initially trained to release a bar when a red dot turned green (Figure 2A). Correct bar releases were rewarded while incorrect ones resulted in the abortion of the trial. After reaching proficiency in this task and after testing all monkeys for sensitivity to reward postponement (Methods), the monkeys were tested in the iDMS task of Figure 2B. In this task, they had to predict the outcomes of individual trials (a reward or a 7s “time-out,” respectively) based on whether two sequentially presented visual stimuli were the same (“match”) or different (“non-match”). Initially, match trials predicted reward and non-match trials predicted a time-out (“reward-if-match” rule). As before, a bar release on green was required to complete each trial, which otherwise was aborted; however, a completed trial would now result in a 7 s time-out (followed by no reward) in 50% of the trials. The monkeys could learn to “skip” time-out trials by not releasing on green in those trials, which required that the monkeys had learned to infer correctly the trial type according to whether a “reward-if-match” rule or a “time-out-if-match” rule was in effect. In scoring the monkey's decisions, we considered “correct” either completed trials that led to reward, or aborted trials that would result in a time-out; “incorrect” trials were either completed trials that led to a time-out, or aborted trials that would lead to a reward. Error trials due to a bar release prior to the occurrence of the test image could not be attributed to a predicted outcome and were not scored.

Initially, the overall percent correct was about 50% for each of the 9 monkeys (Figure 4A). This score originated from their previous testing in the bar release task of Figure 2A, when the monkeys had learned to always release on green to obtain reward. After 6–9 sessions, the percent correct for the control and LPFC monkeys increased so that it was greater than chance (Figure 4A, solid black circles, p < 0.01). The monkeys with OFC lesions took significantly longer (13–21 sessions) to learn to perform the task to criterion (Figures 4, 5).

Figure 4.

Figure 4

Monkey's behavior in the iDMS task. (A) Percent correct vs. session number of all monkeys. Top, middle, and bottom rows display data from control (CTRL), LPFC and OFC monkeys, respectively. In each panel, percent correct with 99% confidence intervals are shown (see section Methods). Filled circles mark sessions with percent correct significantly above 0.5 (dashed lines; p < 0.01). Reversal of the task occurred in correspondence of the vertical lines. (B) “Skip rates” during match (dark bars) vs. non-match trials (lighter bars) in the same data presented in (A). The asterisks mark the first day of statistically significant % correct performance (as defined in A) before (left to the vertical line) and after task reversal.

Figure 5.

Figure 5

Analysis of behavior in the iDMS task. (A) Number of sessions-to-criterion for acquisition (“ACQ”) and reversal (“REV”) of the iDMS task for control, LPFC, and OFC monkeys. A significant difference between groups was found (2-way, mixed design ANOVA, p < 0.02). (B) Interaction plot (protocol × group) of the 2-way ANOVA used to compare mean sessions-to-criterion among control, LPFC, and OFC monkeys shows an impairment of OFC monkeys in acquiring and reversing the task and an impairment of LPFC monkeys in reversing the task, whereas acquisition performance of LPFC monkeys was not significantly different than controls. See the text for details.

The rule-outcome association was then reversed (reversal marked by vertical lines in Figure 4A) between 2 and 5 sessions after acquisition (see Methods section “iDMS task”). Performance dropped to ~50% correct in all monkeys by the end of the first session after reversal (Figure 4A; for monkey SU in the OFC group the overall score of the entire session was significantly different than 50%, but the 50% mark was reached by the 4th quarter of the session; not shown). All 3 control monkeys acquired the reversed task within 5 testing sessions. However, this time both the LPFC and the OFC monkeys took significantly longer to reverse (LPFC monkeys: 10–19 sessions, Figure 4A, middle row; OFC monkeys: 12–21 sessions, Figure 4A, bottom row).

Recall that the monkeys could “skip” a trial in one of two ways: either by releasing the bar too early (within 200 ms from the onset of green), or by not releasing the bar at all (within 1 s after onset of green). The same percent correct performance can result from different patterns of “skip rates.” For example, 50% correct could be the result of never skipping a trial as well as skipping all trials. In Figure 4B we show the skip rates in match vs. non-match trials, before and after reversal. As the monkeys learned the task, skip rates in the rewarded condition tended to decrease while the skip rates in the time-out condition tended to increase. Skip rates decreased to very low values in both conditions during the first session after a reversal, dropping the performance back to ~50% correct, as noted earlier. Despite some idiosyncratic features found in single monkeys (some tend to skip more and some less, in general), these trends were found in monkeys of all groups. Similarly, we found no clear difference in the pattern of “early” (before green) vs. “late” (>1 s after green on) bar releases among different groups (not shown), suggesting a lack of clear difference in how the rules of the task were acquired by monkeys of different groups. When the monkeys started to skip time-out trials, they did so by releasing late, and then switched to releasing early after considerable practice (typically, after the control tasks, when exposed to the iDMS task with short delay—see section Methods for details on the sequence of tasks). This pattern was found in all monkeys.

The numbers of sessions to criterion are summarized in Figure 5A and were subjected to a 2-way, mixed design ANOVA. We found a significant difference among the groups (p < 0.03) and a significant interaction of the effects of group and protocol (i.e., acquisition vs. reversal; p < 0.01). Inspection of the interaction plot (Figure 5B) suggests that significant interaction is to be attributed to the LPFC group. Post-hoc comparisons confirmed that acquisition in this group was similar to controls (p > 0.8, Mann-Whitney test) but reversal learning was significantly impaired (p = 0.0318, Mann-Whitney test, one-tailed). During testing in (but prior to the acquisition of) the iDMS task, the mean reaction times for the treated monkeys were not significantly different from those of control animals (2-way, mixed design ANOVA, no effect of group or group-trial type interaction, p > 0.27), showing that the lesions caused no motor impairment.

Dependence of performance on delay duration and novel stimuli

What strategy did the monkeys use to learn the iDMS task? We sought to answer this question by analyzing how the performance depended on (i) the temporal delay between the visual images or (ii) the use of a new cue set. Regarding (i), if the monkeys were relying on working memory to perform the task, their performance should degrade with lengthened delay duration, as previously reported (Fuster and Bauer, 1974; Bauer and Fuster, 1976; Shindy et al., 1994; Petrides, 2000a). As for (ii), to confirm that the monkeys were using the concept of “same” vs. “different” rather than simply form hundreds of stimulus-outcome associations, we tested their ability to generalize the rule to a new cue set. Given that we used sets of 50 stimuli in each session, there were 2,500 possible combinations of matching and non-matching pairs, leading to the hypothesis that the monkeys used the information contained in the abstract concepts of “same” and “different,” independent of cue identity. Based on this hypothesis, performance with a new cue set should not be different than the previously acquired level.

We tested both hypotheses concurrently by gradually increasing the delay up to 21 s while occasionally introducing a new cue set. Delay duration was increased by 1 s after each day of significant percent correct with the previous delay (see section Methods for details). Performance was idiosyncratically related to delay duration, increasing with duration (and practice) for some of the monkeys and decreasing for others, regardless of group (Figure 6A). The differences in percent correct across groups were not significant (Kruskal-Wallis test, p > 0.39). Thus, in our version of iDMS, increasing the interval between the sample and test images did not necessarily lead to a decline of performance, independently of the manipulation (control, LPF, or OFC lesion).

Figure 6.

Figure 6

Performance vs. delay duration in the iDMS task. (A) Percent correct vs. delay duration for all monkeys (ordered as in Figure 4A). Top, middle, and bottom rows refer to control (CTRL), LPFC and OFC monkeys, respectively. When significant (p < 0.01), regression lines are shown as broken lines (among all significant regressions it was p < 0.0003). See section Methods for details. (B) Percent correct with familiar (circles) and new (dash) cue sets for all monkeys. The novel set was introduced 1 day after familiar one. The numbers above symbols indicate the temporal delay used for both sets (seconds). Vertical broken lines separate data from different monkeys within the same group. See section Methods for details.

While increasing delay duration, new cue sets were occasionally introduced for each monkey, using the same delay as in the immediately preceding session (see section Methods for details). The new cue set did not affect performance (Figure 6B), indicating that the monkeys were using the abstract concepts of “same” and “different” regardless of the individual identity of the visual stimuli. It is also clear, from Figure 6A, that all monkeys attained a similar maximum level of performance, albeit at different points during testing (at the beginning, at the end, or in the middle) within groups as well as across groups. In conclusion, the lesions affected acquisition and/or reversal learning rates (Figure 5), but not the asymptotic levels of performance or the dependence on delay duration. Moreover, the lesions slowed down, but did not abolish, the ability of the monkeys to form an abstract concept, because all monkeys performed well with a new cue set.

The monkey's impairment did not depend on delay

Correct performance of the iDMS task requires several steps: (i) remembering a visual stimulus across a temporal delay (ii) forming the “same” vs. “different” abstract rules, (iii) learning to predict the outcome of each action depending on the rule in effect, and (iv) adopting a suitable behavioral strategy to take advantage of this knowledge. In the previous section, we have investigated the roles of delay duration and the formation of an abstract rule on the monkey's behavior.

However, we have not investigated the role of the delay per se (as opposed to no delay at all) or the role of the abstract concept per se (as opposed to no concept at all). To understand if the observed impairments were due to the presence of a temporal delay, and/or to the necessity to form the abstract concept of “same” vs. “different,” we performed three control experiments, each requiring a subset of the abilities summarized above, but all having the same fundamental task structure (release on green to obtain a reward or a time-out).

In the first control, one of two new visual cues predicted a reward, while the second cue predicted a time-out (Figure 2C). Each cue was present during the entire duration of the trial. Everything else was as in the iDMS task (see section Methods for details). Since the monkeys learned to distinguish between the stimuli within 1 testing session, and also reversed the task in 1 session, we measured performance in terms of the number of trials to criterion, in hopes to resolve differences occurring within a single session. Trials-to-criterion were obtained with an iterative change-point procedure (Gallistel et al., 2001, 2004) (Figure 3 and section Methods) and then analyzed with a 2-way, mixed design ANOVA (Figure 7A). There was a significant group effect (p = 0.048), with the LPFC requiring more trials (93 ± 68, mean ± SD) than either the control or the OFC group (44 ± 26, mean ± SD over both groups). However, there was no difference in the rate of initial acquisition vs. the rate of reversal (acquisition vs. reversal, p > 0.3), nor was there an interaction between rate of acquisition/reversal and group (p > 0.9). Learning took <180 trials across all monkeys and was much faster than in the iDMS task.

Figure 7.

Figure 7

Behavior in control tasks. (A) Number of trials to criterion for the cue-outcome association task of Figure 2C (all monkeys). Dark-gray bars: trials to acquisition; light-gray bars: trials to reversal. See the text for details. (B) Trials to criterion for the delayed cue-outcome association task of Figure 2D (all monkeys; same key as in A).

Qualitatively, these results did not change when a delay was introduced between the cue and the response (Figure 2D). In this variation of the task, the cue or its meaning had to be remembered across the delay. As with the previous task, all monkeys acquired the delayed task within one testing session, and trials-to-criterion were computed as described above (Figure 7B). No significant effect of either group or protocol was found on the number of trials-to-criterion (p > 0.5, 2-way ANOVA, mixed design; 215 ± 150 trials-to-criterion, mean ± SD over all groups and both protocols). Learning was slower than in the simpler cue-outcome association task without delay, but it was still much faster than in the iDMS task.

The monkey's impairments relate to abstract concept formation

Finally, to investigate the role of the abstract concepts of “same” vs. “different,” we re-tested the monkeys in the iDMS task. This time, the mnemonic requirements were kept to a minimum by using only a 100 ms delay between sample and test (see section Methods for details). The monkeys were tested in the same iDMS task to which they had been last exposed (days-to-criterion shown as light bars in Figure 5A and reproduced in Figure 8 as white bars). All 9 monkeys performed well above chance on the first testing day, and then at least 3 sessions were needed to reverse the task (black bars in Figure 8): three sessions for controls, 4–11 sessions for lesioned monkeys, with a small but significant effect of group (p = 0.049, Kruskal-Wallis rank sum test) but no difference between LPF and OFC monkeys (see Figure 8, black bars).

Figure 8.

Figure 8

Behavior in the iDMS task with 100 ms delay and in additional reversals of the iDMS task. Number of sessions required for reversing the iDMS task with 100 ms delay for all monkeys (black bars) and for 3 additional reversals of the iDMS task with 1 s delay, temporally ordered from left to right (gray bars). For comparison, the number of sessions required to reverse the iDMS task with 1 s delay are also reported (white bars, same as light gray bars in Figure 5A). Decrease in learning times with the number of reversals reveals the presence of a learning set. See the text for details.

Differently from the previous tasks based on cue-outcome association rules instead of rules based on abstract concepts, it took longer than one session to reverse the iDMS task with 100 ms delay, with a significant effect of either lesion. This suggests that the necessity to form an abstract concept to perform the task is key to the impairments. We also note, however, that learning to reverse the iDMS task with 100 ms delay was faster than learning to reverse the iDMS with 1 s delay in all monkeys (Figure 8, compare black to white bars). Thus, the presence of a significant temporal delay could still be responsible for the impairments, at least to a partial extent. The alternative hypothesis is that the monkeys were developing a “learning set,” that is, they were learning about the reversals through repeated practice with them. To test this hypothesis, we increased the delay to 1 s and reversed the task three additional times.

For all groups, performance kept improving over the three additional reversals (Figure 8, gray bars), and the difference among groups disappeared (2-way ANOVA, mixed design, p > 0.28). This also shows that despite the difficulty with the initial rule reversals, the ability to form a learning set for reversal of the abstract rule was not impaired in LPFC and OFC monkeys, similar to what found in marmoset monkeys in a simpler association task (Dias et al., 1997).

Taken together, these results imply that the most taxing cognitive component in reversing the iDMS task was learning the new association of the concepts of “same” and “different” with the reward-contingency, with a much smaller role (if any) for short-term memory (Rushworth et al., 1997) or for reversing a cue-outcome association per se.

Discussion

Rule-based behavior is advantageous because no new learning is required when some properties of a task change (here, the identity or the number of the images, and the length of the delay interval), allowing flexibility in mapping circumstances into actions. Here, we have shown that rhesus monkeys can rapidly learn to predict forthcoming behavioral outcomes using a rule based on the abstract concept of “same” vs. “different,” and that their behavior in this task is differentially affected by bilateral removals of the lateral or orbital frontal cortex. Monkeys with bilateral OFC removals learned and reversed the rule significantly more slowly than controls. Perhaps more surprisingly, monkeys with large bilateral LPFC lesions learned the same task just as quickly as unoperated controls, but were significantly slower at carrying out the first rule reversal. Crucial to the impairments observed in lesioned animals was that the rule of the task was based on the abstract concepts of “same” vs. “different,” since the lesions did not affect the ability to learn simpler direct cue-outcome associations that did not require the formation of an abstract concept.

Since it is thought that PFC is essential for rule learning (Burgess, 2000; Procyk et al., 2000; Bussey et al., 2001; Miller and Cohen, 2001; Wallis et al., 2001; Wood and Grafman, 2003; Genovesio et al., 2005; Tanji and Hoshi, 2008; Wise, 2008; Rigotti et al., 2013), our main, and possibly most surprising, result, is the apparent lack of impairment in acquiring the iDMS task following large bilateral LPFC removals. Both the rapid learning rates and the lack of effect of LPFC lesions on the acquisition of our iDMS task seem at odds with previous results. It is unavoidable that different studies differ in the exact extent of the lesions, which is often mentioned as the potential reason for confounding and contradicting results in the literature. In our case, however, we had large LPFC lesions extending to both the superior and inferior convexity and including the principal sulcus, with little opportunity for remaining fragments of LPFC tissue. Such extended lesions should impair the performance of monkeys in matching tasks with or without a spatial component (Passingham, 1975; Mishkin and Manning, 1978), and with (Fuster and Alexander, 1971; Goldman-Rakic, 1995, 1996; Miller et al., 1996; Petrides, 2000b; Curtis and D'Esposito, 2003, 2004), or without (Rushworth et al., 1997), a delay. Thus, we expected that our LPFC monkeys would be impaired in the acquisition of the iDMS task. Instead, we found that the LPFC monkeys were as facile as controls in acquiring the task, and were instead impaired at reversing it. This deficit does not seem to be of a perseverative nature (Dias et al., 1997; Clarke et al., 2008; Buckley et al., 2009), since the monkeys quickly abolished their previously acquired strategies (i.e., release on green in rewarded trials only), and started releasing on green in all trials after the task rule was reversed (cf. Figure 4), presumably to sample as many rule-outcome contingencies as possible.

We believe that at least part of the reason for our unexpected results with LPFC lesions relies in our different training procedure compared to more common implementations of DMS. But before discussing the potential role of task design, we discuss some of the potential roles of the specializations of LPFC, OFC, and connected brain regions, in learning and reversing the iDMS task.

The potential role of functional specializations of the PFC

Given the presence of a delay between the two images in the iDMS task, and between a visual stimulus and the monkey's action in one of our control tasks, a form of short-term memory seems required to learn and perform these tasks. Neural activity related to short-term memory, and particularly working memory, is often seen in the LPFC (Fuster and Alexander, 1971; Funahashi et al., 1989). The principal sulcus especially has been found necessary to support working memory for abstract rules, with the ventrolateral PFC being involved in implementing previously acquired abstract rules in a Wisconsin Card Sorting Test (WCST) analog task (Buckley et al., 2009). It has also been found that neurons in areas 46 and 9/46 of macaque LPFC can maintain representations of uncued rules (Mansouri et al., 2006). These findings resonate with the worsened ability of our LPFC monkeys to reverse the iDMS task, since our task also involves short-term memory for stimuli and uncued task reversals. However, there are important differences between the WCST analog and the iDMS task, especially in the time scale of the reversals, which are rapid in the WCST analog (where they occur multiple times during a session) but slow in our use of the iDMS task (where reversals never occurred within the same session). This difference may explain why Buckley et al. (2009) also found the OFC to be necessary to rapidly update rules on the basis of reinforcement, whereas we found that a rule update over the time course of days (in our control tasks) was not affected by OFC lesions. However, our main results in the iDMS task show that, even in the absence of fast updating, the OFC is required to learn (or reverse) an association between an abstract concept and a rule. Thus, we may be tempted to conclude that OFC is essential either when adapting to a rapidly changing rule, or when learning a rule that depends on the formation of abstract concepts.

Other studies have shown that the LPFC might encode conjunctively the state of the environment and certain attributes associated to it, such as the values of the expected reward (Watanabe and Sakagami, 2007). In the iDMS task, these values would relate to the reward vs. time-out, with positive and negative value, respectively. Given that our LPFC monkeys were impaired in reversing this task, a possible interpretation of our results is that the ability of LPFC to represent states and values of the environment conjunctively seems more essential in changing environments (e.g., after a reversal) than during an initial acquisition.

In contrast to LPFC neurons, there is little evidence so far that neurons in the OFC encode states of the environment together with the values of actions associated with them (Lee et al., 2007). Instead, their activity seems to reflect expected outcome (Tremblay and Schultz, 2000; Wallis and Miller, 2003; Roesch and Olson, 2004) that follows from the monkey's choice (Padoa-Schioppa and Assad, 2006)—as, we would argue, also occurs in our tasks. Our results are in keeping with this view and with the known involvement of the OFC in reversal learning (Dias et al., 1996, 1997; Izquierdo et al., 2004; Walton et al., 2010). However, once again, the agreement of our results with previous findings seems to hold only in the presence of an abstract concept, since we found no impairments in the control tasks. The latter result, on its own, would be compatible with the finding that discrimination learning seems unaffected by OFC lesions (Bussey et al., 2001; Izquierdo et al., 2004).

The potential role of brain areas connected to OFC and LPFC

Although the LPFC lesions did not significantly impact the initial acquisition of the iDMS task, the LPFC monkeys were impaired in reversing the same task, and the OFC monkeys were impaired in both acquisition and reversal. These results may be due to the indirect disruption induced by the lesions in target areas. The dorsolateral PFC is known to be associated with parietal cortices to support attention and executive functions (Chafee and Goldman-Rakic, 2000). For example, Crowe et al. (2013) have shown evidence of flow of rule-based information from dorsolateral PFC to parietal cortices in a decision making task. Thus, it is possible that impaired communication between lateral prefrontal and parietal areas may be responsible for the impairment in reversing the iDMS rule following LPFC lesions. A role could also be played by the disruption of communication between orbitofrontal or lateral prefrontal regions with the anterior cingulate cortex (ACC), which is critical for behavioral flexibility (Shenhav et al., 2013; Kolling et al., 2016; Procyk et al., 2016). In particular, the primate ACC might play a key role in choosing appropriate actions when the environment is uncertain or dynamic (Lee et al., 2007) which, however, is barely the case in our iDMS task. On a related topic, outcome-related activity in ACC has often been linked to the ability to update decision-making strategies after committing an error (Shima and Tanji, 1998; Procyk et al., 2000). According to this view, disruption of pathways between lateral and orbitofrontal cortices and ACC may lead to perseverative errors, which, however, were not observed in any of our tasks. Giving that the primate ACC projects to cortical areas with motor functions (Dum and Strick, 1991; Wang et al., 2001; Luppino et al., 2003), an indirect consequence of disrupting the dialogue between frontal cortices and ACC may result in an inability to encode or update the value of actions (Rushworth et al., 2007), or even in the inability to perform the correct action itself (we did not, however, observe any motor impairment in our experiments).

Similarly, the interaction between OFC and amygdala is critical for updating behavior based on changes in outcome value (Baxter and Murray, 2002; Fiuzat et al., 2017), and they also play complementary roles in processing previously rewarded objects during reversal learning (Rudebeck and Murray, 2008). The OFC is also involved in value-based decision making through its interaction with the perirhinal and entorhinal cortices, especially when monkeys' decisions rely upon contextual or memory information (Clark et al., 2013). The disruption caused in either of these two circuits by OFC lesions could account for the slower adjustment of our animals to the changes in task contingencies.

The potential role of the iDMS task

Assuming that the exact extent of the LPFC lesions is not to blame for the lack of impairment in acquiring the iDMS task, how can one explain those differences with previous findings summarized above? We believe that the specific structure of the tasks used here is the most likely explanation. Unlike procedures involving touching computer screens, or direct manual object displacements as in the Wisconsin General Testing Apparatus (WGTA, see e.g., Moore et al., 2009, 2012 for recent studies), learning DMS in an automated apparatus can be laborious for monkeys, sometimes taking months (Pigarev et al., 2001); yet, our monkeys displayed rapid learning times. This suggests that the monkeys would learn to perform our tasks in a different way compared to other studies. Previous DMS studies, for example, involved an active action selection process, i.e., different motor actions in different conditions (such as match vs. non-match trials in the iDMS task). By removing the action selection process altogether, we stripped the task of any differential motor component. Although the monkeys eventually learned to produce the bar release at different times to complete or “skip” trials, they were initially required to perform only one action: to release the bar on green (see Figure 2A). Although not always the best response in the iDMS task, releasing on green was necessary at first to sample the outcome in each trial type. It was only after having learned the predictive value of each trial type that the monkeys could act on their knowledge and intentionally “skip” time-out trials. The behavior of our control monkeys alone proves that this strategy is fundamentally different from learning, simultaneously, the meaning of trial types and a desired response, eliminating a perennial confounding aspect of these types of tasks and resulting in much faster learning times.

Two factors of our procedures seem most important in producing our results. First, the monkeys are exposed to the iDMS task only after having been taught to release a bar after a red target turned green. The initial training in the bar release task, which is easily learned (~2 weeks, see e.g., Liu et al., 2004), taught the monkeys about the only operant action required to perform our iDMS task. Second, when testing the monkeys in the iDMS task, we capitalized on the tendency of the monkeys to quickly learn the meaning of predictive cues and to abort time-out trials prematurely. As we know from previous studies (Bowman et al., 1996; La Camera and Richmond, 2008; Minamimoto et al., 2009), incurring a time-out period before receiving a reward is a powerful modulator of the monkeys' behavior, that often results in incorrect or premature bar release even when correction trials are required. In our iDMS task, learning this behavior is further facilitated by the fact that incorrectly timed bar release results in an advantageous strategy that the monkeys learned quickly. With different variations of this procedure, we have found that monkeys also learn attentional sets and categories quickly (Lerchner et al., 2007; Minamimoto et al., 2010). This design seems to tap into the monkey's natural ability to infer information from the environment (Bromberg-Martin and Hikosaka, 2009) and take advantage of it, somewhat similarly to what has been observed with the repeat-stay, change-shift strategy (Bussey et al., 2001).

The other key aspect of this study is the direct comparison in tasks that differ only for the presence of a DMS component. This allowed us to disentangle the case where it is an abstract concept (“same” vs. “different”) that is predictive of reward contingency, from the case where it is a single visual stimulus. In addition, since the concepts of “same” vs. “different” had to be formed across a temporal delay (up to 21 s), we could explicitly test the importance of the delay between cue presentation and motor response. Our main finding is that only when the abstract concept is present we detect a clear and statistical significant impairment of the lesioned monkeys. On the one hand, the results on acquisition are in keeping with previous studies finding a largely intact ability of monkeys with OFC damage to make appropriate choices when initially learning the values of available options (Izquierdo et al., 2004; Clarke et al., 2008; Walton et al., 2010). On the other hand, these results are at odds with the known involvement of OFC in reversal learning guided by conditioned stimuli, which is typically accompanied by pathological perseveration (Murray et al., 2007; Clarke et al., 2008). Once again, the most likely explanation is due to the difference in task procedures, which in our case involved a single action that had been previously learned. It is also possible that, having learned the more taxing iDMS task first, our lesioned monkeys found easier to learn the simpler control tasks. Additional experiments are required to resolve this ambiguity—here, we had to test the monkeys in the iDMS task first, to make sure that acquisition of iDMS would not be facilitated by previous testing in the simpler cue-outcome association tasks. If prior learning of the iDMS task did not impact the behavior in the control tasks, the fact that acquisition of our iDMS task was impaired in OFC monkeys, whereas the acquisition of the simpler cue-outcome associations were not, would reveal that OFC is essential for forming an abstract concept and/or for correctly assigning a predicted outcome to it (more work would be required to establish which).

Finally, we must also consider the possibility that our results, and especially the lack of expected impairments, could be the consequence of transfer effects from previous learning experience. Our monkeys had received their ablations in the context of other studies (see section Methods), and had been tested in other tasks prior to their use in this study, which may have resulted in transfer effects of previous learning experience. However, we believe these transfer effects to be minimal, if present at all. The deficits found in OFC monkeys were expected based on previous literature. The lack of impairment in the control tasks is more surprising, but rather than to previous experience, it could be linked to having learned the more difficult iDMS task first, as discussed above. Regarding the LPFC monkeys, they were surprisingly unimpaired in a Wisconsin-analog task. In performing this task, the monkeys were tested for the ability to task-set shifting without the requirement of producing different motor responses to different stimulus attributes (Lerchner et al., 2007). This testing approach is very similar to that used in our iDMS task. The same monkeys were found unimpaired in a rapid visual categorization task (Minamimoto et al., 2010) that followed a similar principle of associating different postponement durations to different attributes of the stimulus (e.g., its being an exemplar of cat or dog). As in our tasks, the task requirements in the Wisconsin-analog and the rapid categorization tasks were atypical and probably responsible for the surprising lack of impairments in those tasks. This is also the view advocated in this article, where we have reached similar conclusions in a task that requires significantly longer learning times and thus is, presumably, a more taxing and substantially different task. As for the impairment of our LPFC monkeys in reversing the iDMS task, we note that it is unlikely to follow from negative transfer effects from a set shifting or a categorization task, where rapid shifting or categorization were not impaired. It could be revealing to test if reversal learning would also be impaired in the rapid categorization task of Minamimoto et al. (2010).

Conclusions

In summary, by comparing the effect of LPFC and OFC lesions in different scenarios, one of which requires forming an abstract concept, we found that both LPFC and OFC are essential for different aspects of learning the association between an abstract concept, a decision, and a predicted outcome. Neither of these brain regions seemed crucial in establishing a learning set for concept-outcome associations. It appears that OFC is needed to form the “same/different” concepts initially, learn the rule and/or associate it with the outcome, whereas LPFC seems to be needed to modify a previously acquired concept-outcome association.

Author contributions

GLC: Designed the behavioral paradigms, carried out the experiments, performed the data analysis, and wrote the paper; SB: Carried out the experiments and wrote the paper; BR: Supervised the study and wrote the paper. All authors contributed to the intellectual development of the study and the manuscript.

Conflict of interest statement

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Acknowledgments

We are indebted to R. C. Saunders and E. A. Murray for performing the lesions. We thank D. P. Soucy and J. Peck for their technical assistance, T. Minamimoto, A. Lerchner, J. M. Simmons, S. Ravel, M. Mishkin, and R. C. Saunders for useful discussions. This study was supported by the Intramural Research Program of the National Institute of Mental Health (project number ZIAMH002032). During the completion of this work GLC was supported by NIH ARRA P30 NS069256 start up funds. The views expressed in this article do not necessarily represent the views of the NIMH, NIH, or the United States Government.

References

  1. Amit D. J., Fusi S., Yakovlev V. (1997). Paradigmatic working memory (attractor) cell in IT cortex. Neural Comput. 9, 1071–1092. 10.1162/neco.1997.9.5.1071 [DOI] [PubMed] [Google Scholar]
  2. Bauer R. H., Fuster J. M. (1976). Delayed-matching and delayed-response deficit from cooling dorsolateral prefrontal cortex in monkeys. J. Comp. Physiol. Psychol. 90, 293–302. 10.1037/h0087996 [DOI] [PubMed] [Google Scholar]
  3. Baxter M. G., Murray E. A. (2002). The amygdala and reward. Nat. Rev. Neurosci. 3, 563–573. 10.1038/nrn875 [DOI] [PubMed] [Google Scholar]
  4. Bouret S., Richmond B. J. (2010). Ventromedial and orbital prefrontal neurons differentially encode internally and externally driven motivational values in monkeys. J. Neurosci. 30, 8591–8601. 10.1523/JNEUROSCI.0049-10.2010 [DOI] [PMC free article] [PubMed] [Google Scholar]
  5. Bowman E. M., Aigner T. G., Richmond B. J. (1996). Neural signals in the monkey ventral striatum related to motivation for juice and cocaine rewards. J. Neurophysiol. 75, 1061–1073. 10.1152/jn.1996.75.3.1061 [DOI] [PubMed] [Google Scholar]
  6. Bromberg-Martin E. S., Hikosaka O. (2009). Midbrain dopamine neurons signal preference for advance information about upcoming rewards. Neuron 63, 119–126. 10.1016/j.neuron.2009.06.009 [DOI] [PMC free article] [PubMed] [Google Scholar]
  7. Brown L. D., Tony Cai T., DasGupta A. (2001). Interval estimation for a binomial proportion. Stat. Sci. 16, 101–133. 10.1214/ss/1009213286 [DOI] [Google Scholar]
  8. Buckley M. J., Mansouri F. A., Hoda H., Mahboubi M., Browning P. G., Kwok S. C., et al. (2009). Dissociable components of rule-guided behavior depend on distinct medial and prefrontal regions. Science 325, 52–58. 10.1126/science.1172377 [DOI] [PubMed] [Google Scholar]
  9. Bunge S. A., Kahn I., Wallis J. D., Miller E. K., Wagner A. D. (2003). Neural circuits subserving the retrieval and maintenance of abstract rules. J. Neurophysiol. 90, 3419–3428. 10.1152/jn.00910.2002 [DOI] [PubMed] [Google Scholar]
  10. Bunge S. A., Wallis J. D., Parker A., Brass M., Crone E. A., Hoshi E., et al. (2005). Neural circuitry underlying rule use in humans and nonhuman primates. J. Neurosci. 25, 10347–10350. 10.1523/JNEUROSCI.2937-05.2005 [DOI] [PMC free article] [PubMed] [Google Scholar]
  11. Burgess P. W. (2000). Strategy application disorder: the role of the frontal lobes in human multitasking. Psychol. Res. 63, 279–288. 10.1007/s004269900006 [DOI] [PubMed] [Google Scholar]
  12. Bussey T. J., Saksida L. M., Murray E. A. (2002). Perirhinal cortex resolves feature ambiguity in complex visual discriminations. Eur. J. Neurosci. 15, 365–374. 10.1046/j.0953-816x.2001.01851.x [DOI] [PubMed] [Google Scholar]
  13. Bussey T. J., Saksida L. M., Murray E. A. (2003). Impairments in visual discrimination after perirhinal cortex lesions: testing ‘declarative’ vs. ‘perceptual-mnemonic’ views of perirhinal cortex function. Eur. J. Neurosci. 17, 649–660. 10.1046/j.1460-9568.2003.02475.x [DOI] [PubMed] [Google Scholar]
  14. Bussey T. J., Wise S. P., Murray E. A. (2001). The role of ventral and orbital prefrontal cortex in conditional visuomotor learning and strategy use in rhesus monkeys (Macaca mulatta). Behav. Neurosci. 115, 971–982. 10.1037/0735-7044.115.5.971 [DOI] [PubMed] [Google Scholar]
  15. Chafee M. V., Goldman-Rakic P. S. (2000). Inactivation of parietal and prefrontal cortex reveals interdependence of neural activity during memory-guided saccades. J. Neurophysiol. 83, 1550–1566. 10.1152/jn.2000.83.3.1550 [DOI] [PubMed] [Google Scholar]
  16. Chudasama Y., Kralik J. D., Murray E. A. (2007). Rhesus monkeys with orbital prefrontal cortex lesions can learn to inhibit prepotent responses in the reversed reward contingency task. Cereb. Cortex 17, 1154–1159. 10.1093/cercor/bhl025 [DOI] [PubMed] [Google Scholar]
  17. Clark A. M., Bouret S., Young A. M., Murray E. A., Richmond B. J. (2013). Interaction between orbital prefrontal and rhinal cortex is required for normal estimates of expected value. J. Neurosci. 33, 1833–1845. 10.1523/JNEUROSCI.3605-12.2013 [DOI] [PMC free article] [PubMed] [Google Scholar]
  18. Clarke H. F., Robbins T. W., Roberts A. C. (2008). Lesions of the medial striatum in monkeys produce perseverative impairments during reversal learning similar to those produced by lesions of the orbitofrontal cortex. J. Neurosci. 28, 10972–10982. 10.1523/JNEUROSCI.1521-08.2008 [DOI] [PMC free article] [PubMed] [Google Scholar]
  19. Crowe D. A., Goodwin S. J., Blackman R. K., Sakellaridi S., Sponheim S. R., MacDonald A. W., III., et al. (2013). Prefrontal neurons transmit signals to parietal neurons that reflect executive control of cognition. Nat. Neurosci. 16, 1484–1491. 10.1038/nn.3509 [DOI] [PMC free article] [PubMed] [Google Scholar]
  20. Curtis C. E., D'Esposito M. (2003). Persistent activity in the prefrontal cortex during working memory. Trends Cogn. Sci. 7, 415–423. 10.1016/S1364-6613(03)00197-9 [DOI] [PubMed] [Google Scholar]
  21. Curtis C. E., D'Esposito M. (2004). The effects of prefrontal lesions on working memory performance and theory. Cogn. Affect. Behav. Neurosci. 4, 528–539. 10.3758/CABN.4.4.528 [DOI] [PubMed] [Google Scholar]
  22. Dias R., Robbins T. W., Roberts A. C. (1996). Dissociation in prefrontal cortex of affective and attentional shifts. Nature 380, 69–72. 10.1038/380069a0 [DOI] [PubMed] [Google Scholar]
  23. Dias R., Robbins T. W., Roberts A. C. (1997). Dissociable forms of inhibitory control within prefrontal cortex with an analog of the Wisconsin Card Sort Test: restriction to novel situations and independence from “on-line” processing. J. Neurosci. 17, 9285–9297. [DOI] [PMC free article] [PubMed] [Google Scholar]
  24. Dum R. P., Strick P. L. (1991). The origin of corticospinal projections from the premotor areas in the frontal lobe. J. Neurosci. 11, 667–689. [DOI] [PMC free article] [PubMed] [Google Scholar]
  25. Fiuzat E. C., Rhodes S. E., Murray E. A. (2017). The role of orbitofrontal-amygdala interactions in updating action-outcome valuations in macaques. J. Neurosci. 37, 2463–2470. 10.1523/JNEUROSCI.1839-16.2017 [DOI] [PMC free article] [PubMed] [Google Scholar]
  26. Funahashi S., Bruce C. J., Goldman-Rakic P. S. (1989). Mnemonic coding of visual space in the monkey's dorsolateral prefrontal cortex. J. Neurophysiol. 61, 331–349. 10.1152/jn.1989.61.2.331 [DOI] [PubMed] [Google Scholar]
  27. Funahashi S., Bruce C. J., Goldman-Rakic P. S. (1993). Dorsolateral prefrontal lesions and oculomotor delayed-response performance: evidence for mnemonic “scotomas.” J. Neurosci. 13, 1479–1497. [DOI] [PMC free article] [PubMed] [Google Scholar]
  28. Fuster J. M. (2008). The Prefrontal Cortex, 4th Edn. Amsterdam; Boston, MA: Academic Press/Elsevier. [Google Scholar]
  29. Fuster J. M., Alexander G. E. (1971). Neuron activity related to short-term memory. Science 173, 652–654. 10.1126/science.173.3997.652 [DOI] [PubMed] [Google Scholar]
  30. Fuster J. M., Bauer R. H. (1974). Visual short-term memory deficit from hypothermia of frontal cortex. Brain Res. 81, 393–400. 10.1016/0006-8993(74)90838-5 [DOI] [PubMed] [Google Scholar]
  31. Gallistel C. R., Fairhurst S., Balsam P. (2004). The learning curve: implications of a quantitative analysis. Proc. Natl. Acad. Sci. U.S.A. 101, 13124–13131. 10.1073/pnas.0404965101 [DOI] [PMC free article] [PubMed] [Google Scholar]
  32. Gallistel C. R., Mark T. A., King A. P., Latham P. E. (2001). The rat approximates an ideal detector of changes in rates of reward: implications for the law of effect. J. Exp. Psychol. Anim. Behav. Process. 27, 354–372. 10.1037/0097-7403.27.4.354 [DOI] [PubMed] [Google Scholar]
  33. Genovesio A., Brasted P. J., Mitz A. R., Wise S. P. (2005). Prefrontal cortex activity related to abstract response strategies. Neuron 47, 307–320. 10.1016/j.neuron.2005.06.006 [DOI] [PMC free article] [PubMed] [Google Scholar]
  34. Goldman-Rakic P. S. (1995). Cellular basis of working memory. Neuron 14, 477–485. 10.1016/0896-6273(95)90304-6 [DOI] [PubMed] [Google Scholar]
  35. Goldman-Rakic P. S. (1996). The prefrontal landscape: implications of functional architecture for understanding human mentation and the central executive. Philos. Trans. R. Soc. Lond. B Biol. Sci. 351, 1445–1453. 10.1098/rstb.1996.0129 [DOI] [PubMed] [Google Scholar]
  36. Hampton A. N., Bossaerts P., O'Doherty J. P. (2006). The role of the ventromedial prefrontal cortex in abstract state-based inference during decision making in humans. J. Neurosci. 26, 8360–8367. 10.1523/JNEUROSCI.1010-06.2006 [DOI] [PMC free article] [PubMed] [Google Scholar]
  37. Huettel S. A., Misiurek J., Jurkowski A. J., McCarthy G. (2004). Dynamic and strategic aspects of executive processing. Brain Res. 1000, 78–84. 10.1016/j.brainres.2003.11.041 [DOI] [PubMed] [Google Scholar]
  38. Izquierdo A., Suda R. K., Murray E. A. (2004). Bilateral orbital prefrontal cortex lesions in rhesus monkeys disrupt choices guided by both reward value and reward contingency. J. Neurosci. 24, 7540–7548. 10.1523/JNEUROSCI.1921-04.2004 [DOI] [PMC free article] [PubMed] [Google Scholar]
  39. Kolling N., Wittmann M. K., Behrens T. E., Boorman E. D., Mars R. B., Rushworth M. F. (2016). Value, search, persistence and model updating in anterior cingulate cortex. Nat. Neurosci. 19, 1280–1285. 10.1038/nn.4382 [DOI] [PMC free article] [PubMed] [Google Scholar]
  40. La Camera G., Richmond B. J. (2008). Modeling the violation of reward maximization and invariance in reinforcement schedules. PLoS Comput. Biol. 4:e1000131. 10.1371/journal.pcbi.1000131 [DOI] [PMC free article] [PubMed] [Google Scholar]
  41. Lebedev M. A., Messinger A., Kralik J. D., Wise S. P. (2004). Representation of attended versus remembered locations in prefrontal cortex. PLoS Biol. 2:e365. 10.1371/journal.pbio.0020365 [DOI] [PMC free article] [PubMed] [Google Scholar]
  42. Lee D., Rushworth M. F., Walton M. E., Watanabe M., Sakagami M. (2007). Functional specialization of the primate frontal cortex during decision making. J. Neurosci. 27, 8170–8173. 10.1523/JNEUROSCI.1561-07.2007 [DOI] [PMC free article] [PubMed] [Google Scholar]
  43. Lerchner A., Minamimoto T., Soucy D. P., Saunders R., Richmond B. J. (2007). Rule-based categorization and extra-dimensional set shifting without lateral prefrontal cortex in monkeys. Program No 638-8 2007 Neuroscience Meeting Planner. San Diego, CA: Society for Neuroscience. [Google Scholar]
  44. Liu Z., Richmond B. J., Murray E. A., Saunders R. C., Steenrod S., Stubblefield B. K., et al. (2004). DNA targeting of rhinal cortex D2 receptor protein reversibly blocks learning of cues that predict reward. Proc. Natl. Acad. Sci. U.S.A. 101, 12336–12341. 10.1073/pnas.0403639101 [DOI] [PMC free article] [PubMed] [Google Scholar]
  45. Luppino G., Rozzi S., Calzavara R., Matelli M. (2003). Prefrontal and agranular cingulate projections to the dorsal premotor areas F2 and F7 in the macaque monkey. Eur. J. Neurosci. 17, 559–578. 10.1046/j.1460-9568.2003.02476.x [DOI] [PubMed] [Google Scholar]
  46. Mansouri F. A., Matsumoto K., Tanaka K. (2006). Prefrontal cell activities related to monkeys' success and failure in adapting to rule changes in a Wisconsin Card Sorting Test analog. J. Neurosci. 26, 2745–2756. 10.1523/JNEUROSCI.5238-05.2006 [DOI] [PMC free article] [PubMed] [Google Scholar]
  47. Mante V., Sussillo D., Shenoy K. V., Newsome W. T. (2013). Context-dependent computation by recurrent dynamics in prefrontal cortex. Nature 503, 78–84. 10.1038/nature12742 [DOI] [PMC free article] [PubMed] [Google Scholar]
  48. McDannald M. A., Lucantonio F., Burke K. A., Niv Y., Schoenbaum G. (2011). Ventral striatum and orbitofrontal cortex are both required for model-based, but not model-free, reinforcement learning. J. Neurosci. 31, 2700–2705. 10.1523/JNEUROSCI.5499-10.2011 [DOI] [PMC free article] [PubMed] [Google Scholar]
  49. Miller E. K., Cohen J. D. (2001). An integrative theory of prefrontal cortex function. Annu. Rev. Neurosci. 24, 167–202. 10.1146/annurev.neuro.24.1.167 [DOI] [PubMed] [Google Scholar]
  50. Miller E. K., Erickson C. A., Desimone R. (1996). Neural mechanisms of visual working memory in prefrontal cortex of the macaque. J. Neurosci. 16, 5154–5167. [DOI] [PMC free article] [PubMed] [Google Scholar]
  51. Miller E. K., Li L., Desimone R. (1993). Activity of neurons in anterior inferior temporal cortex during a short-term memory task. J. Neurosci. 13, 1460–1478. [DOI] [PMC free article] [PubMed] [Google Scholar]
  52. Minamimoto T., La Camera G., Richmond B. J. (2009). Measuring and modeling the interaction among reward size, delay to reward, and satiation level on motivation in monkeys. J. Neurophysiol. 101, 437–447. 10.1152/jn.90959.2008 [DOI] [PMC free article] [PubMed] [Google Scholar]
  53. Minamimoto T., Saunders R. C., Richmond B. J. (2010). Monkeys quickly learn and generalize visual categories without lateral prefrontal cortex. Neuron 66, 501–507. 10.1016/j.neuron.2010.04.010 [DOI] [PubMed] [Google Scholar]
  54. Mishkin M., Manning F. J. (1978). Non-spatial memory after selective prefrontal lesions in monkeys. Brain Res. 143, 313–323. 10.1016/0006-8993(78)90571-1 [DOI] [PubMed] [Google Scholar]
  55. Miyashita Y. (1988). Neuronal correlate of visual associative long-term memory in the primate temporal cortex. Nature 335, 817–820. 10.1038/335817a0 [DOI] [PubMed] [Google Scholar]
  56. Miyashita Y., Chang H. S. (1988). Neuronal correlate of pictorial short-term memory in the primate temporal cortex. Nature 331, 68–70. 10.1038/331068a0 [DOI] [PubMed] [Google Scholar]
  57. Moore T. L., Schettler S. P., Killiany R. J., Rosene D. L., Moss M. B. (2009). Effects on executive function following damage to the prefrontal cortex in the rhesus monkey (Macaca mulatta). Behav. Neurosci. 123, 231–241. 10.1037/a0014723 [DOI] [PubMed] [Google Scholar]
  58. Moore T. L., Schettler S. P., Killiany R. J., Rosene D. L., Moss M. B. (2012). Impairment in delayed nonmatching to sample following lesions of dorsal prefrontal cortex. Behav. Neurosci. 126, 772–780. 10.1037/a0030493 [DOI] [PMC free article] [PubMed] [Google Scholar]
  59. Murray E. A., O'Doherty J. P., Schoenbaum G. (2007). What we know and do not know about the functions of the orbitofrontal cortex after 20 years of cross-species studies. J. Neurosci. 27, 8166–8169. 10.1523/JNEUROSCI.1556-07.2007 [DOI] [PMC free article] [PubMed] [Google Scholar]
  60. Padoa-Schioppa C., Assad J. A. (2006). Neurons in the orbitofrontal cortex encode economic value. Nature 441, 223–226. 10.1038/nature04676 [DOI] [PMC free article] [PubMed] [Google Scholar]
  61. Padoa-Schioppa C., Assad J. A. (2008). The representation of economic value in the orbitofrontal cortex is invariant for changes of menu. Nat. Neurosci. 11, 95–102. 10.1038/nn2020 [DOI] [PMC free article] [PubMed] [Google Scholar]
  62. Passingham R. (1975). Delayed matching after selective prefrontal lesions in monkeys (Macaca mulatta). Brain Res. 92, 89–102. 10.1016/0006-8993(75)90529-6 [DOI] [PubMed] [Google Scholar]
  63. Petrides M. (2000a). Dissociable roles of mid-dorsolateral prefrontal and anterior inferotemporal cortex in visual working memory. J. Neurosci. 20, 7496–7503. [DOI] [PMC free article] [PubMed] [Google Scholar]
  64. Petrides M. (2000b). The role of the mid-dorsolateral prefrontal cortex in working memory. Exp. Brain Res. 133, 44–54. 10.1007/s002210000399 [DOI] [PubMed] [Google Scholar]
  65. Petrides M., Pandya D. N. (1994). Comparative cytoarchitectonic analysis of the human and macaque frontal cortex, in Handbook of Neurophysiology eds Boller F., Grafman J. (Amsterdam: Elsevier; ), 17–58. [Google Scholar]
  66. Pigarev I., Chelvanayagam D. K., Cappello J., Vidyasagar T. R. (2001). Primary visual cortex and memory. Retinal position specificity and lack of size constancy at early stages of learning a visual memory task in the macaque. Exp. Brain Res. 140, 311–317. 10.1007/s002210100825 [DOI] [PubMed] [Google Scholar]
  67. Procyk E., Goldman-Rakic P. S. (2006). Modulation of dorsolateral prefrontal delay activity during self-organized behavior. J. Neurosci. 26, 11313–11323. 10.1523/JNEUROSCI.2157-06.2006 [DOI] [PMC free article] [PubMed] [Google Scholar]
  68. Procyk E., Tanaka Y. L., Joseph J. P. (2000). Anterior cingulate activity during routine and non-routine sequential behaviors in macaques. Nat. Neurosci. 3, 502–508. 10.1038/74880 [DOI] [PubMed] [Google Scholar]
  69. Procyk E., Wilson C. R., Stoll F. M., Faraut M. C., Petrides M., Amiez C. (2016). Midcingulate motor map and feedback detection: converging data from humans and monkeys. Cereb. Cortex 26, 467–476. 10.1093/cercor/bhu213 [DOI] [PMC free article] [PubMed] [Google Scholar]
  70. R Development Core Team (2008). R: A Language and Environment for Statistical Computing. 2.7.0 Mac OS X Edn. R Foundation for Statistical Computing. [Google Scholar]
  71. Rigotti M., Barak O., Warden M. R., Wang X. J., Daw N. D., Miller E. K., et al. (2013). The importance of mixed selectivity in complex cognitive tasks. Nature 497, 585–590. 10.1038/nature12160 [DOI] [PMC free article] [PubMed] [Google Scholar]
  72. Roesch M. R., Olson C. R. (2004). Neuronal activity related to reward value and motivation in primate frontal cortex. Science 304, 307–310. 10.1126/science.1093223 [DOI] [PubMed] [Google Scholar]
  73. Rolls E. T. (2000). The orbitofrontal cortex and reward. Cereb. Cortex 10, 284–294. 10.1093/cercor/10.3.284 [DOI] [PubMed] [Google Scholar]
  74. Rolls E. T., Critchley H. D., Mason R., Wakeman E. A. (1996). Orbitofrontal cortex neurons: role in olfactory and visual association learning. J. Neurophysiol. 75, 1970–1981. 10.1152/jn.1996.75.5.1970 [DOI] [PubMed] [Google Scholar]
  75. Rudebeck P. H., Murray E. A. (2008). Amygdala and orbitofrontal cortex lesions differentially influence choices during object reversal learning. J. Neurosci. 28, 8338–8343. 10.1523/JNEUROSCI.2272-08.2008 [DOI] [PMC free article] [PubMed] [Google Scholar]
  76. Rushworth M. F., Behrens T. E., Rudebeck P. H., Walton M. E. (2007). Contrasting roles for cingulate and orbitofrontal cortex in decisions and social behaviour. Trends Cogn. Sci. 11, 168–176. 10.1016/j.tics.2007.01.004 [DOI] [PubMed] [Google Scholar]
  77. Rushworth M. F., Nixon P. D., Eacott M. J., Passingham R. E. (1997). Ventral prefrontal cortex is not essential for working memory. J. Neurosci. 17, 4829–4838. [DOI] [PMC free article] [PubMed] [Google Scholar]
  78. Salzman C. D., Paton J. J., Belova M. A., Morrison S. E. (2007). Flexible neural representations of value in the primate brain. Ann. N. Y. Acad. Sci. 1121, 336–354. 10.1196/annals.1401.034 [DOI] [PMC free article] [PubMed] [Google Scholar]
  79. Schultz W., Tremblay L., Hollerman J. R. (2000). Reward processing in primate orbitofrontal cortex and basal ganglia. Cereb. Cortex 10, 272–284. 10.1093/cercor/10.3.272 [DOI] [PubMed] [Google Scholar]
  80. Shenhav A., Botvinick M. M., Cohen J. D. (2013). The expected value of control: an integrative theory of anterior cingulate cortex function. Neuron 79, 217–240. 10.1016/j.neuron.2013.07.007 [DOI] [PMC free article] [PubMed] [Google Scholar]
  81. Shima K., Isoda M., Mushiake H., Tanji J. (2007). Categorization of behavioural sequences in the prefrontal cortex. Nature 445, 315–318. 10.1038/nature05470 [DOI] [PubMed] [Google Scholar]
  82. Shima K., Tanji J. (1998). Role for cingulate motor area cells in voluntary movement selection based on reward. Science 282, 1335–1338. 10.1126/science.282.5392.1335 [DOI] [PubMed] [Google Scholar]
  83. Shindy W. W., Posley K. A., Fuster J. M. (1994). Reversible deficit in haptic delay tasks from cooling prefrontal cortex. Cereb. Cortex 4, 443–450. 10.1093/cercor/4.4.443 [DOI] [PubMed] [Google Scholar]
  84. Simmons J. M., Minamimoto T., Murray E. A., Richmond B. J. (2010). Selective ablations reveal that orbital and lateral prefrontal cortex play different roles in estimating predicted reward value. J. Neurosci. 30, 15878–15887. 10.1523/JNEUROSCI.1802-10.2010 [DOI] [PMC free article] [PubMed] [Google Scholar]
  85. Simmons J. M., Richmond B. J. (2008). Dynamic changes in representations of preceding and upcoming reward in monkey orbitofrontal cortex. Cereb. Cortex 18, 93–103. 10.1093/cercor/bhm034 [DOI] [PubMed] [Google Scholar]
  86. Tanji J., Hoshi E. (2008). Role of the lateral prefrontal cortex in executive behavioral control. Physiol. Rev. 88, 37–57. 10.1152/physrev.00014.2007 [DOI] [PubMed] [Google Scholar]
  87. Tanji J., Shima K., Mushiake H. (2007). Concept-based behavioral planning and the lateral prefrontal cortex. Trends Cogn. Sci. 11, 528–534. 10.1016/j.tics.2007.09.007 [DOI] [PubMed] [Google Scholar]
  88. Tremblay L., Schultz W. (1999). Relative reward preference in primate orbitofrontal cortex. Nature 398, 704–708. 10.1038/19525 [DOI] [PubMed] [Google Scholar]
  89. Tremblay L., Schultz W. (2000). Reward-related neuronal activity during go-nogo task performance in primate orbitofrontal cortex. J. Neurophysiol. 83, 1864–1876. 10.1152/jn.2000.83.4.1864 [DOI] [PubMed] [Google Scholar]
  90. Walker A. E. (1940). A cytoarchitectural study of the prefrontal area of the macaque monkey. J. Comp. Neurol. 73, 59–86. 10.1002/cne.900730106 [DOI] [Google Scholar]
  91. Wallis J. D., Anderson K. C., Miller E. K. (2001). Single neurons in prefrontal cortex encode abstract rules. Nature 411, 953–956. 10.1038/35082081 [DOI] [PubMed] [Google Scholar]
  92. Wallis J. D., Miller E. K. (2003). Neuronal activity in primate dorsolateral and orbital prefrontal cortex during performance of a reward preference task. Eur. J. Neurosci. 18, 2069–2081. 10.1046/j.1460-9568.2003.02922.x [DOI] [PubMed] [Google Scholar]
  93. Walton M. E., Behrens T. E., Buckley M. J., Rudebeck P. H., Rushworth M. F. (2010). Separable learning systems in the macaque brain and the role of orbitofrontal cortex in contingent learning. Neuron 65, 927–939. 10.1016/j.neuron.2010.02.027 [DOI] [PMC free article] [PubMed] [Google Scholar]
  94. Wang Y., Shima K., Sawamura H., Tanji J. (2001). Spatial distribution of cingulate cells projecting to the primary, supplementary, and pre-supplementary motor areas: a retrograde multiple labeling study in the macaque monkey. Neurosci. Res. 39, 39–49. 10.1016/S0168-0102(00)00198-X [DOI] [PubMed] [Google Scholar]
  95. Watanabe M., Sakagami M. (2007). Integration of cognitive and motivational context information in the primate prefrontal cortex. Cereb Cortex 17(Suppl. 1), i101–i109. 10.1093/cercor/bhm067 [DOI] [PubMed] [Google Scholar]
  96. Wise S. P. (2008). Forward frontal fields: phylogeny and fundamental function. Trends Neurosci. 31, 599–608. 10.1016/j.tins.2008.08.008 [DOI] [PMC free article] [PubMed] [Google Scholar]
  97. Wise S. P., Murray E. A., Gerfen C. R. (1996). The frontal cortex-basal ganglia system in primates. Crit. Rev. Neurobiol. 10, 317–356. 10.1615/CritRevNeurobiol.v10.i3-4.30 [DOI] [PubMed] [Google Scholar]
  98. Wood J. N., Grafman J. (2003). Human prefrontal cortex: processing and representational perspectives. Nat. Rev. Neurosci. 4, 139–147. 10.1038/nrn1033 [DOI] [PubMed] [Google Scholar]
  99. Yakovlev V., Fusi S., Berman E., Zohary E. (1998). Inter-trial neuronal activity in inferior temporal cortex: a putative vehicle to generate long-term visual associations. Nat. Neurosci. 1, 310–317. 10.1038/1131 [DOI] [PubMed] [Google Scholar]

Articles from Frontiers in Neuroscience are provided here courtesy of Frontiers Media SA

RESOURCES