Skip to main content
Elsevier Sponsored Documents logoLink to Elsevier Sponsored Documents
. 2015 Sep;56:1–14. doi: 10.1016/j.neubiorev.2015.06.015

The rat's not for turning: Dissociating the psychological components of cognitive inflexibility

Simon RO Nilsson a,b,, Johan Alsiö a,b,e, Elizabeth M Somerville c, Peter G Clifton d
PMCID: PMC4726702  PMID: 26112128

Highlights

  • Non-rewarded or irrelevant prior associations are important for flexible responding.

  • Associations of reward and non-reward in reversal learning are neurally dissociable.

  • Disruption of prior irrelevant or rewarded associations cause pathological deficits.

  • Experimental paradigms of cognitive flexibility can be improved to aid translation.

Keywords: Reversal learning, Attentional set shifting, Discrimination learning, Cognitive flexibility, Animal models

Abstract

Executive function is commonly assessed by assays of cognitive flexibility such as reversal learning and attentional set-shifting. Disrupted performance in these assays, apparent in many neuropsychiatric disorders, is frequently interpreted as inability to overcome prior associations with reward.

However, non-rewarded or irrelevant associations may be of considerable importance in both discrimination learning and cognitive flexibility. Non-rewarded associations can have greater influence on choice behaviour than rewarded associations in discrimination learning. Pathology-related deficits in cognitive flexibility can produce selective disruptions to both the processing of irrelevant associations and associations with reward. Genetic and pharmacological animal models demonstrate that modulation of reversal learning may result from alterations in either rewarded or non-rewarded associations.

Successful performance in assays of cognitive flexibility can therefore depend on a combination of rewarded, non-rewarded, and irrelevant associations derived from previous learning, accounting for some inconsistencies observed in the literature. Taking this combination into account may increase the validity of animal models and may also reveal pathology-specific differences in problem solving and executive function.

1. Introduction

Successful goal-directed behaviour requires that discrimination learning builds on stimulus-reward associations that are flexible in response to changing situational demands. This capacity is often referred to as ‘cognitive flexibility’. Disruptions in cognitive flexibility are a common feature of neuropsychiatric disorders that are not alleviated by available drugs (Barch, 2005). The extent of cognitive disruption correlates with long-term health outcome (Green, 2006) and the inability to treat these deficits can act as a rate-limiter of treatment progression, trapping patients within life-long dependency (Geyer and Tamminga, 2004).

Impairments in tasks of cognitive flexibility have been associated with diverse pathologies such as Parkinson's disease, autism, Alzheimer's disease, depression, Huntington's disease, ADHD, and schizophrenia. Table 1 serves to highlight both the broad range of disorders linked to cognitive inflexibility and the failure of extant tasks and paradigms to distinguish between these pathologies: patients with widely disparate diagnoses and symptom profiles often display ‘similar’ deficits. For some neuropsychiatric and neurological disorders, this common behavioural marker could represent similar neurobiological dysfunctions produced by variable aetiologies. Neuropsychological markers of disorders that unite clinical diagnoses (endophenotypes) have previously been reported and have received increased focus through the National Institute of Mental Health's Research Domain Criteria Project (Insel et al., 2010); examples include enhanced ‘model-free’ habit learning in obsessive–compulsive disorder, drug addiction and binge eating disorder (Worbe et al., 2015), or increased impulsivity in drug addiction and attention-deficit hyperactivity disorder (Dalley et al., 2011). However, it is also conceivable that currently prevalent reversal and set-shifting paradigms lack sensitivity to detect dissociable impairments manifested as cognitive inflexibility. A more detailed understanding of the cognitive mechanisms that underpin performance on such tests may thus lead to more sensitive assays that reveal pathology-specific differences and, in turn, to rational research strategies that permit the development of more effective therapeutic interventions. Such knowledge could furthermore promote the identification of endophenotypes where they exist and inform both neuropsychiatric research and investigation into the genetic basis of reversal learning and attentional set shifting (Ouden et al., 2013).

Table 1.

Observations of cognitive inflexibility in psychiatric disorders in reversal learning and attentional set-shifting tasks.

Underlying condition Deficit observed in Reference
Parkinson's disease Spatial reversal learning Freedman and Oscar-Berman (1989)
CANTAB ED perseverance/learned irrelevance probes Owen et al. (1993)
CANTAB ED learned irrelevance probe Slabosz et al. (2006)
Probabilistic visual reversal learning Cools et al. (2001)
CANTAB set-shifting and reversal learning Downes et al. (1989)
Alzheimer's disease Spatial reversal learning Freedman and Oscar-Berman (1989)
Object reversal learning Freedman and Oscar-Berman (1989)
OCD Probabilistic visual reversal learning Remijnse et al. (2006)
CANTAB set-shifting Chamberlain et al. (2006)
CANTAB set-shifting and Go/No-go reversal learning Watkins et al. (2005)
Schizophrenia CANTAB ED perseverance probe Elliott et al., 1995, Elliott et al., 1995
CANTAB set-shifting and reversal learning Ceaser et al. (2008), Jazbec et al. (2007), Leeson et al. (2009), Murray et al. (2008), Pantelis et al. (1999)
Autism Spatial reversal learning Coldren and Halloran (2003)
Probabilistic reversal learning D’Cruz et al. (2013)
CANTAB set-shifting and reversal learning Hughes et al. (1994), Ozonoff et al. (2004)
CANTAB set-shifting Ozonoff et al. (2000)
Unipolar depression Probabilistic visual reversal learning Reischies (1999)
WCST Martínez-Arán et al. (2004), Merriam et al. (1999)
CANTAB set-shifting and reversal learning Taylor Tavares et al. (2007)
Bipolar depression CANTAB set-shifting and reversal learning Clark et al. (2001), McKirdy et al. (2009)
Probabilistic visual reversal learning Gorrindo et al. (2005)
Go/No-go reversal learning Holmes et al. (2008), Murphy et al. (1999)
Attentional set-shifting Clark et al. (2005)
CANTAB set-shifting Clark et al. (2002)
CANTAB reversal learning Dickstein et al. (2004)
WCST Martínez-Arán et al. (2004)
Huntington's disease CANTAB set-shifting and reversal learning Lange et al. (1995), Lawrence et al. (1996)
CANTAB ED perseverance probe Lawrence et al. (1999)
Object reversal learning Oscar-Berman and Zola-Morgan (1980)
CANTAB set-shifting Lawrence et al. (1998)
ADHD Go/No-go reversal learning Itami and Uno (2002)
WCST Reeve and Schandler (2001)
CANTAB set-shifting and reversal Gau and Shang (2010), Kempton et al. (1999)
Cocaine abuse Probabilistic reversal learning Ersche et al. (2008)

Note. This table is not meant as an exhaustive list allowing comparisons between brain functioning and cognitive flexibility. Nor is it meant to stress the importance of cognitive flexibility impairments by ubiquity. Rather, the table draws attention to the non-selectivity of prevalent tasks of cognitive flexibility in discriminating between patient populations.

Here, commonly used assays of cognitive flexibility are reviewed with a particular focus on reversal learning. We define the underlying components of cognitive flexibility, and stress that learning theory can interpret reversal learning as concurrent schedules that may reveal both perseverance and learned non-reward. We review experiments designed to separately assess these mechanisms, and discuss their consequences for interpreting reversal learning performance. Although clinical and preclinical assays have begun to define the anatomical loci and neurochemistry involved in cognitive flexibility, the protocols used to assess flexible responding are varied and this may account for some inconsistency in the literature. We outline some potential benefits of resolving reversal learning into tests of its constituent cognitive components, including the possibility of obtaining clearer indications of pathology-specific differences in problem solving and increased validity of their respective animal models.

2. Measuring cognitive flexibility

Cognitive flexibility is most commonly assessed in reversal learning and attentional set-shifting tasks. Reversal learning can often involve a single perceptual dimension typically containing two different conditioned stimuli (CS's). Responses to one stimulus are rewarded (CS+) while responses to the second stimulus are non-rewarded (CS−). After learning an initial CS+ versus CS− discrimination, the contingencies are reversed. Attentional set-shifting, in contrast, involves at least two superimposed perceptual dimensions, each containing at least two different stimuli. In an initial acquisition phase of attentional set-shifting, two stimuli within one perceptual dimension serve as CS+ and CS− while stimuli within other dimensions are irrelevant. In the following set-shifting phase, the previous CS+ and CS− become irrelevant while stimuli within a previously irrelevant dimension become the relevant CS+ and CS−.

Reversal learning and attentional set-shifting have been incorporated into the intradimensional/extradimensional task (ID/ED-task) of the Cambridge Neuropsychological Test Automated Battery (CANTAB) which has been used to assess primates and human participants (Dias et al., 1996, Leeson et al., 2009). This task consists of a series of components of increasing level of difficulty. Subjects initially learn a two-choice discrimination where responses to one stimulus are perfectly correlated with reward and responses to the second stimulus are perfectly correlated with the absence of reward. This is followed by a compound discrimination, where a second superimposed sensory or cognitive dimension is introduced but the correct and incorrect stimuli remain constant. Next, subjects are faced with an ID-shift where learned stimuli are replaced by novel stimuli, again with the relevant and irrelevant dimensions remaining constant. Once this discrimination has been acquired, subjects are challenged by an ED-shift, where the irrelevant dimension becomes relevant and the relevant dimension becomes irrelevant. Each of these tests is typically followed by a reversal learning test. There are also analogous bowl-digging (McAlonan and Brown, 2003) and operant (Scheggia et al., 2013) versions of this task used with rodents.

Although superficially similar, attentional set-shifting is more cognitively challenging than reversal learning. This was initially attributed to the partial positive reinforcement of a previously rewarded stimulus after the attentional set shift (Buss, 1953); but this hypothesis was falsified by studies demonstrating that attentional set-shifting is more difficult than reversal learning even after controlling for partial reinforcement (Buss, 1956, Harrow and Friedman, 1958, Kendler and D’Amato, 1955). Yet it should be noted that the latter paradigms contain additional confounds, including memory-related constraints, which may severely challenge meaningful comparisons between the reversal learning and attentional set-shifting (Slamecka, 1968). However, the discrepancy in difficulty is most likely produced by the attentional relocation demanded in attentional set-shifting, but not reversal learning, as well as the suppression of associations of non-correlated reinforcement, or learned irrelevance (Bissonette et al., 2008, Dias et al., 1996).

As an assay of cognitive flexibility, reversal learning has been somewhat overshadowed by attentional set-shifting. One reason for this might be the relatively greater difficulty of attentional set-shifting which may increase the likelihood of observing clinically-relevant deficits and their alleviation through treatment. There is nevertheless evidence suggesting that reversal learning may be a more suitable translational model for clinical application. Attentional set-shifting, but not reversal learning, has been linked to verbal ability and intelligence (Laws, 1999). By comparing groups matched for current IQ, Leeson et al. (2009) found that schizophrenic patients do not differ from healthy controls on attentional set-shifting but do show reversal learning impairments in the CANTAB ID/ED-task. Prompting schizophrenic patients to verbalise their decision-making can remediate poor attentional set-shifting performances (Choi and Kurtz, 2009) and when figural stimuli are replaced by verbal stimuli, the performance of schizophrenic patients deteriorates whilst the performance of healthy controls improves (Rossell and David, 1997). Hence, deficits in attentional set-shifting ability may reflect verbal and intelligence deficits in patients while impairments in reversal learning are relatively independent of intelligence (Leeson et al., 2009). Reversal learning is also a better predictor of social functioning (Shamay-Tsoory et al., 2007).

3. Behavioural analyses of discrimination learning

Discrimination learning occurs in response to dissociable reinforcements of stimuli, perceptual dimensions or contexts and the nature of subsequent reversal learning and attentional set-shifting processes largely depend on how these discriminations are acquired. Early views held discrimination learning to be a low-level comparative process where behaviour is controlled by relational features of stimuli and rules are acquired through trial-and-error processes. However, this account had difficulty with phenomena such as stimulus generalisation and reversal following transposition tests (Spence, 1937). Spence suggested that discrimination learning depended on both excitation and inhibition (Spence, 1936). Positive reinforcements increase the excitatory strength of a stimulus and elicit approach, while non-reinforcements decrease the excitatory strength of a stimulus and make approach less likely. While Spence treated non-reinforcement as a non-aversive mechanism inhibiting the excitatory strength and approach tendency associated with a stimulus, others have considered non-reinforcement to result in aversive negative reinforcement (Amsel, 1958, Behar, 1961, Nissen, 1950, Terrace, 1971). However, Spence's conditioned excitatory–inhibition theory predicts that previously non-reinforced stimuli, or stimuli not correlated with reinforcement, will acquire excitatory tendencies at a similar or faster rate than neutral stimuli. The theory therefore fails to explain phenomena such as latent inhibition or learned irrelevance, in which such associations delay learning. The mechanisms of excitation and inhibition are also insufficient to explain phenomena such as the serial-reversal effect (Mackintosh et al., 1968) where later reversals are acquired at a faster rate than earlier reversals, or the overtraining reversal effect, where prolonged discrimination training can facilitate reversal learning (Lovejoy, 1966).

To account for these phenomena, it has been necessary to involve the additional mechanisms of attentional stimulus selection-processes and predictability (Mackintosh, 1983, Pearce and Hall, 1980). Attention is a composite term for processes ensuring appropriate and continued maintenance and selection of stimuli for goal-directed behaviour (Parasuraman, 1998). In the context of discrimination learning and cognitive flexibility, attention is thought of as a determinant of perception allowing stimuli predictive of reinforcement to gain excitatory or inhibitory conditioning while irrelevant information fail to interfere with these processes (Mackintosh, 1965). In attentional theory, it is suggested that subjects attend to those stimuli (Mackintosh, 1975), or stimulus dimensions (Sutherland and Mackintosh, 1971), that are the best predictors of the reward contingencies, and this attentional allocation subsequently drives responding towards the appropriate CS+ and CS−. This approach suggests that although all forms of discrimination learning and cognitive flexibility require attentional allocation, only extradimensional set-shifting require attentional relocation and will therefore be acquired at a slower rate. Overtraining reversal effects (Mackintosh, 1969) and serial reversal effects (Mackintosh et al., 1968) may also be explained by extensive training increasing attentional allocation to particular discriminative features.

In sum, although theoretical approaches to animal learning differ in the value they place on attentional factors and reinforcement contingencies throughout the process of discrimination acquisition, they explain differential responding to distinctive stimuli in terms of independent mechanisms of excitation versus inhibition and consequent changes in attention to the relevant stimuli.

4. Dissociating the components of cognitive flexibility

The implication of this analysis is that pathology-related cognitive inflexibility, revealed by disturbances in reversal learning or attentional set shifting, may be due to abnormalities in one or more separate processes. In reversal learning, the initial two-choice discrimination can be resolved to an excitatory CS – US association, eliciting approach and contact, and an inhibitory CS – ‘no US’ association, eliciting avoidance. After the subsequent contingency shift, the CS predicting the US becomes associated with ‘no US’, a process opposed by perseverance. Conversely, the CS initially predicting ‘no US’ now predicts the US, a process opposed by learned non-reward.

Similarly, in attentional set-shifting, the initial discrimination is likely to depend on conditioned attention towards the relevant dimension and conditioned inattention towards the irrelevant dimension. After the subsequent contingency shift, the relevant dimension becomes irrelevant, a process opposed by perseverance. Conversely, the irrelevant dimension becomes relevant, a process opposed by learned irrelevance. Deficits in reversal learning and attentional set-shifting can therefore be interpreted as failure to dissipate prior associations of previous positive (perseverance) and negative (learned non-reward or learned irrelevance) outcome together with any associated attentional effects.

4.1. Perseverance

‘Perseveration’ describes a range of phenomena related to the inappropriate repetition or maintenance of an activity or an abstract rule (Crider, 1997). This includes, for example, various forms of catatonia-like repetitions of motor-outputs (Freeman and Gathercole, 1966, Helmick and Berg, 1976, Luria, 1965, Sandson and Albert, 1984). In cognitive terms perseveration is most often used to describe an executive functioning deficit causing the repetition or maintenance of abstract information encoding for relationships between stimuli and goals (Garner, 2006). Perseveration could potentially also be affected by episodic retrieval mechanisms of task rules (Frank and Fossella, 2010) and motor impulsivity (Izquierdo and Jentsch, 2012). However, there is little to suggest that these mechanisms would have separable effects on discrimination learning, perseveration and learned non-reward or learned irrelevance in assays of cognitive flexibility. It is notable that the converse of cognitive flexibility – habitual responding – is ubiquitous in animal behaviour and has always been interpreted as critical for allowing efficient and rapid responding to environmental challenges as well as a way of freeing cognitive resource (James, 1890). However, excessive perseverance has also long been recognised as a component of psychopathology (Bleuler, 1950).

Perseveration need not be delimited by the outcome of the association that is inappropriately repeated or maintained. Thus excessive avoidance of a previous CS− and excessive approach of a previous CS+ are both potentially examples of perseveration. Yet discussions of perseveration are typically phrased in terms of an inability to overcome reinforced, rather than irrelevant or non-reinforced associations. The majority of reported manipulations affecting cognitive flexibility are also interpreted as due to effects involving reinforced associations or perseverance (Abdul-Monim et al., 2007, Boulougouris et al., 2008, Floresco et al., 2006).

4.2. Learned non-reward

Learned non-reward is the consequence of the CS—‘no US’ association formed in a two-choice discrimination paradigm. After a contingency shift, learned non-reward is expressed as interference with learning of a new CS–US association after prior experience of non-reward with that CS. Recent use of the term originates from a study describing the role of the orbitofrontal cortex (OFC) in reversal learning (Tait and Brown, 2007) although similar descriptors were used earlier in this context (e.g., learned avoidance of non-reward) (Behar, 1961). It is closely related to the phenomenon of latent inhibition. The difference between latent inhibition and learned non-reward is most marked in simultaneous discrimination and reversal paradigms where the CS+ and CS− are presented concurrently. In successive reversal tasks, such as go/no-go paradigms, the two stimuli are presented independently. In these tasks the presentation of a previously non-rewarded but now rewarded CS can therefore be nearly identical to an appetitive assay of latent inhibition.

The inability to overcome a non-rewarded association in two-choice reversal learning has also been referred to as learned avoidance (Clarke et al., 2007) and learned irrelevance (Boulougouris et al., 2008). However, none of these terms accurately capture the phenomena occurring in appetitive reversal learning (Table 2). Learned non-reward is produced when a stimulus now positively correlated with reward has been previously negatively correlated with reward. Learned avoidance is produced when a now neutral stimulus has been previously correlated with aversive reinforcement. Learned irrelevance is produced when a stimulus now positively correlated with reward has been previously non-correlated with reward.

Table 2.

Stimuli-reinforcement correlations in two-stage discrimination and reversal paradigms.

Conditioning phase
Test phase
Phenomenon Stimulus A Stimulus B Stimulus A Stimulus B
Reversal learning +1.0 −1.0 −1.0 +1.0
 Learned non-reward −1.0 +1.0
Attentional set-shifting +1.0 +0.5 +0.5 +1.0
 Learned irrelevance +0.5 +1.0
Learned avoidance 1.0 0 0 0
Latent inhibition 0 None 1.0 None

In learned non-reward, a stimulus initially negatively correlated with reinforcement becomes positively correlated with reinforcement. In learned irrelevance, a stimulus initially non-correlated with reinforcement becomes correlated with reinforcement. In a typical learned avoidance tasks, a stimulus initially correlated with reinforcement becomes neutral. In an appetitive two-stage latent inhibition, an initially neutral stimulus becomes correlated with reinforcement.

Aversive.

4.3. Errors as measures of performance

The most common approach for assessing perseverative responding in reversal learning is to classify incorrect responses into one of two categories; either as an ‘early-error’ or ‘perseverative error’, or as a ‘late-error’ or ‘regressive error’. For example, in operant tasks, early or perseverative errors have been indexed by repetitive bouts of responding towards the previous CS+ (Boulougouris and Robbins, 2010) or incorrect responses made early in reversal when the majority of responses are to the previous CS+ (Boulougouris et al., 2008). In maze tasks incorrect responses made before making a first response to the new CS+ (McCool et al., 2008), or three such responses in a block of four trials (Dias and Aggleton, 2000, Palencia and Ragozzino, 2004), have been used, depending on the experimental protocol. Other incorrect responses are then classified as late errors.

It is often assumed that the number of early errors reflects the stability of the original CS–US association, or perseverance, while late errors are considered to reflect general cognitive abilities related to attention and the acquisition of an alternative CS–US association. However, in experiments in which early errors and late errors are analysed, previously correct and incorrect CSs are presented during all reversal trials. Previous excitatory and inhibitory conditioning can therefore influence choice-behaviour in both early and late phases of learning. It may also be that the presence of a CS− invigorates the rate of responding to a CS+ (Mackintosh, 1974, Terrace, 1963). Analyses of early errors and late errors may nevertheless be valuable in other ways. Early and late errors may potentially reflect failures in memory processing. For example, Palencia and Ragozzino (2004) discuss, and then reject, the possibility that an increase in regressive, late errors resulting from blockade of dorsomedial striatal NMDA receptors in a place reversal task might have arisen from a disruption of working memory that impaired the ability to recall earlier non-reinforced trials. More generally, similar patterns of error patterns during early and late phases of learning in two different reversal tasks may indicate that the two tasks are solved using similar approaches or that they depend on similar underlying neural mechanisms (Boulougouris et al., 2007, Chudasama and Robbins, 2003a).

4.4. Preclinical assays of perseverance and learned non-reward

Ever since learning theorists stressed that discrimination learning is a two-process phenomenon (Amsel, 1958, Behar, 1961, Nissen, 1950, Spence, 1936), experimental efforts have been made to determine the relative contribution of reinforcement and non-reinforcement in discrimination and reversal learning. The paradigms used in these studies include measures of reversal learning as well as perseverance and learned non-reward and therefore provide superior ‘resolution’ relative to standard discrimination and reversal protocols. Yet results have varied across protocols and species, with data indicating that both reinforced and non-reinforced contingencies can guide responding.

One approach has been to train subjects on the contingency of a single stimulus prior to either a discrimination or reversal challenge (Table 3). CS+ pre-exposure should facilitate performance relative to CS− pre-exposure if responses during reversal learning or discrimination learning are primarily guided by reward. Conversely, CS− pre-exposure should facilitate performance relative to CS+ pre-exposure if responses during reversal learning primarily are guided by non-reward. In these protocols, performances have indeed been found to be best if the subject has been pre-exposed to the CS− prior to discrimination learning (Blomquist et al., 1973, Moss and Harlow, 1947) or been pre-exposed to non-rewarded responses to the previous CS+ prior to reversal learning (Cross and Brown, 1965, Sasaki, 1969). This suggests that associations with non-reward have a greater role in discrimination and reversal learning than associations with reward. Nevertheless, it has been suggested that the superior performance from CS− pre-exposure in studies of discrimination learning may be due to an attraction to novelty (Grabbe and Campione, 1969). Others have also found that pre-reversal CS+ learning has a bigger effect on reversal learning (Vaughter and Cross, 1965). These inconsistencies are likely to be related to the use of different species, different stimuli dimensions, and the number of single-stimulus pre-exposure trials.

Table 3.

Example of tests assessing the role of perseverance and learned non-reward in reversal learning through CS+ or CS− pre-exposure.

Test CS+ CS−
Discrimination learning graphic file with name fx1.gif graphic file with name fx2.gif
Perseverance test
Single-stimulus pre-exposure graphic file with name fx1.gif
Test-phase graphic file with name fx2.gif graphic file with name fx1.gif
Learned non-reward test
Single-stimulus pre-exposure graphic file with name fx2.gif
Test-phase graphic file with name fx2.gif graphic file with name fx1.gif

To test the relative influence of perseverance or learned non-reward in reversal learning, subjects receive forced-choice pre-exposure trials to the reversed contingencies of either the previous CS+ or the previous CS− prior to the two-choice reversal challenge. If response behaviour in reversal learning primarily is guided by perseverance, pre-exposure to the reversed contingency of the previous CS+ should facilitate performance relative to pre-exposure to the reversed contingency of the previous CS−. If response behaviour in reversal learning primarily is guided by learned non-reward, pre-exposure to the reversed contingency of the previous CS− should facilitate performance relative to pre-exposure to the reversed contingency of the previous CS+.

The importance of the CS+ and the CS− in discrimination and reversal learning can also be assessed by replacing or introducing variability in either the CS+ or CS− (Table 4). When the CS+ is replaced or varied, the only reliable predictor of reinforcement across trials is the CS−. When the CS− is replaced or varied, the only reliable predictor of reinforcement across trials is the CS+. Stimulus variability has been introduced during discrimination learning to assess the relative roles of reinforcement and non-reinforcement on discrimination learning, while others have included a contingency shift prior to stimulus replacement to assess the role of perseverance or learned non-reward in reversal learning. Within these paradigms, subjects have often been observed to make more errors with CS− variability than CS+ variability (Goulart et al., 2005, Mullins and Winefield, 1979, Stevens and Fechter, 1968) indicating that that successful performance in discrimination and reversal learning primarily dependent upon avoidance of the CS−. However, reports have not always been consistent, with some protocols showing that CS+ variability can retard performance more than CS− variability (Gardner and Coate, 1965).

Table 4.

Example of tests assessing the role of reinforcement and non-reinforcement in discrimination and reversal learning by varying or replacing the CS+ or CS−.

Stage Test CS+ CS−
Discrimination learning Reinforcement graphic file with name fx3.gif graphic file with name fx4.gif
Non-reinforcement graphic file with name fx5.gif graphic file with name fx6.gif
Reversal learning Initial discrimination learning graphic file with name fx7.gif graphic file with name fx8.gif
Perseverance graphic file with name fx9.gif graphic file with name fx10.gif
Learned non-reward graphic file with name fx8.gif graphic file with name fx2.gif

To assess the role of rewarded associations in discrimination learning, the CS+ is kept constant while the CS− varies across trials. Here, the only stimulus reliably predicting reward across trials is the rewarded stimulus. To assess the role of non-reinforcement in discrimination learning, the CS− is kept constant while the CS+ varies across trials. Here, the only stimulus reliably predicting reward across trials is the CS−. To assess the role of perseverance in reversal learning, the previous CS+ becomes CS−, while a novel CS+ replaces the previous CS. In this test, established non-reinforced association cannot guide responding as the previous CS− has been removed. To assess the role of learned non-reward in reversal learning, the previous CS+ becomes CS−, while a novel CS+ replaces the previous CS−. In this test, established associations of reward cannot guide responding as the previous CS+ has been removed (adapted from Clarke et al., 2007).

Of further note is a 3-choice simultaneous visual reversal task (Jentsch et al., 2002) that has been suggested to give a measure of discernment of stimulus perseveration (Table 5). Here, vervet monkeys were presented with an initial discrimination where one stimulus was designated as CS+ while two other stimuli were designated as CS−’s. Following reversal, the previous CS+ became a CS−, and one of the previous CS−’s became the new CS+. The third stimulus remained non-rewarded during both discrimination and reversal learning. Responses to the previously rewarded but now non-rewarded stimulus were coded as perseverative while responses to the consistently non-rewarded stimulus gave a control for random responses that may be expected during a search for an alternative response strategy, or as result of a more general invigoration of responding. However, in this paradigm, all three stimuli (stimuli that were previously non-rewarded but are now rewarded, stimuli previously rewarded but now non-rewarded, and stimuli remaining non-rewarded) are presented in both the initial discrimination and subsequent reversal trials. This may make interpretations of stimulus avoidance versus approach strategies difficult.

Table 5.

Example of test assessing the role of perseverance using a 3-stimulus simultaneous discrimination and reversal paradigm.

Stage CS+ CS− CS−
Discrimination learning graphic file with name fx1.gif graphic file with name fx9.gif graphic file with name fx8.gif
Reversal learning graphic file with name fx9.gif graphic file with name fx1.gif graphic file with name fx8.gif

In discrimination learning, one stimulus is rewarded while two stimuli are non-rewarded. In the reversal test, the previous CS+ stimulus becomes a CS− and a previous CS− becomes the CS+. The second CS− remains non-rewarded in both discrimination and reversal learning. In this paradigm, responses to the previous CS+ are treated as perseverative while responses to the constant CS− controls for non-perseverative errors that occur when the subject is searching for an alternative response strategy. The positions of the stimuli vary pseudorandomly across trials (Jentsch et al., 2002).

In sum, although interpretations of preclinical data often build on the assumption that reinforced associations guide choice behaviour in reversal paradigms, the above experiments serve to highlight that non-reinforcement also can be of considerable importance.

4.5. Learned irrelevance and perseverance in attentional set-shifting tasks

Learned irrelevance in an attentional set-shifting task is the analogue of learned non-reward. In learned irrelevance, a stimulus dimension initially non-correlated with reinforcement becomes correlated with reinforcement. That is, a stimulus dimension previously rewarded 50% of the time becomes rewarded 100% of the time (Table 2).

Learned irrelevance is more difficult to overcome than latent inhibition and the discrepancy in difficulty between reversal learning and attentional set-shifting may be related to the discrepancy in difficulty between learned non-reward and learned irrelevance (Buss, 1956, Kendler and D’Amato, 1955). Learned irrelevance could be viewed as the product of the simpler mechanisms of pre-exposure to the CS as well as the US, or as representing the combined effect of latent inhibition and US pre-exposure (Rescorla and Holland, 1982). However, later data has favoured learned irrelevance as a non-reducible independent phenomenon (Bennett et al., 1995).

In contrast to the dissociation between learned non-reward and perseverance, the dissociation of learned irrelevance and perseverance in attentional set-shifting has been investigated in clinical populations. The approach has been to modify the CANTAB ID/ED-task. Either the previously relevant dimension is replaced by a novel irrelevant dimension to probe learned irrelevance, or the previously irrelevant dimension is replaced by a novel relevant dimension to probe perseverance (Table 6). Using this method in a non-clinical human group, learned irrelevance has been shown to contribute more than perseverance to the difficulty of attentional acquisition (Maes and Eling, 2009) as well as attentional set-shifting (Maes et al., 2004) paralleling the often more prominent role of learned non-reward in preclinical tests of reversal learning. These studies have also revealed pathology-specific dissociations. Humans with prefrontal damage (Owen et al., 1993), individuals with schizophrenia (Elliott et al., 1995, Elliott et al., 1998) or Huntington's disease (Lawrence et al., 1999) all exhibit perseverative set-shifting deficits, while non-medicated individuals with Parkinson's disease show deficits in both perseverance and learned irrelevance (Owen et al., 1993). However, l-Dopa medication in Parkinson's disease is associated with selective impairments in overcoming learned irrelevance (Owen et al., 1993, Slabosz et al., 2006). This suggests that perseverance, but not learned irrelevance, is related to dopaminergic hypoactivity. In the rodent, this approach to measuring attentional set-shifting has only been reported once, with an assessment of perseverance, but not learned irrelevance, in mice using the bowl-digging task (Garner et al., 2006).

Table 6.

Example of tests separately assessing the role of perseverance and learned irrelevance in attentional set-shifting.

Stage Stimuli Relevant Irrelevant Correct stimulus
IDR graphic file with name fx11.gif graphic file with name fx12.gif Shape Lines graphic file with name fx13.gif
Perseverance test graphic file with name fx14.gif graphic file with name fx15.gif Solidity Shape graphic file with name fx14.gif
Learned irrelevance test graphic file with name fx16.gif graphic file with name fx17.gif Lines Solidity graphic file with name fx18.gif

In a perseverance test of attentional set-shifting, the relevant dimension of the intradimensional reversal stage (IDR) becomes irrelevant while the previously irrelevant dimension is replaced by a novel relevant dimension. In a learned irrelevance test of attentional set-shifting, the irrelevant dimension of the IDR stage becomes relevant while the previously relevant dimension is replaced by a novel irrelevant dimension (adapted from Owen et al., 1993).

4.6. Non-responding in discrimination and reversal tasks

In discrimination and reversal tasks, failures to respond or omissions are often included and discussed as controls for motivational or motoric confounds. Yet in discrimination learning theory, non-responding towards the CS− in a simultaneous discrimination is variously interpreted as a product of the CS− lacking excitatory strength (Spence, 1936), being aversive or acting as a conditioned inhibitor (Amsel, 1958, Behar, 1961, Nissen, 1950, Terrace, 1971), having acquired inattention (Sutherland and Mackintosh, 1971), or resulting from a lack of attention, due to being an accurate predictor of non-reward (Pearce and Hall, 1980). In addition to measuring motivation or motoric factors, non-responding in discrimination and reversal assays could therefore also be seen as a direct consequence of learning, most readily produced by associations with non-reward.

After initial high rates of responding towards the previous CS+ immediately following a contingency shift, the excitatory strength or attentional control of the CS+ decreases and the excitatory strength of both CS+ and CS− should be similarly low. Thus, the previous CS+ becomes associated with non-reward in early reversal learning while the previous CS− remains associated with non-reward from conditioning in discrimination learning and this can cause a high number of non-responses or omissions.

However, some experiments suggest that omissions in reversal learning can also be used as a measure of learning. For example, in serial operant reversal tasks, omissions may show a serial reversal effect and decrease across subsequent reversal tests (Boulougouris et al., 2008, Nilsson et al., 2012) indicating that non-responding is a measure of learning as well as of motivation and motoric factors. Also, in a rat bowl-digging paradigm (Tait and Brown, 2007) and a mouse operant paradigm (Nilsson et al., 2012), omissions are more prominent in a learned non-reward test, where the previous CS− is paired with a novel CS, than in a perseverance test, where the previous CS+ is paired with a novel CS, indicating that omissions in tasks of reversal learning can be specifically related to learning and learned non-reward.

Notably, response omissions are also a primary measure of learning in other tasks assessing cognitive flexibility, including successive reversal learning paradigms (Burke et al., 2009, McEnaney and Butter, 1969, Nonkes et al., 2011, Schoenbaum et al., 2003) and latent inhibition (Lubow, 1989). Thus omissions should be considered relevant to learning in simultaneous discrimination and reversal tasks and not simply discussed as potential controls for motor or motivational side effects of an experimental manipulation.

4.7. Novelty confounds

A novel stimulus or a novel stimulus dimension is present in most studies that feature separate tests of perseverance and learned non-reward or learned irrelevance in simultaneous discrimination procedures. Changes in novelty attraction or avoidance could therefore confound any interpretation regarding the effect of an experimental manipulation on learning. In a perseverance test, a novel rewarded stimulus or dimension is paired with a previously rewarded but now non-rewarded stimulus or dimension. In this test increased novelty attraction could be misinterpreted as facilitated learning while increased novelty avoidance could be misinterpreted as retarded learning. In a learned non-reward or learned irrelevance test, a novel non-rewarded stimulus or dimension is paired with a previously non-rewarded but now rewarded stimulus or dimension. In this test increased novelty attraction would be observed as retarded learning while increased novelty avoidance would be observed as facilitated learning.

One intrinsic control for a manipulation of novelty attraction or avoidance in these tasks is that it would cause opposing effects on measures of learning in the perseverance and learned non-reward or learned irrelevance tests (Clarke et al., 2007). For example, a manipulation-induced or pathology-related increase in novelty attraction would improve performance in a perseverance test, in which the novel CS is correct, but retard performance in a learned non-reward test, in which the novel CS is incorrect.

Performance in a reversal learning test could also control for effects on novelty attraction or novelty recognition as no novel stimuli are presented in this test. If an effect of a manipulation is observed in perseverance and/or learned non-reward tests where novelty is a feature, as well as in a reversal learning test which lacks novelty, a plausible interpretation would be that the effects are related to shared features of the tests and unrelated to test differences in the presentation of a novel stimulus.

A further approach to account for novelty-related effects is to add control tests of discrimination learning where increases in perseverance and learned irrelevance would facilitate learning. In a perseverance control condition, this can be done by replacing the CS− with a novel CS− while the CS+ remains constant. In a learned non-reward control condition, the CS+ is replaced by a novel CS+, while the CS− remains constant (Gauntlett-Gilbert et al., 1999). Notably, these discrimination tests are identical to the tests designed to assess the role of negative and positive associations in discrimination learning described above (Goulart et al., 2005, Mullins and Winefield, 1979, Stevens and Fechter, 1968). Importantly, this experimental configuration still allows for a novelty confound, as it involves a choice between a previously relevant or irrelevant dimension and a novel dimension (Slabosz et al., 2006). This form of novelty control has as yet only been performed in the context of attentional set-shifting (Gauntlett-Gilbert et al., 1999) but could also be useful in reversal learning.

Novelty confounds in tests of perseverance and learned non-reward can also be investigated using independent tests of spontaneous recognition memory (Nilsson et al., 2013). These tests are non-reinforced and novelty preference is studied in the absence of interference from reinforcement learning. As no overtly reinforced learning is involved, previous conditioning does not affect responses to novelty. Such tests consist of a sample-phase and a test-phase. In the sample-phase, the two stimuli featuring in discrimination learning are presented and approaches are measured. In the test-phase, one of the two stimuli is replaced by a novel stimulus and novelty preference is measured by the ratio of approach to the novel stimulus relative to the total approaches to both stimuli.

We recently developed an additional approach to investigate learned non-reward and stimulus-perseveration in a two-choice visual discrimination and reversal learning task for rats, by interleaving the visual discrimination (CS+ vs. CS−) trials with trials in which either the CS+ or the CS− was paired to a third, neutral stimulus (Table 8; Alsiö, Mar, Nilsson, and Robbins, unpublished). Responses to the CS+ are rewarded, responses to the CS− are non-rewarded, and responses to the third stimulus (CS50/50) during the probe trials are rewarded on 50% of the trials. Animals thus learn to choose the CS50/50 over the CS−, and to choose the CS+ when this stimulus is presented with the CS50/50; each of these pairs is presented once every 10 trials, with the remaining 8 trials being standard CS+ vs. CS−.

Table 8.

Assessing stimulus perseverance and learned non-reward in visual touchscreen reversal learning in the rat using interleaved probe trials.

Stimuli
Stimulus A (reward contingency) Stimulus B (reward contingency)
Discrimination learning
 Pair 1 graphic file with name fx19.gif graphic file with name fx20.gif
 Pair 2 graphic file with name fx19.gif graphic file with name fx21.gif
 Pair 3 graphic file with name fx21.gif graphic file with name fx22.gif
Reversal learning
 Full reversal test graphic file with name fx23.gif graphic file with name fx24.gif
 Learned non-reward test graphic file with name fx22.gif graphic file with name fx21.gif
 Perseverance test graphic file with name fx24.gif graphic file with name fx21.gif

Animals initially acquire three separate stimulus reward contingences using three two-choice discriminations. Responses to the CS+ are rewarded, responses to the CS− are non-rewarded, and responses to the CS50/50 are rewarded on 50% of the trials. In the subsequent reversal stage, the previous CS+ becomes CS− and the previous CS− becomes CS+ while the CS50/50 remains unchanged. In probe trials of full reversal learning, the CS+ is paired with the CS−. On every 5th trial, animals are presented with a probe trial of perseverance or learned non-reward. In probe trials of learned non-reward, the CS+ (previous CS−) is paired with the CS50/50. Impaired performance in this condition indicates that previous negative associations guide responding. In probe trials of stimulus perseverance, the CS− (previous CS+) is paired with the CS50/50. Impaired performance in this condition indicates that previous positive associations guide responding.

Importantly, in the subsequent reversal stage, the previous CS+ becomes CS− and the previous CS− becomes CS+ while the CS50/50 remains unchanged. In trials of full reversal learning, the previous CS− is paired with the previous CS+. The two different probe trials are still presented once each every 10 trials, but now represent learned non-reward (previous CS− vs. CS50/50) and perseverance (previous CS+ vs. CS50/50), allowing within-subject comparisons of these two processes. This paradigm offers a significant advantage over stimulus-replacement designs, as novelty attraction and novelty avoidance cannot confound data interpretations as no novel stimulus is presented. The data (Fig. 1) indicate that previous positive and negative associations both guide responding in reversal learning, and also suggest that there is individual variability in approach and avoidance behaviour in the rat, which is in agreement with work from discrimination learning using the probabilistic selection task (see below).

Fig. 1.

Fig. 1

Performance of Lister Hooded rats on a touchscreen reversal learning task with interleaved CS+ vs. CS− presentations (visual discrimination, VD) and probe trials investigating perseverance and learned non-reward. Retention (Ret.) represents performance on the last 100 trials preceding reversal. ‘Early’ represents the first 100 trials after reversal, ‘Random’ represents 100 trials when performance on the VD trials have reached 50%, and ‘Late’ represents 100 trials when the rats have reached 80% accuracy on the VD trials. (a) Reversal-learning performance in the probe task does not differ from control rats tested on the VD trials only (test group: F1,10 = .046, p = .834, group × group p = F9,90 = .714, p = .695). (b) Accuracy on probe trials, i.e. choosing CS50/50 over CS− and choosing CS+ over CS50/50, is significantly above chance (50%) during retention (one-sample t-test, t6 = 2.703, p = .035) and late reversal phases (t6 = 3.532, p = .0123), whereas performance drops below chance during early reversal (t6 = 2.677, p = .0367). (c) Number of errors on perseverance and learned non-reward probe trials does not differ (paired t-test, t6 = 0.242, p = .817), indicating that the task equally assesses previous positive and negative associations in reversal learning.

5. Neural mechanisms of association-dependent responding in two-choice discrimination learning and reversal learning

The neural substrates of discrimination and reversal learning have been the subject of extensive investigation (Ragozzino, 2007). Such studies have highlighted the involvement of prefrontal, especially orbitofrontal, cortex, dorsal striatum and amygdala (Frank and O’Reilly, 2006, Keeler and Robbins, 2011), and of a variety of neurotransmitter systems, including 5-HT (Daw et al., 2002, Roberts, 2011), dopamine (Frank and O’Reilly, 2006), and acetylcholine (Robbins and Roberts, 2007). Moreover, attentional set-shifting and reversal learning has been shown to anatomically dissociable; while attentional set-shifting primarily depend upon activity along the prefrontal medial wall (the primate dorsolateral prefrontal cortex or rodent medial prefrontal cortex), the OFC is required for reversal learning (Keeler and Robbins, 2011, Tait et al., 2014). Electrophysiological recordings and imaging studies have also revealed spatial and neural segregations on responses to punishments and rewards (Kringelbach, 2005, Schultz, 1998, Ursu and Carter, 2005), yet explicit investigation of the neural mechanisms underpinning association-dependent responding in tasks of cognitive flexibility has been relatively rare.

5.1. Non-human preclinical approaches

Lesions or inactivation of the OFC retard reversal learning in the rodent (Bissonette et al., 2008, Bussey et al., 1997, Chudasama and Robbins, 2003b, Ghods-Sharifi et al., 2008, Graybeal et al., 2011, Kim and Ragozzino, 2005, McAlonan and Brown, 2003, Schoenbaum et al., 2002). This deficit has been further explored in a bowl-digging paradigm separately probing perseverance and learned non-reward by challenging rats in conditions where either the previously correct or incorrect olfactory or somatosensory stimulus reverses contingency and is paired with a novel stimulus of the opposing contingency. In this paradigm, OFC-lesions have been observed to impair performance in a learned non-reward test, but facilitated performance in a perseverance test (Tait and Brown, 2007). Similar opposing effects on learned non-reward and perseverance has been observed in an egocentric spatial task in mice challenged with the 5-HT2C receptor (5-HT2CR) antagonist SB242084 (Nilsson et al., 2013) which has been shown to affect reversal learning through activity in the OFC (Boulougouris and Robbins, 2010). The effect of 5-HT2CR antagonism on cognitive flexibility also appears selective to reversal learning, as SB242084 fails to affect attentional set-shifting (Baker et al., 2011).

Furthermore, OFC-specific 5,7-DHT-induced 5-HT depletions impaired visual reversal learning in the marmoset (Clarke et al., 2007). 5-HT depleted animals showed deficits in a perseverance test, where the previous CS+ becomes incorrect and is paired with a novel CS+, but performed as well as controls on a learned non-reward test where the previous CS− became correct and was paired with a novel CS− (Clarke et al., 2007). Also, using an analogous visuospatial operant assay, 5-HT2CR KO mice and mice systemically treated with the 5-HT2CR antagonist SB242084 showed facilitated reversal learning and decreased learned non-reward, but did not differ from controls in a perseverance test (Nilsson et al., 2012).

A further method to dissociate perseverance and learned non-reward is to use Pavlovian successive reversal tasks. This approach has been used with blocked presentations of either CS+ or CS− followed by reward, which thus potentially gives insight into perseverance and learned non-reward. In this task, presumed OFC-inactivation, with intracerebral infusion of a baclofen/muscimol mixture, impaired reversal learning (Burke et al., 2009). Importantly, OFC-inactivation did not affect the rats’ ability to extinguish responding during the previous CS+, but retarded their ability to start responding during the previous CS−. This suggests that activation of the OFC is required for overcoming learned non-reward rather than perseverance (Burke et al., 2009). Conversely, the 5-HT transporter KO rat showed facilitated performance in a similar, though non-blocked, Pavlovian two-choice auditory reversal task (Nonkes et al., 2011). The mutant strain developed faster responding towards the previous CS− (opposed by learned non-reward) but did not differ from wild-type animals in responding towards the previous CS+ (opposed by perseverance). The improvement therefore appeared to be due to enhanced suppression of learned non-reward rather than increased ability to overcome previous positive associations leading to perseverance.

Although the role of the OFC in reversal learning has recently been questioned (Stalnaker et al., 2015), with particular reference to a reported dissociation of the effects of aspiration and excitotoxic lesions on reversal learning in the macaque (Rudebeck et al., 2013), it is not obvious how the results of the studies using temporary OFC activation or intracerebral administration of serotonergic agents reviewed earlier would be accommodated by this hypothesis.

5.2. Human preclinical and clinical approaches

Clinical studies of mechanisms underpinning association-dependent responding in two-choice discrimination and reversal learning have typically used either the unexpected outcome task (Cools, 2006) or the probabilistic stimulus selection (PSS) task (Frank et al., 2004). These tasks are dissimilar to standard preclinical tasks in the use of probabilistic rather than fully predictive cues (the PSS task) and one-trial Pavlovian learning rather than prolonged instrumental learning (the unexpected outcome task). Nevertheless, data derived from these tasks support the idea that pharmacological effects, genetic functions and pathology-related dysfunctions in discrimination and reversal learning can depend on the selective processing of positive or negative outcomes.

Furthermore, data in both paradigms has most often been interpreted the context of striatal dopamine signalling. The positive prediction error represented by phasic dopamine release from fast-spiking ventral tegmental cells may increase activity at the D1 receptor (D1R) and promote stimulus approach learning, while the negative prediction error represented by decreased tonic dopamine levels attenuate activity at the D2R and promote stimulus avoidance learning. It has been argued that the two processes are necessary for cognitive flexibility and sufficient to explain reinforcement and non-reinforcement learning (Frank and O’Reilly, 2006).

The unexpected outcome task taps stimulus avoidance and approach strategies in Pavlovian serial reversal learning (Cools, 2006). Participants are presented with a simultaneous two-choice discrimination with one stimulus highlighted by a black border. After a series of correct responses, a contingency reversal is signalled by an unexpected rewarded trial to the previously incorrect stimulus or unexpected loss of reward to previously rewarded stimuli. The ability to overcome previous positive associations is measured by performance on trials immediately preceding unexpected loss of reward, whereas the ability to overcome previously negative associations is measured by performance on the trials immediately preceding unexpected rewards.

In this task, central 5-HT and dopamine signalling appear to have dissociable effects on learned non-reward and stimulus perseveration. Acute tryptophan depletion causes selective increases in the prediction of loss of reward (Cools et al., 2008) and depressed patients show decreased accuracies and anterior ventrolateral putamen blood oxygen level-dependent (BOLD) response following unexpected reward but normal performances following unexpected loss of reward (Robinson et al., 2012). Unexpected rewards but not unexpected loss of reward are related to increased BOLD activation in the posterior dorsolateral striatum (Robinson et al., 2010a). Conversely, l-Dopa and the D3R agonist pramipexole impair learning from unexpected loss of reward but not unexpected rewards in individuals with mild Parkinson's disease The D2R agonist bromocriptine improves learning following unexpected loss of reward but not unexpected rewards in individuals with high striatal dopamine synthesis capacity (Cools et al., 2009) and reduced dopamine synthesis through acute tyrosine and phenylalanine depletion can improve learning from unexpected loss of reward (Robinson et al., 2010b). Thus, central 5-HT depletion and depression appear to selectively affect learned non-reward while central dopaminergic manipulations have more prominent effects on perseveration. These data are in general agreement with preclinical findings from constitutive or pharmacological 5-HT manipulations affecting responding to previously incorrect but not previously correct stimuli in rodent reversal learning (Nilsson et al., 2012, Nonkes et al., 2012)) and discussions implicating dopamine signalling in perseverative responding using instrumental reversal learning (Clarke et al., 2011, Clatworthy et al., 2009) and attentional set-shifting tasks (Owen et al., 1993).

In contrast to the unexpected outcome task, the PSS task contrasts response strategies only in discrimination learning by employing transitive inference of six probabilistic reward contingencies acquired in separate two-choice discriminations (Frank et al., 2004, Frank et al., 2007b). As shown in Table 7, participants initially learn three two-choice discriminations (AB, CD, EF). In the AB discrimination, stimulus A is rewarded on 80% of trials while stimulus B is rewarded on 20% of the trials. In the CD and EF discriminations, stimulus C and stimulus E are rewarded 70% and 60% of the trials while stimulus D and stimulus F are rewarded 30% and 40% of the trials, respectively. In subsequent probe tests, stimulus A and B are paired with stimuli C, D, E, and F to contrast stimulus approach learning (‘Choose A’ discriminations: AC, AD, AE, AF) with stimulus avoidance learning (‘Avoid B’ discriminations: BC, BD, BE, BF). If responses primarily are guided by reward, performance should be better in the ‘Choose A’ discrimination as these include the stimulus with the greatest reward predictability. If responses primarily are guided by avoidance, performance should be better in the ‘Avoid B’ discrimination as these include the stimulus with the lowest reward predictability. This is consistent with the suggestion that two processes are required for reversal learning (Frank and O’Reilly, 2006); stimulus approach learning is needed to overcome previous negative associations (learned non-reward) while stimulus avoidance learning is required to overcome previous positive associations (perseverance). Notably, these transitive inference tests also contain shared features complicating performance interpretations. Only probe tests of approach learning consist of discriminations between stimuli of positive probabilistic reward contingencies (discriminations: AC and AE) and only probe tests of avoidance learning consist of discriminations between stimuli of negative probabilistic reward contingencies (discriminations: BF, BE). Nevertheless, both tests of reward learning and avoidance learning consist of discriminations between stimuli of positive and negative reward contingencies (discriminations: AD, AF and BF, BE). Hence general deficits in reinforcement learning would produce decreased accuracy in both the ‘Choose A’ and ‘Avoid B’ and conditions, and it may be erroneous to conclude that this is due to disparate deficits in two processes. General deficits in reinforcement learning may also mask more selective effects in approach versus avoidance learning.

Table 7.

Assessing stimulus avoidance and stimulus approach strategies in two-choice discrimination learning using the probabilistic stimulus selection task.

Stimuli
Stimulus A (reward contingency) Stimulus B (reward contingency)
Discrimination learning
 Pair 1 ま (80%/20%) み (20%/80%)
 Pair 2 そ (70%/30%) の (30%/70%)
 Pair 3 ら (60%/40%) わ (40%/60%)
Test approach learning
 Pair 1
 Pair 2
 Pair 3
 Pair 4
Test of ‘avoidance’ learning
 Pair 1
 Pair 2
 Pair 3
 Pair 4

The table depict the principles of the probabilistic stimulus selection task (Frank et al., 2004, Frank et al., 2007b) and is not representative of individual studies using the task. Subjects initially acquire six separate stimulus reward contingences using three two-choice discriminations. Once acquired, the stimuli from Pair 1 are paired with the stimuli from Pair 2 and Pair 3 to form eight novel two-choice discriminations. Four discriminations tap stimulus approach learning and involve the stimulus most predictive of reward, while another four discriminations tap avoidance learning and involve the stimulus most predictive of non-reward. No feedback is given during discrimination tests of avoidance and approach learning.

Some studies using this paradigm have reported dopaminergic-related functions in avoidance but not approach learning. This includes the single-nucleotide polymorphisms affecting D2R mRNA and DARPP-32 mRNA expression and D2R density (Cavanagh et al., 2014, Frank and Hutchison, 2009, Frank et al., 2007a). Medication in Parkinson's disease can also have a greater effect on avoidance than approach learning (Frank et al., 2007b) which would be in accordance with the selective effects of l-dopa on perseverative responding in an attentional set-shifting task using a stimulus replacement design (Owen et al., 1993).

There are several reports of selective genetic, pharmacological and pathological effects on reward relative to avoidance learning. Relative to carriers of the DARPP-32 G allele, DARPP-32A/A carriers show improved reward relative to punishment learning (Frank et al., 2007a, Frank and Hutchison, 2009) and the D2R antagonist amisulpride improves (Jocham et al., 2011) while the the D2R agonist cabergoline impairs (Frank and O’Reilly, 2006) reward learning without affecting avoidance learning. Similarly, schizophrenic patients show impaired reward learning but intact avoidance learning (Gold et al., 2012, Waltz et al., 2007). This effect may be associated with negative symptom expression (Gold et al., 2012, Waltz et al., 2007), which would be in accordance with the learned non-reward and reversal learning deficits observed in depressed patients using the unexpected outcome task (Cools et al., 2008, Robinson et al., 2012). Moreover, ADHD patients off medication show deficits in both avoidance and approach learning while ADHD patients on medication show selective deficits in avoidance learning (Frank et al., 2006). This would suggest that stimulants improve approach learning in ADHD patients, while avoidance learning manifested by perseverative responding in tasks of cognitive flexibility (Frank and O’Reilly, 2006) is not addressed by currently available drugs. Lastly, the COMT single-nucleotide polymorphisms fail to affect both approach and avoidance learning in the PSS task (Frank et al., 2007a) despite their clinical (Malhotra et al., 2002) and preclinical (Scheggia et al., 2013) linkage to cognitive flexibility. However, the PSS task measures approach and avoidance strategies in discrimination learning and inferences regarding its relevance for stimulus perseveration and learned non-reward in later reversal learning (Frank and O’Reilly, 2006) remain tentative. Moreover, the PSS task has generally been used to assess global pharmacological and genetic effects on discrimination learning. Yet the data from these studies are typically interpreted in the very detailed contexts of striatal D1R and D2R signalling. To enable conclusions regarding circuit-specific effects on stimulus perseverance and learned non-reward in reversal learning, such as the roles of D1R vs. D2R signalling, a preclinical paradigm suitable for non-human animals is required.

Finally, there are obvious methodological differences in the discussed studies that may account for inconsistencies observed in the literature. This includes the use of reversal learning assays underpinned by either Pavlovian learning or instrumental responding. Rodent reversal assays can also depend on instrumental responding with a strong spatial component as opposed to non-human primate and human tasks in which a wide array of visual discriminative stimuli are employed. The further development of touchscreen tasks in rodents provides one solution to this dilemma (Horner et al., 2013, Mar et al., 2013) though there has to be a residual concern about the ecological validity of this approach given the extended training that is necessary. Notably, task differences also appear to affect the main association guiding responding in discrimination and reversal learning. Responding can be primarily guided by the rewarded association (Nilsson et al., 2013, Robinson et al., 2010b, Tait and Brown, 2007, Waltz et al., 2007), the non-rewarded association (Jocham et al., 2011, Nilsson et al., 2012) or equally dependent on rewarded and non-rewarded associations (Clarke et al., 2007, Cools et al., 2008). Thus, slight protocol differences and species are likely to affect the relative guidance of approach and avoidance strategies.

In sum, investigations of neural substrates underpinning reinforced and non-reinforced associations in discrimination and reversal learning have used a wide variety of disparate paradigms, which is likely to influence the relative roles of approach and avoidance strategies and the effect of experimental manipulations. Nevertheless, the data show that perseverance and learned non-reward can be dissociable on anatomical, pharmacological, genetic and pathological levels, and that both perseverance and learned non-reward can contribute to the effect of an experimental manipulation on reversal learning performance.

6. Conclusions

Despite clear differences in symptom profiles across pathologies, most human psychiatric patients and animal disease models show relatively similar impairments on standard tests of cognitive flexibility. Current reversal learning assays may be rather crude measures of cognitive functioning that give few indications of possible pathology-specific deficits and therapeutic approaches. The Wisconsin Card Sorting Task (WCST) was recognised as a crude measure of cognitive flexibility and the task was successfully modified into a series of tests assessing the separate cognitive mechanisms required for attentional set-shifting in the more refined CANTAB ID/ED-task. Separate tests of perseverance and learned non-reward may have similar potential in delineating the neural basis of flexible associations with reward and non-reward in reversal learning.

Although manipulations of cognitive flexibility have traditionally been interpreted as manipulations of perseverance, the effects of learned non-reward and learned irrelevance must also be considered. This suggestion is supported both by empirical data and theoretical considerations from learning theory indicating that irrelevant or non-rewarded associations can be a major determinant of choice-behaviour in discrimination learning, reversal learning and attentional set-shifting tasks. Further work suggests that impairments of cognitive flexibility in different psychopathologies have different underlying behavioural mechanisms, and that the cognitive components of perseverance, learned non-reward, and learned irrelevance may be anatomically and neurochemically dissociable. This has important implications for the current understanding of reversal learning and attentional set-shifting and also has the potential to increase the validity of preclinical models of cognitive inflexibility.

Thus the resolution of reversal learning into separate tests of perseverance and non-reward will increase preclinical validity by clarifying species and task related differences in cognitive flexibility. A preclinical reversal learning task where choice-behaviour primarily is guided by associations of stimuli with non-reward is likely to have limited validity if choice behaviour in the clinical context is primarily guided by associations of stimuli with reward. Conversely, the demonstration that animal and human subjects solve analogous tasks using similar cognitive strategies by assigning comparable attention to cues of positive and negative association with reward would increase the validity of the task.

Thus assays of cognitive flexibility should be designed to separate effects of perseverance, learned non-reward and learned irrelevance. Dissociating these cognitive components into separate tests will enhance the validity of preclinical assays of cognitive flexibility and enhance the potential contribution of preclinical findings to improved mental health.

Acknowledgements

Supported by BBSRC and Eli Lilly through CASE studentship (BB/F529054/1). J.A. was supported by the Swedish Research Council (350-2012-230). The Behavioural and Clinical Neuroscience Institute is co-funded by the Medical Research Council and the Wellcome Trust.

Footnotes

The title of this paper refers to a phrase that defined the politics of Margaret Thatcher, UK prime minister between 1979 and 1990. In October 1980 she made a speech responding to critics of her economic policy in which she said ‘To those waiting with bated breath for that favourite media catchphrase, the ‘U-turn’, I have only one thing to say: You turn if you want to. The lady's not for turning.’.

References

  1. Abdul-Monim Z., Neill J.C., Reynolds G.P. Sub-chronic psychotomimetic phencyclidine induces deficits in reversal learning and alterations in parvalbumin-immunoreactive expression in the rat. J. Psychopharmacol. 2007;21:198–205. doi: 10.1177/0269881107067097. [DOI] [PubMed] [Google Scholar]
  2. Amsel A. The role of frustrative nonreward in noncontinuous reward situations. Psychol. Bull. 1958;55:102–119. doi: 10.1037/h0043125. [DOI] [PubMed] [Google Scholar]
  3. Baker P.M., Thompson J.L., Sweeney J.A., Ragozzino M.E. Differential effects of 5-HT(2A) and 5-HT(2C) receptor blockade on strategy-switching. Behav. Brain Res. 2011;219:123–131. doi: 10.1016/j.bbr.2010.12.031. [DOI] [PMC free article] [PubMed] [Google Scholar]
  4. Barch D.M. The cognitive neuroscience of schizophrenia. Annu. Rev. Clin. Psychol. 2005;1:321–353. doi: 10.1146/annurev.clinpsy.1.102803.143959. [DOI] [PubMed] [Google Scholar]
  5. Behar I. Learned avoidance of nonreward. Psychol. Rep. 1961;9:34–52. [Google Scholar]
  6. Bennett C.H., Maldonado A., Mackintosh N.J. Learned irrelevance is not the sum of exposure to CS and US. Q. J. Exp. Psychol., B. 1995;48:117–128. [PubMed] [Google Scholar]
  7. Bissonette G.B., Martins G.J., Franz T.M., Harper E.S., Schoenbaum G., Powell E.M. Double dissociation of the effects of medial and orbital prefrontal cortical lesions on attentional and affective shifts in mice. J. Neurosci. Res. 2008;28:11124–11130. doi: 10.1523/JNEUROSCI.2820-08.2008. [DOI] [PMC free article] [PubMed] [Google Scholar]
  8. Bleuler E. International Universities Press; New York, NY: 1950. Dementia Praecox or the Group of Schizophrenias. [Google Scholar]
  9. Blomquist A.J., Deets A.C., Harlow H.F. Effects of list length and first-trial reward upon concurrent discrimination performance. Learn. Motiv. 1973;4:28–39. [Google Scholar]
  10. Boulougouris V., Dalley J.W., Robbins T.W. Effects of orbitofrontal, infralimbic and prelimbic cortical lesions on serial spatial reversal learning in the rat. Behav. Brain Res. 2007;179:219–228. doi: 10.1016/j.bbr.2007.02.005. [DOI] [PubMed] [Google Scholar]
  11. Boulougouris V., Glennon J.C., Robbins T.W. Dissociable effects of selective 5-HT2A and 5-HT2C receptor antagonists on serial spatial reversal learning in rats. Neuropsychopharmacology. 2008;33:2007–2019. doi: 10.1038/sj.npp.1301584. [DOI] [PubMed] [Google Scholar]
  12. Boulougouris V., Robbins T.W. Enhancement of spatial reversal learning by 5-HT2C receptor antagonism is neuroanatomically specific. J. Neurosci. 2010;30:930–938. doi: 10.1523/JNEUROSCI.4312-09.2010. [DOI] [PMC free article] [PubMed] [Google Scholar]
  13. Burke K.A., Takahashi Y.K., Correll J., Leon Brown P., Schoenbaum G. Orbitofrontal inactivation impairs reversal of Pavlovian learning by interfering with “disinhibition” of responding for previously unrewarded cues. Eur. J. Neurosci. 2009;30:1941–1946. doi: 10.1111/j.1460-9568.2009.06992.x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  14. Buss A.H. Rigidity as a function of reversal and non-reversal shifts in the learning of successive discriminations. J. Exp. Psychol. 1953;45:75–81. doi: 10.1037/h0054027. [DOI] [PubMed] [Google Scholar]
  15. Buss A.H. Reversal and nonreversal shifts in concept formation with partial reinforcement eliminated. J. Exp. Psychol. 1956;52:162–166. doi: 10.1037/h0048669. [DOI] [PubMed] [Google Scholar]
  16. Bussey T.J., Everitt B.J., Robbins T.W. Dissociable effects of cingulate and medial frontal cortex lesions on stimulus-reward learning using a novel Pavlovian autoshaping procedure for the rat: implications for the neurobiology of emotion. Behav. Neurosci. 1997;111:908–919. doi: 10.1037//0735-7044.111.5.908. [DOI] [PubMed] [Google Scholar]
  17. Cavanagh J.F., Masters S.E., Bath K., Frank M.J. Conflict acts as an implicit cost in reinforcement learning. Nat. Commun. 2014;5:5394. doi: 10.1038/ncomms6394. [DOI] [PubMed] [Google Scholar]
  18. Ceaser A.E., Goldberg T.E., Egan M.F., McMahon R.P., Weinberger D.R., Gold J.M. Set-shifting ability and schizophrenia: a marker of clinical illness or an intermediate phenotype? Biol. Psychiatry. 2008;64:782–788. doi: 10.1016/j.biopsych.2008.05.009. [DOI] [PMC free article] [PubMed] [Google Scholar]
  19. Chamberlain S.R., Fineberg N.A., Blackwell A.D., Robbins T.W., Sahakian B.J. Motor inhibition and cognitive flexibility in obsessive–compulsive disorder and trichotillomania. Am. J. Psychiatry. 2006;163:1282–1284. doi: 10.1176/ajp.2006.163.7.1282. [DOI] [PubMed] [Google Scholar]
  20. Choi J., Kurtz M.M. A comparison of remediation techniques on the Wisconsin Card Sorting Test in schizophrenia. Schizophrn. Res. 2009;107:76–82. doi: 10.1016/j.schres.2008.09.017. [DOI] [PMC free article] [PubMed] [Google Scholar]
  21. Chudasama Y., Robbins T.W. Dissociable contributions of the orbitofrontal and infralimbic cortex to pavlovian autoshaping and discrimination reversal learning: further evidence for the functional heterogeneity of the rodent frontal cortex. J. Neurosci. 2003;23:8771–8780. doi: 10.1523/JNEUROSCI.23-25-08771.2003. [DOI] [PMC free article] [PubMed] [Google Scholar]
  22. Chudasama Y., Robbins T.W. Dissociable contributions of the orbitofrontal and infralimbic cortex to pavlovian autoshaping and discrimination reversal learning: further evidence for the functional heterogeneity of the rodent frontal cortex. J. Neurosci. 2003;23:8771–8780. doi: 10.1523/JNEUROSCI.23-25-08771.2003. [DOI] [PMC free article] [PubMed] [Google Scholar]
  23. Clark L., Iversen S.D., Goodwin G.M. A neuropsychological investigation of prefrontal cortex involvement in acute mania. Am. J. Psychiatry. 2001;158:1605–1611. doi: 10.1176/appi.ajp.158.10.1605. [DOI] [PubMed] [Google Scholar]
  24. Clark L., Iversen S.D., Goodwin G.M. Sustained attention deficit in bipolar disorder. Br. J. Psychiatry. 2002;180:313–319. doi: 10.1192/bjp.180.4.313. [DOI] [PubMed] [Google Scholar]
  25. Clark L., Sarna A., Goodwin G.M. Impairment of executive function but not memory in first-degree relatives of patients with bipolar I disorder and in euthymic patients with unipolar depression. Am. J. Psychiatry. 2005;162:1980–1982. doi: 10.1176/appi.ajp.162.10.1980. [DOI] [PubMed] [Google Scholar]
  26. Clarke H.F., Hill G.J., Robbins T.W., Roberts A.C. Dopamine, but not serotonin, regulates reversal learning in the marmoset caudate nucleus. J. Neurosci. 2011;31:4290–4297. doi: 10.1523/JNEUROSCI.5066-10.2011. [DOI] [PMC free article] [PubMed] [Google Scholar]
  27. Clarke H.F., Walker S.C., Dalley J.W., Robbins T.W., Roberts A.C. Cognitive inflexibility after prefrontal serotonin depletion is behaviorally and neurochemically specific. Cereb. Cortex. 2007;17:18–27. doi: 10.1093/cercor/bhj120. [DOI] [PubMed] [Google Scholar]
  28. Clatworthy P.L., Lewis S.J.G., Brichard L., Hong Y.T., Izquierdo D., Clark L., Cools R., Aigbirhio F.I., Baron J.C., Fryer T.D., Robbins T.W. Dopamine release in dissociable striatal subregions predicts the different effects of oral methylphenidate on reversal learning and spatial working memory. J. Neurosci. 2009;29:4690–4696. doi: 10.1523/JNEUROSCI.3266-08.2009. [DOI] [PMC free article] [PubMed] [Google Scholar]
  29. Coldren J.T., Halloran C. Spatial reversal as a measure of executive functioning in children with autism. J. Genet. Psychol. 2003;164:29–41. doi: 10.1080/00221320309597501. [DOI] [PubMed] [Google Scholar]
  30. Cools R. Dopaminergic modulation of cognitive function-implications for l-DOPA treatment in Parkinson's disease. Neurosci. Biobehav. Rev. 2006;30:1–23. doi: 10.1016/j.neubiorev.2005.03.024. [DOI] [PubMed] [Google Scholar]
  31. Cools R., Barker R.A., Sahakian B.J.B., Robbins T.W. Enhanced or impaired cognitive function in Parkinson's disease as a function of dopaminergic medication and task demands. Cereb. Cortex. 2001;11:1136–1143. doi: 10.1093/cercor/11.12.1136. [DOI] [PubMed] [Google Scholar]
  32. Cools R., Frank M.J., Gibbs S.E., Miyakawa A., Jagust W., D’Esposito M. Striatal dopamine predicts outcome-specific reversal learning and its sensitivity to dopaminergic drug administration. J. Neurosci. 2009;29:1538–1543. doi: 10.1523/JNEUROSCI.4467-08.2009. [DOI] [PMC free article] [PubMed] [Google Scholar]
  33. Cools R., Robinson O.J., Sahakian B. Acute tryptophan depletion in healthy volunteers enhances punishment prediction but does not affect reward prediction. Neuropsychopharmacology. 2008;33:2291–2299. doi: 10.1038/sj.npp.1301598. [DOI] [PubMed] [Google Scholar]
  34. Crider A. Perseveration in schizophrenia. Schizophr. Bull. 1997;23:63–74. doi: 10.1093/schbul/23.1.63. [DOI] [PubMed] [Google Scholar]
  35. Cross H.A., Brown L.T. Discrimination reversal learning in squirrel monkeys as a function of number of acquisition trials and prereversal experience. J. Comp. Physiol. Psychol. 1965;59:429–431. doi: 10.1037/h0022056. [DOI] [PubMed] [Google Scholar]
  36. D’Cruz A.-M., Ragozzino M.E., Mosconi M.W., Shrestha S., Cook E.H., Sweeney J.A. Reduced behavioral flexibility in autism spectrum disorders. Neuropsychology. 2013;27:152–160. doi: 10.1037/a0031721. [DOI] [PMC free article] [PubMed] [Google Scholar]
  37. Dalley J.W., Everitt B.J., Robbins T.W. Impulsivity, compulsivity, and top-down cognitive control. Neuron. 2011;69:680–694. doi: 10.1016/j.neuron.2011.01.020. [DOI] [PubMed] [Google Scholar]
  38. Daw N.D., Kakade S., Dayan P. Opponent interactions between serotonin and dopamine. Neural Netw. 2002;15:603–616. doi: 10.1016/s0893-6080(02)00052-7. [DOI] [PubMed] [Google Scholar]
  39. Dias R., Aggleton J.P. Effects of selective excitotoxic prefrontal lesions on acquisition of nonmatching- and matching-to-place in the T-maze in the rat: differential involvement of the prelimbic-infralimbic and anterior cingulate cortices in providing behavioural flexibility. Eur. J. Neurosci. 2000;12:4457–4466. doi: 10.1046/j.0953-816x.2000.01323.x. [DOI] [PubMed] [Google Scholar]
  40. Dias R., Robbins T.W., Roberts A.C. Dissociation in prefrontal cortex of affective and attentional shifts. Nature. 1996;380:69–72. doi: 10.1038/380069a0. [DOI] [PubMed] [Google Scholar]
  41. Dickstein D.P., Treland J.E., Snow J., McClure E.B., Mehta M.S., Towbin K.E., Pine D.S., Leibenluft E. Neuropsychological performance in pediatric bipolar disorder. Biol. Psychiatry. 2004;55:32–39. doi: 10.1016/s0006-3223(03)00701-7. [DOI] [PubMed] [Google Scholar]
  42. Downes J.J., Roberts A.C., Sahakian B.J. Impaired extra-dimensional shift performance in medicated and unmedicated Parkinson's disease: evidence for a specific attentional dysfunction. Neuropsychologia. 1989;27:1329–1343. doi: 10.1016/0028-3932(89)90128-0. [DOI] [PubMed] [Google Scholar]
  43. Elliott R., McKenna P.J., Robbins T.W. Specific neuropsychological deficits in schizophrenic patients with preserved intellectual function. Cognit. Neuropsychiatry. 1998;3:45–70. [Google Scholar]
  44. Elliott R., McKenna P.J., Robbins T.W., Sahakian B.J. Neuropsychological evidence for frontostriatal dysfunction in schizophrenia. Psychol. Med. 1995;25:619–630. doi: 10.1017/s0033291700033523. [DOI] [PubMed] [Google Scholar]
  45. Ersche K.D., Roiser J.P., Robbins T.W., Sahakian B.J. Chronic cocaine but not chronic amphetamine use is associated with perseverative responding in humans. Psychopharmacology. 2008;197:421–431. doi: 10.1007/s00213-007-1051-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
  46. Floresco S.B., Ghods-Sharifi S., Vexelman C., Magyar O. Dissociable roles for the nucleus accumbens core and shell in regulating set shifting. J. Neurosci. 2006;26:2449–2457. doi: 10.1523/JNEUROSCI.4431-05.2006. [DOI] [PMC free article] [PubMed] [Google Scholar]
  47. Frank M.J., Fossella J.A. Neurogenetics and pharmacology of learning, motivation, and cognition. Neuropsychopharmacology. 2010;36:133–152. doi: 10.1038/npp.2010.96. [DOI] [PMC free article] [PubMed] [Google Scholar]
  48. Frank M.J., Hutchison K. Genetic contributions to avoidance-based decisions: striatal D2 receptor polymorphisms. Neuroscience. 2009;164:131–140. doi: 10.1016/j.neuroscience.2009.04.048. [DOI] [PMC free article] [PubMed] [Google Scholar]
  49. Frank M.J., Moustafa A.A., Haughey H.M., Curran T., Hutchison K.E. Genetic triple dissociation reveals multiple roles for dopamine in reinforcement learning. Proc. Natl. Acad. Sci. U.S.A. 2007;104:16311–16316. doi: 10.1073/pnas.0706111104. [DOI] [PMC free article] [PubMed] [Google Scholar]
  50. Frank M.J., O’Reilly R.C. A mechanistic account of striatal dopamine function in human cognition: Psychopharmacological studies with cabergoline and haloperidol. Behav. Neurosci. 2006;120:497–517. doi: 10.1037/0735-7044.120.3.497. [DOI] [PubMed] [Google Scholar]
  51. Frank M.J., Samanta J., Moustafa A.A., Sherman S.J. Hold your horses: impulsivity, deep brain stimulation, and medication in parkinsonism. Science. 2007;318:1309–1312. doi: 10.1126/science.1146157. [DOI] [PubMed] [Google Scholar]
  52. Frank M.J., Santamaria A., O’Reilly R.C., Willcutt E. Testing computational models of dopamine and noradrenaline dysfunction in attention deficit/hyperactivity disorder. Neuropsychopharmacology. 2006;32:1583–1599. doi: 10.1038/sj.npp.1301278. [DOI] [PubMed] [Google Scholar]
  53. Frank M.J., Seeberger L.C., O’Reilly R.C. By carrot or by stick: cognitive reinforcement learning in parkinsonism. Science. 2004;306:1940–1943. doi: 10.1126/science.1102941. [DOI] [PubMed] [Google Scholar]
  54. Freedman M., Oscar-Berman M. Spatial and visual learning deficits in Alzheimer's and Parkinson's disease. Brain. Cognit. 1989;11:114–126. doi: 10.1016/0278-2626(89)90009-2. [DOI] [PubMed] [Google Scholar]
  55. Freeman T., Gathercole C.E. Perseveration the clinical symptoms in chronic schizophrenia and organic dementia. Br. J. Psychiatry. 1966;112:27–32. doi: 10.1192/bjp.112.482.27. [DOI] [PubMed] [Google Scholar]
  56. Gardner R.A., Coate W.B. Reward versus nonreward in a simultaneous discrimination. J. Exp. Psychol. 1965;69:579–582. doi: 10.1037/h0022057. [DOI] [PubMed] [Google Scholar]
  57. Garner J.P. Perseveration and stereotypy—systems-level insights from clinical psychology. In: Mason G., Rushen J., editors. Stereotypic Animal Behaviour—Fundamentals and Applications for Welfare. CABI; Trowbridge, UK: 2006. pp. 121–142. [Google Scholar]
  58. Garner J.P., Thogerson C.M., Würbel H., Murray J.D., Mench J.A. Animal neuropsychology: validation of the intra-dimensional extra-dimensional set shifting task for mice. Behav. Brain Res. 2006;173:53–61. doi: 10.1016/j.bbr.2006.06.002. [DOI] [PubMed] [Google Scholar]
  59. Gau S.S.-F., Shang C.-Y. Executive functions as endophenotypes in ADHD: evidence from the Cambridge Neuropsychological Test Battery (CANTAB) J. Child Psychol. Psychiatry. 2010;51:838–849. doi: 10.1111/j.1469-7610.2010.02215.x. [DOI] [PubMed] [Google Scholar]
  60. Gauntlett-Gilbert J., Roberts R.C.R., Brown V.J. Mechanisms underlying attentional set-shifting in Parkinson's disease. Neuropsychologia. 1999;37:605–616. doi: 10.1016/s0028-3932(98)00049-9. [DOI] [PubMed] [Google Scholar]
  61. Geyer M.A., Tamminga C.A. Measurement and treatment research to improve cognition in schizophrenia: neuropharmacological aspects. Psychopharmacology. 2004;174:1–2. [Google Scholar]
  62. Ghods-Sharifi S., Haluk D.M., Floresco S.B. Differential effects of inactivation of the orbitofrontal cortex on strategy set-shifting and reversal learning. Neurobiol. Learn. Mem. 2008;89:567–573. doi: 10.1016/j.nlm.2007.10.007. [DOI] [PubMed] [Google Scholar]
  63. Gold J.M., Waltz J.A., Matveeva T.M., Kasanova Z., Strauss G.P., Herbener E.S., Collins A.G.E., Frank M.J. Negative symptoms and the failure to represent the expected reward value of actions: behavioral and computational modeling evidence. Arch. Gen. Psychiatry. 2012;69:129–138. doi: 10.1001/archgenpsychiatry.2011.1269. [DOI] [PMC free article] [PubMed] [Google Scholar]
  64. Gorrindo T., Blair K.S., Budhani S., Dickstein D.P., Pine D.S., Leibenluft E. Deficits on a probabilistic response-reversal task in patients with pediatric bipolar disorder. Am. J. Psychiatry. 2005;162:1975–1977. doi: 10.1176/appi.ajp.162.10.1975. [DOI] [PubMed] [Google Scholar]
  65. Goulart P.R.K., Mendonça M.B., Barros R.S., Galvão O.F., McIlvane W.J. A note on select- and reject-controlling relations in the simple discrimination of capuchin monkeys (Cebus apella) Behav. Processes. 2005;69:295–302. doi: 10.1016/j.beproc.2004.12.005. [DOI] [PubMed] [Google Scholar]
  66. Grabbe W., Campione J.C. A novelty interpretation of the Moss–Harlow effect in preschool children. Child Dev. 1969;40:1077–1084. doi: 10.1111/j.1467-8624.1969.tb04566.x. [DOI] [PubMed] [Google Scholar]
  67. Graybeal C., Feyder M., Schulman E., Saksida L.M., Bussey T.J., Brigman J.L., Holmes A. Paradoxical reversal learning enhancement by stress or prefrontal cortical damage: rescue with BDNF. Nat. Neurosci. 2011;14:1507–1509. doi: 10.1038/nn.2954. [DOI] [PMC free article] [PubMed] [Google Scholar]
  68. Green M.F. Cognitive impairment and functional outcome in schizophrenia and bipolar disorder. J. Clin. Psychiatry. 2006;67(supl 9):3–8. [PubMed] [Google Scholar]
  69. Harrow M., Friedman G.B. Comparing reversal and nonreversal shifts in concept formation with partial reinforcement controlled. J. Exp. Psychol. 1958;55:592–598. doi: 10.1037/h0043893. [DOI] [PubMed] [Google Scholar]
  70. Helmick J.W., Berg C.B. Perseveration in brain-injured adults. J. Commun. Disord. 1976;9:143–156. doi: 10.1016/0021-9924(76)90006-x. [DOI] [PubMed] [Google Scholar]
  71. Holmes M.K., Erickson K., Luckenbaugh D.A., Drevets W.C., Bain E.E., Cannon D.M., Snow J., Sahakian B.J., Manji H.K., Zarate C.A., Jr. A comparison of cognitive functioning in medicated and unmedicated subjects with bipolar depression. Bipolar Disord. 2008;10:806–815. doi: 10.1111/j.1399-5618.2008.00628.x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  72. Horner A.E., Heath C.J., Hvoslef-Eide M., Kent B.A., Kim C.H., Nilsson S., Alsiö J., Oomen C.A., Holmes A., Saksida L.M., Bussey T.J. The touchscreen operant platform for testing learning and memory in rats and mice. Nat. Protoc. 2013;8:1961–1984. doi: 10.1038/nprot.2013.122. [DOI] [PMC free article] [PubMed] [Google Scholar]
  73. Hughes C., Russell J., Robbins T.W. Evidence for executive dysfunction in autism. Neuropsychologia. 1994;32:477–492. doi: 10.1016/0028-3932(94)90092-2. [DOI] [PubMed] [Google Scholar]
  74. Insel T., Cuthbert B., Garvey M., Heinssen R., Pine D.S., Quinn K., Sanislow C., Wang P. Research domain criteria (RDoC): toward a new classification framework for research on mental disorders. Am. J. Psychiatry. 2010;167:748–751. doi: 10.1176/appi.ajp.2010.09091379. [DOI] [PubMed] [Google Scholar]
  75. Itami S.S., Uno H.H. Orbitofrontal cortex dysfunction in attention-deficit hyperactivity disorder revealed by reversal and extinction tasks. NeuroReport. 2002;13:2453–2457. doi: 10.1097/00001756-200212200-00016. [DOI] [PubMed] [Google Scholar]
  76. Izquierdo A., Jentsch J.D. Reversal learning as a measure of impulsive and compulsive behavior in addictions. Psychopharmacology. 2012;219:607–620. doi: 10.1007/s00213-011-2579-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
  77. James W. Henry Holt & CO; New York, NY: 1890. The Principles of Psychology. [Google Scholar]
  78. Jazbec S., Pantelis C., Robbins T.W., Weickert T., Weinberger D.R., Goldberg T.E. Intra-dimensional/extra-dimensional set-shifting performance in schizophrenia: impact of distractors. Schizophr. Res. 2007;89:339–349. doi: 10.1016/j.schres.2006.08.014. [DOI] [PubMed] [Google Scholar]
  79. Jentsch J.D., Olausson P., La Garza, De R., Taylor J.R. Impairments of reversal learning and response perseveration after repeated, intermittent cocaine administrations to monkeys. Neuropsychopharmacology. 2002;26:183–190. doi: 10.1016/S0893-133X(01)00355-4. [DOI] [PubMed] [Google Scholar]
  80. Jocham G., Klein T.A., Ullsperger M. Dopamine-mediated reinforcement learning signals in the striatum and ventromedial prefrontal cortex underlie value-based choices. J. Neurosci. 2011;31:1606–1613. doi: 10.1523/JNEUROSCI.3904-10.2011. [DOI] [PMC free article] [PubMed] [Google Scholar]
  81. Keeler J.F., Robbins T.W. Translating cognition from animals to humans. Biochem. Pharmacol. 2011;81:1356–1366. doi: 10.1016/j.bcp.2010.12.028. [DOI] [PubMed] [Google Scholar]
  82. Kempton S., Vance A., Maruff P., Luk E., Costin J., Pantelis C. Executive function and attention deficit hyperactivity disorder: stimulant medication and better executive function performance in children. Psychol. Med. 1999;29:527–538. doi: 10.1017/s0033291799008338. [DOI] [PubMed] [Google Scholar]
  83. Kendler H.H., D’Amato M.F. A comparison of reversal shifts and nonreversal shifts in human concept formation behavior. J. Exp. Psychol. 1955;49:165. doi: 10.1037/h0049296. [DOI] [PubMed] [Google Scholar]
  84. Kim J., Ragozzino M.E. The involvement of the orbitofrontal cortex in learning under changing task contingencies. Neurobiol. Learn. Mem. 2005;83:125–133. doi: 10.1016/j.nlm.2004.10.003. [DOI] [PMC free article] [PubMed] [Google Scholar]
  85. Kringelbach M.L. The human orbitofrontal cortex: linking reward to hedonic experience. Nat. Rev. Neurosci. 2005;6:691–702. doi: 10.1038/nrn1747. [DOI] [PubMed] [Google Scholar]
  86. Lange K.W., Sahakian B.J., Quinn N.P., Marsden C.D., Robbins T.W. Comparison of executive and visuospatial memory function in Huntington's disease and dementia of Alzheimer type matched for degree of dementia. J. Neurol. Neurosurg. Psychiatry. 1995;58:598–606. doi: 10.1136/jnnp.58.5.598. [DOI] [PMC free article] [PubMed] [Google Scholar]
  87. Lawrence A.D., Hodges J.R., Rosser A.E. Evidence for specific cognitive deficits in preclinical Huntington's disease. Brain. 1998;121(Pt 7):1329–1341. doi: 10.1093/brain/121.7.1329. [DOI] [PubMed] [Google Scholar]
  88. Lawrence A.D., Sahakian B.J., Hodges J.R. Executive and mnemonic functions in early Huntington's disease. Brain. 1996;119(Pt 5):1633–1645. doi: 10.1093/brain/119.5.1633. [DOI] [PubMed] [Google Scholar]
  89. Lawrence A.D., Sahakian B.J., Rogers R.D., Hodge J.R., Robbins T.W. Discrimination, reversal, and shift learning in Huntington's disease: mechanisms of impaired response selection. Neuropsychologia. 1999;37:1359–1374. doi: 10.1016/s0028-3932(99)00035-4. [DOI] [PubMed] [Google Scholar]
  90. Laws K.R. A meta-analytic review of wisconsin card sort studies in schizophrenia: general intellectual deficit in disguise? Cognit. Neuropsychiatry. 1999;4:1–30. doi: 10.1080/135468099396025. [DOI] [PubMed] [Google Scholar]
  91. Leeson V.C., Robbins T.W., Matheson E., Hutton S.B., Ron M.A., Barnes T.R.E., Joyce E.M. Discrimination learning, reversal, and set-shifting in first-episode schizophrenia: stability over six years and specific associations with medication type and disorganization syndrome. Biol. Psychiatr. 2009;66:586–593. doi: 10.1016/j.biopsych.2009.05.016. [DOI] [PMC free article] [PubMed] [Google Scholar]
  92. Lovejoy E. Analysis of the overlearning reversal effect. Psychol. Rev. 1966;73:87–103. doi: 10.1037/h0022687. [DOI] [PubMed] [Google Scholar]
  93. Lubow R.E. Cambridge University Press; Cambridge, UK: 1989. Latent Inhibition and Conditioned Attention Theory. [Google Scholar]
  94. Luria A.R. Two kinds of motor perseveration in massive injury of the frontal lobes. Brain. 1965;88:1–10. doi: 10.1093/brain/88.1.1. [DOI] [PubMed] [Google Scholar]
  95. Mackintosh N.J. Selective attention in animal discrimination learning. Psychol. Bull. 1965;64:124. doi: 10.1037/h0022347. [DOI] [PubMed] [Google Scholar]
  96. Mackintosh N.J. Further analysis of the overtraining reversal effect. J. Comp. Physiol. Psychol. 1969;67:Suppl–Suppl18. doi: 10.1037/h0026784. [DOI] [PubMed] [Google Scholar]
  97. Mackintosh N.J. A search for contrast effects in discrete-trial discrimination learning by pigeons. Learn. Motiv. 1974;5:311–327. [Google Scholar]
  98. Mackintosh N.J. A theory of attention: Variations in the associability of stimuli with reinforcement. Psychol. Rev. 1975;82:276–298. [Google Scholar]
  99. Mackintosh N.J. Oxford University Press; Oxford, UK: 1983. Conditioning and Associative Learning. [Google Scholar]
  100. Mackintosh N.J., McGonigle B., Holgate V., Vanderver V. Factors underlying improvement in serial reversal learning. Can. J. Psychol. 1968;22:85–95. doi: 10.1037/h0082753. [DOI] [PubMed] [Google Scholar]
  101. Maes J.H.R., Damen M.D.C., Eling P.A.T.M. More learned irrelevance than perseveration errors in rule shifting in healthy subjects. Brain Cognit. 2004;54:201–211. doi: 10.1016/j.bandc.2004.01.003. [DOI] [PubMed] [Google Scholar]
  102. Maes J.H.R., Eling P.A.T.M. Do learned irrelevance and perseveration play a role during discrimination learning? Learn. Motiv. 2009;40:274–283. [Google Scholar]
  103. Malhotra A.K., Kestler L.J., Mazzanti C., Bates J.A., Goldberg T.E., Goldman D. A functional polymorphism in the COMT gene and performance on a test of prefrontal cognition. Am. J. Psychiatry. 2002;159:652–654. doi: 10.1176/appi.ajp.159.4.652. [DOI] [PubMed] [Google Scholar]
  104. Mar A.C., Horner A.E., Nilsson S., Alsiö J., Kent B.A., Kim C.H., Holmes A., Saksida L.M., Bussey T.J. The touchscreen operant platform for assessing executive function in rats and mice. Nat. Protoc. 2013;8:1985–2005. doi: 10.1038/nprot.2013.123. [DOI] [PMC free article] [PubMed] [Google Scholar]
  105. Martínez-Arán A., Vieta E., Reinares M., Colom F., Torrent C., Sánchez-Moreno J., Benabarre A., Goikolea J.M., Comes M., Salamero M. Cognitive function across manic or hypomanic, depressed, and euthymic states in bipolar disorder. Am. J. Psychiatry. 2004;161:262–270. doi: 10.1176/appi.ajp.161.2.262. [DOI] [PubMed] [Google Scholar]
  106. McAlonan K., Brown V.J. Orbital prefrontal cortex mediates reversal learning and not attentional set shifting in the rat. Behav. Brain Res. 2003;146:97–103. doi: 10.1016/j.bbr.2003.09.019. [DOI] [PubMed] [Google Scholar]
  107. McCool M.F., Patel S., Talati R., Ragozzino M.E. Differential involvement of M1-type and M4-type muscarinic cholinergic receptors in the dorsomedial striatum in task switching. Neurobiol. Learn. Mem. 2008;89:114–124. doi: 10.1016/j.nlm.2007.06.005. [DOI] [PMC free article] [PubMed] [Google Scholar]
  108. McEnaney K.W., Butter C.M. Perseveration of responding and nonresponding in monkeys with orbital frontal ablations. J. Comp. Physiol. Psychol. 1969;68:558–561. doi: 10.1037/h0027639. [DOI] [PubMed] [Google Scholar]
  109. McKirdy J., Sussmann J.E.D., Hall J., Lawrie S.M., Johnstone E.C., McIntosh A.M. Set shifting and reversal learning in patients with bipolar disorder or schizophrenia. Psychol. Med. 2009;39:1289–1293. doi: 10.1017/S0033291708004935. [DOI] [PubMed] [Google Scholar]
  110. Merriam E.P., Thase M.E., Haas G.L., Keshavan M.S., Sweeney J.A. Prefrontal cortical dysfunction in depression determined by Wisconsin Card Sorting Test performance. Am. J. Psychiatry. 1999;156:780–782. doi: 10.1176/ajp.156.5.780. [DOI] [PubMed] [Google Scholar]
  111. Moss E.M., Harlow H.F. The role of reward in discrimination learning in monkeys. J. Comp. Physiol. Psychol. 1947;40:333–342. doi: 10.1037/h0058469. [DOI] [PubMed] [Google Scholar]
  112. Mullins G.P., Winefield A.H. The relative importance of responses to S+ and S− in simultaneous discrimination learning. Q. J. Exp. Psychol. 1979;31:329–338. doi: 10.1080/14640747908400731. [DOI] [PubMed] [Google Scholar]
  113. Murphy F.C., Sahakian B.J., Rubinsztein J.S., Michael A., Rogers R.D., Robbins T.W., Paykel E.S. Emotional bias and inhibitory control processes in mania and depression. Psychol. Med. 1999;29:1307–1321. doi: 10.1017/s0033291799001233. [DOI] [PubMed] [Google Scholar]
  114. Murray G.K., Cheng F., Clark L., Barnett J.H., Blackwell A.D., Fletcher P.C., Robbins T.W., Bullmore E.T., Jones P.B. Reinforcement and reversal learning in first-episode psychosis. Schizophr. Bull. 2008;34:848–855. doi: 10.1093/schbul/sbn078. [DOI] [PMC free article] [PubMed] [Google Scholar]
  115. Nilsson S., Ripley T.L., Somerville E.M., Clifton P.G. Reduced activity at the 5-HT(2C) receptor enhances reversal learning by decreasing the influence of previously non-rewarded associations. Psychopharmacology. 2012;224:241–254. doi: 10.1007/s00213-012-2746-5. [DOI] [PubMed] [Google Scholar]
  116. Nilsson S., Somerville E.M., Clifton P.G. Dissociable effects of 5-HT2C receptor antagonism and genetic inactivation on perseverance and learned non-reward in an egocentric spatial reversal task. PLoS ONE. 2013;8:e77762. doi: 10.1371/journal.pone.0077762. [DOI] [PMC free article] [PubMed] [Google Scholar]
  117. Nissen H.W. Description of the learned response in discrimination behavior. Psychol. Rev. 1950;57:121–131. [Google Scholar]
  118. Nonkes L.J.P., Maes J.H.R., Homberg J.R. Improved cognitive flexibility in serotonin transporter knockout rats is unchanged following chronic cocaine self-administration. Addict Biol. 2011 doi: 10.1111/j.1369-1600.2011.00351.x. [DOI] [PubMed] [Google Scholar]
  119. Nonkes L.J.P., van de Vondervoort I.I.G.M., de Leeuw M.J.C., Wijlaars L.P., Maes J.H.R., Homberg J.R. Serotonin transporter knockout rats show improved strategy set-shifting and reduced latent inhibition. Learn. Mem. 2012;19:190–193. doi: 10.1101/lm.025908.112. [DOI] [PubMed] [Google Scholar]
  120. Oscar-Berman M., Zola-Morgan S.M. Comparative neuropsychology and Korsakoff's syndrome, I.—Spatial and visual reversal learning. Neuropsychologia. 1980;18:499–512. doi: 10.1016/0028-3932(80)90152-9. [DOI] [PubMed] [Google Scholar]
  121. Ouden, den H.E.M., Daw N.D., Fernandez G., Elshout J.A., Rijpkema M., Hoogman M., Franke B., Cools R. Dissociable effects of dopamine and serotonin on reversal learning. Neuron. 2013;80:1090–1100. doi: 10.1016/j.neuron.2013.08.030. [DOI] [PubMed] [Google Scholar]
  122. Owen A.M., Roberts A., Hodges J.R., Summers B., Polkey C.E., Robbins T.W. Contrasting mechanisms of impaired attentional set-shifting in patients with frontal lobe damage or Parkinson's disease. Brain. 1993:1159–1175. doi: 10.1093/brain/116.5.1159. [DOI] [PubMed] [Google Scholar]
  123. Ozonoff S., Cook I., Coon H., Dawson G., Joseph R.M., Klin A., McMahon W.M., Minshew N., Munson J.A., Pennington B.F., Rogers S.J., Spence M.A., Tager-Flusberg H., Volkmar F.R., Wrathall D. Performance on Cambridge Neuropsychological Test Automated Battery subtests sensitive to frontal lobe function in people with autistic disorder: evidence from the collaborative programs of excellence in autism network. J. Autism Dev. Disord. 2004;34:139–150. doi: 10.1023/b:jadd.0000022605.81989.cc. [DOI] [PubMed] [Google Scholar]
  124. Ozonoff S., South M., Miller J.N. DSM-IV-defined Asperger syndrome: cognitive, behavioral and early history differentiation from high-functioning autism. Autism. 2000;4:29–46. [Google Scholar]
  125. Palencia C.A., Ragozzino M.E. The influence of NMDA receptors in the dorsomedial striatum on response reversal learning. Neurobiol. Learn. Mem. 2004;82:81–89. doi: 10.1016/j.nlm.2004.04.004. [DOI] [PubMed] [Google Scholar]
  126. Pantelis C., Barber F.Z., Barnes T.R., Nelson H.E., Owen A.M., Robbins T.W. Comparison of set-shifting ability in patients with chronic schizophrenia and frontal lobe damage. Schizophrn. Res. 1999;37:251–270. doi: 10.1016/s0920-9964(98)00156-x. [DOI] [PubMed] [Google Scholar]
  127. Parasuraman R. The attentive brain: issues and prospects. In: Parasuraman R., editor. The Attentive Brain. MIT Press; Cambridge: 1998. pp. 3–16. [Google Scholar]
  128. Pearce J.M., Hall G. A model for Pavlovian learning: variations in the effectiveness of conditioned but not of unconditioned stimuli. Psychol. Rev. 1980;87:532–552. [PubMed] [Google Scholar]
  129. Ragozzino M.E. The contribution of the medial prefrontal cortex, orbitofrontal cortex, and dorsomedial striatum to behavioral flexibility. Ann. N.Y. Acad. Sci. 2007;1121:355–375. doi: 10.1196/annals.1401.013. [DOI] [PubMed] [Google Scholar]
  130. Reeve W.V., Schandler S.L. Frontal lobe functioning in adolescents with attention deficit hyperactivity disorder. Adolescence. 2001;36:749–765. [PubMed] [Google Scholar]
  131. Reischies F.M. Pattern of disturbance of different ventral frontal functions in organic depression. Ann. N.Y. Acad. Sci. 1999;877:775–780. doi: 10.1111/j.1749-6632.1999.tb09321.x. [DOI] [PubMed] [Google Scholar]
  132. Remijnse P.L., Nielen M.A., van Balkom A.J.L.M., Cath D.C., van Oppen P., Uylings H.B.M., Veltman D.J. Reduced orbitofrontal–striatal activity on a reversal learning task in obsessive–compulsive disorder. Arch. Gen. Psychiatry. 2006;63:1225–1236. doi: 10.1001/archpsyc.63.11.1225. [DOI] [PubMed] [Google Scholar]
  133. Rescorla R.A., Holland P.C. Behavioral studies of associative learning in animals. Annu. Rev. Psychol. 1982;33:265–308. [Google Scholar]
  134. Robbins T.W., Roberts A.C. Differential regulation of fronto-executive function by the monoamines and acetylcholine. Cereb. Cortex. 2007;17(Suppl 1):i151–i160. doi: 10.1093/cercor/bhm066. [DOI] [PubMed] [Google Scholar]
  135. Roberts A.C. The importance of serotonin for orbitofrontal function. Biol. Psychiatry. 2011;69:1185–1191. doi: 10.1016/j.biopsych.2010.12.037. [DOI] [PubMed] [Google Scholar]
  136. Robinson O.J., Cools R., Carlisi C.O., Sahakian B.J., Drevets W.C. Ventral striatum response during reward and punishment reversal learning in unmedicated major depressive disorder. Am. J. Psychiatry. 2012;169:152–159. doi: 10.1176/appi.ajp.2011.11010137. [DOI] [PMC free article] [PubMed] [Google Scholar]
  137. Robinson O.J., Frank M.J., Sahakian B.J., Cools R. Dissociable responses to punishment in distinct striatal regions during reversal learning. NeuroImage. 2010;51:1459–1467. doi: 10.1016/j.neuroimage.2010.03.036. [DOI] [PMC free article] [PubMed] [Google Scholar]
  138. Robinson O.J., Standing H.R., DeVito E.E., Cools R., Sahakian B.J. Dopamine precursor depletion improves punishment prediction during reversal learning in healthy females but not males. Psychopharmacology. 2010;211:187–195. doi: 10.1007/s00213-010-1880-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
  139. Rossell S.L.S., David A.S.A. Improving performance on the WCST: variations on the original procedure. Schizophrn. Res. 1997;28:63–76. doi: 10.1016/s0920-9964(97)00093-5. [DOI] [PubMed] [Google Scholar]
  140. Rudebeck P.H., Saunders R.C., Prescott A.T., Chau L.S., Murray E.A. Prefrontal mechanisms of behavioral flexibility, emotion regulation and value updating. Nat. Publ. Group. 2013;16:1140–1145. doi: 10.1038/nn.3440. [DOI] [PMC free article] [PubMed] [Google Scholar]
  141. Sandson J., Albert M.L. Varieties of perseveration. Neuropsychologia. 1984;22:715–732. doi: 10.1016/0028-3932(84)90098-8. [DOI] [PubMed] [Google Scholar]
  142. Sasaki M. Influence of prereversal experience on reversal learning in white rats. Annu. Anim. Psychol. 1969;19:17–28. [Google Scholar]
  143. Scheggia D., Bebensee A., Weinberger D.R., Papaleo F. The ultimate intra-/extra-dimensional attentional set-shifting task for mice. Biol. Psychiatry. 2013:1–11. doi: 10.1016/j.biopsych.2013.05.021. [DOI] [PubMed] [Google Scholar]
  144. Schoenbaum G., Nugent S.L., Saddoris M.P., Setlow B. Orbitofrontal lesions in rats impair reversal but not acquisition of go, no-go odor discriminations. NeuroReport. 2002;13:885–890. doi: 10.1097/00001756-200205070-00030. [DOI] [PubMed] [Google Scholar]
  145. Schoenbaum G., Setlow B., Nugent S.L., Saddoris M.P., Gallagher M. Lesions of orbitofrontal cortex and basolateral amygdala complex disrupt acquisition of odor-guided discriminations and reversals. Learn. Mem. 2003;10:129–140. doi: 10.1101/lm.55203. [DOI] [PMC free article] [PubMed] [Google Scholar]
  146. Schultz W. Predictive reward signal of dopamine neurons. J. Neurophysiol. 1998;80:1–27. doi: 10.1152/jn.1998.80.1.1. [DOI] [PubMed] [Google Scholar]
  147. Shamay-Tsoory S.G., Shur S., Harari H., Levkovitz Y. Neurocognitive basis of impaired empathy in schizophrenia. Neuropsychology. 2007;21:431–438. doi: 10.1037/0894-4105.21.4.431. [DOI] [PubMed] [Google Scholar]
  148. Slabosz A., Lewis S.J.G., Smigasiewicz K., Szymura B., Barker R.A., Owen A.M. The role of learned irrelevance in attentional set-shifting impairments in Parkinson's disease. Neuropsychology. 2006;20:578–588. doi: 10.1037/0894-4105.20.5.578. [DOI] [PubMed] [Google Scholar]
  149. Slamecka N.J. A methodological analysis of shift paradigms in human discrimination learning. Psychol. Bull. 1968;69:423–438. doi: 10.1037/h0025762. [DOI] [PubMed] [Google Scholar]
  150. Spence K.W. The nature of discrimination learning in animals. Psychol. Rev. 1936;43:427–449. [Google Scholar]
  151. Spence K.W. The differential response in animals to stimuli varying within a single dimension. Psychol. Rev. 1937;44:430. [Google Scholar]
  152. Stalnaker T.A., Cooch N.K., Schoenbaum G. What the orbitofrontal cortex does not do. Nat. Neurosci. 2015;18:620–627. doi: 10.1038/nn.3982. [DOI] [PMC free article] [PubMed] [Google Scholar]
  153. Stevens D.A., Fechter L.D. Relative strengths of approach and avoidance tendencies in discrimination learning of rats trained under two types of reinforcement. J. Exp. Psychol. 1968;76:489–491. doi: 10.1037/h0025553. [DOI] [PubMed] [Google Scholar]
  154. Sutherland N.S., Mackintosh N.J. Academic Press; London: 1971. Mechanisms of Animal Discrimination Learning. [Google Scholar]
  155. Tait D.S., Brown V.J. Difficulty overcoming learned non-reward during reversal learning in rats with ibotenic acid lesions of orbital prefrontal cortex. Ann. N.Y. Acad. Sci. 2007;1121:407–420. doi: 10.1196/annals.1401.010. [DOI] [PubMed] [Google Scholar]
  156. Tait D.S., Chase E.A., Brown V.J. Attentional set-shifting in rodents: a review of behavioural methods and pharmacological results. Curr. Pharm. Des. 2014;20:5046–5059. doi: 10.2174/1381612819666131216115802. [DOI] [PubMed] [Google Scholar]
  157. Taylor Tavares J.V., Clark L., Cannon D.M., Erickson K., Drevets W.C., Sahakian B.J. Distinct profiles of neurocognitive function in unmedicated unipolar depression and bipolar II depression. Biol. Psychiatry. 2007;62:917–924. doi: 10.1016/j.biopsych.2007.05.034. [DOI] [PubMed] [Google Scholar]
  158. Terrace H.S. Escape from S−. Learn. Motiv. 1971;2:148–163. [Google Scholar]
  159. Terrace H.S.H. Discrimination learning with and without “errors”. J. Exp. Anal. Behav. 1963;6:1–27. doi: 10.1901/jeab.1963.6-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
  160. Ursu S., Carter C.S. Outcome representations, counterfactual comparisons and the human orbitofrontal cortex: implications for neuroimaging studies of decision-making. Cognit. Brain Res. 2005;23:51–60. doi: 10.1016/j.cogbrainres.2005.01.004. [DOI] [PubMed] [Google Scholar]
  161. Vaughter R.M., Cross H.A. Discrimination reversal performance in children as a function of prereversal experience and overlearning. Psychon. Sci. 1965;2:363–364. [Google Scholar]
  162. Waltz J.A., Frank M.J., Robinson B.M., Gold J.M. Selective reinforcement learning deficits in schizophrenia support predictions from computational models of striatal-cortical dysfunction. Biol. Psychiatry. 2007;62:756–764. doi: 10.1016/j.biopsych.2006.09.042. [DOI] [PMC free article] [PubMed] [Google Scholar]
  163. Watkins L.H., Sahakian B.J., Robertson M.M., Veale D.M., Rogers R.D., Pickard K.M., Aitken M.R.F., Robbins T.W. Executive function in Tourette's syndrome and obsessive–compulsive disorder. Psychol. Med. 2005;35:571–582. doi: 10.1017/s0033291704003691. [DOI] [PubMed] [Google Scholar]
  164. Worbe Y., Palminteri S., Savulich G., Daw N.D., Fernandez-Egea E., Robbins T.W., Voon V. Valence-dependent influence of serotonin depletion on model-based choice strategy. Mol. Psychiatry. 2015 doi: 10.1038/mp.2015.46. [DOI] [PMC free article] [PubMed] [Google Scholar]

RESOURCES