Abstract
Functions associated with processing reward-related information are fundamental drivers of motivation, learning, and goal-directed behavior. Such functions have been classified as the positive valence system under the Research Domain and Criteria (RDoC) criteria and are negatively impacted across a range of psychiatric disorders and mental illnesses. The positive valence system is composed of three comprehensive categories containing related but dissociable functions that are organized into either Reward Responsiveness, Reward Learning, or Reward Valuation. The presence of overlapping behavioral dysfunction across diagnostic mental disorders is in-part what motivated the RDoC initiative, which emphasized that the study of mental illness focus on investigating relevant behavior and cognitive functions and their underlying mechanisms, rather than separating efforts on diagnostic categories (i.e., transdiagnostic). Moreover, the RDoC approach is well-suited for preclinical neuroscience research, as the rise in genetic toolboxes and associated neurotechnologies enables researchers to probe specific cellular targets with high specificity. Thus, there is an opportunity to dissect whether behaviors and cognitive functions are supported by shared or distinct neural mechanisms. For preclinical research to effectively inform our understandings of human behavior however, the cognitive and behavioral paradigms should have predictive, neurobiological, and pharmacological predictive validity to the human test. Touchscreen-based testing systems provide a further advantage for this endeavor enabling tasks to be presented to animals using the same media and task design as in humans. Here, we outline the primary categories of the positive valence system and review the work that has been done cross-species to investigate the neurobiology and neurochemistry underlying reward-related functioning. Additionally, we provide clinical tasks outlined by RDoC, along with validity and/or need for further validation for analogous rodent paradigms with a focus on implementing the touchscreen-based cognitive testing systems.
Graphical abstract text:

This work reviews cross-species paradigms with validity for testing positive valence systems used in clinical populations, identified by the National Institute for Mental Health (NIMH) Research Domain Criterion (RDoC) initiative. Tests available for rats and mice are discussed, including whether they have evidence for face, predictive, or neurobiological validities. A focus is placed on touchscreen tests, but all operant tasks are included in the discussion, and future tests are recommended.
Introduction
Implementing RDoC in preclinical research
Psychiatric disorders continue to be amongst the leading causes of morbidity in today’s society, contributing to the diminished quality of life for millions of afflicted individuals and their caregivers (Rapaport, Clary, Fayyad, & Endicott, 2005; Vigo, Thornicroft, & Atun, 2016). Such conditions, also referred to as mental disorders or mental illness, are traditionally described using the diagnostic categories included in the Diagnostic and Statistical Manual of Mental Disorders (DSM), currently in its fifth edition (DSM-V) (Diagnostic and statistical manual of mental disorders: DSM-5™, 5th ed, 2013). These conditions include schizophrenia, bipolar disorder, major depression disorder, attention deficit hyperactive disorder, and anxiety disorder. While the DSM has been fundamental in identifying sufferers and their treatment, research, and acknowledgement of their legitimate healthcare concerns, the nature of psychiatric disturbances provides a number of challenges for assignment into distinct diagnostic categories. Firstly, diagnoses in psychiatry rely exclusively on the observation of behavioral/cognitive symptoms and subjective reporting, as there are currently no physical or chemical signatures of neural insult that are diagnostic (Insel & Cuthbert, 2015). Such diagnoses are complicated by the multi-faceted nature of symptomology between disorders, which includes a high heterogeneity of symptom presentation and severity within individual diagnostic categories. Furthermore, psychiatric disorders display extensive comorbidity, and the presence of shared symptomology further blurs the line between conditions (S. E. Morris & Cuthbert, 2012). For example, cognitive impairments such as executive dysfunction and abnormalities in reward processing (e.g., anhedonia) are present to some degree in most if not all psychiatric disorders. Even though psychotic symptoms are the characteristic manifestation of schizophrenia, they are not unique to the disorder (Dunayevich & Keck Jr, 2000). For example, over half of individuals with bipolar disorder will also experience psychotic episodes (Pope & Lipinski, 1978{Goodwin, 2007 #967). Distinct diagnostic categorization does not therefore necessarily equate to distinct behavioral profiles between groups.
Unsurprisingly, multiple genome-wide association studies (GWAS) and cross-disorder gene analyses also revealed shared genetic risk factors for psychiatric disorders that span across the boundaries of diagnostic categories (Andreassen, Hindley, Frei, & Smeland, 2023; B. Consortium et al., 2018; C.-D. G. o. t. P. G. Consortium, 2013; O'Donovan & Owen, 2016). Additionally, the presence of certain copy number variant (CNV) mutations confers a substantial risk factor for developing schizophrenia, and to a lesser degree bipolar disorder (W. T. C. C. Consortium et al., 2009; Grozeva, 2010) but have also been linked to neurodevelopmental disorders such as autism spectrum disorder and intellectual disability (Craddock & Owen, 2010; Fiksinski et al., 2017; Kushima et al., 2018). Moreover, there is evidence for shared pathophysiological mechanisms, such as abnormal protein expression, across psychiatric and neurodegenerative diseases (e.g. Alzheimer’s, Parkinson’s, schizophrenia, bipolar disorder), raising the question of whether research and treatment should be focusing on establishing the relationship between general pathophysiology and function outcome, rather than distinct diagnostic categories (Wingo et al., 2022).
These concerns were the genesis of the Research Domain and Criteria (RDoC) initiative, which emphasized focus on the mechanisms underlying cognitive, behavioral, and emotional functions, irrespective of specific diagnostic categories (Cuthbert, 2022; Cuthbert & Kozak, 2013; Insel et al., 2010). Instead, the RDoC matrix outlines multiple functional domains, including cognitive, arousal and regulatory, sensorimotor, social, negative valence, and positive valence systems; the final category being the focus of this review. The development of RDoC represents a conceptual paradigm shift, which transitions emphasis from disorder-based/diagnostic categories to characterizing psychiatric abnormalities in relation to neurological functions (Ayoub, Noback, Ahern, & Young, 2024; Insel et al., 2010; S. E. Morris & Cuthbert, 2012). Thus, by elucidating how aberrant genes or circuitry contribute to specific behavioral/emotional/cognitive functions, it may be possible to shed light onto how genetic predispositions may render vulnerability to a number of [often overlapping] mental illnesses (Kozak & Cuthbert, 2016; Sanislow et al., 2010). This approach would facilitate the characterization of how cognitive functions and behaviors are similar from a mechanistic point of view, but also how they differ, providing the opportunity to identify potential biomarkers of symptomology in disease (Nusslock & Alloy, 2017). From a preclinical research perspective, this approach likely provides a more sustainable avenue for successful translation of findings (Young, 2023; Young, Winstanley, Brady, & Hall, 2017). As informed by GWAS, genetically modified rodent models are developed to recapitulate identified genetic risk factors for psychiatric disorders and are not expected to replicate the entire symptomology of a given diagnostic category (Geyer & Markou, 1995; Markou, Chiamulera, Geyer, Tricklebank, & Steckler, 2009). For example, psychotic symptoms such as delusions and hallucinations are staples of psychiatric diseases and are currently unable to be accurately assessed in rodents (Kotov et al., 2024). Despite this limitation, identified genetic risk factors can be reproduced in animal models to assess their impact on other afflicted domains, such as cognitive or negative symptoms. Thus, preclinical models are an effective tool for dissecting the functional outputs of certain genes or circuit manipulations, despite not being able to replicate all of the facets of clinical disorders. This review provides an overview of the positive valence systems, their dysfunction in disease, and how the component functions are assessed in clinical and preclinical research. Although RDoC’s domains include various paradigms for assessing sub-constructs of the positive valence system, we will focus our discussion on the human paradigms that have, or have the potential to develop, analogous paradigms that available for rodents, thus allowing for cross-species validation to be assessed. Given the rise of touchscreen-based cognitive testing in preclinical research, we evaluate how these systems and similar operant paradigms can be implemented to facilitate cross-species cognitive testing in the positive valence domain.
Positive valence systems
Assessing valence systems in preclinical research
A fundamental driver of animal behavior is to seek out experiences that are pleasurable or desirable and avoid those that are painful or aversive (Skinner, 1963; Thorndike, 1911). Reward and punishment exist as opposing behavioral outcomes and along a continuum of intensity, enabling animals to form associations between stimuli, actions, and outcomes to inform future decisions and make predictions about the external world. In the simplest sense, behaviors are reinforced if they lead to favorable outcomes (positive reinforcement) and suppressed if they lead to negative outcomes (positive punishment) (Keller & Schoenfeld, 1950). Alternatively, behaviors can be suppressed under anticipation so that they will result in the loss of a reward (negative punishment) or facilitated if they will aid in avoiding punishment (negative reinforcement) (Gray, Stafford, & Tallman, 1991; Holland & Skinner, 1961; Murray Sidman, 1962; M. Sidman, 2006). As knowledge of past experiences is integrated with internal drives, valence (i.e., the subjective value) is attributed to environmental stimuli and used to guide goal-directed behavior (Miller & Cohen, 2001). The constructs that underlie responses to aversive and reinforcing stimuli are organized into two categories, the negative and positive valence systems, respectively. The negative valence systems, also known as avoidance systems, refer to functions associated with avoiding and preventing the experience of aversive stimuli. Behaviors associated with processing negative valence include acute, chronic, and future threat avoidance (e.g., fear and anxiety, respectively); as well as the experience of negative emotions (e.g., grief), and the loss of rewarding experiences. Opposingly, positive valence systems, or approach systems, include constructs that promote seeking rewarding experiences, the positive emotions associated with them, and the use of reward to facilitate learning and habit formation (Böttger et al., 2023; Taylor, Pearlstein, & Stein, 2020; Woody & Gibb, 2015).
Dysfunctional processing of positive and/or aversive stimuli are believed to underlie depression-like symptoms observed across mood and psychiatric disorders, where a reduction in positive feelings and motivation to obtain rewards may stem from increases in negative affect and/or a loss of positive affect (Dieterich et al., 2019). Due to the widespread prevalence and societal burden of mood dysregulation, determining the underlying neurobiology and potential therapeutic options is a massive initiative in preclinical research. Our focus will be on functions associated with the positive valence system such as perceiving and integrating reward value, using value representations to learn and make decisions, and the motivation to obtain rewards, which offer an alternative avenue for investigating the mechanism underlying valence processing in a preclinical setting.
The positive valence system as outlined by RDoC consists of three interconnected primary categories that are each composed of multiple sub-constructs. The overarching domains include reward responsiveness, reward learning, and reward valuation, which cover the underlying physiological processing and behavioral outputs that facilitate the motivation for obtaining, and the response to, positive stimuli. If the behavioral endpoint is to seek out or approach positive experiences appropriately, each component of the positive valence system yields an essential function that is integrated amongst the other constructs to drive behavior (Fig. 1). An individual must process positive stimuli as such and have a desire to obtain it, which compose the fundamental building blocks of reinforcement, “liking” and “wanting”, respectively. These constructs are included in reward responsiveness, which provides the foundation for positive stimuli to be used to guide learning and decision-making. Through reward learning, external stimuli are attributed salience as associations are made between actions and rewarding outcomes, and representations are formed that enable predictions to be made about future rewarding situations. Representations of a reward’s “worth”, which is also influenced by temporary internal states (e.g., hungry or satiated), will further inform the degree to which an individual will work to obtain a given reward, which composes reward valuation. A reward’s value guide cost versus benefit-based decision-making, which is necessary for optimizing behavior based on the given context (Salamone & Correa, 2002). While the categories are distinct in the functions they describe, they do not operate in isolation, and the interplay between these constructs is necessary for optimal behavior (Figure 1). Thus, investigating positive valence in preclinical research requires sensitive and selective behavioral paradigms that are able to tease apart discrete behavioral manifestations of the underlying processes, especially if they are to inform human testing.
Figure 1.

Outline of the positive valence system constructs and associated tasks outlined by RDoC. The three primary categories are organized by color and include the primary sub-constructs in corresponding-colored boxes (Reward responsiveness – Blue; Reward learning – Green; Reward valuation – Red). The hierarchal organization (i.e., Top to bottom) represents the dependence that higher order constructs (e.g., reward learning and valuation) have on the foundation of fundamental responsiveness to rewarding stimuli. The direction of the dark arrows shows constructs that directly influence the receiving function. For example, the degree to which one is satiated with a particular reward will influence both the intensity of reward anticipation and how the individual will respond to presentation of the reward. The dotted arrows demonstrate mechanisms of feedback that can influence future responses to reward or updating reward-related representations. For example, changes in the environmental context can lead to reward prediction errors, which are used through reinforcement learning to update the relationship between relevant environmental stimuli and their association with rewarding outcomes and will influence future interactions with situations that were previously associated with reward. * denote preclinical paradigms that require further validation.
Implementing translational cognitive paradigms to probe positive valence systems
To study the mechanisms underlying cognition and behavior, conducting preclinical research in rodents provides an unmatched avenue to probe specific genes and circuits with high specificity and temporal precision to determine their roles in behavior. For this potential to be realized however, tasks used in preclinical research must be translatable to those used in human clinical settings (Grottick et al., 2021; Palmer et al., 2021). This concept, particularly reinforced by the high failure rate from preclinical to clinical stages of drug treatments for brain disorders (Mullard, 2016), has inspired the development of cross-species cognitive and behavioral paradigms that emphasize translational validity. The evolution of clinical testing apparatuses has progressed from pen and paper-based testing to primarily computerized automated systems, utilizing mouse and keyboards, joysticks, and touchscreens. Similarly, the development of automated behavioral paradigms for rodents enables behavioral testing with minimal experimenter influence and protocol deviation, utilizing systems with levers, nose-holes, or touchscreens. For cross-species task development, computerized systems can be used to adapt human paradigms for testing in rodents (reverse translation e.g., (Perry et al., 2009)) or create human variants of animal tasks (forward translation e.g.,(Young, Light, Marston, Sharp, & Geyer, 2009)).
The RDoC matrix includes six domains that are composed of distinct, but related, functional categories, aligning with the concept that certain mental processes possess overlap in terms of underlying neural mechanisms and their contributions to functional behavior. By having standardized testing equipment that can run various cognitive tests on the same platform, task batteries can be developed that probe distinct cognitive and behavioral functions with minimal off-target differences across tasks. A number of cross-species task batteries have already been developed and proposed, including the Cambridge Automated Neuropsychological Test Battery (CANTAB) (T. W. Robbins et al., 1998; T.W. Robbins et al., 1994), the NIMH’s Cognitive Neuroscience Treatment Research to Improve Cognition in Schizophrenia (CNTRICS) (Barch et al., 2009; Carter et al., 2011; Young, Jentsch, Bussey, Wallace, & Hutcheson, 2013), the Novel Methods leading to New Medications in Depression and Schizophrenia (NEWMEDS) (Hvoslef-Eide et al., 2015), and the EMOTICOM test battery (Roiser & Sahakian, 2013). (Bland et al., 2016) (Savulich et al., 2018). Although rodent operant and touchscreen systems offer similar advantages in terms of standardization, automation, and opportunities for task development, the rise of touchscreen-based technology may offer substantial benefits for minimizing task differences across species (See Box 1).
Box 1:
Aspects of the touchscreen systems that are well-suited for studies aimed at investigating positive valence system functions.
One of the largest advantages of touchscreen-based systems is the ability to provide more varied stimuli and additionally identical information to human and rodents subjects (e.g., visual images) that require similar motor responses (i.e., finger touch or nose-poke), maximizing face validity (Bussey et al., 2008). An additional feature of the touchscreen apparatus that is well-suited to assessing positive valence functions in preclinical research in particular is the low baseline motoric demand that is associated with nose-poking the touch-sensitive screen. The systems are equipped with a series of infra-red beams that line the front of the screen, and simply require a mouse to break the beam path to generate a response. The low physical requirements have been previously described as beneficial for assessing cognition in models of neurodegeneration or damage that are associated with impaired motor function, such as stroke, Parkinson’s or Huntington’s disease (Bussey et al., 2008; Dumont, Salewski, & Beraldo, 2021; Houlton, Barwick, & Clarkson, 2021; Morton, Skillings, Bussey, & Saksida, 2006; Oomen et al., 2013). In relation to reward-related processes, a key domain is the degree to which one will exert effort to achieve rewarding stimuli, based on the internal reward valuation. As a result, lower baseline effort demands enables a more gradual progression span from low to high demands, providing a larger scale for the assessment of behavior. Moreover, the touchscreens provide flexibility for how effort can be modulated, such as increasing the number of screen responses to get a reward (e.g., Progressive ratio breakpoint (Hailwood, Heath, Robbins, Saksida, & Bussey, 2018; Heath, Bussey, & Saksida, 2015) or by making the stimuli more challenging to interact with by adjusting its location on the screen (e.g., Rearing effort discounting (Lopez-Cruz et al., 2024)) (See category 3). In contrast, a paradigm with a high baseline effort requirement, such as climbing over a barrier to collect a reward, may be less sensitive to subtle changes in motivation. Furthermore, if effort to achieve reward is not the primary measure of focus, higher physical requirements (e.g., learning a location-reward association in a maze) may act as a confound for manipulations aimed at alternative reward functions and not match onto human efforts. As a result, it may be more challenging to identify if a manipulation is affecting reward learning processes (i.e., the intended function) or instead altering the degree of effort an animal is willing to expend to explore the maze and acquire the association, particularly for relevant to human testing. While various behavioral apparatuses have been instrumental for informing our understanding of reward-related functions, the touchscreen systems are well equipped to investigate effort-related reward valuation (Heath, Phillips, Bussey, & Saksida, 2016; Markou et al., 2013; Phillips, Lopez-Cruz, et al., 2018; Young & Markou, 2015).
By maximizing the translational potential of behavioral paradigms, there is opportunity to implement manipulations and recording techniques to identify and characterize neural biomarkers across species underlying disease-relevant functions (Cavanagh et al., 2021; Cavanagh et al., 2022; Noback et al., 2024). Although it is challenging to probe specific circuits and cellular populations in humans, the availability of techniques with cellular and temporal precision readily available for use in rodents. For cellular monitoring, optical techniques enable proxies of neural activity, such as genetically encoded calcium indicators (GECI) (Cui et al., 2013; Flusberg et al., 2008; Lütcke et al., 2010), voltage indicators (GEVIs) (Akemann et al., 2013; Dimitrov et al., 2007), biosensors of neuromodulators (Jing et al., 2018; Patriarchi et al., 2018; H. Wang, Jing, & Li, 2018), and in vivo electrophysiology, to be readily imaged in freely moving rodents. For cellular manipulations, the toolbox of available optogenetic opsins and methods of expressing in target cell populations has grown exponentially, enabling cellular inhibition and activation with various degrees of membrane kinetics and technical application (Bernstein & Boyden, 2011; Bernstein, Garrity, & Boyden, 2012; Förster, Dal Maschio, Laurell, & Baier, 2017; Madisen et al., 2012; Rost, Schneider-Warme, Schmitz, & Hegemann, 2017). Thus, there are a number of avenues for investigating correlative and causative relationships between cellular targets and temporally-specific behavioral events, which is strengthened through combination with automated paradigms possessing cross-species validity. Importantly, such an integrated approach aligns with RDoC’s emphasis on investigating the mechanisms underlying distinct cognitive and behavioral functions, providing insight into potential cellular dysfunctions relevant to psychiatric disorders (See Table 1).
Table 1.
Overview of studies that have implemented optical-based techniques to study mechanisms of the positive valence system in freely moving mice.
| Techniques | Positive Valence Domain | Cellular Target | Specificity | Findings |
|---|---|---|---|---|
|
| ||||
| Calcium Imaging | Reward Learning (Pavlovian) | D1R and D2R medium spiny neurons (NAc) | Population level recording | D1R late learning: ↓ activity during S+ approach and ↑ activity following reward consumption. |
| D2R late learning: ↑ activity following reward delivery. (Skirzewski et al., 2022). | ||||
| Reward Learning (Operant conditioning) | CAMKii-expressing neurons (BLA) | Population level recording | Early learning: ↑ BLA learning at reward deliverxy | |
| Late learning: ↑ BLA activity at reward-predicting cue (Crouse et al., 2020). | ||||
| Reward learning (Probabilistic) | Glutamatergic neurons (vmOFC) | Population level recording | Early learning:↑ activity following informed (reward or not) feedback. | |
| Late learning: ↓ activity following informed feedback (Barnes et al., 2023). | ||||
|
| ||||
| Neurotransmitter Imaging (Fiber photometry) | Reward Learning (Pavlovian) | Dopamine (NAc) | Population level recording | Early learning: ↑ DA release following reward delivery. |
| Late learning: ↑ DA activity during S+ approach. ↓ DA release following reward delivery (Skirzewski et al. 2022). | ||||
| Acetylcholine (NAc) | Population level recording | Late learning: ↓ Ach release following reward (Skirzewski et al. 2022). | ||
| Reward Learning (Operant conditioning) | Acetylcholine (BLA) | Population level recording | Early learning: ↑ Ach activity at reward delivery. | |
| Late learning: ↑ Ach activity at reward-predicting cue (Crouse et al. 2020). | ||||
|
| ||||
| Mini-scopes | Reward Learning (Simple discrimination) | D1R and D2R medium spiny neurons (NAc) | Individual cellular resolution | Heterogeneity within both populations, particularly in D1R neurons. D2R neuron activity primarily ↑ following errors, and ↓ following correct responses. Both signals were strengthened through learning (Nishioka et al., 2023). |
| Reward Valuation (Operant progressive lever pressing) | ACC excitatory neurons | Individual cellular resolution | ↓ ACC activity prior to progressive lever pressing when rewards are freely available (E. E. Hart, Blair, O'Dell, Blair, & Izquierdo, 2020). | |
|
| ||||
| Optogenetics | Reward Valuation (Delayed discounting) | Aldh1a1 GABAergic interneurons (VTA) | ↑ Input to VTA | ↑ Preference for larger, delayed reward (X. Li et al., 2021). |
| Reward Learning (Simple discrimination) | D1R medium spiny neurons (NAc) | Direct Activation | ↑ Preference for stimulus paired with NAc activation (Attachaipanich, Ozawa, Macpherson, & Hikida, 2023) | |
| Reward Learning (Deterministic learning and reversal) | VTA neurons projecting to NAc core | Direct Activation | ↓ Early reversals; ↓ win-stay, ↑ lose-shift, ↑ lose-stay (A. K. Radke et al., 2019) | |
| D1R and D2R medium spiny neurons (NAc) | Direct Inhibition during 1) Post-errors; 2) ITI | D2R post-error: ↓ correct choices; D1R during ITI: ↓ correct choices (Niskioka et al. 2023). | ||
| Reward Learning (Operant conditioning) | Cholinergic NBM-BLA terminals | Direct activation during learning | ↑ Correct choices throughout task learning (Crouse et al. 2020). | |
| Reward Learning (Probabilistic) | Glutamatergic neurons (vmOFC) | Direct activation | ↓ Total number of reversals and ↓ win-stay choices. ↑ Total number of reversals (Barnes et al., 2023). ↑ Number of lever presses to obtain a reward (i.e.., ↑ breakpoint) (Zhou et al., 2022). |
|
In this review, we will discuss the three functional categories that comprise the positive valence system and how these functions are or can be tested using touchscreen-based cognitive tasks in humans and rodents. To facilitate the translation of cognitive and behavioral paradigms across species, tasks must provide multiple forms of validity. Most, if not all, will have face validity (phenomenological similarity), more importantly, however, are predictive (including pharmacological), and neurobiological validities (see (Young, 2023)). Predictive validity is the capacity by which a task can predict the outcome of a manipulation in animals as it would occur in humans. Pharmacological predictive validity is more specific and refers to the ability of a task to recreate drug-induced changes in performance consistently across animals and humans. Neurobiological validity refers to confirmation that the same neurobiological circuits underlie the performance of a task in animals and humans. Thus, we will describe the evidence for the validity of cross-species tasks to assess domains within the positive valence system with an emphasis on touchscreen-based paradigms. Finally, we will describe the neurobiology and chemistry underlying the various functions associated with reward-related processing and how dysfunction of these systems can manifest as pathological cognition and behavior.
Category 1: Reward Responsiveness
Deciphering the neural processes involved in reward responsiveness requires tasks that are less cognitively based so-as to refrain from conflation of data with responses elicited by other positive valence domains (i.e., reward learning). Importantly, the components of reward responsiveness provide the foundation for all functions in which reward can be used to modulate current and future behavior, making the ability to properly isolate and test these constructs imperative for research.
Reward anticipation
Reward anticipation shapes future goal-directed behavior, as evidenced by drug-paired cues driving drug-taking and drug-seeking responses across species. The RDoC initiative recommends the computerized monetary incentive delay (MID) task (Knutson, Westdorp, Kaiser, & Hommer, 2000) to measure reward anticipation as it effectively segregates such responses from other reward processes. The MID task was created to dissect neural responses at distinct reward stages in participants performing a simple (i.e., low cognitive demand) instrumental response (Knutson et al., 2000). In this task dissociable anticipatory win-cues signal trials where a participant can earn money after a simple reaction-timed response (Figure 2). Neutral cues instead signal trials where reward cannot be earned, though participants still perform the reaction-timed response. Responses are followed by outcome feedback (e.g., win or loss), enabling distinct incentive -anticipation and -outcome phases. Importantly, the response window is adjusted based on individual reaction-times to reduce potential performance and/or practice effects (Knutson et al., 2000). Thus, while anticipatory win-cues reduce reaction-time relative to neutral cues (Dhingra et al., 2020; Knutson, Fong, Adams, Varner, & Hommer, 2001), studies utilizing this task largely focus on neural recordings, rather than behavior. The inherent low motor and cognitive demand required makes the MID task easy to perform during more complex neurophysiological assessment.
Figure 2.
The monetary incentive delay (MID) task for humans (left) and its’ equivalent for rodents (right). Measuring reward anticipation requires that MID tasks be paired with neurological recording measures, such as fMRI in humans and regional oxygen or calcium imaging in rodents. Subjects learn to associate a win-cue with a rewarding outcome, providing an anticipatory window between cue and reward receipt. Thus, while MID tasks include a component of reward learning, the objective measurement is the neurological correlates of an anticipated reward elicited by an associative cue, rather than specific behavioral endpoints. For rodent research, MIDs can be translated using discriminative stimulus tasks, where an animal is trained to associate a discriminable cue with a rewarding outcome and neural activity can be measured during the anticipatory period between cue and reward receipt.
Early work utilizing the MID task in combination with functional magnetic resonance imagining (fMRI) revealed activation of the ventral striatum and ventromedial prefrontal cortex (vmPFC) by reward anticipation and reward outcome, respectively (Knutson et al., 2001) (Table 2). These data called attention to differential regional recruitment during distinct phases of reward processing and highlighted the ability to temporarily sparse apart neural correlates of reward anticipation. Despite these distinctions, however, subsequent work suggested these processes are entwined, since people with vmPFC lesions demonstrated reduced ventral striatal volume and responsivity to reward anticipation when compared to healthy controls (Pujara, Philippi, Motzkin, Baskaya, & Koenigs, 2016). The MID task has since been adapted to explore the anticipation of other incentives, including social reward (Martins et al., 2021; Rademacher et al., 2010; Spreckelmeyer et al., 2009), food reward (J. J. Simon et al., 2015a), drugs (Nestor & Ersche, 2023), and sexually explicit content (Knutson, Wimmer, Kuhnen, & Winkielman, 2008; Markert, Klein, Strahler, Kruse, & Stark, 2021), emphasizing the flexibility of this task to probe reward anticipation across a variety of reinforcing stimuli. Further fMRI in the these tasks consistently highlighted striatal activation in response to anticipatory win-cues, while also extending more widespread cortical and subcortical activation to reward anticipation, including areas of the cingulate cortex, insula, thalamus, and amygdala (Chen, Chaudhary, & Li, 2022; Dhingra et al., 2020; Dugre, Dumais, Bitar, & Potvin, 2018; Oldham et al., 2018; Wilson et al., 2018). MID task variations include monetary loss trials signaled by alternative cues, creating an opportunity to explore loss anticipation (Knutson et al., 2000). Efforts to parse apart distinct networks associated with incentive versus loss anticipation have revealed overlapping regional recruitment across positive and negative valences, with some contrasts (Chen et al., 2022; Oldham et al., 2018; Wilson et al., 2018). Given our focus on positive valence systems, loss anticipation will not be further discussed.
Table 2.
Evidence for cross-species validation of tasks that assess reward anticipation
| Reward Responsiveness – Reward Anticipation | ||
|---|---|---|
|
| ||
| Validity Domain | Human Task – Monetary Incentive Delay | Rodent Task – Discriminative Stimulus Task |
|
| ||
| Predictive |
Bipolar and stimulant use disorders: ↓ striatal activity, ↓ fronto-temporal-limbic connectivity during anticipation (Nestor et al. 2023; Johnson et al. 2019; Patel et al. 2013; Wakatsuki et al. 2022). Schizophrenia: ↓ striatal activation during anticipation. |
N/A |
|
| ||
| Neurobiological |
During anticipation: ↑ activity in NAc (Knutson et al. 2001), striatum, cingulate cortex, insula, thalamus, amygdala (Dhingra et al. 2020; Chen et al. 2022; Oldham et al. 2018; Wilson et al. 2018; Dugre et al. 2018), substantia nigra, and VTA (Schott et al. 2018). During anticipation: ↑ DA release in ventral striatum (Schott et al. 2008). |
Following S+ response: ↑ O2 signals in NAc (Francois et al. 2012). During S+ approach: ↑ NAc DA release, tuned during learning (Skirzewski et al. 2022). |
|
| ||
| Pharmacological |
Amphetamine: Prolongs ventral striatal tonic DA release; ↓ phasic DA release (Knutson et al. 2004). Ketamine: ↓ ventral striatum response during anticipation (Francois et al. 2016). |
Ketamine: ↓ ventral striatum response during anticipation (Francois et al. 2016). |
Providing clinical sensitivity, the MID task captures altered neural responses during reward anticipation in neuropsychiatric populations with known reward processing disturbances. For example, compared to healthy participants, people with bipolar or stimulant-use disorder exhibited blunted striatal activation and reduced fronto-temporal-limbic connectivity during reward anticipation (S. L. Johnson, Mehta, Ketter, Gotlib, & Knutson, 2019; Nestor & Ersche, 2023; Patel et al., 2013; Wakatsuki et al., 2022). Similarly, striatal activation during reward anticipation correlated with psychometric symptomology in people with schizophrenia (Juckel et al., 2006; Nielsen et al., 2012; Zeng et al., 2022) and was inversely related with impulsivity as measured by the Barratt Impulsiveness Scale in pathological gamblers and detoxified problem drinkers (Balodis et al., 2012; Beck et al., 2009). As RDoC highlights, abnormalities in reward-related behavioral functions and associated neurophysiological processes span across diagnostic categories in support for a systems and circuitry-based approach for understanding behavioral dysfunction. Hence, the MID task can be a powerful transdiagonistic tool to study the mechanisms and treatment of reward anticipation dysregulation across neuropsychiatric disorders.
Mesolimbic dopamine is highly implicated in reward anticipation. Early studies in nonhuman primates suggested repeated exposure to a reward shifts phasic midbrain dopamine release following the reward to cues that predict reward (W. Schultz, Dayan, & Montague, 1997), replicated in rodents (Day, Roitman, Wightman, & Carelli, 2007). Similarly, in humans performing the MID task, combined fMRI and positron emission tomography (PET) revealed reward anticipatory cues activated dopamine-rich midbrain regions which correlated with ventral striatal dopamine release upon reward receipt (Schott et al., 2008). Additionally, amphetamine hemodynamically altered striatal activation during reward anticipation in healthy participants, in a manner that suggested a blunted phasic, yet prolonged tonic, dopaminergic response (Knutson et al., 2004). Because this task has minimal behavioral requirements, temporal precision to track sub-second fluctuations in neural dynamics is required to delineate neural structures further. These temporally precise neurotechnologies are invasive, however, requiring animal models. For example, fiber photometry and optogenetics could determine distinct neuromodulators and implicated receptor systems that contribute to reward anticipation, and permutations relative to models of neuropsychiatric disorder. Further, integration of neural recordings with pharmacology manipulations can inform the development of targeted therapeutics to treat aberrant neural signaling.
There is no standardized animal equivalent to the human MID task, however discriminative stimulus (DS) tasks may be an avenue for the development of an analog. Importantly, food rewards in animals should not cause concern for interpretation of data across tasks, since the anticipation of food reward elicits striatal activity similar to monetary rewards in humans (J. J. Simon et al., 2015b). In DS tasks, animals are trained to discriminate between two cues that predict reward delivery (CS+), or non-reward (CS-) (Figure 2). Variations of these tasks either require an instrumental response following CS+ presentation to earn reward (Francois, Conway, Lowry, Tricklebank, & Gilmour, 2012) or they non-contingently deliver reward following CS+ presentation (Day et al., 2007), the former more closely resembling the MID task. For example, Francois and colleagues (Francois et al., 2012) trained rats to press a lever that extended into a chamber following the presentation of a CS+. Thus, like the MID task the anticipatory reward cue (CS+) preceded both the response for reward, as well as its delivery. Regional oxygen (O2) amperometry in this task revealed rats exhibited increased nucleus accumbens (NAc) O2 signals following a lever-press during the CS+, but not the CS-. Differential O2 signaling between the CS+ and CS- suggested this activity likely did not reflect motoric consequences of the task. Further, CS- responses did not result in reward delivery, thus reward receipt did not likely contribute to altered NAc activity. These results somewhat match those obtained by human fMRI MID task studies wherein reward anticipation increased striatal activation – highlighting cross-species neurobiological validity. In the same rodent DS task acute sub-anesthetic doses of ketamine attenuated ventral striatum responses to “reward anticipation” in rats, while also reducing such neural activation in humans performing the MID task (Francois et al., 2016), providing cross-species pharmacological predictive validity. Hence, DS tasks may provide a rodent task equivalent to the human MID task.
Importantly, DS tasks can be easily adapted for touchscreen applications in rodents, to further enhance research utility. In a touchscreen version of the task, animals could be exposed to visual cues, as done in humans with the MID, rather than auditory cues to enhance consistency across species in this task. Further, the strength (i.e. Resemblance) of cues to the anticipatory win cues could be altered to permit parametric validations of DS tasks and to probe incentive conditioning generalization. A similar paradigm available in the touchscreen systems is the autoshaping task, in which animals are trained to discriminate between spatially distinct stimuli for reward delivery (C+) or no reward (C-) (Bussey, Everitt, & Robbins, 1997). Thus, time-locked neuronal recording or manipulation techniques to the C+ approach behavior can provide an index of reward anticipation, though could be conflated with motoric activity. Of course, data from these tasks can equally inform reward learning (discussed below). A recent study demonstrated the value of integrating temporally and cellular precise techniques during the C+ approach phase, revealing the interplay between NAc dopamine, cholinergic interneurons, and medium spiny neuron (MSNs) activity during stimulus-reward associations (Skirzewski et al., 2022). Specifically, NAc dopamine responses became specifically tuned during acquisition when mice approached the C+ and was robustly suppressed by reward, corroborating the strong recruitment of the striatum in reward anticipation consistent with human MID task performance. Hence, MID-like tasks for rodents using touchscreen systems can be integrated with neural recording techniques to identify relevant underlying mechanisms of reward anticipation.
Initial Response to Reward
Similar to reward anticipation, consummatory responding may alter the motivation to seek reward and guide goal-oriented behavior. As mentioned, the MID task contains an outcome feedback phase which can be used to probe initial responses to reward. Unlike the widespread cortical and subcortical activation patterns observed during reward anticipation, reward receipt in the MID task more selectively recruits the orbitofrontal cortex (OFC) and prefrontal cortex (PFC) (Knutson et al., 2001; Knutson, Fong, Bennett, Adams, & Hommer, 2003; Oldham et al., 2018). Further, unlike the indistinguishable engagement of brain regions by reward win- and loss-anticipation (see above), the OFC and PFC may differentially respond to win and loss outcomes thereby acting as neural valence markers. For example, medial PFC (mPFC) activity was elevated following the receipt of an anticipated win but lowered when an anticipated win did not occur (Knutson et al., 2003). Similarly, oscillatory patterns, such as the reward positivity (RewP) event related potential (ERP), differentially respond to the receipt of reward versus non-reward, see details below (Broyd et al., 2012; Novak & Foti, 2015).
The RDoC initiative recommended computerized Simple Guessing Tasks to inform initial responses to reward since observations in the MID task may be confounded both by incentive anticipatory responding and practice effects. In one, participants are instructed to guess whether the number between 1 and 9 on a playing card is higher or lower than 5 for a monetary reward, or loss, outcome (Delgado, Nystrom, Fissell, Noll, & Fiez, 2000). Similarly, in the Doors Task participants are asked to choose between two doors to win or avoid monetary loss. Across tasks, outcomes are randomized and not contingent on a response, thereby eliminating potential learning and practice effects. Consistent with the MID task, heightened mesocorticolimbic activation and RewP ERPs were associated with win over loss outcomes in Simple Guessing Tasks (Carlson, Foti, Mujica-Parodi, Harmon-Jones, & Hajcak, 2011; Delgado et al., 2000; Tricomi, Delgado, McCandliss, McClelland, & Fiez, 2006) (Table 3). Instead, loss feedback decreased striatal responses below baseline levels and attenuated reward-delivery ERP signals (Carlson et al., 2011; Delgado et al., 2000; Foti, Weinberg, Dien, & Hajcak, 2011; Tricomi et al., 2006). Thus, striatal activation and reward-related ERP measures act as neural valence indicators (win/lose) of outcomes. Importantly, blunted RewP ERPs during reward receipt appear to be transdiganostic as they occur in individuals with depression (Proudfit, 2015; Weinberg, 2023), attention-deficit/hyperactivity disorder (ADHD) (Kallen, Perkins, Klawohn, & Hajcak, 2020) and in healthy populations following an acute stressor (Burani et al., 2021; Porcelli, Lewis, & Delgado, 2012). Within depressed populations, reduced RewP activity has predicted both subsequent remission, and as well as treatment responsivity (Burkhouse et al., 2018; Klawohn, Brush, & Hajcak, 2021). Further, attenuated neural differentiation between gains and losses as determined by RewP during Simple Guessing Tasks were associated with risk propensity extremes as measured by the Balloon-Analog Risk Task (Huggins, Weinberg, Gorka, & Shankman, 2019). Hence, this altered neural activity pattern may participate in maladaptive decision-making. Thus, RewP may be a prospective biomarker for disordered reward processing across neuropsychiatric disorders, which can be leveraged to develop targeted treatments.
Table 3.
Evidence for cross-species validation of tasks that assess initial response to reward.
| Reward Responsiveness – Initial Response to Reward | ||
|---|---|---|
|
| ||
| Validity Domain | Human Task – Simple Guessing Task | Rodent Task – T-Maze Guessing Task |
|
| ||
| Predictive | ↓ RewP ERPs observed in depression (Weinberg et al. 2022; Proudfit et al. 2015), ADHD (Kallen et al. 2020), and following stress (Burani et al. 2021; Porcelli et al. 2012). | N/A |
|
| ||
| Neurobiological |
Following win: ↑ mesocorticolimbic activation; ↑ RewP ERPs (Carlson et al. 2011; Delago et al. 2000; Tricomi et al. 2006). Following loss: ↓ striatal responses; ↓ reward-delivery ERP signals (Carlson et al. 2011; Tricomi et al. 2006; Delgado et al. 2000; Foti et al 2011). During outcome: ↑ vmPFC activity (Knutson et al. 2001). During reward: ↑ OFC activation (Knutson et al. 2001; Knutson et al. 2003; Oldham et al. 2018). |
N/A |
|
| ||
| Pharmacological | N/A | N/A |
Simple Guessing Tasks have been utilized in animal studies, though traditionally for assessing stereotypic responding. For example, the T-Maze apparatus is utilized wherein one of two goal boxes are randomly baited with reward. However, unlike humans who receive instruction to randomly choose, animals may form associations between choice behavior (movement to left or right goal post) and reward outcome. While chance reward outcomes prevent associations, neural activity underlying attempted learning may differ from known guessing, a potential translational confound. Simple Guessing Tasks could be adapted to rodent touchscreens, using distinct visual or spatial stimuli that are randomly rewarded throughout a session. In fact, this approach has been used in probabilistic learning tasks (see below). Given the lack of cognitive demand in Simple Guessing Tasks (e.g., absence of learning), such paradigms may be better suited for assessment of initial reward responding. Combining photometry and intracranial drug delivery in these prospective touchscreen tasks would inform both the neural systems and pharmacology relevant to this domain. Validity studies would be required, however, to recreate such RewP outcomes, as seen in Simple Guessing Tasks and in some probabilistic learning tasks (Cavanagh et al., 2021).
Reward Satiation
Responsiveness to a reward is influenced by the degree to which that reward holds subjective value. Reward satiation refers to the (often temporary) loss of incentive value for a reward as a result of experiencing it over time. Thus, reward satiation is tightly linked to the hedonic pleasure associated with reward, signaling the endpoint at which a rewarding stimulus is no longer valued. Satiety is often discussed in the context of consummatory behaviors, where hunger drive is diminished while eating or food preference is reduced when it is overconsumed. The latter refers to reward-specific satiety, resulting in a steeper reduction in the value of a specific reward that does not always generalize (Berridge & Kringelbach, 2008; Rolls, Rolls, & Rowe, 1983), though can occur across sensory domains, including for olfactory, auditory, visual, and tactile stimuli (Cabanac, 1971; Rolls & Rolls, 1997; Triscoli, Ackerley, & Sailer, 2014). Satiety also represents the end of a reward-related cycle incorporating the additional reward responsiveness constructs. Regarding consumption, food is sought out via environmental predictions (i.e., reward anticipation) and consumed (i.e., initial response to reward) until the hedonic value of consumption is depleted (i.e., satiety) (Morten L. Kringelbach, Stein, & Van Hartevelt, 2012). Importantly, satiation reflects an altered internal state that influences how a reward guides future behavior, including its ability to reinforce learning and the motivation to obtain that reward.
In humans, satiation is typically measured using self-reports of their internal state and feelings toward reward-related stimuli. In non-human primates, however, non-verbal paradigms train animals to associate specific food rewards with discrete stimuli, enabling researchers to assess reward preference as satiety occurs (Pastor-Bernier, Stasiak, & Schultz, 2021). Changes in satiety have been observed using neuroimaging, e.g., fMRI revealed those in a hungry state exhibited elevated amygdala, parahippocampal gyrus, anterior fusiform gyrus, OFC, and hypothalamus activity in response to food stimuli (Haase, Cerf-Ducastel, & Murphy, 2009; LaBar et al., 2001; Siep et al., 2009). Alternatively, a state of satiety is associated with increased blood oxygen level–dependent (BOLD) signals in the dorsolateral PFC (dlPFC), and a suppression of activity in reward processing regions, such as the vmPFC, NAc, hypothalamus, insula (Thomas et al., 2015), OFC, and amygdala (Gottfried, O'Doherty, & Dolan, 2003). The OFC is a key region for representing reward valuation (see above) and the association of pleasant or positive feelings towards sensory stimuli in humans (Morten L Kringelbach, 2005). In non-human primates, the OFC contains the secondary taste- and olfactory- cortices where local neurons are activated by the sight and smell of food when hungry, but decrease their firing during satiation (Critchley & Rolls, 1996; Rolls, Sienkiewicz, & Yaxley, 1989). Further, disrupting communication between the OFC and amygdala impairs the natural devaluation of a food reward typically induced by satiation (Baxter, Parker, Lindner, Izquierdo, & Murray, 2000). Thus, changes in satiety can be reflected in changes in neuronal activation.
Satiation is influenced by central and peripheral nervous system signaling molecules and is regulated by both homeostatic and hedonic processes. Interestingly, appetite increased in individuals using antihistamines with antiserotonergic properties (Stone, Wenger, Ludden, Stavorski, & Ross, 1961; Voigt & Fink, 2015). In contrast, serotonergic activation via selective serotonin reuptake inhibitors (SSRIs) or serotonergic agonists reduced food intake and are posited as anti-obesity compounds (Foltin, Haney, Comer, & Fischman, 1996; Hill & Blundell, 1990). Further, activation of serotonin (5-HT)2C receptors reduced PFC and striatal activation in response to food stimuli, potentially indicative of reduced value representations (Thomas et al., 2018). Targeted hypothalamic 5-HT2A/2B receptor activation reduced feeding duration while not altering the number of feeding bouts (Gruninger, LeBoeuf, Liu, & Rene Garcia, 2007; Shor-Posner, Grinker, Marinescu, Brown, & Leibowitz, 1986). In humans, 5-HT activation reduced food intake and consummatory rate, while dopaminergic activation also reduced feeding duration but increased feeding bouts and consummatory rate. Hence, dopaminergic modulation may alter the hedonic aspects of feeding while serotonin regulates satiation (Blundell & Latham, 1978; Halford & Harrold, 2012).
In rodents, satiety can be assessed in operant (including touchscreen)-based fixed-ratio (FR) responding paradigms, where animals are required to make a single response for a single food reward throughout the session (Phillips, Heath, Ossowska, Bussey, & Saksida, 2017). Importantly, trial block analyses enable assessment of response-rates across the session, to determine whether a specific manipulation is altering the hedonic value of a reward (responses during early blocks) compared to satiation (responses during later blocks). Although touchscreen FR paradigms remains to be validated, nutritional content modulates satiation rate during performance. Specifically, rewards with a high sugar content resulted in a more rapid decay in response-rate throughout a session (i.e. satiation) compared to a lower sugar content reward (Kim et al., 2017). Interestingly, higher caloric, but not sugar, content increased an animal’s responses to receive a reward (see progressive-ratio task in reward valuation). Thus, the FR paradigm enables identifying factors that modulate sensitivity to satiety that do not necessarily reflect changes in motivational state. In addition to the FR paradigm, most if not all touchscreen paradigms include ancillary measures that can be evaluated as indices of whether satiation may be influencing primary outcomes (e.g., number of trials/responses completed, latencies to make screen responses, and collect rewards). Serotonin depletion modulates various reward-related measures in touchscreen paradigms, e.g., both latency to respond and collect rewards (Alsiö et al., 2021; Desrochers & Nautiyal, 2022), which, may reflect an increase food drive and suppression of satiety. Alternatively, high doses of the serotonin antagonist lorcaserin reduced screen responding and slowed response and reward collection latencies, which coincided with reduced responding in a fixed-ratio operant task (P. J. Fletcher et al., 2023). Finally, global serotonin depletion facilitated the acquisition of autoshaping (described above), sped response latency, and maintained responding to a devalued target (Winstanley, Dalley, Theobald, & Robbins, 2004). Thus, some pharmacological and predictive validity exists for such behavioral measures and the validation of a fixed-ratio paradigm would further aid in identifying if abnormal satiety responses contribute to the behavioral profile of a given manipulation or animal model.
Category 2: Reward learning
Reward learning comprises the ability to form associations between environmental stimuli and rewarding outcomes, and update these associations in a dynamic environment (NIMH, 2023). Impaired reward learning is a common feature of several neuropsychiatric disorders, including schizophrenia (Lancaster et al., 2016), substance use disorders (Knabbe et al., 2022), major depressive disorder (Liverant, Arditte Hall, Wieman, Pineles, & Pizzagalli, 2021), post-traumatic stress disorder (B. H. Morris, Bylsma, Yaroslavsky, Kovacs, & Rottenberg, 2015), and anxiety disorders (Hein, de Fockert, & Ruiz, 2021). Importantly, the presentation and severity of reward learning deficits can differ both across and within diagnostic categories (Der-Avakian, Barnes, Markou, & Pizzagalli, 2016; Kirschner et al., 2024). Furthermore, the ability to use reinforcement feedback is strongly dependent on functions outlined in reward responsiveness, making it important to be able to detect potential impairments in functions such as hedonic processing, simple habit formation, processing unpredictable feedback, or general motivation. The RDoC framework subdomains of reward learning are probabilistic and reinforcement learning, reward prediction error, and habit formation.
Reinforcement learning
The most fundamental form of reward learning is operant (instrumental) conditioning, where subjects learn to make associations between a stimulus, a voluntary response, and a rewarding outcome. The ability to make these initial associations is necessary to perform any task that implements reinforcement for correct responses, regardless of whether those paradigms are assessing executive function, motivation, or memory. Given the ease of establishing a single response when feedback is accurate and consistent (deterministic learning) for humans, these types of paradigms are less commonly implemented than probabilistic learning tasks but are extensively used in animal research. Deterministic learning paradigms typically include a reversal component, where subjects are initially required to learn that a specific stimulus is rewarded (S+), while another is not (S-), which is followed by a reward contingency change. As a result, these paradigms can evaluate initial reward association learning and the ability to adjust behavior in response to a contingency reversal (Butter, 1969; Jones & Mishkin, 1972). While both initial association and contingency reversals require a subject to acquire a new association, the difference is reflected in behavioral deficits observed in clinical populations and the underlying neurobiology. Individuals with schizophrenia exhibit deficits in initial discrimination learning while simple contingency reversals are unaffected (Murray et al., 2008). Alternatively, individuals with depression exhibit a learning deficit following reversal, particularly following unexpected positive feedback (Robinson, Cools, Carlisi, Sahakian, & Drevets, 2012).
Establishing reward-based associations recruits corticostriatal circuits that enable previously irrelevant information to gain salience and are used to drive future behavior (Fuster, 2001; Miller & Cohen, 2001). Importantly, deconstructing the neurobiology underlying reinforcement learning encompasses various reward-related processes such as anticipation and reward prediction/prediction errors (discussed below) that are formed through learning. In humans, ventral striatal activation occurs during responses in reward in the early phases of learning, but then shifts to reward delivery predictive cues once associations are established (McClure, Berns, & Montague, 2003; J. P. O'Doherty, Dayan, Friston, Critchley, & Dolan, 2003), which is also observed in dynamic dopamine activity in non-human primates (Ljungberg, Apicella, & Schultz, 1992; Mirenowicz & Schultz, 1994). When contingencies are reversed, unexpected reward delivery is associated with increased striatal activity, illustrating a central role of striatal dopamine in updating associations (Robinson, Frank, Sahakian, & Cools, 2010). Moreover, this elevation in striatal activity during reversal learning was attenuated in depression, reflecting a possible insensitivity to rewarding feedback (Robinson, Cools, Carlisi, Sahakian, & Drevets, 2012). The differences in initial association and reversal learning are also reflected in the OFC, where damage resulted in deficits after a contingency reversal, while initial reward association was unaffected (Hornak et al., 2004; J. O'Doherty, Kringelbach, Rolls, Hornak, & Andrews, 2001). Human neuroimaging studies reveal that vmPFC activation correlates with the expected value of reward outcome and during reward receipt across learning (Daw, O'Doherty, Dayan, Seymour, & Dolan, 2006; Gläscher, Hampton, & O'Doherty, 2009; Liu, Hairston, Schrier, & Fan, 2011; Tanaka et al., 2016), together suggesting that while the vmPFC appeared to monitor and update reward value, its contributions to learning are necessary only when outcomes are unpredictable. Further, while vmPFC damage impaired reversal learning, damage to the dlPFC did not affect initial association nor contingency reversals (Fellows & Farah, 2003). PFC damage was associated with reduced reward sensitivity, which could underlie deficits in reward-related learning (Manohar & Husain, 2016). The neurochemical contributions to reward learning in humans in the context of probabilistic learning are discussed in more detail below, given its more abundant use in the field.
There has been extensive characterization of the underlying neurobiology of deterministic reversal learning using rodent touchscreen systems. As in the human reversal learning task, rodents learn that responding to one visual stimulus consistently yields a reward (S+) while a second stimulus does not (S-). Following the acquisition of this association, the contingencies are reversed. Dorsal striatum activity (e.g., c-fos imaging) was elevated by late-stage initial discrimination, decreased during early reversal, and increased with the new S+ acquisition. Further, in vivo single unit recordings demonstrated that during late reversal, dorsal striatum neurons were strongly suppressed following correct choices, followed by a sharp increase in activity at reward delivery. Ablation of the GluN2B NMDA receptor subunits from corticostriatal neurons also impaired initial discrimination (Brigman et al., 2013). Consistent with humans, disruption of the OFC impaired reversal, but not initial learning (Alsio et al., 2021; Chudasama & Robbins, 2003; Graybeal et al., 2011; Hervig et al., 2020; Izquierdo, Brigman, Radke, Rudebeck, & Holmes, 2017). During initial discrimination, OFC neurons increased activity on win-stay responses. Immediately after the contingency reversal however, OFC activity was significantly higher during perseverative responses, linked to increased c-fos expression during early reversal (Brigman et al., 2013) and returned to tracking win-stay responses after reversal was learned (Marquardt, Sigdel, & Brigman, 2017). These findings suggest that dynamic regulation of OFC activity is necessary for disengagement of the previous S+ stimulus and biasing responding towards the new S+ stimulus. Fluorescence in situ hybridization revealed that dorsolateral striatum (DLS) D1R MSNs were significantly more active during early versus late discrimination learning. In contrast, optogenetic D1R MSNs inhibition early in learning facilitated correct responses (and improved decision-making metrics), where the opposite effect was observed late in learning. Meanwhile, D2R MSNs suppression reduced errors early in learning and increased correct choices later in learning (Bergstrom et al., 2018). Optogenetically silencing dorsolateral striatal neurons (i.e., via the CAG promoter) increased errors during reversal learning, demonstrating a dissociable role of this region during learning after a contingency reversal (Bergstrom, Lieberman, Graybeal, Lipkin, & Holmes, 2020). Phasic NAc dopamine release occurred after correct responses during initial discrimination, while elevated transient dopamine release was observed prior to reward delivery early in reversal (See reward anticipation and prediction error). Optogenetic inhibition of NAc D2R, but not D1R, MSNs during the outcome period of error trials significantly impaired task performance during initial discrimination phases (Nishioka et al., 2023). Further, silencing VTA projections to the NAc core during early reversal, which corresponded with the temporal responsiveness of NAc dopamine, increased the number of errors (Radke, Zweifel, & Holmes, 2019). Hence, mesolimbic dopamine has been shown to facilitate reward learning across species, while OFC activity is consistently observed to track the value of stimuli associated with reward, and damage to this region lead to a rigidity in updating new value information. Thus, deterministic learning and reversal paradigms show established neurobiological validity across species.
Pharmacological predictive validity is seen in the reward learning task whereby global 5-HT depletion slowed both initial and reversal learning, consistent with human observations (Rogers et al., 1999). More specifically, mPFC serotonin depletion via 7-dihydroxytryptamine (DHT) impaired initial discrimination, while similarly targeted depletion of serotonin in the OFC specifically impaired reversal learning (Alsio et al., 2021). Reversal learning was facilitated by intra-OFC infusions of a 5-HT2C-receptor antagonist (Alsio et al., 2015; Boulougouris & Robbins, 2010). Finally, amphetamine increased errors early in reversal in rats, and increased sensitivity to positive feedback as indexed by responses following correct, but not incorrect, outcomes (Stolyarova, O'Dell, Marshall, & Izquierdo, 2014). This finding has been confirmed in humans, with increased self-reported euphoria following amphetamine use correlating with increased neural activation following reward in a monetary reward task (Crane et al., 2018). Thus, the reward learning task has both neurobiological and pharmacological predictive validity across species.
Probabilistic Reinforcement learning
In contrast to deterministic learning, probabilistic learning paradigms introduce response-outcome uncertainty, which represents the uncertain nature of a real environment where a given stimulus does not always predict a consistent outcome. Probabilistic reinforcement learning is disrupted in multiple psychiatric conditions, including major depressive disorder (Klein et al., 1996), schizophrenia (Park et al., 2015), post-traumatic stress disorder (D. T. Acheson et al., 2022), and anxiety disorders (Taylor, Hoffman, & Khan, 2022). Probabilistic reinforcement learning paradigms offer an avenue for investigating whether abnormal reward learning is a function of deficits in processing unpredicted reward-related feedback.
The probabilistic reversal learning task (PRLT) is commonly used in humans and non-human animals, presenting two stimuli, one commonly (S+: 80% reward) and one uncommonly (S-: 20%) rewarded (Figure 3). Thus, subjects learn to optimize their responding to the S+. The reversal component implements a contingency shift of the S- to the S+, following a number of S+ responses (Cools, Clark, Owen, & Robbins, 2002). The PRLT provides metrics including the number of trials required for initial acquisition (reward learning), the number of reversals achieved (cognitive flexibility), win-stay (repeated response after reward) and lose-shift (shift choice after a non-reward).
Figure 3.
The probabilistic reversal learning task (PRLT) for humans (left) and its equivalent in rodents (right). The PRLT provides a cross-species paradigm where participants learn that one stimulus conveys a higher probability of yielding a reward than a concurrently presented less frequently rewarded stimulus. In addition to the initial learning phase, the PRLT contains a reversal component, where the previously attributed reward probabilities are switched between stimuli, allowing for prediction error and cognitive flexibility to be measured. The objective measures of the PRLT include learning which stimulus is the optimal choice to maximize rewards, total reversals, plus indices of sensitivity to feedback including the likelihood of changing responding following a reward (win-stay, win-shift) or the absence of reward (lose-stay, lose-shift) (i.e., the “punishment” included in the rodent task). Stimuli for the clinical task (left) taken from (Dahlia Mukherjee, Filipowicz, Vo, Satterthwaite, & Kable, 2020).
In the probabilistic reward task (PRT; or response-bias probabilistic reward task), participants learn to select an optimal (rich) over a nearly-indistinguishable suboptimal (lean) stimulus (Pizzagalli, Jahn, & O'Shea, 2005). Importantly, the participant does not always receive feedback when responding to a stimulus, with the optimal stimulus simply providing rewarding feedback more often (Figure 4). With such implicit (rather than explicit as above) feedback, healthy individuals develop a bias towards responding to the rich stimulus, despite no improvement in task accuracy (Pizzagalli, Goetz, Ostacher, Iosifescu, & Perlis, 2008; Pizzagalli et al., 2005). Importantly, the development of bias for the rich stimulus is attenuated in depression (Pizzagalli, Iosifescu, Hallett, Ratner, & Fava, 2008). Both probabilistic learning tasks and their respective validations have been recently reviewed in detail (Young, 2023), highlighting the difference of explicit feedback, provided in the PRLT, from the implicit feedback provided by the PRT.
Figure 4.
The probabilistic reversal task (PRT) for humans and its equivalent for rodents. The PRT is another cross-species paradigm for assessing probabilistic reward learning, and similarly requires subjects to respond to a “rich” stimulus that is more frequently reward then a “lean”, less frequently rewarded, concurrently presented stimulus. Similar to the PRLT, the PRT also measures the ability to develop an optimal response strategy (i.e., biasing responding to the rich stimulus), as well as assesses sensitivity to positive (win-stay, win-shift) and negative outcomes (lose-stay, lose-shift). An important component of the PRT is that subjects are not provided feedback on every trial, requiring them to implicitly learn the associations, which contrasts from the explicit learning employed in the PRLT.
Interestingly, people with depression (Esfand, Null, Duda, de Leeuw, & Pizzagalli, 2024; K. Fletcher et al., 2015; W. H. Liu et al., 2011; Pizzagalli, Iosifescu, et al., 2008; Vrieze et al., 2013) and BD (Pizzagalli, Goetz, et al., 2008) exhibit impaired PRLT and PRT performance, but people with schizophrenia only exhibit PRLT deficits (Culbreth, Westbrook, Daw, Botvinick, & Barch, 2016; Reddy, Waltz, Green, Wynn, & Horan, 2016; Waltz & Gold, 2007), not in PRT (Ahnallen et al., 2012; Barch et al., 2017; Heerey, Bell-Warren, & Gold, 2008) (See tables 4 and 5). Relevant to a transdiagnostic approach for PRLT use, people with depression or social anxiety disorder showed higher self-reported symptoms of anhedonia were associated with PRLT impairments relative to healthy participants, but diagnostic categories were not (E. E. Reilly et al., 2020). The PRLT is also sensitive to individuals diagnosed with other psychiatric disorders, such as depression, bipolar disorder, and obsessive-compulsive disorder (OCD) (Young, 2023). Hence, while anhedonia-related symptoms are seen in every group with psychiatric conditions using the PRLT, deficits in implicit reward-learning in the PRT were not.
Table 4.
Evidence for cross-species validation of the probabilistic reversal learning task (PRLT).
| Reward Learning – Probabilistic Reinforcement Learning #1 | ||
|---|---|---|
| Validity Domain | Human Task – Probabilistic Reversal Learning Task (PRLT) | Rodent Task – Probabilistic Reversal Learning Task (PRLT) |
| Predictive | Early-life stress exposure induces PRLT deficits (Wilkinson 2021) | Early-life stress exposure impairs PRLT performance in females more strongly than in males (Dutcher 2023, Zuhlsdorff 2023) |
| Neurobiological | Activation of anterior cingulate cortex Activation of orbitofrontal cortex Activation of ventral medial PFC Impairments associated with reduced striatal D2R expression (Jocham 2009) |
Orbitofrontal cortex lesion impairs reversal learning (reviewed in Izquierdo 2018) |
| Damage to PFC and temporal cortex impairs PRLT performance following reversal (Swainson 2000) | mPFC lesions do not affect PRLT performance (Birrell and Brown, 2000; Bissonette 2008; Floresco 2008; Churchwell 2009; Cordova 2014) mPFC lesion or stress may enhance PRLT performance (Salazar 2004; Graybeal 2011; Bryce and Howland 2015) |
|
| Frontal midline delta band power associated with reward positivity (Cavanagh 2021) | Frontal midline delta band power associated with reward positivity (Cavanagh 2021) | |
| Pharmacological | Citalopram impaired performance in PRLT (increased errors and response latency) (Chamberlain 2006) | Win-stay behaviors affected by antidepressants Reboxetine: reduced win-stay Venlafaxine: no effect Citalopram: increased win-stay (Wilkinson 2020) |
| Ketamine and scopolamine reduced win-stay behaviors and number of reversals achieved (Wilkinson 2020) | ||
| Impairments in rule learning following serotonin depletion; facilitation of PRLT learning by serotonergic agonism (Kanen 2020) | Reduced win-stay rate following 5-HT2C antagonism Increased win-stay rate following 5-HT2C agonism (Phillips 2018) |
|
| Reduced D2R function impairs PRLT performance (Jocham 2009) | D2R antagonism impairs win-stay and lose-shift behavior (Alsio 2019) | |
| Initial learning improved by nicotine (Milienne-Petiot, 2018) | ||
| Impairments in PRLT performance associated with methamphetamine usage (Robinson 2021) | Impairments in PRLT following methamphetamine treatment (Groman 2018) | |
| Clinical sensitivity | Impairments in schizophrenia (Waltz and Gold, 2007; Culbreth 2016; Reddy 2016) | |
Table 5.
Evidence for cross-species validation of the probabilistic reward task (PRT)
| Reward Learning – Probabilistic Reinforcement Learning #2 | ||
|---|---|---|
| Validity Domain | Human Task – Probabilistic Reward Task (PRT) | Rodent Task – Probabilistic Reward Task (PRT) |
| Predictive | Blunted response in PRT is indicative of future anhedonia (Kangas 2022) | Early-life stress impairs PRT performance (Lamontagne 2021; Hisey 2023) |
| Major depressive disorder impairs PRT performance (Hisey 2023) | ||
| Neurobiological | Improved PRT learning associated with decreased dopamine transporter binding potential (Kaiser 2018) | D2R agonism decreases response bias; D2R antagonism increases response bias (Der-Avakian 2013) |
| Increased dopaminergic activity at reward delivery (Santesso 2009; Bogdan 2011) | Increased activity in anterior cingulate and NAc at reward delivery (Iturra-Mena, 2023) | |
| Pharmacological | No effect of amphetamine on PRT performance (Soder 2021) | Amphetamine increases rich stimulus bias (Kangas 2020; Lamontagne 2021) |
| Pramipexole inhibited reward learning (Santesso 2009) | Pramipexole reduces rich response bias (Lamontagne 2021) | |
| Scopolamine increases rich response bias Oxycodone does not affect rich response bias (Kangas 2020) |
||
| Clinical sensitivity | No impairments detected in people with schizophrenia (Heerey 2008; Ahnallen 2012; Barch 2017) | |
The use of both positive and negative feedback in the PRLT and PRT can dissociate abnormalities in positive and negative valence processing. For example, task deficits in individuals with schizophrenia appear to result from an inability to optimally use both positive feedback (win-stay) and negative feedback (lose-shift) (Reddy et al., 2016), while people with depression seem to be specifically impaired in using positive feedback (reduced win-stay with intact lose-shift) (D. Mukherjee, van Geen, & Kable, 2023) in the PRLT. This difference may stem from variance in the recruitment of underlying mechanisms to support distinct forms of learning. Striatal dopaminergic activity signals receipt of unexpected rewards, which occurs early in primary contingency learning or when contingencies shift. Reciprocal feedback loops with the frontal regions like the dlPFC and OFC facilitate the formation of value representations. In the PRLT, human fMRI revealed elevated activity of ACC, OFC, vmPFC and ventral striatum during final errors (last error before acquiring the new S+) (Cools et al., 2002; Dodds et al., 2008; Jocham et al., 2009). Interestingly, individuals with D2R polymorphisms that result in reduced striatal D2R expression are impaired in acquiring the new S+ following reversal accompanied by a lack of ventral striatal and OFC activation following final errors (Jocham et al., 2009). Alternatively, unexpected negative feedback reduced OFC and amygdala activation, whereas individuals with bipolar disorder exhibit an attenuation in such reduced activity (Linke et al., 2012). Finally, PFC and temporal cortex damage impaired PRLT performance, especially following reversal (Swainson et al., 2000). While these circuits may underlie the integration and use of explicit feedback, the human neurobiology underlying the implicit learning used in PRT remains to be fully characterized. One study used PET imaging to demonstrate that individuals with lower dopamine transporter binding potential exhibited better PRT learning, accompanied by increased functional connectivity between the nAc and OFC (Kaiser et al., 2018). As is central to the RDoC framework, identifying distinct mechanisms underlying PRLT and PRT task performance may yield neural circuit candidates that distinguish reward-related dysfunction across psychiatric populations. Thus, differences in pathophysiology may manifest as impairments depending on the type of learning being taxed by the given task. Hence, both paradigms offer unique applications for targeting differences in dysfunctional reward learning.
Both the PRLT and PRT have been developed as touchscreen-based tasks for use in rodents. While non-touchscreen PRLT tasks have been extensively used and validated (Young, 2023), validation of the touchscreen-based PRLT, particularly neurobiological validity, is still underway. In a study of several commonly-used SSRIs, reboxetine reduced win-stay behaviors, venlafaxine had no effects, and citalopram increased win-stay behavior and facilitated rule learning in rats (Wilkinson, Grogan, Mellor, & Robinson, 2020). Both acute ketamine and scopolamine reduced win-stay behaviors and the number of reversals of rats (Wilkinson et al., 2020). Additionally, blockade and activation of 5-HT2C receptors by the 5-HT2CR antagonist SB 242084 reduced and increased win-stay behaviors, respectively (Phillips, Dewan, et al., 2018), which aligned with a facilitation of PRLT by serotonergic activation in human tasks (Kanen et al., 2020). Finally, D2R blockade via the D2R antagonist raclopride disrupted performance on a touchscreen PRLT variant, reducing both win-stay and lose-shift behaviors (Alsio et al., 2019), which is similar to reductions observed in humans with reduced D2R functioning (Jocham et al., 2009). Additionally, exposure to early-life stress (ELS) (e.g., maternal separation) induced PRLT deficits with sex-specific response profiles (stronger in females) (Dutcher et al., 2023; Zuhlsdorff, Dalley, Robbins, & Morein-Zamir, 2023), which corresponds human PRLT deficits following ELS in humans (Wilkinson, Slaney, Mellor, & Robinson, 2021), providing predictive validity for the PRLT. Disrupting glutamate neurotransmission in early postnatal neurodevelopment (postnatal phencyclidine treatment), a common manipulation to approximate schizophrenia (Grayson et al., 2016), impaired performance in the PRLT (Tranter, Aggarwal, Young, Dillon, & Barnes, 2023). These deficits consisted of an impairment in win-stay responding and a reduction in the rate at which reward prediction errors (RPEs, see below) updated value estimates. Moreover, these impairments were reproduced in a follow-up study that also demonstrated that they may result from hyperactivation of glutamate neurons within the ventromedial OFC (Tranter et al., 2024). Hence, utility of the animal PRLT has enabled the investigation of mechanisms relevant to PRLT including potential mechanisms involved in deficient performance in patient populations.
Cross-species neurobiological validation has been shown for the PRT, as significant increases in ACC and NAc activity were observed in rats at reward delivery on either correct rich or lean responses, which was followed by sharp suppression of activity (Iturra-Mena, Kangas, Luc, Potter, & Pizzagalli, 2023), resembling ERPs observed during the human PRT measured via skin conductance before and after reward delivery (Bogdan, Santesso, Fagerness, Perlis, & Pizzagalli, 2011; Santesso et al., 2009). In the PRT, rich stimulus bias was increased by amphetamine in rats (Kangas, Wooldridge, Luc, Bergman, & Pizzagalli, 2020; Lamontagne, Melendez, & Olmstead, 2018), though it did not affect human PRT performance (Soder et al., 2021). Rich response bias was reduced by the D2/3 receptor (D2/3R) agonist pramipexole and glucocorticoid dexamethasone in rats (Lamontagne et al., 2018), and increased by scopolamine and unaffected by oxycodone in rats (Kangas et al., 2020), providing targets for validation in human studies. Further, the PRT similarly demonstrates predictive validity for ELS manipulations, with mice raised under ELS conditions showing impairments in PRT performance consistent with those seen in humans with MDD (Hisey et al., 2023). Finally, nicotine withdrawal affects performance on the PRT similarly in humans and in rats, with a significant reduction in reward responsiveness following 24 hours of withdrawal in both species (Pergadia et al., 2014). Hence, the PRT has some evidence for translational validity though more explicit testing is required.
Together, the touchscreen systems offer multiple paradigms reverse translated from human tasks to determine mechanisms underlying probabilistic reward learning. Future studies of the underlying circuitry will aid in dissecting the mechanisms underlying these different forms of reward learning and developing targeted treatments.
Reward prediction error
The reward learning subdomain of reward prediction error (RPE) concerns the mechanisms involved when an expected reward does not occur. RPE can be categorized as positive error, when an outcome is better than expected, or negative error, when an outcome is worse than expected. A fundamental mechanism by which RPEs are encoded involves changes in the activity of dopaminergic neurons (Wolfram Schultz, 2016; Watabe-Uchida, Eshel, & Uchida, 2017). When an outcome is better than expected, and a positive RPE is elicited, there is an increase in dopamine cell firing. By contrast, when an expected reward is omitted there is a brief pause in dopamine cell activity, encoding a negative RPE. In humans, microelectrode recordings in the substantia nigra were taken while subjects performed a probability learning task (Zaghloul et al., 2009). Unexpected rewards were associated with significantly greater neuron firing than unexpected losses, with no differences following expected outcomes. Dopaminergic signaling is propagated through regions involved in reward learning, including the striatum, amygdala, and frontal cortex, resulting in responsiveness to the RPE signal (McClure et al., 2003; Wolfram Schultz, 2016; Thut et al., 1997). Thus, dopaminergic activity reflects the difference between the expected and observed reward magnitudes. Potentials of dopaminergic neurons therefore can provide a useful biomarker to combine with cognitive and behavioral assessment for RPE.
Tasks such as the PRLT can be implemented to investigate RPE testing, as the shifts in reward contingencies still occur with an unexpected change in stimulus outcome. RDoC, however, highlights Bandit tasks as paradigms for testing RPE, as they assess a participant’s ability to respond to shifting reward contingencies. These tasks are similar in concept to probabilistic reinforcement tasks, but the changes in reward contingencies are not affected by a participant’s responses, instead changing independently over the course of a session (Daw et al., 2006). RPE outcomes are based on the striatum and VTA, as well regions in the vmPFC via dopaminergic projections when making reward-related decisions as shown by fMRI (Bartra, McGuire, & Kable, 2013; Daw et al., 2006; Hare, O'doherty, Camerer, Schultz, & Rangel, 2008) (Table 6). Bandit tasks utilize similar behavioral measures to probabilistic learning paradigms, and can assess distinct behavioral strategies underlying decision-making. Performance in bandit tasks is sensitive to the trade-off between exploiting high-valued choices and exploring alternative options to gain additional information. Optimal task performance requires these mutually exclusive strategies to be balanced, and extremes in either strategies can lead to maladaptive behaviors evident in mental illnesses (Addicott, Pearson, Sweitzer, Barack, & Platt, 2017). Win-stay/lose-shift behavior is improved by baseline anxiety state (Harle, Guo, Zhang, Paulus, & Yu, 2017) but impaired by methamphetamine dependence (Harle et al., 2015). Moreover, anhedonia was associated with an increase in exploration that reduced the preference for the most rewarded option (Harle et al., 2017). In people with schizophrenia, reduced win-stay behavior on probabilistic tasks was also associated with lower RPE signaling in the ventral striatum (Katthagen, Kaminski, Heinz, Buchert, & Schlagenhauf, 2020; Schlagenhauf et al., 2014) and reduced dlPFC activation by positive RPEs (Ermakova et al., 2018) compared to healthy controls. Additionally, a recent meta-analysis reported reduced activation in various brain regions associated with reward learning during RPE processing in schizophrenia, including the striatum, amygdala, thalamus and ACC (X. Yang, Song, Zou, Li, & Zeng, 2024). While additional studies are required to establish associations between predispositions to behavioral response to RPEs (e.g., more or less exploratory or exploitive) and the underlying neural responses, abnormalities in midbrain dopamine signaling and the corresponding RPE disruptions likely contribute to the deficits in probabilistic learning deficits in psychiatric disorders.
Table 6.
Evidence for cross-species validity of the tasks that assess reward prediction error (RPE).
| Reward Learning – Reward Prediction Error | ||
|---|---|---|
| Validity Domain | Human Task – Bandit tasks | Rodent Task – Bandit tasks |
| Neurobiological | Sex differences in spatial restless bandit task – males spend more time exploring the environment (Chen et al 2021) | |
| Reward prediction error encoded by dopaminergic neurons (Schultz 2016) | Reward prediction error encoded by dopaminergic neurons (Schultz 2016) | |
| Clinical sensitivity | People with schizophrenia showed impaired performance on a bandit task (Cathomas 2021) | |
Habit learning
Goal-directed behavior utilizes the learned associations between behavioral actions and their predicted outcomes enabling the flexible navigation of dynamic environments to achieve desirable outcomes (S. de Wit & Dickinson, 2009; Dickinson, 1985). When a given response is continuously associated with an outcome, that specific action becomes more dependent on the situation or stimulus by which it is elicited, rather than an action. The behavior then transitions from intentionally goal-directed to an unconscious habit, which generates optimal behavior at a lower cognitive cost (Balleine & Dickinson, 1998). Over time, formed habits can become rigid and inflexible, but can be shifted by focusing on goal-directed actions.
A fundamental challenge in studying habitual behavior is ensuring the distinction between habit and behaviors that implement goal-oriented strategies. Two paradigms, first described by (Dickinson, 1985), were developed to address these concerns, both removing the rewarding component of a behavioral response over time. The first is reward devaluation, in which the reinforcing component of the rewarding stimulus provided to the behavior is gradually reduced over time. Second is contingency degradation, where the association between a given behavior and a reward is altered, such as randomized schedule of delivering reward following the behavior response (reviewed in (Mendelsohn, 2019)). Similar to the latter is extinction, in which a reward is completely removed from the behavioral response. Essentially, a behavior can be considered habitual if it persists in the absence of the pre-existing reward conditions. Thus, cross-species habit tasks need an initial phase of procedural learning to establish stimulus-outcome associations, followed by a phase that substantially changes the response-reward association.
A commonly-used human paradigm is the habit learning task (Sjoerds et al., 2013), and presents participants with a stimulus (S) to which they respond (R) to achieve an outcome (O). The S-R-O flow utilizes goal-oriented behavior, as individuals provide a designated response to achieve a rewarding outcome. Once the S-R-O relationship is established, the flow is simplified to an S-R strategy (habitual response) (Figure 5). Cognitive tasks sensitive to changes in habit learning are designed around gauging an individual’s ability to shift between habitual control and goal-oriented behavior, as well as their ability to develop new habits.
Figure 5.
The habit learning task for humans (left) and the visuomotor conditional learning task (VMCL) for rodents (right). Habit learning tasks consists of two primary phases. The first requires goal-directed reward learning, where subjects learn to perform a behavior (e.g., which hand to move) in response to a stimulus (e.g., a picture of a fruit in the human task) that is associated with a specific outcome (e.g., the presentation of a paired picture for humans, or a food reward for rodents). The ability to form habits is then assessed by the removal of a goal-directed outcome, where the presentation of a stimulus now indistinctive elicits the habitual response. For rodent paradigms, such as the visuomotor conditional learning tasks (VMCL), animals are similarly trained to associate an image with a response to a square located on the right or left of the screen. Habit formation can be tested using a devaluation or extinction protocol, where the outcome of correctly responding to a stimulus (i.e., a reward) is removed or diluted, and researchers can test whether the habitual behavior (i.e., response to the stimulus) is maintained. Thus, habit learning paradigms allow researchers to tease apart goal-directed learning from habitual responding. Schematic for the habit learning task (left) adapted from (Sjoerds et al., 2013).
Abnormalities in habit learning can bias individuals towards increased habitual behaviors or disrupt flexibility. Individuals with substance abuse disorders appear to be more rigid in habitual drug-seeking behaviors, and are insensitive to devaluations in the rewarding properties of a given drug (Dickinson, Wood, & Smith, 2002). Similarly, pathological habit can manifest as compulsion (as in OCD), resulting from an abnormal strengthening of stimulus-response associations and an insensitivity to devaluation, resulting in deficits in balancing habitual and goal-directed behavior (Gillan et al., 2011; Maia, Cooney, & Peterson, 2008). Diseases with striatal degeneration such as Parkinson’s disease may have impaired procedural learning, preventing the formation of initial associations (Hay, Moscovitch, & Levine, 2002). Their deficits manifest as difficulty in shifting from S-R to R-O strategies, and in appropriately responding to devalued outcomes. This pattern indicates an impairment in goal-directed behavior in substance use disorders in combination with an overreliance on previously-developed habitual responses. People with alcohol dependence show difficulty shifting away from habitual control to goal-oriented behavior (Sjoerds et al., 2013), suggesting that the neural mechanisms underlying habit learning are negatively affected by long-term alcohol use. The striatum is engaged during habitual behavior (J. Li & Daw, 2011) (Table 7), with distinct recruitment of prefrontal and striatal activity during habit-formation tasks. Striatal activity was only associated with behavior at stimulus presentation, while PFC was active both when subjects performed an action and during response outcome. These differences may relate to habitual responding and goal-directed activity, respectively (McNamee, Liljeholm, Zika, & O'Doherty, 2015). A recent study demonstrated that while individuals with vmPFC lesions were able to readily establish a habitual association, they were unable to optimize responding once the reinforcer was devalued compared to healthy controls (Reber et al., 2017). The importance of the vmPFC was consistent where significantly increased vmPFC activity during goal-directed compared to habitual responding on an S-R-O task (S. de Wit, Corlett, Aitken, Dickinson, & Fletcher, 2009). Similarly, striatal activity strongly modulated the formation of initial associations, while vmPFC activity is necessary for updating previously established associations. Thus, habit formation can be assessed in people with clinically sensitive tasks, with targets for validation in rodent tasks.
Table 7.
Evidence for cross-species translation for tasks that measure habit formation.
| Reward Learning – Habit Formation | ||
|---|---|---|
| Validity Domain | Human Task – Habit Learning Task | Rodent Task – Visuomotor Conditional Learning (VMCL) |
| Predictive |
Substance-abuse disorder: ↓ sensitivity to devaluation OCD: Intact reward learning but ↓ sensitivity to devaluation (Gillan et al. 2011). Parkinson’s disease: ↓ ability to form procedural habits (Hay et al. 2002). |
|
| Neurobiological | ↑ striatal activity during habitual responding (Li & Daw, 2011; Graybel & Grafton, 2015). ↓ vmPFC activity during habitual compared to goal-directed responding (de Wit et al. 2019). vmPFC lesions ↓ shifting from habitual to goal-directed responding (Reber et al. 2017). |
↓ task acquisition following dorsal striatum lesion (Delotterie, 2015). ↓ task acquisition following posterior cingulate cortex lesion (Bussey 1997). |
The touchscreen-based visuomotor conditional learning (VMCL) is a procedural learning paradigm (Horner et al., 2013) to assess habitual learning in rodents. In the VMCL task, the animal is presented with one of two easily-discriminable cue images in the center of the screen, followed by the presentation of two white squares on either side. They learn that the presentation of one image is associated with a reward after response at the right stimulus, while the other is associated with a reward after responses to the left (Figure 5). Similar to the human habit learning tasks, it is expected that animals will adopt an S-R habitual strategy over training. Reversal of the contingencies (the correct location associated with the initial cue stimulus) can be used to shift animals back to an R-O strategy. Posterior cingulate cortex lesions impaired mouse performance (Bussey, Everitt, & Robbins, 1997) while performance was independent of the ACC (Bussey, Muir, Everitt, & Robbins, 1997), perirhinal cortex or fornix (Bussey, Duck, Muir, & Aggleton, 2000), the prelimbic cortex, hippocampus (Delotterie et al., 2015), anterior or mediodorsal thalamus (Chudasama & Muir, 2001). Acquisition of the stimulus-association was ablated in animals with dorsal striatum lesions (Delotterie et al., 2015). Thus, striatal dependence in the absence of frontal cortex involvement aligns well with the neurobiological basis of habit formation in humans, providing neurobiological validity. To our knowledge, no publications have implemented a reward devaluation procedure for the VMCL task, which remains to be a critical component for establishing it as a habitual learning paradigm.
Devaluation procedures have been conducted in other touchscreen tasks, including assessments of effortful motivation (Stewart et al., 2022), visual discrimination, and simple stimulus responding (Favier et al., 2020). Impairing striatal cholinergic tone biased animals towards habit formation, as indicated by sustained responding following devaluation, while disrupting striatal glutamate release favored goal-directed behavior (Favier et al., 2020). Thus, optimizing the VMCL task with a devaluation procedure will aid in investigating the mechanisms underlying habitual behavior, providing insight into dissociable contributions to habit formation and goal-directed reward learning. Certainly, a great deal of validation efforts still need to be conducted to assess habit formation across species.
Category 3: Reward Valuation
The construct of reward valuation comprises three sub-constructs pertaining to the computation of a reinforcer’s value as a function of its magnitude and various devaluing factors surrounding its acquisition. Numerous laboratory assessments quantify the degree to which these factors—probability, delay to receipt, and requisite effort—impact reward value in various contexts and clinical populations. Most are decision-based, presenting a choice between rewards of different magnitudes at varying degrees of likelihood or cost. A subset of these assessments (called “discounting” tasks in reference to the devaluation of a reward by one of the above factors) can be readily administered to both humans and rodents. Other, similarly translatable assessments require participants to complete a single repetitive action for reward. Though traditionally administered to rodents using nosepoke- or lever-based operant boxes, these paradigms can be readily implemented in touchscreens. Furthermore, the versatility in stimulus delivery provided by touchscreens enables development of novel paradigms (e.g., Rearing-Effort Discounting task; (Lopez-Cruz et al., 2024)) that mitigate some of the confounds of earlier tasks.
Reward probability
Probabilistic discounting (PD) tasks present choices between small, certain rewards and larger, uncertain ones (Rachlin, Raineri, & Cross, 1991). Magnitudes and probabilities of larger rewards vary across trials and are signaled before choices are made. The large-reward options ultimately become disadvantageous at lower probabilities, and thus their selection decreases concurrently. The resultant “discounting curve” is altered in psychiatric populations (Bernhardt et al., 2017; Dai, Harrow, Song, Rucklidge, & Grace, 2016; Garami & Moustafa, 2019; K. L. Hart, Brown, Roffman, & Perlis, 2019; Schluter & Hodgins, 2021), and cigarette smokers (Reynolds, Richards, Horn, & Karraker, 2004) (but see: (Mitchell, 1999)), supporting the clinical sensitivity of this assay. Shallower discounting curves or overall reductions in PD (i.e., elevated preferences for large, unlikely rewards, indicating less devaluation by unfavorable probability) are often interpreted as impulsivity or risk insensitivity in clinical contexts (Bernhardt et al., 2017; Dai et al., 2016; Schluter & Hodgins, 2021). Conversely, reduced preference for probabilistic, large-reward options (i.e., enhanced PD) is typically interpreted as risk aversion, especially when these options have favorable odds (Garami & Moustafa, 2019; K. L. Hart et al., 2019).
Clinical findings implicate a network of brain regions in PD, including vmPFC, OFC, and ventral striatum (Miedl, Peters, & Büchel, 2012; Mok et al., 2021; Peters & Büchel, 2009; Peters & D'Esposito, 2020; Seaman et al., 2018; P. Wang et al., 2023) (Table 8). FMRI studies identified these structures as non-specific value encoders across several value-based decision making tasks (i.e., probabilistic, delay, and/or effort discounting, see below) (Mok et al., 2021; Peters & Büchel, 2009; Seaman et al., 2018); indeed, people with vmPFC and/or OFC damage exhibited reduced PD in a paradigm incorporating hypothetical rewards and no trial-by-trial feedback (Mok et al., 2021; Peters & D'Esposito, 2020). Meanwhile, trend-level PD deficits in people with gambling disorder were accompanied by diminished positive correlations between subjective value (advantageousness of the probabilistic option) and OFC and ventral striatum activation (Miedl et al., 2012). Pharmacological sensitivity is less supported, however, as the few studies reporting such manipulations describe limited effects. Namely, alcohol dose- and time point-dependently reduced PD in healthy adults (Bidwell et al., 2013) (but not adolescent males at the group level (Bernhardt et al., 2019)), while acute benzodiazepine (A. Acheson, Reynolds, Richards, & de Wit, 2006), opioid (Zacny & de Wit, 2009), cannabinoid (McDonald, Schleifer, Richards, & de Wit, 2003), and psychostimulant (A. Acheson & de Wit, 2008) administration failed to affect this behavior.
Table 8.
Evidence for cross-species validity of tasks that assess probability-based reward valuation.
| Reward Valuation – Reward Probability | ||
|---|---|---|
| Validity Domain | Human Task – Probabilistic Reversal Learning Task (PRLT) | Rodent Task – Probabilistic Reversal Learning Task (PRLT) |
| Predictive | Early-life stress exposure induces PRLT deficits (Wilkinson 2021) | Early-life stress exposure impairs PRLT performance in females more strongly than in males (Dutcher 2023, Zuhlsdorff 2023) |
| Neurobiological | Activation of anterior cingulate cortex Activation of orbitofrontal cortex Activation of ventral medial PFC Impairments associated with reduced striatal D2R expression (Jocham 2009) |
Orbitofrontal cortex lesion impairs reversal learning (reviewed in Izquierdo 2018) |
| Damage to PFC and temporal cortex impairs PRLT performance following reversal (Swainson 2000) | mPFC lesions do not affect PRLT performance (Birrell and Brown, 2000; Bissonette 2008; Floresco 2008; Churchwell 2009; Cordova 2014) mPFC lesion or stress may enhance PRLT performance (Salazar 2004; Graybeal 2011; Bryce and Howland 2015) |
|
| Frontal midline delta band power associated with reward positivity (Cavanagh 2021) | Frontal midline delta band power associated with reward positivity (Cavanagh 2021) | |
| Pharmacological | Citalopram impaired performance in PRLT (increased errors and response latency) (Chamberlain 2006) | Win-stay behaviors affected by antidepressants Reboxetine: reduced win-stay Venlafaxine: no effect Citalopram: increased win-stay (Wilkinson 2020) |
Rodent PD tasks operate on similar contingencies as the human task and are traditionally administered in lever- (Cardinal & Howes, 2005) and nosepoke-based (A. Acheson, Farrar, et al., 2006) operant systems. Touchscreen PD programs exist, although they are mostly similar to earlier operant versions and are rarely used (Abela & Chudasama, 2013) (Rojas, Curry-Pochy, Chen, Heller, & Grissom, 2022). Consistent with human assessments, rodents repeatedly choose between one option that always yields a small reward and another option that yields a large, probabilistic reward. Whereas probabilities of the large-reward option are modulated from trial to trial in human tasks, rodent testing sessions are typically organized block-wise, with probabilities held static for several consecutive trials (i.e., trial blocks). Reward likelihood of the probabilistic option steadily increases or decreases across blocks (e.g., (Cardinal & Howes, 2005)). These procedural differences must be considered when comparing findings across species (see below).
Neurobiological validity of rodent PD tasks is generally supported, with rodent OFC and ventral striatum playing roles in subjective value calculation consistent with clinical data. Inactivation of medial OFC (but not OFC in general; (St Onge & Floresco, 2010)) reduced rat PD (Stopper, Green, & Floresco, 2014), consistent with observations from human studies (Peters & D'Esposito, 2020) (Table 8). Interestingly, however, the only extant touchscreen-based study incorporating neurological manipulations reported increased PD in complete OFC-lesioned rats, albeit in a rarely used paradigm that changed probabilistic contingencies between sessions (Abela & Chudasama, 2013). Importantly, these rats were also not trained on the task until after surgery, raising the potential confound of non-specific learning effects of the lesion. The inconsistency between touchscreen PD data and those of clinical and lever-based paradigms may therefore have arisen from procedural deviations rather than the touchscreen system itself. Meanwhile, in a non-touchscreen-based paradigm, inactivation of the NAc generally reduced choice of the probabilistic option during trial blocks in which that option was advantageous or neutral (i.e., had high subjective value) (Stopper & Floresco, 2011). There is therefore some neurobiological validity between human and rodent PD tasks in general, although further validation is necessary for touchscreen-based paradigms. The effects of mPFC manipulations are less straightforward, however. Specifically, mPFC inactivation exerted opposite effects on discounting of the large reward depending on whether its probability increased or decreased across blocks (St Onge & Floresco, 2010). These data indicate a role for mPFC in tracking contingency changes, but do not align with outcomes of clinical tasks that modulate probabilities between individual trials. While functional differences between rodent and human cortical subregions should be noted when extrapolating findings across species (Seamans, Lapish, & Durstewitz, 2008; Uylings, Groenewegen, & Kolb, 2003), such parameter effects indicate deeper issues with the translatability of the task itself.
Pharmacological predictive validity of rodent PD tasks is partially supported by findings that acute ethanol reduced discounting in male rats but not females (Wallin-Miller, Chesley, Castrillon, & Wood, 2017), despite no gender-specific effects in human PD (Bidwell et al., 2013). Meanwhile, another study detected no significant effects of acute ethanol on PD in intact adult male rats, although intermittent adolescent ethanol exposure reduced PD in adulthood (Boutros, Semenova, Liu, Crews, & Markou, 2014). Finally, acute and chronic amphetamine consistently reduced PD in rats (Floresco & Whelan, 2009; Islas-Preciado et al., 2020; Ozga-Hess & Anderson, 2019; St Onge, Chiu, & Floresco, 2010; St Onge & Floresco, 2009; Yates et al., 2020) but not humans (A. Acheson & de Wit, 2008), further challenging the assay’s pharmacological predictive validity.
Inconsistencies between rodent and human findings may be partially explained by differences in testing procedure. For example, rodents must be well-trained in the PD task, exhibiting stable baseline performance prior to assessment, whereas clinical tests are administered in single sessions. By the time rodents reach stability, they have consolidated the reward contingencies (becoming habitual, see above) and can likely predict probability changes (St Onge & Floresco, 2010), thus raising the issue of differentiating actual decision-making from habitual response patterns. This complication arises from the block-wise structure of the preclinical task. When combined with daily training, this predictability of contingency shifts enables animals to refine performance strategies across several sessions, possibly to the detriment of the task’s predictive validity. Indeed, the discounting curve is less steep when the large-reward lever’s probability does not consistently ascend or descend across blocks, but instead undergoes a “mixed” block-wise progression (i.e., 100%, 12.5%, 25%, 50%; (St Onge et al., 2010)). Rat PD is therefore facilitated when the change in this lever’s utility is unidirectional and easier to learn across training sessions. An additional concern is “anchoring,” whereby response patterns are unduly influenced by the original contingency of the probabilistic lever. Anchoring is particularly persistent in mouse PD, evident even after months of training in a touchscreen-based task (Rojas et al., 2022). Thus, procedural differences across species may underlie the lack of consistent neurobiological and pharmacological validation.
Additional disparities in pharmacological effects across human and rodent PD may arise partly from the specifications of clinical assessments. The task utilized in the human amphetamine study (as well as the benzodiazepine, opioid, and cannabinoid studies; see above) presented PD trials interleaved with delay discounting trials (i.e., choice between small, immediate versus larger, delayed rewards), with no trial-by-trial feedback provided (i.e., reward versus no reward) until all trials had been completed (Richards, Zhang, Mitchell, & de Wit, 1999). While it is unclear whether the juxtaposition of probability and delay discounting trials would affect participants’ behavior in either condition, this design cannot be administered to rodents as it is unlikely that they can be trained to alternate between these trial types within a single session. The absence of immediate reinforcement in the clinical task is similarly problematic. Furthermore, rodent paradigms can rarely defer reward delivery until the end of the session since consistent reinforcement is necessary to maintain responding. Deferred reinforcement in clinical testing therefore removes the construct of reward processing that is central to animal assessments, whereas the effects of certain drugs on animal PD may be confounded by their concurrent effects on reward processing (especially amphetamine; (Leith & Barrett, 1976)). Finally, pharmacological manipulation of rodent PD is also confounded by the block-wise structure of the task. Similar to mPFC inactivation (St Onge & Floresco, 2010), amphetamine exerted opposite effects on PD depending on whether the odds for the large-reward lever ascended or descended across the session (with no effect at all in a “mixed progression” PD task) (St Onge et al., 2010). Hence, task parameters clearly impact outcomes as rodents are tested in traditional operant chambers.
The versatility in stimulus presentation offered by touchscreens may provide an opportunity to mitigate some of these shortcomings. For example, given that earlier operant systems could only deliver choice stimuli that were qualitatively identical, any within-session contingency modulation had to be signaled via several “forced-choice” trials preceding blocks of identical “free-choice” trials (Cardinal & Howes, 2005). Touchscreen systems may enable rodents to be trained to associate individual, visually distinct stimuli with specific rewards and probabilities that are presentable successively alongside other stimuli in a single, non-block-arranged PD assessment. The ability to modulate reward contingencies across individual trials similarly to clinical assessments would circumvent the confounds associated with block-wise task structures.
Delayed reward
Delay discounting (DD) tasks quantify the degree to which the subjective value of a reward is reduced by the time required to wait for its receipt. Similar to PD tasks, subjects are presented with options yielding rewards of different magnitudes, with the devaluation of the larger reward resulting from its delayed delivery. Human DD tasks (Scheres, de Water, & Mies, 2013) are typically hypothetical in nature, offering participants the choice between immediate versus delayed sums of imaginary money. Certain variants do offer actual money (or other reinforcers), meanwhile, honoring the conditions of every trial (defined as “real” or “experiential” tasks (Reynolds & Schiffbauer, 2004)) or of one or more random trials (defined as “potentially real” tasks (M. W. Johnson & Bickel, 2002); c.f. “hypothetical” tasks, in which all rewards are imaginary) (Figure 6). Interestingly, humans perform questionnaire-based DD tasks similarly across “potentially real” and hypothetical reward conditions (M. W. Johnson & Bickel, 2002). Meanwhile, operant choice procedures providing real-time reinforcement (i.e., seconds or minutes after each decision; (M. W. Johnson, 2012) produce faster rates of DD in humans than questionnaire-based assessments utilizing longer time-scales (i.e., days or months after completion of the session) (Navarick, 2004). Thus, trial-by-trial reinforcement influences human performance of these tasks.
Figure 6.
The delay discounting (DD) task for humans (left) and the touchscreen equivalent developed for rodents (right). DD tasks assess whether subjects would rather wait to obtain a large reward than receive a small reward immediately, providing measures of impulsivity and the degree to which a reward is devalued by the interval to its delivery. Human subjects are provided with verbal prompts while rodents must learn reward contingencies over several training sessions. Performance is measured either by: (1) overall percentage of choices of the delayed versus immediate reward; or (2) a discounting curve illustrating the step-wise reduction in large-reward choices as the delay to its receipt increases. The clinical task depicted here (developed by Johnson, 2012; left) is a rare example of an operant choice DD task utilizing experiential rewards. While this design optimizes cross-species translatability of findings, most clinical studies of DD utilize questionnaire-based assessments and hypothetical or “potentially real” rewards.
DD is commonly interpreted as a measure of impulsivity and often assessed alongside PD in clinical populations with impaired impulse control. Elevated DD (preference for small, immediate rewards) is observed across several such disorders (Jackson & MacKillop, 2016) (Ahn et al., 2011) (Petry, 2001) (Heerey, Robinson, McMahon, & Gold, 2007) (Pennisi et al., 2023) (Amlung, Vedelago, Acker, Balodis, & MacKillop, 2017; Phung, Snider, Tegge, & Bickel, 2019) (Table 9). Neurobiological study of human DD emphasizes cortical and striatal areas, with hippocampal and insular contributions. Consistent with the previously discussed roles of vmPFC, OFC, and ventral striatum in encoding value during decision making (Mok et al., 2021; Peters & Büchel, 2009; Seaman et al., 2018), people with vmPFC and/or OFC lesions and consequent PD deficits were also impaired in DD (hypothetical and “potentially real” paradigms, no trial-by-trial feedback; (Mok et al., 2021) (Peters & D'Esposito, 2020)). Transcranial direct current stimulation over the OFC or dlPFC significantly modulated choice behavior for hypothetical monetary gains and losses during DD (He et al., 2016; Hecht, Walsh, & Lavidor, 2013; Moro et al., 2023; Shen et al., 2016; Xiong et al., 2019). Ventral striatal dopamine D2/3R binding is also implicated in DD, as demonstrated by an inverse relationship between DD and [11C]raclopride binding potential in people with gambling disorder (questionnaire, hypothetical rewards) (Joutsa et al., 2015). Meanwhile, people with insula damage demonstrated reduced DD for hypothetical monetary rewards (i.e., elevated preference for larger, delayed rewards) (Sellitto, Ciaramelli, Mattioli, & di Pellegrino, 2015). Finally, the hippocampus may also contribute to DD, potentially via simulating future outcomes and thereby informing subjective value signaling in prefrontal areas (Peters & Büchel, 2010). This theory arises from evidence that individuals with hippocampal and associated temporal lobe damage are only impaired in DD while: (1) actively simulating hypothetical monetary rewards (engaging in “episodic future thinking,” or visualizing themselves spending imaginary money; (Kwan et al., 2012; Kwan, Craver, Green, Myerson, & Rosenbaum, 2013; Palombo, Keane, & Verfaellie, 2015; Patt, Hunsberger, Jones, & Verfaellie, 2023)); or (2) playing for non-monetary experiential rewards (Patt et al., 2023). Therefore, while overall choice behavior may not appreciably differ between hypothetical and real-reward DD paradigms in healthy populations (M. W. Johnson & Bickel, 2002), additional brain regions and neural processes are nevertheless active when outcomes are projected and/or expected during decision-making. This distinction must therefore be considered when assessing cross-species disparities in the effects of experimental manipulations.
Table 9.
Evidence for cross-species validity of tasks that assess delay-based reward valuation.
| Reward Valuation – Reward Delay | ||
|---|---|---|
| Validity Domain | Human Task – Delay Discounting (DD) | Rodent Task – Delay Discounting (DD) |
| Predictive | ↑ DD in people with ADHD (meta-analysis) (Jackson and MacKillop, 2016). |
↑ DD in spontaneously hypertensive rats (Wooters and Bardo, 2011). ↑ DD in spontaneously hypertensive rats (Fox et al., 2023). ↑ DD in frontal dopamine-depleted rats; Freund et al., 2014 |
| ↑ DD in people with more severe alcohol use disorder than those with less severe alcohol use disorder (Phung et al., 2019). ↑ DD across alcohol, tobacco, cannabis, stimulant, opiate, and gambling addictions (meta-analysis) (Amlung et al., 2017) |
↑ DD in rats bred for high alcohol drinking (Wilhelm and Mitchell, 2008). ↑ DD in cocaine-self-administered rats (Mendez et al., 2010). |
|
| Neurobiological | OFC damage ↑ DD (Peters et al., 2020). Transcranial direct current stimulation over OFC ↓ DD (Moro et al., 2023). |
Medial and lateral OFC lesions respectively ↓ and ↑ DD (Mar et al., 2011). |
| Bi-frontal direct current stimulation (left dlPFC facilitated and right dlPFC inhibited) ↓ DD (Hecht et al., 2013). Anodal stimulation left dlPFC ↓ DD (He et al., 2016). Anodal stimulation of left dlPFC ↓ DD, cathodal stimulation left dlPFC ↑ DD (no effect of either on right dlPFC) (Shen et al., 2016). vmPFC damage ↑ DD (Mok et al., 2021). |
mPFC inactivation ↑ DD (Churchwell et al., 2009). | |
| Inverse relationship between DD and ventral striatal [11C]raclopride binding potential in people with gambling disorder (Joutsa et al., 2015). | NAc D2/3R availability inversely correlated with DD (Barlow et al., 2018). | |
| Insula damage ↓ DD (Sellitto et al., 2015) | Intra-insula D1R blockade ↑ DD (Pattij et al., 2014). | |
| Medial temporal lobe damage did not affect DD (Kwan et al., 2012). Medial temporal lobe damage did not affect DD (Kwan et al., 2013). Medial temporal lobe damage ↑ DD specifically during concomitant episodic future thinking (Palombo et al., 2015). Medial temporal lobe damage ↓ DD in experiential but not hypothetical assessments (Patt et al., 2023). |
Hippocampus lesion ↑ DD (Cheung and Cardinal, 2005). Ventral hippocampus lesion ↑ DD (Abela and Chudasama, 2013). |
|
| Pharmacological | Amphetamine (10 and 20 mg, acute) ↓ DD (de Wit et al., 2002). | Amphetamine (1–2.3 mg/kg, i.p., acute) ↓ DD (Winstanley et al., 2003). Amphetamine (0.25 mg/kg, i.p., acute) ↓ DD (Floresco et al., 2007). Amphetamine respectively ↑ and ↓ DD in cued (1.0 and 1.6 mg/kg, i.p., acute) and un-cued paradigms (0.3 mg/kg, i.p., acute) (Cardinal et al., 2000). Amphetamine (1.0 and 1.7 mg/kg, i.p., acute) ↑ DD (Slezak and Anderson, 2009). Amphetamine (0.80 and 1.20 mg/kg, i.p., acute) ↑ DD (adjusting amount procedure) (Helms et al., 2006). |
| Haloperidol (2 mg, acute) ↓ DD (Wagner et al., 2020). Haloperidol (1.5 mg, acute) did not affect DD (Pine et al., 2010). |
Haloperidol (.01–0.1 mg/kg, i.p., acute) ↑ DD (Boomhower and Rasmussen, 2014). | |
| Acute testosterone gel (150 mg) ↑ DD in healthy men (Wu et al., 2020). Acute testosterone gel (50 mg) did not affect DD in healthy men (Ortner et al., 2013). |
Chronic testosterone (7.5 mg/kg, s.c., 15 weeks) ↓ DD in male rats (Wood et al., 2013). | |
| Diazepam (20 mg, acute) did not affect DD (Acheson et al., 2006). | Diazepam (1–10 mg/kg, i.p., acute) dose- and strain-dependently modulated DD (Huskinson and Anderson, 2012). | |
Clinical DD tasks are somewhat pharmacologically sensitive, although effects are not always reproducible. For example, d-amphetamine reduced DD in healthy humans (H. de Wit, Enggasser, & Richards, 2002) (but not in a paradigm presenting interleaved DD and PD trials (A. Acheson & de Wit, 2008)), as did the dopamine D2R antagonist haloperidol (at 2 mg (Wagner, Clos, Sommer, & Peters, 2020), but not 1.5 mg (Pine, Shiner, Seymour, & Dolan, 2010)). These findings may indicate a U-shaped relationship between dopamine tone and delay-based reward valuation in humans whereby both increases and decreases in dopamine transmission increase tolerance to delayed delivery of large rewards. Additionally, acute application of testosterone gel elevated DD in healthy men (at 150 (Wu et al., 2020) but not 50 mg (Ortner et al., 2013)), with no effects of acute diazepam (A. Acheson, Reynolds, et al., 2006), oxycodone (Zacny & de Wit, 2009), THC (McDonald et al., 2003), or bupropion (A. Acheson & de Wit, 2008) on DD trials interspersed with PD trials. Thus, few extant studies describe drug effects on DD (and even fewer do so separately from concurrent PD assessment), rendering the pharmacological sensitivity of these assays inconclusive.
Rodent DD procedures are largely identical to those of preclinical PD tasks and exclusively operate on “experiential,” or real-reward contingencies, versus the hypothetical and “potentially real” designs common in clinical study. Touchscreen DD paradigms exist (Abela & Chudasama, 2013) (Figure 6), although lever- and nosepoke-based assessments are more commonly used and are largely identical in terms of procedure. As with PD paradigms, rodent DD tasks typically proceed block-wise, with delays associated with the larger reward progressively increasing or decreasing across testing sessions. Hence, rodents must be well-trained on the DD task prior to assessment, producing similar limitations as those described above for PD tasks (e.g., habitual responding; (Slezak & Anderson, 2009)). Consistently, certain pharmacological effects are dependent upon order of delay presentation (i.e., ascending versus descending; (Tanno, Maguire, Henson, & France, 2014)), although anchoring of response patterns based on original task contingencies is less persistent in mouse DD than PD (Rojas et al., 2022). Alternatively, rodents may be tested in a less common “adjusting amount” procedure, which: (1) adaptively changes magnitude of the immediate reinforcer in response to animals’ last choice and; (2) eschews block-wise organization by instead modulating delay duration across several testing days (Richards, Mitchell, de Wit, & Seiden, 1997).
Increased DD was detected in the spontaneously hypertensive (Fox et al., 2023; Wooters & Bardo, 2011) and frontal dopamine depletion (Freund et al., 2014) rat models of ADHD, as well as in rat models of stimulant and alcohol use (Mendez et al., 2010; Wilhelm & Mitchell, 2008), supporting the predictive validity of rodent DD tasks (Table 9). Neurobiological validity of rodent DD tasks is also supported, although little work has been done using touchscreens. While inactivating rodent OFC in general produced mixed results on DD performance (Abela & Chudasama, 2013; Stopper et al., 2014; Winstanley, Theobald, Cardinal, & Robbins, 2004), lesions of discrete medial versus lateral subregions respectively increased and decreased DD (Mar, Walker, Theobald, Eagle, & Robbins, 2011). Meanwhile, inactivating rodent mPFC (Seamans et al., 2008; Uylings et al., 2003) increased choice of the small, immediate reward (Churchwell, Morris, Heurtelou, & Kesner, 2009) (but see: (Cardinal, Pennicott, Sugathapala, Robbins, & Everitt, 2001)), consistent with dlPFC involvement in human DD (see above). Ventral striatal dopamine D2/3R binding is similarly involved across clinical and rodent DD tasks, with accumbal D2/3R -availability inversely correlated with DD in rats (Barlow et al., 2018). Furthermore, while lesion/inactivation studies targeting insular regions have yet to be conducted on rodents, intra-insular dopamine D1R blockade increased DD in rats (Pattij, Schetters, & Schoffelmeer, 2014), demonstrating this structure’s involvement in DD across species. Contrary to clinical findings however, hippocampal lesions increased DD in rats (Abela & Chudasama, 2013; Cheung & Cardinal, 2005) (including in a touchscreen paradigm; (Abela & Chudasama, 2013)), although both the specificity of the respective lesions in human versus rodent studies as well as the contingency of clinical findings on such factors as “episodic future thinking” must be considered when comparing findings across species.
Pharmacological validity of preclinical DD tasks is difficult to determine due to the relative scarcity of human pharmacological studies. Nevertheless, inter-species pharmacological effects on DD studies tend to be inconsistent. For example, while the reduction of human DD by amphetamine (H. de Wit et al., 2002) has been reproduced in rats (Floresco, Tse, & Ghods-Sharifi, 2008; Winstanley, Dalley, Theobald, & Robbins, 2003), the opposite effect was reported in rats and mice tested in less common paradigms utilizing cued delays and adjusting amount procedures (Cardinal, Robbins, & Everitt, 2000; Helms, Reeves, & Mitchell, 2006; Slezak & Anderson, 2009). Cross-species disparity is also evident from other studies. For example, DD was increased by haloperidol in rats (Boomhower & Rasmussen, 2014) but decreased (or unaffected; (Pine et al., 2010)) in humans (Wagner et al., 2020). Meanwhile, chronic testosterone treatment increased preference for larger, delayed rewards in male rats (Wood et al., 2013), contrary to the effects of acute testosterone in men (Wu et al., 2020), while diazepam dose- and strain-dependently modulated DD in rats (Huskinson & Anderson, 2012) but did not affect humans (concurrent DD/PD assessment; (A. Acheson, Reynolds, et al., 2006)). Thus, pharmacological predictive validity of preclinical DD tasks is currently lacking.
The question of reinforcement complicates comparison of DD findings across species. A relatively small portion of extant human DD literature utilizes explicit (i.e., non-hypothetical) reinforcers, and of these studies, only a small subset deliver reinforcement following individual trials within the session itself (Reynolds & Schiffbauer, 2004), versus hypothetical days or months. As with certain PD variants discussed above, the majority of human DD paradigms thereby remove the component of reward receipt/consumption from the assessment, a fundamental component of preclinical operant paradigms that can also influence human performance of non-monetary DD tasks (Navarick, 2004). While touchscreen operant systems may conceivably enable eschewal of block-wise session structures in discounting tasks and thereby prevent perseverative responding (see “Reward Probability”), they are unlikely to facilitate replacement of real-time reinforcement. Cross-species consistency may therefore be best enhanced by simply standardizing testing procedures within clinical and preclinical realms, ultimately facilitating translation by reducing the number of task variants.
Effort for reward
Progressive ratio breakpoint
Progressive ratio breakpoint tasks (PRBTs) require participants to complete a gradually increasing number of effortful actions (cognitive or physical) in exchange for a fixed reward (Strauss et al., 2016; Wolf et al., 2014). The task ends when the participant determines that the requisite effort exceeds the value of the reward and stops responding, i.e., reaches their “breakpoint”. While the PRBT has long-been used in animal research to study effortful motivation (Hodos, 1961) and characterize the reinforcing properties of drugs (Brady & Griffiths, 1976), these paradigms have only recently begun to gain footing in clinical studies. Task procedures are highly consistent across species, the primary difference being the modality of response (Figure 7). Moreover, all clinical PRBTs incorporate some form of response feedback (with many delivering actual rewards), thereby addressing limitations pertaining to reward processing discussed for PD and DD tasks.
Figure 7.
The progressive ratio breakpoint task (PRBT) for humans (left) and the operant and touchscreen-based equivalents for rodents (middle and right respectively). The PRBT is a cross-species paradigm that measures the degree to which a subject will expend physical effort to achieve a reward. In humans, effort is typically quantified by joystick rotations or button presses, while rodents are required to nosepoke into an aperture (middle) or on a touchscreen (right). The number of responses required to earn a reward is progressively increased across the session. The primary outcome variable is the point at which the subject stops responding, i.e., the “breakpoint”.
The PRBT demonstrates both clinical and preclinical sensitivity. Reduced breakpoints are reliably detected in people with schizophrenia (Strauss et al., 2016; Wolf et al., 2014) and unipolar or bipolar depression (Hershenberg et al., 2016), which have been recreated in animal models (Tam et al., 2010) (Dieterich, Liu, & Samuels, 2021; Dieterich et al., 2019; Picard et al., 2021) (Young et al., 2015) (Table 10). Meanwhile, a touchscreen paradigm detected reduced breakpoints in a mouse model of Huntington’s disease, consistent with clinical findings (Heath et al., 2019). Few clinical studies to date have collected neurophysiological measures alongside PRBT assessment, although some convergent findings across species support neurobiological validity of the rodent tasks. The ventral striatum is implicated in both human and rodent performance, with lower activation correlating with lower breakpoints in people with schizophrenia (Wolf et al., 2014) and NAc inactivation reducing breakpoints in rats (Moscarello, Ben-Shahar, & Ettenberg, 2010). Increased parietal alpha power was also observed in humans and mice as they neared their breakpoints (Cavanagh et al., 2021; Noback et al., 2024), constituting a common electrophysiological biomarker of performance across species. Finally, pharmacological predictive validity is supported whereby human and rodent breakpoints were similarly elevated by acute amphetamine (Bensadoun, Brooks, & Dunnett, 2004; Hailwood et al., 2018; Heath et al., 2015; Noback et al., 2024).
Table 10.
Evidence for cross-species validity of tasks that assess effort-based reward valuation – Progressive ratio breakpoint.
| Reward Valuation – Effort #1 | ||
|---|---|---|
| Validity Domain | Human Task – Progressive Ratio Breakpoint (PRBT) | Rodent Task – Progressive Ratio Breakpoint (PRBT) |
| Predictive | ↓ breakpoints in people with schizophrenia (Wolf et al., 2014). ↓ breakpoints in people with schizophrenia (Strauss et al., 2016). |
↓ breakpoints in Sp4 hypomorphic mice (trend-level) (Young et al., 2015). |
| ↓ breakpoints in people with unipolar or bipolar depression (Hershenberg et al., 2016). | ↓ breakpoints in mice following chronic corticosterone (Dieterich et al., 2019). ↓ breakpoints in mice following social defeat stress (Dietrich and Samuels, 2021). ↓ breakpoints in mice following chronic unpredictable mild stress (Picard et al., 2021). |
|
| ↓ breakpoints in people with Huntington’s disease (Heath et al., 2019). | ↓ breakpoints in R6/1 mice (Heath et al., 2019). | |
| Neurobiological | Lower ventral striatal activation correlates with lower breakpoints in people with schizophrenia (Wolf et al., 2014). | NAc inactivation ↓ breakpoint (Moscarello et al., 2010). |
| ↑ parietal alpha power approaching breakpoint (Cavanagh et al., 2021). ↑ parietal alpha power approaching breakpoint (Noback et al., 2024). |
↑ parietal alpha power approaching breakpoint (Cavanagh et al., 2021). ↑ parietal alpha power approaching breakpoint (Noback et al., 2024). |
|
| Pharmacological | Amphetamine (10 and 20 mg) ↑ breakpoint (Noback et al., 2024). | Amphetamine (1 mg/kg, acute) ↑ breakpoint (Bensadoun et al., 2004). Amphetamine (1 mg/kg, i.p., acute) ↑ breakpoint (Heath et al., 2015). Amphetamine (1 mg/kg, i.p., acute) ↑ breakpoint (Hailwood et al., 2018). Amphetamine (0.3 mg/kg, i.p., acute) ↑ breakpoint (Noback et al., 2024). |
As mentioned above, a touchscreen version of the PRBT is available that is capable of sustaining vigorous responding despite the lower level of tactile feedback provided by this system relative to more traditional operant boxes (Hailwood et al., 2018; Heath et al., 2015) (Figure 7). The touchscreen PRBT is sensitive to both pharmacological (Hailwood et al., 2018; Heath et al., 2015; Heath et al., 2016) and genetic manipulations (Heath et al., 2019) and the electrophysiological biomarker, parietal alpha, also increases approaching breakpoint (Noback et al., 2024) (Table 10). Interestingly, the touchscreen task detected a greater amphetamine-induced increase in breakpoint than had been previously reported using a nosepoke aperture-based system (Bensadoun et al., 2004; Heath et al., 2015), suggesting an increased sensitivity to performance-enhancing manipulations (Heath et al., 2015).
Incorporation of this task into a battery of touchscreen-based paradigms would enable assessment of effort-based reward valuation/effortful motivation alongside other cognitive domains in individual cohorts of animals (Heath et al., 2015). Such a design would in turn facilitate multifaceted characterization of animal models and determination of any motivational contributions to other behaviors (e.g., (Kwiatkowski et al., 2020; Roberts et al., 2021; Roberts et al., 2019)).
Effort-Based Decision-Making Tasks
While the PRBT enables assessment of effortful motivation across species, it is limited in that its outcome is unitary (i.e., simply when the participant chooses to stop responding). However, other tasks exist which offer choices between exerting effort for larger rewards versus nominal effort for smaller rewards. The archetypal clinical test of effort-based decision-making is the effort expenditure for rewards task (EEfRT; (Treadway, Buckholtz, Schwartzman, Lambert, & Zald, 2009)). The EEfRT is a multi-trial paradigm in which participants repeatedly choose between fulfilling easy versus difficult physical response criteria in exchange for “potentially real” monetary rewards of varying amounts (defined in Delayed reward). Reward magnitudes associated with the difficult option are modulated across trials, as is the probability that either choice would be rewarded at all. Furthermore, real-time feedback is provided during and after responding, indicating: (1) percent completion of the trial’s response criterion; (2) attainment or failure of the criterion; and (3) reward or non-reward (Figure 8).
Figure 8.
The effort-expenditure for rewards task (EEfRT) for humans (left) and the operant and touchscreen-based equivalent tasks for rodents (right). In the human task, subjects decide whether to make more button presses with their non-dominant hand (more challenging) for a large reward or fewer button presses with their dominant hand (easier) for a smaller reward. Both reward magnitude and probability of reward delivery are modulated from trial to trial. In the rodent effort discounting (ED) task (middle), rodents choose between completing many nosepokes for a large reward or a single nosepoke for a smaller reward. The number of responses required for the larger reward increase across trial blocks. The rearing effort discounting (RED) task (right) is a recently developed touchscreen-based effort-based decision-making task that provides the choice between earning small and large rewards by responding to, respectively, a stimulus presented at floor level (easy option) and one presented higher on the screen that requires the animal to rear up in order to reach it (hard option). Primary outcome measures for these tasks are the number of choices of the more challenging option versus the easier one, while ancillary measures such as attrition across the task provide indices of motivational failure over time.
Clinical sensitivity of the EEfRT is well-established, with deficits identified in people with major depression (Treadway, Bossaller, Shelton, & Zald, 2012), schizophrenia (e.g., (Barch, Treadway, & Schoen, 2014)), and autism spectrum disorder (Damiano, Aloi, Treadway, Bodfish, & Dichter, 2012) (Table 11). As with PD and DD tasks, converging evidence implicates frontal and striatal areas in EEfRT performance. Resting left frontal cortical activity predicted greater effort expenditure when probability of reward was sub-optimal (Hughes, Yates, Morton, & Smillie, 2015; Kaack, Chae, Shadli, & Hillman, 2020). Accordingly, transcranial direct current stimulation over the left dlPFC selectively increased effortful choice in trials with low reward probability or high potential reward (Ohmann, Kuper, & Wacker, 2018). Additionally, dopamine response in vmPFC and left caudate nucleus to amphetamine administration positively correlated with effortful choice in low-probability trials (Treadway, Buckholtz, et al., 2012). Meanwhile, NAc activity positively correlated with effortful choice in intermediate- and high-probability trials (Huang et al., 2016), while caudate activity was negatively correlated with high-effort selection during high-probability trials (X. H. Yang et al., 2016). While pharmacological studies utilizing the EEfRT are few, amphetamine (but not caffeine, nicotine, naltrexone, THC, or CBD; (Lawn et al., 2016; Nunez et al., 2022; Wardle, Treadway, & de Wit, 2012)) consistently increased effortful choice in healthy participants (Soder et al., 2021; Wardle, Treadway, Mayo, Zald, & de Wit, 2011), as seen in the PRBT.
Table 11.
Evidence for cross-species validity of tasks that assess effort-based reward valuation – Effort Expenditure for Rewards Task.
| Reward Valuation – Effort #2 | |||
|---|---|---|---|
| Validity Domain | Human Task – Effort Expenditure for Rewards Task (EEfRT) | Rodent Task – Effort Discounting (ED) | Rodent Touchscreen Task – Rearing Effort Discounting (RED) |
| Predictive | ↓ effortful choice in people with major depression (Treadway et al., 2012). |
↑ ED in mice following acute restraint stress (Shafiei et al., 2012). ↑ ED in rats following social stress (Lemon and Del Arco, 2022). ↑ ED in mice gestated under winter-like light cycle (Roberts and O’ Connor et al., 2023). |
N/A |
| ↓ effortful choice in people with schizophrenia (Barch et al., 2014). | ↑ ED in mice gestated under winter-like light cycle (Roberts and O’ Connor et al., 2023). | ||
| Neurobiological | Transcranial direct current stimulation over left dlPFC ↑ effortful choice (low-probability or high-reward trials) (Ohmann et al., 2018). Resting state left frontal cortex activity predicted effort expenditure (low-probability trials) (Hughes et al., 2015). Resting state left frontal cortex activity predicted effort expenditure (low-probability trials) (Kaack et al., 2020). vmPFC dopamine responsivity positively correlated with effortful choice (low-probability trials) (Treadway et al., 2012). |
mPFC lesions ↑ ED (T-maze paradigm) (Walton et al., 2002). ↓ mPFC D2R expression associated with ↓ ED (Simon et al., 2013). |
N/A |
| NAc activity positively correlated with effortful choice in people (intermediate- and high-probability trials) (Huang et al., 2016). Caudate activity negatively correlated with effortful choice (high-probability trials) (Yang et al., 2016). Left caudate dopamine responsivity positively correlated with effortful choice (low-probability trials) (Treadway et al., 2012). |
NAc core inactivation ↑ ED (Ghods-Sharifi et al., 2010). Intra-NAc dopamine D2/3R agonist ↑ ED (Bryce and Floresco, 2019). |
||
| Pharmacological | Amphetamine (20 mg, acute) ↑ effortful choice (Wardle et al., 2011). Amphetamine (20 mg, acute) ↑ effortful choice (Soder et al., 2021). |
Amphetamine (0.5 mg/kg, i.p., acute) ↑ ED (↓ effortful choice) (Floresco et al., 2007). Amphetamine (0.1 mg/kg, i.p., acute) ↓ ED (↑ effortful choice) in mice gestated under winter-like light cycles (but ↑ ED in controls) (Roberts and O’Connor et al., 2023). |
Amphetamine (0.25 and 0.50 mg/kg, acute) ↓ ED (↑ effortful choice) in haloperidol-treated (0.5–1.5 mg/kg) mice (Lopez-Cruz et al., 2024). |
Similar to the PRBT, the EEfRT was preceded by several similar rodent tasks. An early such task utilized a modified T-maze paradigm whereby rats were simultaneously offered obstructed access to a preferred reinforcer (blocked by a scalable barrier) and unobstructed access to a smaller quantity of standard lab chow (Salamone, Cousins, & Bucher, 1994). Development of an operant effort discounting (ED) task enabled more trials and removed the confound of repeatedly handling animals during assessment. The ED task utilizes a similar two-lever design as PD and DD paradigms, whereby a large reward is offered at a progressively increasing or decreasing response cost alongside a smaller reward available at a consistently low cost (Floresco et al., 2008). As with the human EEfRT, the primary outcome is choice of high versus low effort options throughout the session, albeit in a systematic block-wise fashion (Figure 8).
Preclinical ED tasks detected effortful choice deficits in mice gestated under a winter-like light cycle (a risk factor for schizophrenia, depression, and bipolar disorder; (Disanto et al., 2012)) (Roberts, O'Connor, Kenton, Barnes, & Young, 2023), as well as in rodent stress models (Lemon & Del Arco, 2022; Shafiei, Gray, Viau, & Floresco, 2012), supporting preclinical sensitivity (Table 11). The effects of rodent neurological manipulations largely converge with clinical findings, demonstrating overall neurobiological validity of these assays. For example, medial frontal cortex lesions biased rats towards low-effort options in the T-maze paradigm (Walton, Bannerman, & Rushworth, 2002), and inactivating the NAc core reduced effortful choice in the operant ED task (Ghods-Sharifi & Floresco, 2010). Furthermore, application of the dopamine D2/3R agonist quinpirole to either NAc core or shell similarly increased rat ED (Bryce & Floresco, 2019), while lower mPFC D2R expression was associated with less ED in rats (N. W. Simon et al., 2013). Additionally, pharmacological predictive validity of the rodent ED task is supported by findings that amphetamine consistently increases effortful choice across humans and rodents (Floresco et al., 2008; Soder et al., 2021; Wardle et al., 2011) (Roberts et al., 2023) (but see: (Floresco et al., 2008)), as seen in the PRBT.
While the convergence in neurobiological substrates of effort-based reward valuation is promising, it should be noted that most of the clinical imaging findings reported above were specific to conditions of variably low, intermediate, or high reward probability. For example, dopamine function was most strongly associated with willingness to expend effort during low-probability EEfRT trials, in which both effort and probabilistic response costs were high (Treadway, Buckholtz, et al., 2012). Varying reward probabilities in the EEfRT was partly intended to ensure that subjects’ trial-to-trial performance reflected spontaneous decision making, as this added complexity would likely make calculation of long-term response strategies untenable (Treadway et al., 2009); however, this design may conflate effort- and probability-based reward valuation. While the relative contributions of effort versus probability discounting to EEfRT performance may be parsed by computational modeling (Soder et al., 2021), cross-species comparison of findings is nevertheless complicated by the fact that no existing preclinical paradigms modulate both dimensions simultaneously. Even with the advantages offered by touchscreen operant systems, such a complicated, multi-variable procedure may be beyond the ability of rodents to learn.
Some shortcomings of the rodent ED task are addressable by touchscreen paradigms. For example, completion of high-effort response criteria in this task carries a concurrent temporal cost, as it takes more time to complete multiple lever presses than one. A given manipulation’s effects on ED can therefore be conflated with its effects on DD. Indeed, 0.25 mg/kg and 0.50 mg/kg amphetamine originally produced opposite effects in a standard rat ED paradigm; however, follow-up testing using an “equivalent delay” ED task (with time to reward delivery normalized across low- and high-effort options) and a separate DD task parsed this effect profile into separate, dose-specific actions on these two distinct forms of decision making (Floresco et al., 2008). Furthermore, the block-wise task organization complicates ED assessment in a similar manner as PD and DD, with the order of response costs (ascending versus descending) strongly influencing the shape of the resultant discounting curve (Floresco et al., 2008). Both of these limitations are addressed by the novel touchscreen Rearing-Effort Discounting (RED) task (Lopez-Cruz et al., 2024). The RED task presents rodents with two stimuli requiring individual nosepokes; one stimulus yields a small reward and is readily accessible (floor-level), while the other yields a larger reward, but is positioned such that the rodent must rear to reach it. The height of the large-reward stimulus (hence requisite effort), is modulated from trial to trial, although the temporal cost remains consistent. The spatial arrangement of the stimuli enables explicit signaling of effort costs within individual trials, enabling circumvention of trial blocks. Both greater rearing distance and dopamine D2R blockade (haloperidol) reduced selection of the high-effort stimulus while amphetamine rescued the haloperidol-induced deficit, demonstrating parametric and pharmacological validation (Lopez-Cruz et al., 2024). Although additional validation is required, the RED task is a prime example of how touchscreen operant systems may be leveraged to eliminate confounding factors and off-target cognitive demands in the assessment of reward valuation.
Discussion
Here, we reviewed the positive valence system as outlined by the RDoC framework and evaluated the potential of preclinical rodent research to investigate the neurobiology underlying these functions in non-pathological and disease states. Given the potential for cross-species translation, we also identified the utility and validity of touchscreen-based rodent tasks that assess such positive valence domains. Touchscreen-based tasks have been most commonly implemented to investigate aspects of reward learning with evidence for neurobiological, pharmacological, and predictive validity. Although the touchscreen systems have been less extensively used to assess functions of reward valuation, there are available paradigms for discounting (both probability and delay) and effort-based motivation (e.g., PRBT and RED task), along with an extensive literature using operant paradigms. Reward responsiveness is the least characterized of the RDoC domains in preclinical research, which may in part be due to the lack of relevant behavioral endpoints in paradigms such as the MID (i.e., reward anticipation) or simple guessing task (i.e., initial response to reward), thus necessitating the incorporation of neural recoding measures to capture reward-related activity. Nonetheless, equivalent task designs can be easily adapted for rodents and pairing in vivo techniques such as imaging calcium or voltage indicators, electrophysiology, or mini-scopes could provide a precision approach for investigating the mechanisms underlying reward anticipation and initial response (see Table 1). Additionally, measures of reward satiation can be assessed in essentially all operant and touchscreen paradigms, represented in the response rate and latencies at the end of the session compared to the start, or separately in fixed ratio paradigms. Overall, there are a number of established and recently developed rodent paradigms suitable for investigating the positive valence systems, and future validation will aid in establishing paradigms that can determine neural circuitry, characterize clinically informed disease models, and test potential therapeutic compounds.
As discussed, there is an extensive amount of shared neurobiological circuitry across the RDoC categories, as evidenced by imaging and lesion studies in both humans and rodents. Commonly recruited brain structures include the OFC - strongly implicated in reward value representation and updating - and the striatum - dopaminergic signaling facilitates the anticipation of reward, the detection of unexpected reward, and using contextual associations to promote reward-related actions. Thus, future efforts aimed at dissociating whether neurobiology is similarly recruited by related neural computations (e.g., anticipating reward and prediction errors), as opposed to behavioral paradigms requiring various reward-related functions (e.g., probabilistic reversal learning), will be important for identifying mechanisms that are shared and distinct across reward-related functions. Relatedly, there is tremendous overlap in the reward-related behavioral abnormalities that are present across human psychiatric disorders. For example, individuals with schizophrenia and depression both show impairments in probabilistic learning, exerting effort to obtain rewards, and discounting future rewards, which may be the result of reductions in hedonic capacity and/or reward-seeking motivation (Höflich, Michenthaler, Kasper, & Lanzenberger, 2019; Oorschot et al., 2013; Treadway & Zald, 2011), but not when that learning is implicit (Danion, Meulemans, Kauffmann-Muller, & Vermaat, 2001; Perry, Light, Davis, & Braff, 2000; Soler, Ruiz, Dasí, & Fuentes-Durá, 2015; Sponheim, Steele, & McGuire, 2004). Importantly, the degree of individual symptom severity, rather than diagnostic category, appears to be a strong predictor of reward-related dysfunction as observed in depression, anxiety, and schizophrenia studies (Lewandowski et al., 2016; Erin E Reilly et al., 2020; Strauss et al., 2015). The degree of overlap in neurobiology underlying the fundamental behavior and symptomology in disease not only supports a transdiagnostic approach for clinical populations but highlights the necessity of preclinical approaches with high specificity, both in the behavior being assessed and the neural circuits being targeted. Together, the touchscreen systems may provide a valuable platform for applying the RDoC framework to preclinical research. This approach can be maximized by combining translational tasks developed by forward or reverse translation with recording and manipulation techniques possessing high cellular specificity and temporal precision, enabling researchers to dissect the mechanisms underlying positive valence functions.
Future Directions
It is important to note the potential limitations and considerations that should be addressed in future research investigating positive valence systems in rodents. First, while visual stimuli can be similarly presented to humans and rodents on a computerized screen, the method of task interaction differs across the human paradigms described in this review. While rodents can nosepoke or press at a touch-sensitive screen, humans may instead use a joystick, keyboard, or mouse clicks, all of which are more common than an actual touchscreen (though they can be used). Using the same presentation media (e.g., a screen), is however, an important component of establishing tasks with cross-species translation, increasing the likelihood of connecting visual and cortical regions (Norman et al., 2021). Moreover, tasks that are designed for rodents can be forward translated for use in humans, facilitating increased use of human touchscreen tasks. Such an approach was taken with the paired associative learning task (PAL), which identified identical deficits in humans possessing a genetic risk factor for schizophrenia and mice with the corresponding mutation (Nithianantharajah et al., 2013; Nithianantharajah et al., 2015). Finally, the global availability of smart phones enables implementing touchscreens in human testing that can take place outside of a laboratory utilizing ecological momentary assessment approaches (Barch et al., 2023; Bomyea et al., 2021; Titone et al., 2022). Thus, touchscreen use could provide an avenue for delineating neural mechanisms underlying positive valence across species, with applicability to clinical testing.
Additional limitations for evaluating positive valence system functions in rodents, are also worth mentioning. The degree of variance in movement requirements within cross-species tasks is an important consideration given the corticostriatal dopaminergic contributions to both reward-related functions and action initiation/inhibition. While chamber systems offer a lower physical requirement compared to maze-based paradigms, differences in effort to achieve rewards between species may be a factor when comparing human and rodent results, even in comparable tasks. The touchscreen PRBT of motivational effort, however, was successful in identifying motivational deficits in humans with Huntington’s disease (Heath et al., 2019) and schizophrenia (Bismark et al., 2018; Young et al., 2015) and their corresponding mouse models, in addition to other disease states. Additionally, the potential for movement artifacts to interfere with neural activity related to reward processes during testing combined with recording techniques must always be considered and can be minimized with appropriate signal filtering and baselining. Furthermore, causal manipulations (e.g., optogenetics), that target corticostriatal circuitry should consider temporal epochs in which stimulation may be delivered outside of movement windows to avoid off-target behavioral perturbations. Lastly is the nature of the rewarding stimulus, which is typically monetary incentive in humans and food reward in rodents. As mentioned, there is evidence that anticipating food or money rewards elicits similar striatal activity in humans (J. J. Simon et al., 2015a). Similarly, when individuals valued food, money, and social rewards equally, the rewards were indistinguishable in their ability to drive stimulus-association learning (Lehner, Balsters, Herger, Hare, & Wenderoth, 2017). That said, ancillary reward-related functions such as reward satiation may be less of a consideration for monetary rewards during a testing session compared to food stimuli, again emphasizing the importance of implementing trial analyses in rodent paradigms to assess loss of reward valuation across a session. Lastly, the requirement of food restriction in rodent-based paradigms is an important consideration, as perpetual states of hunger may be associated with elevated levels of stress hormones and altered reward seeking, and may not be represented in the reward drive being assessed in humans (e.g., need for financial compensation) (Mallien et al., 2016; Quante et al., 2023).
While the touchscreen systems provide an unparalleled ability to present various forms of basic and complex stimuli to rodents, one limiting aspect in regard to assessing positive valence systems is the inability to test social cognition. One of the most debilitating consequences of reward-related dysfunction in psychiatric disease is the loss of interest and/or motivation to engage in social interactions, impeding development of interpersonal relationships and the ability to use social stimuli to navigate the world (Bellack, Morrison, Wixted, & Mueser, 1990). Unsurprisingly, deficits in reward valuation or motivation that are detected using other rewards (e.g., money) can also manifest as reduced social drive, which is persistent in conditions such as schizophrenia (Gard et al., 2014; Mueser, Bellack, Douglas, & Morrison, 1991), depression (Hirschfeld et al., 2000; Rhebergen et al., 2010), and PTSD (Brancu et al., 2014; Scoglio et al., 2022). Rodents are social animals and rats and mice will exert effort on a novel lever-pressing PRBT to achieve access to a social interaction (Lee, Venniro, Shaham, Hope, & Ramsey, 2024; Ramsey, Holloman, Hope, Shaham, & Venniro, 2022; Ramsey, Holloman, Lee, & Venniro, 2023; Venniro & Shaham, 2020). While such a task would require modification of the touchscreen chamber to include a compartment for the social partner, the PRBT component is already established in the touchscreens. Alternatively, human social tasks generally present participants with images of social stimuli and touchscreens do afford the ability to present essentially any visual stimulus to the rodents, including videos or 3-dimensional dynamic graphics. Interestingly, previous studies have shown that rats exhibit a place preference for a chamber equipped with a video recording of a conspecific over a barren chamber (Yakura et al., 2018). While the absence of olfactory and tactile interaction may limit the motivational capacity of virtual stimuli to drive rodent behavior, it would be interesting to investigate if pharmacological compounds that increase sociability in humans and rodents could modulate rodent’s behavioral responses to social stimuli presented on the touchscreen.
Overall, abnormalities in positive valence system functioning are a consistent and debilitating collection of symptoms present in various psychiatric disorders. In-line with the RDoC framework, studies in humans and rodents demonstrate a high degree of overlap in terms of the underlying neurobiology that supports these functions, as well as the consequences of pharmacological intervention. As efforts continue to further characterize the neurobiology and neurochemistry that facilitate reward-related processing, validated cross-species paradigms are a necessity for maximizing the translational potential of findings from preclinical to clinical settings. The touchscreen systems provide a valuable resource for this endeavor, enabling researchers to administer extremely similar, if not identical, tasks to humans and rodents. The success of translation, however, is dependent on establishing pharmacological, neurobiological, and predictive validity for rodent paradigms to ensure that analogous behavioral functions and neurological systems are being measured across species.
--Acknowledgments--
This work was funded by National Institute on Drug Abuse, (Grant / Award Number: 'T32DA031098 ') National Institute of Mental Health, (Grant / Award Number: 'R01MH128869 ','R01MH134175 ','R25MH081482 ') (grant number ): This information is usually included already, but please add to the Acknowledgments if not.
We would like to thank Ms. Mahalah R. Buell, Mr. Richard F. Sharp, and Dr. Susan B. Powell for all their support on the production of this review. We also thank support from NIH funding R01MH134175 (JY), R01MH128869 (JY), R25MH081482 (SMB/TD), and T32DA031098 (MN).
Abbreviations
- 5-HT
Serotonin
- ADHD
Attention-deficit/hyperactivity disorder
- ACC
Anterior cingulate cortex
- ACh
Acetylcholine
- BOLD
Blood oxygen level–dependent
- BLA
Basolateral amygdala
- CBD
Cannabidiol
- CNTRICS
Cognitive Neuroscience Treatment Research to Improve Cognition in Schizophrenia
- CNV
Copy number variant
- D1R
Dopamine receptor D1
- D2R
Dopamine receptor D2
- D2/3R
Dopamine receptor D2/3
- DA
Dopamine
- DD
Delay discounting
- dlPFC
Dorsolateral prefrontal cortex
- DLS
Dorsolateral striatum
- DS
Discriminative stimulus
- DSM
Diagnostic and statistical manual of mental disorders
- EEfRT
Effort expenditure for rewards task
- ED
Effort discounting
- ELS
Early life stress
- ERP
Event related potential
- fMRI
Functional magnetic resonance imagining
- FR
Fixed ratio
- GABA
Gamma-aminobutyric acid
- GECI
Genetically encoded calcium indicators
- GEVI
Genetically encoded voltage indicators
- GWAS
Genome wide association studies
- MID
Monetary incentive delay
- mPFC
Medial prefrontal cortex
- MSN
Medium spiny neurons
- NAc
Nucleus accumbens
- NBM
Nucleus basalis magnocellularis
- NEWMEDS
Novel Methods leading to New Medications in Depression and Schizophrenia
- NIMH
National Institute of Mental Health
- NMDA
N-methyl-D-aspartate
- O2
Regional oxygen
- OCD
Obsessive compulsive disorder
- OFC
Orbitofrontal cortex
- PAL
Paired associative learning
- PD
Probabilistic discounting
- PET
Positron emission tomography
- PFC
Prefrontal cortex
- PRBT
Progressive ratio breakpoint task
- PRLT
Probabilistic reversal learning task
- PRT
Probabilistic reward task
- RDoC
Research Domain and Criteria
- RED
Rearing-effort discounting
- RewP
Reward positivity
- RPE
Reward prediction error
- S-R-O
Stimulus – response – outcome
- SSRI
Selective serotonin re-uptake inhibitor
- THC
Delta-9-tetrahydrocannabinol
- vmOFC
Ventromedial orbitofrontal cortex
- vmPFC
Ventromedial prefrontal cortex
- VMCL
Visuomotor conditional learning
- VTA
Ventral tegmental area
Footnotes
ARRIVE guidelines have been followed:
=> if it is a Review or Editorial, skip complete sentence => if No, include a statement in the "Conflict of interest disclosure" section: "ARRIVE guidelines were not followed for the following reason:
"
(edit phrasing to form a complete sentence as necessary).
=> if Yes, insert in the "Conflict of interest disclosure" section:
"All experiments were conducted in compliance with the ARRIVE guidelines." unless it is a Review or Editorial
Conflicts of interest:
=> if 'none', insert "The authors have no conflict of interest to declare."
=> else insert info unless it is already included
--Human subjects --
Involves human subjects:
If yes: Informed consent & ethics approval achieved:
=> if yes, please ensure that the info "Informed consent was achieved for all subjects, and the experiments were approved by the local ethics committee." is included in the Methods.
Conflict of Interest
The authors report no conflict of interest
References
- Abela AR, & Chudasama Y. (2013). Dissociable contributions of the ventral hippocampus and orbitofrontal cortex to decision-making with a delayed or uncertain outcome. Eur J Neurosci, 37(4), 640–647. doi: 10.1111/ejn.12071 [DOI] [PubMed] [Google Scholar]
- Acheson A, & de Wit H. (2008). Bupropion improves attention but does not affect impulsive behavior in healthy young adults. Exp Clin Psychopharmacol, 16(2), 113–123. doi: 10.1037/1064-1297.16.2.113 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Acheson A, Farrar AM, Patak M, Hausknecht KA, Kieres AK, Choi S, … Richards JB. (2006). Nucleus accumbens lesions decrease sensitivity to rapid changes in the delay to reinforcement. Behav Brain Res, 173(2), 217–228. doi: 10.1016/j.bbr.2006.06.024 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Acheson A, Reynolds B, Richards JB, & de Wit H. (2006). Diazepam impairs behavioral inhibition but not delay discounting or risk taking in healthy adults. Exp Clin Psychopharmacol, 14(2), 190–198. doi: 10.1037/1064-1297.14.2.190 [DOI] [PubMed] [Google Scholar]
- Acheson DT, Vinograd M, Nievergelt CM, Yurgil KA, Moore TM, Risbrough VB, & Baker DG. (2022). Prospective examination of pre-trauma anhedonia as a risk factor for post-traumatic stress symptoms. Eur J Psychotraumatol, 13(1), 2015949. doi: 10.1080/20008198.2021.2015949 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Addicott MA, Pearson JM, Sweitzer MM, Barack DL, & Platt ML. (2017). A primer on foraging and the explore/exploit trade-off for psychiatry research. Neuropsychopharmacology, 42(10), 1931–1939. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Ahn WY, Rass O, Fridberg DJ, Bishara AJ, Forsyth JK, Breier A, … O'Donnell BF. (2011). Temporal discounting of rewards in patients with bipolar disorder and schizophrenia. J Abnorm Psychol, 120(4), 911–921. doi: 10.1037/a0023333 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Ahnallen CG, Liverant GI, Gregor KL, Kamholz BW, Levitt JJ, Gulliver SB, … Kaplan GB. (2012). The relationship between reward-based learning and nicotine dependence in smokers with schizophrenia. Psychiatry Res, 196(1), 9–14. doi: 10.1016/j.psychres.2011.09.011 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Akemann W, Sasaki M, Mutoh H, Imamura T, Honkura N, & Knöpfel T. (2013). Two-photon voltage imaging using a genetically encoded voltage indicator. Scientific reports, 3(1), 2231. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Alsio J, Lehmann O, McKenzie C, Theobald DE, Searle L, Xia J, … Robbins TW. (2021). Serotonergic Innervations of the Orbitofrontal and Medial-prefrontal Cortices are Differentially Involved in Visual Discrimination and Reversal Learning in Rats. Cereb Cortex, 31(2), 1090–1105. doi: 10.1093/cercor/bhaa277 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Alsiö J, Lehmann O, McKenzie C, Theobald DE, Searle L, Xia J, … Robbins TW. (2021). Serotonergic innervations of the orbitofrontal and medial-prefrontal cortices are differentially involved in visual discrimination and reversal learning in rats. Cerebral Cortex, 31(2), 1090–1105. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Alsio J, Nilsson SR, Gastambide F, Wang RA, Dam SA, Mar AC, … Robbins TW. (2015). The role of 5-HT2C receptors in touchscreen visual reversal learning in the rat: a cross-site study. Psychopharmacology (Berl), 232(21–22), 4017–4031. doi: 10.1007/s00213-015-3963-5 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Alsio J, Phillips BU, Sala-Bayo J, Nilsson SRO, Calafat-Pla TC, Rizwand A, … Robbins TW. (2019). Dopamine D2-like receptor stimulation blocks negative feedback in visual and spatial reversal learning in the rat: behavioural and computational evidence. Psychopharmacology (Berl), 236(8), 2307–2323. doi: 10.1007/s00213-019-05296-y [DOI] [PMC free article] [PubMed] [Google Scholar]
- Amlung M, Vedelago L, Acker J, Balodis I, & MacKillop J. (2017). Steep delay discounting and addictive behavior: a meta-analysis of continuous associations. Addiction, 112(1), 51–62. doi: 10.1111/add.13535 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Andreassen OA, Hindley GF, Frei O, & Smeland OB. (2023). New insights from the last decade of research in psychiatric genetics: discoveries, challenges and clinical implications. World Psychiatry, 22(1), 4–24. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Attachaipanich S, Ozawa T, Macpherson T, & Hikida T. (2023). Dual roles for nucleus accumbens core dopamine D1-expressing neurons projecting to the substantia nigra pars reticulata in limbic and motor control in male mice. Eneuro, 10(6). [DOI] [PMC free article] [PubMed] [Google Scholar]
- Ayoub SM, Noback MN, Ahern J, & Young JW. (2024). Using cross-species behavioral tools to determine mechanisms contributing to HIV-associated neurocognitive disorder and comorbid substance use. In HIV-Associated Neurocognitive Disorders (pp. 503–524): Elsevier. [Google Scholar]
- Balleine BW, & Dickinson A. (1998). Goal-directed instrumental action: contingency and incentive learning and their cortical substrates. Neuropharmacology, 37(4–5), 407–419. doi: 10.1016/s0028-3908(98)00033-1 [DOI] [PubMed] [Google Scholar]
- Balodis IM, Kober H, Worhunsky PD, Stevens MC, Pearlson GD, & Potenza MN. (2012). Diminished frontostriatal activity during processing of monetary rewards and losses in pathological gambling. Biol Psychiatry, 71(8), 749–757. doi: 10.1016/j.biopsych.2012.01.006 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Barch DM, Berman MG, Engle R, Jones JH, Jonides J, MacDonald A, … Sponheim SR. (2009). CNTRICS Final Task Selection: Working Memory. Schizophrenia Bulletin, 35(1), 136–152. doi: 10.1093/schbul/sbn153 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Barch DM, Carter CS, Gold JM, Johnson SL, Kring AM, MacDonald AW, … Strauss ME. (2017). Explicit and implicit reinforcement learning across the psychosis spectrum. J Abnorm Psychol, 126(5), 694–711. doi: 10.1037/abn0000259 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Barch DM, Culbreth AJ, Zeev DB, Campbell A, Nepal S, & Moran EK. (2023). Dissociation of Cognitive Effort–Based Decision Making and Its Associations With Symptoms, Cognition, and Everyday Life Function Across Schizophrenia, Bipolar Disorder, and Depression. Biological Psychiatry, 94(6), 501–510. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Barch DM, Treadway MT, & Schoen N. (2014). Effort, anhedonia, and function in schizophrenia: reduced effort allocation predicts amotivation and functional impairment. J Abnorm Psychol, 123(2), 387–397. doi: 10.1037/a0036299 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Barlow RL, Gorges M, Wearn A, Niessen HG, Kassubek J, Dalley JW, & Pekcec A. (2018). Ventral Striatal D2/3 Receptor Availability Is Associated with Impulsive Choice Behavior As Well As Limbic Corticostriatal Connectivity. Int J Neuropsychopharmacol, 21(7), 705–715. doi: 10.1093/ijnp/pyy030 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Barnes SA, Dillon DG, Young JW, Thomas ML, Faget L, Yoo JH, … Ramanathan DS. (2023). Modulation of ventromedial orbitofrontal cortical glutamatergic activity affects the explore-exploit balance and influences value-based decision-making. Cereb Cortex, 33(10), 5783–5796. doi: 10.1093/cercor/bhac459 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Bartra O, McGuire JT, & Kable JW. (2013). The valuation system: a coordinate-based meta-analysis of BOLD fMRI experiments examining neural correlates of subjective value. Neuroimage, 76, 412–427. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Baxter MG, Parker A, Lindner CC, Izquierdo AD, & Murray EA. (2000). Control of response selection by reinforcer value requires interaction of amygdala and orbital prefrontal cortex. Journal of Neuroscience, 20(11), 4311–4319. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Beck A, Schlagenhauf F, Wustenberg T, Hein J, Kienast T, Kahnt T, … Wrase J. (2009). Ventral striatal activation during reward anticipation correlates with impulsivity in alcoholics. Biol Psychiatry, 66(8), 734–742. doi: 10.1016/j.biopsych.2009.04.035 [DOI] [PubMed] [Google Scholar]
- Bellack AS, Morrison RL, Wixted JT, & Mueser KT. (1990). An analysis of social competence in schizophrenia. British Journal of Psychiatry, 156(6), 809–818. [DOI] [PubMed] [Google Scholar]
- Bensadoun JC, Brooks SP, & Dunnett SB. (2004). Free operant and discrete trial performance of mice in the nine-hole box apparatus: validation using amphetamine and scopolamine. Psychopharmacology (Berl), 174(3), 396–405. doi: 10.1007/s00213-003-1751-0 [DOI] [PubMed] [Google Scholar]
- Bergstrom HC, Lieberman AG, Graybeal C, Lipkin AM, & Holmes A. (2020). Dorsolateral striatum engagement during reversal learning. Learning & Memory, 27(10), 418–422. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Bergstrom HC, Lipkin AM, Lieberman AG, Pinard CR, Gunduz-Cinar O, Brockway ET, … Holmes A. (2018). Dorsolateral Striatum Engagement Interferes with Early Discrimination Learning. Cell Rep, 23(8), 2264–2272. doi: 10.1016/j.celrep.2018.04.081 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Bernhardt N, Nebe S, Pooseh S, Sebold M, Sommer C, Birkenstock J, … Smolka MN. (2017). Impulsive Decision Making in Young Adult Social Drinkers and Detoxified Alcohol-Dependent Patients: A Cross-Sectional and Longitudinal Study. Alcohol Clin Exp Res, 41(10), 1794–1807. doi: 10.1111/acer.13481 [DOI] [PubMed] [Google Scholar]
- Bernhardt N, Obst E, Nebe S, Pooseh S, Wurst FM, Weinmann W, … Zimmermann US. (2019). Acute alcohol effects on impulsive choice in adolescents. J Psychopharmacol, 33(3), 316–325. doi: 10.1177/0269881118822063 [DOI] [PubMed] [Google Scholar]
- Bernstein JG, & Boyden ES. (2011). Optogenetic tools for analyzing the neural circuits of behavior. Trends in cognitive sciences, 15(12), 592–600. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Bernstein JG, Garrity PA, & Boyden ES. (2012). Optogenetics and thermogenetics: technologies for controlling the activity of targeted cells within intact neural circuits. Current Opinion in Neurobiology, 22(1), 61–71. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Berridge KC, & Kringelbach ML. (2008). Affective neuroscience of pleasure: reward in humans and animals. Psychopharmacology, 199, 457–480. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Bidwell LC, MacKillop J, Murphy JG, Grenga A, Swift RM, & McGeary JE. (2013). Biphasic effects of alcohol on delay and probability discounting. Exp Clin Psychopharmacol, 21(3), 214–221. doi: 10.1037/a0032284 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Bismark AW, Thomas ML, Tarasenko M, Shiluk AL, Rackelmann SY, Young JW, & Light GA. (2018). Relationship between effortful motivation and neurocognition in schizophrenia. Schizophr Res, 193, 69–76. doi: 10.1016/j.schres.2017.06.042 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Bland AR, Roiser JP, Mehta MA, Schei T, Boland H, Campbell-Meiklejohn DK, … Elliott R. (2016). EMOTICOM: A Neuropsychological Test Battery to Evaluate Emotion, Motivation, Impulsivity, and Social Cognition. Frontiers in Behavioral Neuroscience, 10. doi: 10.3389/fnbeh.2016.00025 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Blundell J, & Latham C. (1978). Pharmacological manipulation of feeding behavior: Possible influences of serotonin and dopamine on food intake. Central mechanisms of anorectic drugs, 83–109. [Google Scholar]
- Bogdan R, Santesso DL, Fagerness J, Perlis RH, & Pizzagalli DA. (2011). Corticotropin-releasing hormone receptor type 1 (CRHR1) genetic variation and stress interact to influence reward learning. J Neurosci, 31(37), 13246–13254. doi: 10.1523/JNEUROSCI.2661-11.2011 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Bomyea JA, Parrish EM, Paolillo EW, Filip TF, Eyler LT, Depp CA, & Moore RC. (2021). Relationships between daily mood states and real-time cognitive performance in individuals with bipolar disorder and healthy comparators: a remote ambulatory assessment study. Journal of clinical and experimental neuropsychology, 43(8), 813–824. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Boomhower SR, & Rasmussen EB. (2014). Haloperidol and rimonabant increase delay discounting in rats fed high-fat and standard-chow diets. Behav Pharmacol, 25(8), 705–716. doi: 10.1097/fbp.0000000000000058 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Böttger SJ, Förstner BR, Szalek L, Koller-Schlaud K, Rapp MA, & Tschorn M. (2023). Mood and anxiety disorders within the Research Domain Criteria framework of Positive and Negative Valence Systems: a scoping review. Frontiers in Human Neuroscience, 17, 1184978. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Boulougouris V, & Robbins TW. (2010). Enhancement of spatial reversal learning by 5-HT2C receptor antagonism is neuroanatomically specific. J Neurosci, 30(3), 930–938. doi: 10.1523/JNEUROSCI.4312-09.2010 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Boutros N, Semenova S, Liu W, Crews FT, & Markou A. (2014). Adolescent intermittent ethanol exposure is associated with increased risky choice and decreased dopaminergic and cholinergic neuron markers in adult rats. Int J Neuropsychopharmacol, 18(2). doi: 10.1093/ijnp/pyu003 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Brady JV, & Griffiths RR. (1976). Behavioral procedures for evaluating the relative abuse potential of CNS drugs in primates. Fed Proc, 35(11), 2245–2253. [PubMed] [Google Scholar]
- Brancu M, Thompson NL, Beckham JC, Green KT, Calhoun PS, Elbogen EB, … Wagner HR. (2014). The impact of social support on psychological distress for US Afghanistan/Iraq era veterans with PTSD and other psychiatric diagnoses. Psychiatry research, 217(1–2), 86–92. [DOI] [PubMed] [Google Scholar]
- Brigman JL, Daut RA, Wright T, Gunduz-Cinar O, Graybeal C, Davis MI, … Holmes A. (2013). GluN2B in corticostriatal circuits governs choice learning and choice shifting. Nat Neurosci, 16(8), 1101–1110. doi: 10.1038/nn.3457 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Broyd SJ, Richards HJ, Helps SK, Chronaki G, Bamford S, & Sonuga-Barke EJ. (2012). An electrophysiological monetary incentive delay (e-MID) task: a way to decompose the different components of neural response to positive and negative monetary reinforcement. J Neurosci Methods, 209(1), 40–49. doi: 10.1016/j.jneumeth.2012.05.015 [DOI] [PubMed] [Google Scholar]
- Bryce CA, & Floresco SB. (2019). Alterations in effort-related decision-making induced by stimulation of dopamine D. Psychopharmacology (Berl), 236(9), 2699–2712. doi: 10.1007/s00213-019-05244-w [DOI] [PubMed] [Google Scholar]
- Burani K, Gallyer A, Ryan J, Jordan C, Joiner T, & Hajcak G. (2021). Acute stress reduces reward-related neural activity: Evidence from the reward positivity. Stress, 24(6), 833–839. doi: 10.1080/10253890.2021.1929164 [DOI] [PubMed] [Google Scholar]
- Burkhouse KL, Gorka SM, Klumpp H, Kennedy AE, Karich S, Francis J, … Phan KL. (2018). Neural Responsiveness to Reward as an Index of Depressive Symptom Change Following Cognitive-Behavioral Therapy and SSRI Treatment. J Clin Psychiatry, 79(4). doi: 10.4088/JCP.17m11836 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Bussey TJ, Duck J, Muir JL, & Aggleton JP. (2000). Distinct patterns of behavioural impairments resulting from fornix transection or neurotoxic lesions of the perirhinal and postrhinal cortices in the rat. Behav Brain Res, 111(1–2), 187–202. doi: 10.1016/s0166-4328(00)00155-8 [DOI] [PubMed] [Google Scholar]
- Bussey TJ, Everitt BJ, & Robbins TW. (1997). Dissociable effects of cingulate and medial frontal cortex lesions on stimulus-reward learning using a novel Pavlovian autoshaping procedure for the rat: implications for the neurobiology of emotion. Behav Neurosci, 111(5), 908–919. doi: 10.1037//0735-7044.111.5.908 [DOI] [PubMed] [Google Scholar]
- Bussey TJ, Muir JL, Everitt BJ, & Robbins TW. (1997). Triple dissociation of anterior cingulate, posterior cingulate, and medial frontal cortices on visual discrimination tasks using a touchscreen testing procedure for the rat. Behav Neurosci, 111(5), 920–936. doi: 10.1037//0735-7044.111.5.920 [DOI] [PubMed] [Google Scholar]
- Bussey TJ, Padain TL, Skillings EA, Winters BD, Morton AJ, & Saksida LM. (2008). The touchscreen cognitive testing method for rodents: How to get the best out of your rat. Learning & Memory, 15(7), 516–523. doi: 10.1101/lm.987808 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Butter CM. (1969). Perseveration in extinction and in discrimination reversal tasks following selective frontal ablations in Macaca mulatta. Physiology & Behavior, 4(2), 163–171. [Google Scholar]
- Cabanac M. (1971). Physiological Role of Pleasure: A stimulus can feel pleasant or unpleasant depending upon its usefulness as determined by internal signals. Science, 173(4002), 1103–1107. [DOI] [PubMed] [Google Scholar]
- Cardinal RN, & Howes NJ. (2005). Effects of lesions of the nucleus accumbens core on choice between small certain rewards and large uncertain rewards in rats. BMC Neurosci, 6, 37. doi: 10.1186/1471-2202-6-37 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Cardinal RN, Pennicott DR, Sugathapala CL, Robbins TW, & Everitt BJ. (2001). Impulsive choice induced in rats by lesions of the nucleus accumbens core. Science, 292(5526), 2499–2501. doi: 10.1126/science.1060818 [DOI] [PubMed] [Google Scholar]
- Cardinal RN, Robbins TW, & Everitt BJ. (2000). The effects of d-amphetamine, chlordiazepoxide, alpha-flupenthixol and behavioural manipulations on choice of signalled and unsignalled delayed reinforcement in rats. Psychopharmacology (Berl), 152(4), 362–375. doi: 10.1007/s002130000536 [DOI] [PubMed] [Google Scholar]
- Carlson JM, Foti D, Mujica-Parodi LR, Harmon-Jones E, & Hajcak G. (2011). Ventral striatal and medial prefrontal BOLD activation is correlated with reward-related electrocortical activity: a combined ERP and fMRI study. Neuroimage, 57(4), 1608–1616. doi: 10.1016/j.neuroimage.2011.05.037 [DOI] [PubMed] [Google Scholar]
- Carter CS, Barch DM, Bullmore E, Breiling J, Buchanan RW, Butler P, … Wykes T. (2011). Cognitive Neuroscience Treatment Research to Improve Cognition in Schizophrenia II: Developing Imaging Biomarkers to Enhance Treatment Development for Schizophrenia and Related Disorders. Biological Psychiatry, 70(1), 7–12. doi: 10.1016/j.biopsych.2011.01.041 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Cavanagh JF, Gregg D, Light GA, Olguin SL, Sharp RF, Bismark AW, … Young JW. (2021). Electrophysiological biomarkers of behavioral dimensions from cross-species paradigms. Transl Psychiatry, 11(1), 482. doi: 10.1038/s41398-021-01562-w [DOI] [PMC free article] [PubMed] [Google Scholar]
- Cavanagh JF, Olguin SL, Talledo JA, Kotz JE, Roberts BZ, Nungaray JA, … Brigman JL. (2022). Amphetamine alters an EEG marker of reward processing in humans and mice. Psychopharmacology, 239(3), 923–933. doi: 10.1007/s00213-022-06082-z [DOI] [PMC free article] [PubMed] [Google Scholar]
- Chen Y, Chaudhary S, & Li CR. (2022). Shared and distinct neural activity during anticipation and outcome of win and loss: A meta-analysis of the monetary incentive delay task. Neuroimage, 264, 119764. doi: 10.1016/j.neuroimage.2022.119764 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Cheung TH, & Cardinal RN. (2005). Hippocampal lesions facilitate instrumental learning with delayed reinforcement but induce impulsive choice in rats. BMC Neurosci, 6, 36. doi: 10.1186/1471-2202-6-36 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Chudasama Y, & Muir JL. (2001). Visual attention in the rat: a role for the prelimbic cortex and thalamic nuclei? Behav Neurosci, 115(2), 417–428. [PubMed] [Google Scholar]
- Chudasama Y, & Robbins TW. (2003). Dissociable contributions of the orbitofrontal and infralimbic cortex to pavlovian autoshaping and discrimination reversal learning: further evidence for the functional heterogeneity of the rodent frontal cortex. J Neurosci, 23(25), 8771–8780. doi: 10.1523/JNEUROSCI.23-25-08771.2003 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Churchwell JC, Morris AM, Heurtelou NM, & Kesner RP. (2009). Interactions between the prefrontal cortex and amygdala during delay discounting and reversal. Behav Neurosci, 123(6), 1185–1196. doi: 10.1037/a0017734 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Consortium B, Anttila V, Bulik-Sullivan B, Finucane HK, Walters RK, Bras J, … Gormley P. (2018). Analysis of shared heritability in common disorders of the brain. Science, 360(6395), eaap8757. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Consortium, C.-D. G. o. t. P. G. (2013). Identification of risk loci with shared effects on five major psychiatric disorders: a genome-wide analysis. The Lancet, 381(9875), 1371–1379. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Consortium WTCC, McCarthy SE, Makarov V, Kirov G, Addington AM, McClellan J, … Sebat J. (2009). Microduplications of 16p11.2 are associated with schizophrenia. Nature Genetics, 41(11), 1223–1227. doi: 10.1038/ng.474 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Cools R, Clark L, Owen AM, & Robbins TW. (2002). Defining the neural mechanisms of probabilistic reversal learning using event-related functional magnetic resonance imaging. J Neurosci, 22(11), 4563–4567. doi: 10.1523/JNEUROSCI.22-11-04563.2002 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Craddock N, & Owen MJ. (2010). The Kraepelinian dichotomy – going, going … but still not gone. British Journal of Psychiatry, 196(2), 92–95. doi: 10.1192/bjp.bp.109.073429 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Crane NA, Gorka SM, Weafer J, Langenecker SA, de Wit H, & Phan KL. (2018). Neural activation to monetary reward is associated with amphetamine reward sensitivity. Neuropsychopharmacology, 43(8), 1738–1744. doi: 10.1038/s41386-018-0042-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Critchley HD, & Rolls ET. (1996). Hunger and satiety modify the responses of olfactory and visual neurons in the primate orbitofrontal cortex. Journal of neurophysiology, 75(4), 1673–1686. [DOI] [PubMed] [Google Scholar]
- Crouse RB, Kim K, Batchelor HM, Girardi EM, Kamaletdinova R, Chan J, … Talmage DA. (2020). Acetylcholine is released in the basolateral amygdala in response to predictors of reward and enhances the learning of cue-reward contingency. Elife, 9, e57335. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Cui G, Jun SB, Jin X, Pham MD, Vogel SS, Lovinger DM, & Costa RM. (2013). Concurrent activation of striatal direct and indirect pathways during action initiation. Nature, 494(7436), 238–242. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Culbreth AJ, Westbrook A, Daw ND, Botvinick M, & Barch DM. (2016). Reduced model-based decision-making in schizophrenia. J Abnorm Psychol, 125(6), 777–787. doi: 10.1037/abn0000164 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Cuthbert BN. (2022). Research Domain Criteria (RDoC): Progress and Potential. Current Directions in Psychological Science, 31(2), 107–114. doi: 10.1177/09637214211051363 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Cuthbert BN, & Kozak MJ. (2013). Constructing constructs for psychopathology: The NIMH research domain criteria. Journal of Abnormal Psychology, 122(3), 928–937. doi: 10.1037/a0034028 [DOI] [PubMed] [Google Scholar]
- Dai Z, Harrow SE, Song X, Rucklidge JJ, & Grace RC. (2016). Gambling, Delay, and Probability Discounting in Adults With and Without ADHD. J Atten Disord, 20(11), 968–978. doi: 10.1177/1087054713496461 [DOI] [PubMed] [Google Scholar]
- Damiano CR, Aloi J, Treadway M, Bodfish JW, & Dichter GS. (2012). Adults with autism spectrum disorders exhibit decreased sensitivity to reward parameters when making effort-based decisions. J Neurodev Disord, 4(1), 13. doi: 10.1186/1866-1955-4-13 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Danion J-M, Meulemans T, Kauffmann-Muller F, & Vermaat H. (2001). Intact implicit learning in schizophrenia. American Journal of Psychiatry, 158(6), 944–948. [DOI] [PubMed] [Google Scholar]
- Daw ND, O'Doherty JP, Dayan P, Seymour B, & Dolan RJ. (2006). Cortical substrates for exploratory decisions in humans. Nature, 441(7095), 876–879. doi: 10.1038/nature04766 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Day JJ, Roitman MF, Wightman RM, & Carelli RM. (2007). Associative learning mediates dynamic shifts in dopamine signaling in the nucleus accumbens. Nat Neurosci, 10(8), 1020–1028. doi: 10.1038/nn1923 [DOI] [PubMed] [Google Scholar]
- de Wit H, Enggasser JL, & Richards JB. (2002). Acute administration of d-amphetamine decreases impulsivity in healthy volunteers. Neuropsychopharmacology, 27(5), 813–825. doi: 10.1016/s0893-133x(02)00343-3 [DOI] [PubMed] [Google Scholar]
- de Wit S, Corlett PR, Aitken MR, Dickinson A, & Fletcher PC. (2009). Differential engagement of the ventromedial prefrontal cortex by goal-directed and habitual behavior toward food pictures in humans. Journal of Neuroscience, 29(36), 11330–11338. [DOI] [PMC free article] [PubMed] [Google Scholar]
- de Wit S, & Dickinson A. (2009). Associative theories of goal-directed behaviour: a case for animal-human translational models. Psychol Res, 73(4), 463–476. doi: 10.1007/s00426-009-0230-6 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Delgado MR, Nystrom LE, Fissell C, Noll DC, & Fiez JA. (2000). Tracking the hemodynamic responses to reward and punishment in the striatum. J Neurophysiol, 84(6), 3072–3077. doi: 10.1152/jn.2000.84.6.3072 [DOI] [PubMed] [Google Scholar]
- Delotterie DF, Mathis C, Cassel JC, Rosenbrock H, Dorner-Ciossek C, & Marti A. (2015). Touchscreen tasks in mice to demonstrate differences between hippocampal and striatal functions. Neurobiol Learn Mem, 120, 16–27. doi: 10.1016/j.nlm.2015.02.007 [DOI] [PubMed] [Google Scholar]
- Der-Avakian A, Barnes SA, Markou A, & Pizzagalli DA. (2016). Translational Assessment of Reward and Motivational Deficits in Psychiatric Disorders. Curr Top Behav Neurosci, 28, 231–262. doi: 10.1007/7854_2015_5004 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Desrochers SS, & Nautiyal KM. (2022). Serotonin 1B receptor effects on response inhibition are independent of inhibitory learning. Neurobiology of Learning and Memory, 187, 107574. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Dhingra I, Zhang S, Zhornitsky S, Le TM, Wang W, Chao HH, … Li CR. (2020). The effects of age on reward magnitude processing in the monetary incentive delay task. Neuroimage, 207, 116368. doi: 10.1016/j.neuroimage.2019.116368 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Diagnostic and statistical manual of mental disorders: DSM-5™, 5th ed. (2013). Arlington, VA, US: American Psychiatric Publishing, Inc. [Google Scholar]
- Dickinson A. (1985). Actions and Habits - the Development of Behavioral Autonomy. Philosophical Transactions of the Royal Society of London Series B-Biological Sciences, 308(1135), 67–78. doi:DOI 10.1098/rstb.1985.0010 [DOI] [Google Scholar]
- Dickinson A, Wood N, & Smith JW. (2002). Alcohol seeking by rats: action or habit? Q J Exp Psychol B, 55(4), 331–348. doi: 10.1080/0272499024400016 [DOI] [PubMed] [Google Scholar]
- Dieterich A, Liu T, & Samuels BA. (2021). Chronic non-discriminatory social defeat stress reduces effort-related motivated behaviors in male and female mice. Transl Psychiatry, 11(1), 125. doi: 10.1038/s41398-021-01250-9 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Dieterich A, Srivastava P, Sharif A, Stech K, Floeder J, Yohn SE, & Samuels BA. (2019). Chronic corticosterone administration induces negative valence and impairs positive valence behaviors in mice. Transl Psychiatry, 9(1), 337. doi: 10.1038/s41398-019-0674-4 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Dimitrov D, He Y, Mutoh H, Baker BJ, Cohen L, Akemann W, & Knöpfel T. (2007). Engineering and characterization of an enhanced fluorescent protein voltage sensor. PLoS One, 2(5), e440. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Disanto G, Morahan JM, Lacey MV, DeLuca GC, Giovannoni G, Ebers GC, & Ramagopalan SV. (2012). Seasonal distribution of psychiatric births in England. PLoS One, 7(4), e34866. doi: 10.1371/journal.pone.0034866 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Dodds CM, Muller U, Clark L, van Loon A, Cools R, & Robbins TW. (2008). Methylphenidate has differential effects on blood oxygenation level-dependent signal related to cognitive subprocesses of reversal learning. J Neurosci, 28(23), 5976–5982. doi: 10.1523/JNEUROSCI.1153-08.2008 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Dugre JR, Dumais A, Bitar N, & Potvin S. (2018). Loss anticipation and outcome during the Monetary Incentive Delay Task: a neuroimaging systematic review and meta-analysis. PeerJ, 6, e4749. doi: 10.7717/peerj.4749 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Dumont JR, Salewski R, & Beraldo F. (2021). Critical mass: The rise of a touchscreen technology community for rodent cognitive testing. Genes, Brain and Behavior, 20(1), e12650. doi: 10.1111/gbb.12650 [DOI] [PubMed] [Google Scholar]
- Dunayevich E, & Keck PE Jr (2000). Prevalence and description of psychotic features in bipolar mania. Current psychiatry reports, 2(4), 286–290. [DOI] [PubMed] [Google Scholar]
- Dutcher EG, Lopez-Cruz L, Pama EAC, Lynall ME, Bevers ICR, Jones JA, … Dalley JW. (2023). Early-life stress biases responding to negative feedback and increases amygdala volume and vulnerability to later-life stress. Transl Psychiatry, 13(1), 81. doi: 10.1038/s41398-023-02385-7 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Ermakova AO, Knolle F, Justicia A, Bullmore ET, Jones PB, Robbins TW, … Murray GK. (2018). Abnormal reward prediction-error signalling in antipsychotic naive individuals with first-episode psychosis or clinical risk for psychosis. Neuropsychopharmacology, 43(8), 1691–1699. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Esfand SM, Null KE, Duda JM, de Leeuw J, & Pizzagalli DA. (2024). Lifetime history of major depressive disorder is associated with decreased reward learning: Evidence from a novel online version of the probabilistic reward task. J Affect Disord, 350, 1007–1015. doi: 10.1016/j.jad.2024.01.133 [DOI] [PubMed] [Google Scholar]
- Favier M, Janickova H, Justo D, Kljakic O, Runtz L, Natsheh JY, … El Mestikawy S. (2020). Cholinergic dysfunction in the dorsal striatum promotes habit formation and maladaptive eating. J Clin Invest, 130(12), 6616–6630. doi: 10.1172/JCI138532 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Fellows LK, & Farah MJ. (2003). Ventromedial frontal cortex mediates affective shifting in humans: evidence from a reversal learning paradigm. Brain, 126(8), 1830–1837. [DOI] [PubMed] [Google Scholar]
- Fiksinski AM, Breetvelt EJ, Duijff SN, Bassett AS, Kahn RS, & Vorstman JAS. (2017). Autism Spectrum and psychosis risk in the 22q11.2 deletion syndrome. Findings from a prospective longitudinal study. Schizophrenia Research, 188, 59–62. doi: 10.1016/j.schres.2017.01.032 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Fletcher K, Parker G, Paterson A, Fava M, Iosifescu D, & Pizzagalli DA. (2015). Anhedonia in melancholic and non-melancholic depressive disorders. J Affect Disord, 184, 81–88. doi: 10.1016/j.jad.2015.05.028 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Fletcher PJ, Rahbarnia A, Li Z, Ji X, Higgins GA, Funk D, & Lê A. (2023). Effects of 5-HT2C receptor stimulation in male mice on behaviour and Fos expression: Feeding, reward and impulsivity. Behavioural Brain Research, 447, 114438. [DOI] [PubMed] [Google Scholar]
- Floresco SB, Tse MT, & Ghods-Sharifi S. (2008). Dopaminergic and glutamatergic regulation of effort- and delay-based decision making. Neuropsychopharmacology, 33(8), 1966–1979. doi: 10.1038/sj.npp.1301565 [DOI] [PubMed] [Google Scholar]
- Floresco SB, & Whelan JM. (2009). Perturbations in different forms of cost/benefit decision making induced by repeated amphetamine exposure. Psychopharmacology (Berl), 205(2), 189–201. doi: 10.1007/s00213-009-1529-0 [DOI] [PubMed] [Google Scholar]
- Flusberg BA, Nimmerjahn A, Cocker ED, Mukamel EA, Barretto RP, Ko TH, … Schnitzer MJ. (2008). High-speed, miniaturized fluorescence microscopy in freely moving mice. Nature Methods, 5(11), 935–938. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Foltin RW, Haney M, Comer SD, & Fischman MW. (1996). Effect of fenfluramine on food intake, mood, and performance of humans living in a residential laboratory. Physiology & Behavior, 59(2), 295–305. [DOI] [PubMed] [Google Scholar]
- Förster D, Dal Maschio M, Laurell E, & Baier H. (2017). An optogenetic toolbox for unbiased discovery of functionally connected cells in neural circuits. Nature communications, 8(1), 116. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Foti D, Weinberg A, Dien J, & Hajcak G. (2011). Event-related potential activity in the basal ganglia differentiates rewards from nonrewards: temporospatial principal components analysis and source localization of the feedback negativity. Hum Brain Mapp, 32(12), 2207–2216. doi: 10.1002/hbm.21182 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Fox AE, Nicholson AM, Singha D, Thieret BAS, Ortiz M, & Visser EJ. (2023). Timing and delay discounting in attention-deficit/hyperactivity disorder: A translational approach. Dev Psychobiol, 65(5), e22399. doi: 10.1002/dev.22399 [DOI] [PubMed] [Google Scholar]
- Francois J, Conway MW, Lowry JP, Tricklebank MD, & Gilmour G. (2012). Changes in reward-related signals in the rat nucleus accumbens measured by in vivo oxygen amperometry are consistent with fMRI BOLD responses in man. Neuroimage, 60(4), 2169–2181. doi: 10.1016/j.neuroimage.2012.02.024 [DOI] [PubMed] [Google Scholar]
- Francois J, Grimm O, Schwarz AJ, Schweiger J, Haller L, Risterucci C, … Meyer-Lindenberg A. (2016). Ketamine Suppresses the Ventral Striatal Response to Reward Anticipation: A Cross-Species Translational Neuroimaging Study. Neuropsychopharmacology, 41(5), 1386–1394. doi: 10.1038/npp.2015.291 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Freund N, MacGillivilray HT, Thompson BS, Lukkes JL, Stanis JJ, Brenhouse HC, & Andersen SL. (2014). Sex-dependent changes in ADHD-like behaviors in juvenile rats following cortical dopamine depletion. Behav Brain Res, 270, 357–363. doi: 10.1016/j.bbr.2014.05.024 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Fuster JM. (2001). The prefrontal cortex—an update: time is of the essence. Neuron, 30(2), 319–333. [DOI] [PubMed] [Google Scholar]
- Garami J, & Moustafa AA. (2019). Probability discounting of monetary gains and losses in opioid-dependent adults. Behav Brain Res, 364, 334–339. doi: 10.1016/j.bbr.2019.02.017 [DOI] [PubMed] [Google Scholar]
- Gard DE, Sanchez AH, Cooper K, Fisher M, Garrett C, & Vinogradov S. (2014). Do people with schizophrenia have difficulty anticipating pleasure, engaging in effortful behavior, or both? Journal of Abnormal Psychology, 123(4), 771–782. doi: 10.1037/abn0000005 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Geyer MA, & Markou A. (1995). Animal models of psychiatric disorders. Psychopharmacology: the fourth generation of progress, 787, 798. [Google Scholar]
- Ghods-Sharifi S, & Floresco SB. (2010). Differential effects on effort discounting induced by inactivations of the nucleus accumbens core or shell. Behav Neurosci, 124(2), 179–191. doi: 10.1037/a0018932 [DOI] [PubMed] [Google Scholar]
- Gillan CM, Papmeyer M, Morein-Zamir S, Sahakian BJ, Fineberg NA, Robbins TW, & de Wit S. (2011). Disruption in the balance between goal-directed behavior and habit learning in obsessive-compulsive disorder. American Journal of Psychiatry, 168(7), 718–726. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Gläscher J, Hampton AN, & O'Doherty JP. (2009). Determining a role for ventromedial prefrontal cortex in encoding action-based value signals during reward-related decision making. Cerebral Cortex, 19(2), 483–495. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Gottfried JA, O'Doherty J, & Dolan RJ. (2003). Encoding predictive reward value in human amygdala and orbitofrontal cortex. Science, 301(5636), 1104–1107. [DOI] [PubMed] [Google Scholar]
- Gray LN, Stafford MC, & Tallman I. (1991). Rewards and punishments in complex human choices. Social psychology quarterly, 318–329. [Google Scholar]
- Graybeal C, Feyder M, Schulman E, Saksida LM, Bussey TJ, Brigman JL, & Holmes A. (2011). Paradoxical reversal learning enhancement by stress or prefrontal cortical damage: rescue with BDNF. Nat Neurosci, 14(12), 1507–1509. doi: 10.1038/nn.2954 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Grayson B, Barnes S, Markou A, Piercy C, Podda G, & Neill J. (2016). Postnatal phencyclidine (PCP) as a neurodevelopmental animal model of schizophrenia pathophysiology and symptomatology: a review. Neurotoxin Modeling of Brain Disorders—Life-long Outcomes in Behavioral Teratology, 403–428. [DOI] [PubMed] [Google Scholar]
- Grottick AJ, MacQueen DL, Barnes SA, Carroll C, Sanabria EK, Bobba V, & Young JW. (2021). Convergent observations of MK-801-induced impairment in rat 5C-CPT performance across laboratories: reversal with a D. Psychopharmacology (Berl), 238(4), 979–990. doi: 10.1007/s00213-020-05744-0 [DOI] [PubMed] [Google Scholar]
- Grozeva D. (2010). Rare Copy Number VariantsA Point of Rarity in Genetic Risk for Bipolar Disorder and SchizophreniaRare Copy Number Variants. Archives of General Psychiatry, 67(4), 318. doi: 10.1001/archgenpsychiatry.2010.25 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Gruninger TR, LeBoeuf B, Liu Y, & Rene Garcia L. (2007). Molecular signaling involved in regulating feeding and other mitivated behaviors. Molecular neurobiology, 35, 1–19. [DOI] [PubMed] [Google Scholar]
- Haase L, Cerf-Ducastel B, & Murphy C. (2009). Cortical activation in response to pure taste stimuli during the physiological states of hunger and satiety. Neuroimage, 44(3), 1008–1021. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Hailwood JM, Heath CJ, Robbins TW, Saksida LM, & Bussey TJ. (2018). Validation and optimisation of a touchscreen progressive ratio test of motivation in male rats. Psychopharmacology, 235(9), 2739–2753. doi: 10.1007/s00213-018-4969-6 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Halford JC, & Harrold JA. (2012). 5-HT 2C receptor agonists and the control of appetite. Appetite Control, 349–356. [DOI] [PubMed] [Google Scholar]
- Hare TA, O'doherty J, Camerer CF, Schultz W, & Rangel A. (2008). Dissociating the role of the orbitofrontal cortex and the striatum in the computation of goal values and prediction errors. Journal of Neuroscience, 28(22), 5623–5630. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Harle KM, Guo D, Zhang S, Paulus MP, & Yu AJ. (2017). Anhedonia and anxiety underlying depressive symptomatology have distinct effects on reward-based decision-making. PloS one, 12(10), e0186473. doi: 10.1371/journal.pone.0186473 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Harle KM, Zhang S, Schiff M, Mackey S, Paulus MP, & Yu AJ. (2015). Altered Statistical Learning and Decision-Making in Methamphetamine Dependence: Evidence from a Two-Armed Bandit Task. Front Psychol, 6, 1910. doi: 10.3389/fpsyg.2015.01910 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Hart EE, Blair GJ, O'Dell TJ, Blair HT, & Izquierdo A. (2020). Chemogenetic modulation and single-photon calcium imaging in anterior cingulate cortex reveal a mechanism for effort-based decisions. Journal of Neuroscience, 40(29), 5628–5643. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Hart KL, Brown HE, Roffman JL, & Perlis RH. (2019). Risk tolerance measured by probability discounting among individuals with primary mood and psychotic disorders. Neuropsychology, 33(3), 417–424. doi: 10.1037/neu0000506 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Hay JF, Moscovitch M, & Levine B. (2002). Dissociating habit and recollection: evidence from Parkinson's disease, amnesia and focal lesion patients. Neuropsychologia, 40(8), 1324–1334. doi: 10.1016/s0028-3932(01)00214-7 [DOI] [PubMed] [Google Scholar]
- He Q, Chen M, Chen C, Xue G, Feng T, & Bechara A. (2016). Anodal Stimulation of the Left DLPFC Increases IGT Scores and Decreases Delay Discounting Rate in Healthy Males. Front Psychol, 7, 1421. doi: 10.3389/fpsyg.2016.01421 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Heath CJ, Bussey TJ, & Saksida LM. (2015). Motivational assessment of mice using the touchscreen operant testing system: effects of dopaminergic drugs. Psychopharmacology, 232(21–22), 4043–4057. doi: 10.1007/s00213-015-4009-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Heath CJ, O'Callaghan C, Mason SL, Phillips BU, Saksida LM, Robbins TW, … Sahakian BJ. (2019). A Touchscreen Motivation Assessment Evaluated in Huntington's Disease Patients and R6/1 Model Mice. Front Neurol, 10, 858. doi: 10.3389/fneur.2019.00858 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Heath CJ, Phillips BU, Bussey TJ, & Saksida LM. (2016). Measuring Motivation and Reward‐Related Decision Making in the Rodent Operant Touchscreen System. Current Protocols in Neuroscience, 74(1). doi: 10.1002/0471142301.ns0834s74 [DOI] [PubMed] [Google Scholar]
- Hecht D, Walsh V, & Lavidor M. (2013). Bi-frontal direct current stimulation affects delay discounting choices. Cogn Neurosci, 4(1), 7–11. doi: 10.1080/17588928.2011.638139 [DOI] [PubMed] [Google Scholar]
- Heerey EA, Bell-Warren KR, & Gold JM. (2008). Decision-making impairments in the context of intact reward sensitivity in schizophrenia. Biol Psychiatry, 64(1), 62–69. doi: 10.1016/j.biopsych.2008.02.015 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Heerey EA, Robinson BM, McMahon RP, & Gold JM. (2007). Delay discounting in schizophrenia. Cogn Neuropsychiatry, 12(3), 213–221. doi: 10.1080/13546800601005900 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Hein TP, de Fockert J, & Ruiz MH. (2021). State anxiety biases estimates of uncertainty and impairs reward learning in volatile environments. Neuroimage, 224, 117424. doi: 10.1016/j.neuroimage.2020.117424 [DOI] [PubMed] [Google Scholar]
- Helms CM, Reeves JM, & Mitchell SH. (2006). Impact of strain and D-amphetamine on impulsivity (delay discounting) in inbred mice. Psychopharmacology (Berl), 188(2), 144–151. doi: 10.1007/s00213-006-0478-0 [DOI] [PubMed] [Google Scholar]
- Hershenberg R, Satterthwaite TD, Daldal A, Katchmar N, Moore TM, Kable JW, & Wolf DH. (2016). Diminished effort on a progressive ratio task in both unipolar and bipolar depression. J Affect Disord, 196, 97–100. doi: 10.1016/j.jad.2016.02.003 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Hervig ME, Fiddian L, Piilgaard L, Bozic T, Blanco-Pozo M, Knudsen C, … Robbins TW. (2020). Dissociable and Paradoxical Roles of Rat Medial and Lateral Orbitofrontal Cortex in Visual Serial Reversal Learning. Cereb Cortex, 30(3), 1016–1029. doi: 10.1093/cercor/bhz144 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Hill A, & Blundell J. (1990). Sensitivity of the appetite control system in obese subjects to nutritional and serotoninergic challenges. International Journal of Obesity, 14(3), 219–233. [PubMed] [Google Scholar]
- Hirschfeld R, Montgomery SA, Keller MB, Kasper S, Schatzberg AF, Hans-Jurgen M, … Versiani M. (2000). Social functioning in depression: a review. Journal of Clinical Psychiatry, 61(4), 268–275. [DOI] [PubMed] [Google Scholar]
- Hisey EE, Fritsch EL, Newman EL, Ressler KJ, Kangas BD, & Carlezon WA Jr. (2023). Early life stress in male mice blunts responsiveness in a translationally-relevant reward task. Neuropsychopharmacology, 48(12), 1752–1759. doi: 10.1038/s41386-023-01610-7 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Hodos W. (1961). Progressive ratio as a measure of reward strength. Science, 134(3483), 943–944. doi: 10.1126/science.134.3483.943 [DOI] [PubMed] [Google Scholar]
- Höflich A, Michenthaler P, Kasper S, & Lanzenberger R. (2019). Circuit mechanisms of reward, anhedonia, and depression. International Journal of Neuropsychopharmacology, 22(2), 105–118. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Holland JG, & Skinner BF. (1961). The Analysis of Behavior: A Program for Self‐Instruction. American Anthropologist, 65(1), 179–183. doi: 10.1525/aa.1963.65.1.02a00410 [DOI] [Google Scholar]
- Hornak J, O'doherty J, Bramham J, Rolls ET, Morris RG, Bullock P, & Polkey C. (2004). Reward-related reversal learning after surgical excisions in orbito-frontal or dorsolateral prefrontal cortex in humans. Journal of cognitive neuroscience, 16(3), 463–478. [DOI] [PubMed] [Google Scholar]
- Horner AE, Heath CJ, Hvoslef-Eide M, Kent BA, Kim CH, Nilsson SRO, … Bussey TJ. (2013). The touchscreen operant platform for testing learning and memory in rats and mice. Nature Protocols, 8(10), 1961–1984. doi: 10.1038/nprot.2013.122 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Houlton J, Barwick D, & Clarkson AN. (2021). Frontal cortex stroke-induced impairment in spatial working memory on the trial-unique nonmatching-to-location task in mice. Neurobiology of Learning and Memory, 177, 107355. doi: 10.1016/j.nlm.2020.107355 [DOI] [PubMed] [Google Scholar]
- Huang J, Yang XH, Lan Y, Zhu CY, Liu XQ, Wang YF, … Chan RC. (2016). Neural substrates of the impaired effort expenditure decision making in schizophrenia. Neuropsychology, 30(6), 685–696. doi: 10.1037/neu0000284 [DOI] [PubMed] [Google Scholar]
- Huggins AA, Weinberg A, Gorka SM, & Shankman SA. (2019). Blunted neural response to gains versus losses associated with both risk-prone and risk-averse behavior in a clinically diverse sample. Psychophysiology, 56(6), e13342. doi: 10.1111/psyp.13342 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Hughes DM, Yates MJ, Morton EE, & Smillie LD. (2015). Asymmetric frontal cortical activity predicts effort expenditure for reward. Soc Cogn Affect Neurosci, 10(7), 1015–1019. doi: 10.1093/scan/nsu149 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Huskinson SL, & Anderson KG. (2012). Effects of acute and chronic administration of diazepam on delay discounting in Lewis and Fischer 344 rats. Behav Pharmacol, 23(4), 315–330. doi: 10.1097/FBP.0b013e3283564da4 [DOI] [PubMed] [Google Scholar]
- Hvoslef-Eide M, Mar AC, Nilsson SRO, Alsiö J, Heath CJ, Saksida LM, … Bussey TJ. (2015). The NEWMEDS rodent touchscreen test battery for cognition relevant to schizophrenia. Psychopharmacology, 232(21–22), 3853–3872. doi: 10.1007/s00213-015-4007-x [DOI] [PubMed] [Google Scholar]
- Insel T, & Cuthbert B. (2015). Brain disorders? Precisely: Precision medicine comes to psychiatry. Science, 348(6234), 499–500. doi: 10.1126/science.aab2358 [DOI] [PubMed] [Google Scholar]
- Insel T, Cuthbert B, Garvey M, Heinssen R, Pine DS, Quinn K, … Wang P. (2010). Research Domain Criteria (RDoC): Toward a New Classification Framework for Research on Mental Disorders. American Journal of Psychiatry, 167(7), 748–751. doi: 10.1176/appi.ajp.2010.09091379 [DOI] [PubMed] [Google Scholar]
- Islas-Preciado D, Wainwright SR, Sniegocki J, Lieblich SE, Yagi S, Floresco SB, & Galea LAM. (2020). Risk-based decision making in rats: Modulation by sex and amphetamine. Horm Behav, 125, 104815. doi: 10.1016/j.yhbeh.2020.104815 [DOI] [PubMed] [Google Scholar]
- Iturra-Mena AM, Kangas BD, Luc OT, Potter D, & Pizzagalli DA. (2023). Electrophysiological signatures of reward learning in the rodent touchscreen-based Probabilistic Reward Task. Neuropsychopharmacology, 48(4), 700–709. doi: 10.1038/s41386-023-01532-4 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Izquierdo A, Brigman JL, Radke AK, Rudebeck PH, & Holmes A. (2017). The neural basis of reversal learning: An updated perspective. Neuroscience, 345, 12–26. doi: 10.1016/j.neuroscience.2016.03.021 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Jackson JN, & MacKillop J. (2016). Attention-Deficit/Hyperactivity Disorder and Monetary Delay Discounting: A Meta-Analysis of Case-Control Studies. Biol Psychiatry Cogn Neurosci Neuroimaging, 1(4), 316–325. doi: 10.1016/j.bpsc.2016.01.007 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Jing M, Zhang P, Wang G, Feng J, Mesik L, Zeng J, … Guagliardo NA. (2018). A genetically encoded fluorescent acetylcholine indicator for in vitro and in vivo studies. Nature biotechnology, 36(8), 726–737. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Jocham G, Klein TA, Neumann J, von Cramon DY, Reuter M, & Ullsperger M. (2009). Dopamine DRD2 polymorphism alters reversal learning and associated neural activity. J Neurosci, 29(12), 3695–3704. doi: 10.1523/JNEUROSCI.5195-08.2009 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Johnson MW. (2012). An efficient operant choice procedure for assessing delay discounting in humans: initial validation in cocaine-dependent and control individuals. Experimental and Clinical Psychopharmacology, 20(3), 191. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Johnson MW, & Bickel WK. (2002). Within-subject comparison of real and hypothetical money rewards in delay discounting. J Exp Anal Behav, 77(2), 129–146. doi: 10.1901/jeab.2002.77-129 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Johnson SL, Mehta H, Ketter TA, Gotlib IH, & Knutson B. (2019). Neural responses to monetary incentives in bipolar disorder. NeuroImage: Clinical, 24, 102018. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Jones B, & Mishkin M. (1972). Limbic lesions and the problem of stimulus—reinforcement associations. Experimental neurology, 36(2), 362–377. [DOI] [PubMed] [Google Scholar]
- Joutsa J, Voon V, Johansson J, Niemelä S, Bergman J, & Kaasinen V. (2015). Dopaminergic function and intertemporal choice. Transl Psychiatry, 5(1), e491. doi: 10.1038/tp.2014.133 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Juckel G, Schlagenhauf F, Koslowski M, Wustenberg T, Villringer A, Knutson B, … Heinz A. (2006). Dysfunction of ventral striatal reward prediction in schizophrenia. Neuroimage, 29(2), 409–416. doi: 10.1016/j.neuroimage.2005.07.051 [DOI] [PubMed] [Google Scholar]
- Kaack I, Chae J, Shadli SM, & Hillman K. (2020). Exploring approach motivation: Correlating self-report, frontal asymmetry, and performance in the Effort Expenditure for Rewards Task. Cogn Affect Behav Neurosci, 20(6), 1234–1247. doi: 10.3758/s13415-020-00829-x [DOI] [PubMed] [Google Scholar]
- Kaiser RH, Treadway MT, Wooten DW, Kumar P, Goer F, Murray L, … Pizzagalli DA. (2018). Frontostriatal and Dopamine Markers of Individual Differences in Reinforcement Learning: A Multi-modal Investigation. Cereb Cortex, 28(12), 4281–4290. doi: 10.1093/cercor/bhx281 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kallen AM, Perkins ER, Klawohn J, & Hajcak G. (2020). Cross-sectional and prospective associations of P300, RewP, and ADHD symptoms in female adolescents. Int J Psychophysiol, 158, 215–224. doi: 10.1016/j.ijpsycho.2020.08.017 [DOI] [PubMed] [Google Scholar]
- Kanen JW, Arntz FE, Yellowlees R, Cardinal RN, Price A, Christmas DM, … Robbins TW. (2020). Probabilistic reversal learning under acute tryptophan depletion in healthy humans: a conventional analysis. J Psychopharmacol, 34(5), 580–583. doi: 10.1177/0269881120907991 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kangas BD, Wooldridge LM, Luc OT, Bergman J, & Pizzagalli DA. (2020). Empirical validation of a touchscreen probabilistic reward task in rats. Transl Psychiatry, 10(1), 285. doi: 10.1038/s41398-020-00969-1 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Katthagen T, Kaminski J, Heinz A, Buchert R, & Schlagenhauf F. (2020). Striatal dopamine and reward prediction error signaling in unmedicated schizophrenia patients. Schizophrenia Bulletin, 46(6), 1535–1546. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Keller FS, & Schoenfeld WN. (1950). Principles of Psychology (Vol. 46): Appleton-Century-Crofts, Inc. [Google Scholar]
- Kim EW, Phillips BU, Heath CJ, Cho SY, Kim H, Sreedharan J, … Kim CH. (2017). Optimizing reproducibility of operant testing through reinforcer standardization: identification of key nutritional constituents determining reward strength in touchscreens. Molecular brain, 10, 1–10. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kirschner H, Nassar MR, Fischer AG, Frodl T, Meyer-Lotz G, Frobose S, … Ullsperger M. (2024). Transdiagnostic inflexible learning dynamics explain deficits in depression and schizophrenia. Brain, 147(1), 201–214. doi: 10.1093/brain/awad362 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Klawohn J, Brush CJ, & Hajcak G. (2021). Neural responses to reward and pleasant pictures prospectively predict remission from depression. J Abnorm Psychol, 130(7), 702–712. doi: 10.1037/abn0000696 [DOI] [PubMed] [Google Scholar]
- Klein DN, Kocsis JH, McCullough JP, Holzer CE 3rd, Hirschfeld RM, & Keller MB. (1996). Symptomatology in dysthymic and major depressive disorder. Psychiatr Clin North Am, 19(1), 41–53. doi: 10.1016/s0193-953x(05)70272-0 [DOI] [PubMed] [Google Scholar]
- Knabbe J, Protzmann J, Schneider N, Berger M, Dannehl D, Wei S, … Cambridge SB. (2022). Single-dose ethanol intoxication causes acute and lasting neuronal changes in the brain. Proc Natl Acad Sci U S A, 119(25), e2122477119. doi: 10.1073/pnas.2122477119 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Knutson B, Bjork JM, Fong GW, Hommer D, Mattay VS, & Weinberger DR. (2004). Amphetamine modulates human incentive processing. Neuron, 43(2), 261–269. doi: 10.1016/j.neuron.2004.06.030 [DOI] [PubMed] [Google Scholar]
- Knutson B, Fong GW, Adams CM, Varner JL, & Hommer D. (2001). Dissociation of reward anticipation and outcome with event-related fMRI. Neuroreport, 12(17), 3683–3687. doi: 10.1097/00001756-200112040-00016 [DOI] [PubMed] [Google Scholar]
- Knutson B, Fong GW, Bennett SM, Adams CM, & Hommer D. (2003). A region of mesial prefrontal cortex tracks monetarily rewarding outcomes: characterization with rapid event-related fMRI. Neuroimage, 18(2), 263–272. doi: 10.1016/s1053-8119(02)00057-5 [DOI] [PubMed] [Google Scholar]
- Knutson B, Westdorp A, Kaiser E, & Hommer D. (2000). FMRI visualization of brain activity during a monetary incentive delay task. Neuroimage, 12(1), 20–27. doi: 10.1006/nimg.2000.0593 [DOI] [PubMed] [Google Scholar]
- Knutson B, Wimmer GE, Kuhnen CM, & Winkielman P. (2008). Nucleus accumbens activation mediates the influence of reward cues on financial risk taking. Neuroreport, 19(5), 509–513. doi: 10.1097/WNR.0b013e3282f85c01 [DOI] [PubMed] [Google Scholar]
- Kotov R, Carpenter WT, Cicero DC, Correll CU, Martin EA, Young JW, … Jonas KG. (2024). Psychosis superspectrum II: neurobiology, treatment, and implications. Molecular Psychiatry, 1–17. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kozak MJ, & Cuthbert BN. (2016). The NIMH Research Domain Criteria Initiative: Background, Issues, and Pragmatics. Psychophysiology, 53(3), 286–297. doi: 10.1111/psyp.12518 [DOI] [PubMed] [Google Scholar]
- Kringelbach ML. (2005). The human orbitofrontal cortex: linking reward to hedonic experience. Nature Reviews Neuroscience, 6(9), 691–702. [DOI] [PubMed] [Google Scholar]
- Kringelbach ML, Stein A, & Van Hartevelt TJ. (2012). The functional human neuroanatomy of food pleasure cycles. Physiology & Behavior, 106(3), 307–316. doi: 10.1016/j.physbeh.2012.03.023 [DOI] [PubMed] [Google Scholar]
- Kushima I, Aleksic B, Nakatochi M, Shimamura T, Okada T, Uno Y, … Ozaki N. (2018). Comparative Analyses of Copy-Number Variation in Autism Spectrum Disorder and Schizophrenia Reveal Etiological Overlap and Biological Insights. Cell Reports, 24(11), 2838–2856. doi: 10.1016/j.celrep.2018.08.022 [DOI] [PubMed] [Google Scholar]
- Kwan D, Craver CF, Green L, Myerson J, Boyer P, & Rosenbaum RS. (2012). Future decision-making without episodic mental time travel. Hippocampus, 22(6), 1215–1219. doi: 10.1002/hipo.20981 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kwan D, Craver CF, Green L, Myerson J, & Rosenbaum RS. (2013). Dissociations in future thinking following hippocampal damage: evidence from discounting and time perspective in episodic amnesia. J Exp Psychol Gen, 142(4), 1355–1369. doi: 10.1037/a0034001 [DOI] [PubMed] [Google Scholar]
- Kwiatkowski MA, Cope ZA, Lavadia ML, van de Cappelle CJA, Dulcis D, & Young JW. (2020). Short-active photoperiod gestation induces psychiatry-relevant behavior in healthy mice but a resiliency to such effects are seen in mice with reduced dopamine transporter expression. Sci Rep, 10(1), 10217. doi: 10.1038/s41598-020-66873-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
- LaBar KS, Gitelman DR, Parrish TB, Kim Y-H, Nobre AC, & Mesulam M. (2001). Hunger selectively modulates corticolimbic activation to food stimuli in humans. Behavioral neuroscience, 115(2), 493. [DOI] [PubMed] [Google Scholar]
- Lamontagne SJ, Melendez SI, & Olmstead MC. (2018). Investigating dopamine and glucocorticoid systems as underlying mechanisms of anhedonia. Psychopharmacology (Berl), 235(11), 3103–3113. doi: 10.1007/s00213-018-5007-4 [DOI] [PubMed] [Google Scholar]
- Lancaster TM, Ihssen N, Brindley LM, Tansey KE, Mantripragada K, O'Donovan MC, … Linden DE. (2016). Associations between polygenic risk for schizophrenia and brain function during probabilistic learning in healthy individuals. Hum Brain Mapp, 37(2), 491–500. doi: 10.1002/hbm.23044 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Lawn W, Freeman TP, Pope RA, Joye A, Harvey L, Hindocha C, … Curran HV. (2016). Acute and chronic effects of cannabinoids on effort-related decision-making and reward learning: an evaluation of the cannabis 'amotivational' hypotheses. Psychopharmacology (Berl), 233(19–20), 3537–3552. doi: 10.1007/s00213-016-4383-x [DOI] [PMC free article] [PubMed] [Google Scholar]
- Lee SS, Venniro M, Shaham Y, Hope BT, & Ramsey LA. (2024). Operant social self-administration in male CD1 mice. Psychopharmacology, 1–12. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Lehner R, Balsters JH, Herger A, Hare TA, & Wenderoth N. (2017). Monetary, food, and social rewards induce similar Pavlovian-to-instrumental transfer effects. Frontiers in Behavioral Neuroscience, 10, 247. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Leith NJ, & Barrett RJ. (1976). Amphetamine and the reward system: evidence for tolerance and post-drug depression. Psychopharmacologia, 46(1), 19–25. doi: 10.1007/bf00421544 [DOI] [PubMed] [Google Scholar]
- Lemon C, & Del Arco A. (2022). Intermittent social stress produces different short- and long-term effects on effort-based reward-seeking behavior. Behav Brain Res, 417, 113613. doi: 10.1016/j.bbr.2021.113613 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Lewandowski KE, Whitton AE, Pizzagalli DA, Norris LA, Ongur D, & Hall M-H. (2016). Reward learning, neurocognition, social cognition, and symptomatology in psychosis. Frontiers in Psychiatry, 7, 190080. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Li J, & Daw ND. (2011). Signals in human striatum are appropriate for policy update rather than value prediction. J Neurosci, 31(14), 5504–5511. doi: 10.1523/JNEUROSCI.6316-10.2011 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Li X, Chen W, Huang X, Jing W, Zhang T, Yu Q, … Ding Y. (2021). Synaptic dysfunction of Aldh1a1 neurons in the ventral tegmental area causes impulsive behaviors. Molecular Neurodegeneration, 16, 1–23. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Linke J, King AV, Rietschel M, Strohmaier J, Hennerici M, Gass A, … Wessa M. (2012). Increased medial orbitofrontal and amygdala activation: evidence for a systems-level endophenotype of bipolar I disorder. American Journal of Psychiatry, 169(3), 316–325. [DOI] [PubMed] [Google Scholar]
- Liu WH, Chan RC, Wang LZ, Huang J, Cheung EF, Gong QY, & Gollan JK. (2011). Deficits in sustaining reward responses in subsyndromal and syndromal major depression. Prog Neuropsychopharmacol Biol Psychiatry, 35(4), 1045–1052. doi: 10.1016/j.pnpbp.2011.02.018 [DOI] [PubMed] [Google Scholar]
- Liu X, Hairston J, Schrier M, & Fan J. (2011). Common and distinct networks underlying reward valence and processing stages: a meta-analysis of functional neuroimaging studies. Neuroscience & Biobehavioral Reviews, 35(5), 1219–1236. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Liverant GI, Arditte Hall KA, Wieman ST, Pineles SL, & Pizzagalli DA. (2021). Associations between insomnia and reward learning in clinical depression. Psychol Med, 1–10. doi: 10.1017/S003329172100026X [DOI] [PubMed] [Google Scholar]
- Ljungberg T, Apicella P, & Schultz W. (1992). Responses of monkey dopamine neurons during learning of behavioral reactions. Journal of neurophysiology, 67(1), 145–163. [DOI] [PubMed] [Google Scholar]
- Lopez-Cruz L, Phillips BU, Hailwood JM, Saksida LM, Heath CJ, & Bussey TJ. (2024). Refining the study of decision-making in animals: differential effects of d-amphetamine and haloperidol in a novel touchscreen-automated Rearing-Effort Discounting (RED) task and the Fixed-Ratio Effort Discounting (FRED) task. Neuropsychopharmacology, 49(2), 422–432. doi: 10.1038/s41386-023-01707-z [DOI] [PMC free article] [PubMed] [Google Scholar]
- Lütcke H, Murayama M, Hahn T, Margolis DJ, Astori S, Meyer S, … Kügler S. (2010). Optical recording of neuronal activity with a genetically-encoded calcium indicator in anesthetized and freely moving mice. Frontiers in neural circuits, 4, 1467. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Madisen L, Mao T, Koch H, Zhuo J. m., Berenyi A, Fujisawa S, … Zanella S. (2012). A toolbox of Cre-dependent optogenetic transgenic mice for light-induced activation and silencing. Nature neuroscience, 15(5), 793–802. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Maia TV, Cooney RE, & Peterson BS. (2008). The neural bases of obsessive–compulsive disorder in children and adults. Development and psychopathology, 20(4), 1251–1283. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Mallien AS, Palme R, Richetto J, Muzzillo C, Richter SH, Vogt MA, … Gass P. (2016). Daily exposure to a touchscreen-paradigm and associated food restriction evokes an increase in adrenocortical and neural activity in mice. Hormones and behavior, 81, 97–105. [DOI] [PubMed] [Google Scholar]
- Manohar SG, & Husain M. (2016). Human ventromedial prefrontal lesions alter incentivisation by reward. Cortex, 76, 104–120. doi: 10.1016/j.cortex.2016.01.005 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Mar AC, Walker AL, Theobald DE, Eagle DM, & Robbins TW. (2011). Dissociable effects of lesions to orbitofrontal cortex subregions on impulsive choice in the rat. J Neurosci, 31(17), 6398–6404. doi: 10.1523/jneurosci.6620-10.2011 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Markert C, Klein S, Strahler J, Kruse O, & Stark R. (2021). Sexual incentive delay in the scanner: Sexual cue and reward processing, and links to problematic porn consumption and sexual motivation. J Behav Addict, 10(1), 65–76. doi: 10.1556/2006.2021.00018 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Markou A, Chiamulera C, Geyer MA, Tricklebank M, & Steckler T. (2009). Removing obstacles in neuroscience drug discovery: the future path for animal models. Neuropsychopharmacology, 34(1), 74–89. doi: 10.1038/npp.2008.173 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Markou A, Salamone JD, Bussey TJ, Mar AC, Brunner D, Gilmour G, & Balsam P. (2013). Measuring reinforcement learning and motivation constructs in experimental animals: Relevance to the negative symptoms of schizophrenia. Neuroscience & Biobehavioral Reviews, 37(9), 2149–2165. doi: 10.1016/j.neubiorev.2013.08.007 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Marquardt K, Sigdel R, & Brigman JL. (2017). Touch-screen visual reversal learning is mediated by value encoding and signal propagation in the orbitofrontal cortex. Neurobiol Learn Mem, 139, 179–188. doi: 10.1016/j.nlm.2017.01.006 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Martins D, Rademacher L, Gabay AS, Taylor R, Richey JA, Smith DV, … Paloyelis Y. (2021). Mapping social reward and punishment processing in the human brain: A voxel-based meta-analysis of neuroimaging findings using the social incentive delay task. Neurosci Biobehav Rev, 122, 1–17. doi: 10.1016/j.neubiorev.2020.12.034 [DOI] [PubMed] [Google Scholar]
- McClure SM, Berns GS, & Montague PR. (2003). Temporal prediction errors in a passive learning task activate human striatum. Neuron, 38(2), 339–346. [DOI] [PubMed] [Google Scholar]
- McDonald J, Schleifer L, Richards JB, & de Wit H. (2003). Effects of THC on behavioral measures of impulsivity in humans. Neuropsychopharmacology, 28(7), 1356–1365. doi: 10.1038/sj.npp.1300176 [DOI] [PubMed] [Google Scholar]
- McNamee D, Liljeholm M, Zika O, & O'Doherty JP. (2015). Characterizing the associative content of brain structures involved in habitual and goal-directed actions in humans: a multivariate FMRI study. Journal of Neuroscience, 35(9), 3764–3771. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Mendelsohn AI. (2019). Creatures of Habit: The Neuroscience of Habit and Purposeful Behavior. Biol Psychiatry, 85(11), e49–e51. doi: 10.1016/j.biopsych.2019.03.978 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Mendez IA, Simon NW, Hart N, Mitchell MR, Nation JR, Wellman PJ, & Setlow B. (2010). Self-administered cocaine causes long-lasting increases in impulsive choice in a delay discounting task. Behav Neurosci, 124(4), 470–477. doi: 10.1037/a0020458 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Miedl SF, Peters J, & Büchel C. (2012). Altered neural reward representations in pathological gamblers revealed by delay and probability discounting. Arch Gen Psychiatry, 69(2), 177–186. doi: 10.1001/archgenpsychiatry.2011.1552 [DOI] [PubMed] [Google Scholar]
- Miller EK, & Cohen JD. (2001). An integrative theory of prefrontal cortex function. Annu Rev Neurosci, 24, 167–202. doi: 10.1146/annurev.neuro.24.1.167 [DOI] [PubMed] [Google Scholar]
- Mirenowicz J, & Schultz W. (1994). Importance of unpredictability for reward responses in primate dopamine neurons. Journal of neurophysiology, 72(2), 1024–1027. [DOI] [PubMed] [Google Scholar]
- Mitchell SH. (1999). Measures of impulsivity in cigarette smokers and non-smokers. Psychopharmacology (Berl), 146(4), 455–464. doi: 10.1007/pl00005491 [DOI] [PubMed] [Google Scholar]
- Mok JNY, Green L, Myerson J, Kwan D, Kurczek J, Ciaramelli E, … Rosenbaum SR. (2021). Does Ventromedial Prefrontal Cortex Damage Really Increase Impulsiveness? Delay and Probability Discounting in Patients with Focal Lesions. J Cogn Neurosci, 33(9), 1–19. doi: 10.1162/jocn_a_01721 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Moro AS, Saccenti D, Vergallito A, Scaini S, Malgaroli A, Ferro M, & Lamanna J. (2023). Transcranial direct current stimulation (tDCS) over the orbitofrontal cortex reduces delay discounting. Front Behav Neurosci, 17, 1239463. doi: 10.3389/fnbeh.2023.1239463 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Morris BH, Bylsma LM, Yaroslavsky I, Kovacs M, & Rottenberg J. (2015). Reward learning in pediatric depression and anxiety: preliminary findings in a high-risk sample. Depress Anxiety, 32(5), 373–381. doi: 10.1002/da.22358 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Morris SE, & Cuthbert BN. (2012). Research Domain Criteria: cognitive systems, neural circuits, and dimensions of behavior. Dialogues in clinical neuroscience, 14(1), 29–37. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Morton AJ, Skillings E, Bussey TJ, & Saksida LM. (2006). Measuring cognitive deficits in disabled mice using an automated interactive touchscreen system. Nature Methods, 3(10), 767–767. doi: 10.1038/nmeth1006-767 [DOI] [PubMed] [Google Scholar]
- Moscarello JM, Ben-Shahar O, & Ettenberg A. (2010). External incentives and internal states guide goal-directed behavior via the differential recruitment of the nucleus accumbens and the medial prefrontal cortex. Neuroscience, 170(2), 468–477. doi: 10.1016/j.neuroscience.2010.07.002 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Mueser KT, Bellack AS, Douglas MS, & Morrison RL. (1991). Prevalence and stability of social skill deficits in schizophrenia. Schizophrenia Research, 5(2), 167–176. [DOI] [PubMed] [Google Scholar]
- Mukherjee D, Filipowicz AL, Vo K, Satterthwaite TD, & Kable JW. (2020). Reward and punishment reversal-learning in major depressive disorder. Journal of Abnormal Psychology, 129(8), 810. [DOI] [PubMed] [Google Scholar]
- Mukherjee D, van Geen C, & Kable J. (2023). Leveraging Decision Science to Characterize Depression. Curr Dir Psychol Sci, 32(6), 462–470. doi: 10.1177/09637214231194962 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Mullard A. (2016). Parsing clinical success rates. Nature Reviews Drug Discovery, 15(7), 447–448. [DOI] [PubMed] [Google Scholar]
- Murray GK, Cheng F, Clark L, Barnett JH, Blackwell AD, Fletcher PC, … Jones PB. (2008). Reinforcement and reversal learning in first-episode psychosis. Schizophr Bull, 34(5), 848–855. doi: 10.1093/schbul/sbn078 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Navarick DJ. (2004). Discounting of Delayed Reinforcers: Measurement by Questionnaires Versus Operant Choice Procedures. The Psychological Record, 54, 85–94. [Google Scholar]
- Nestor LJ, & Ersche KD. (2023). Abnormal Brain Networks Related to Drug and Nondrug Reward Anticipation and Outcome Processing in Stimulant Use Disorder: A Functional Connectomics Approach. Biol Psychiatry Cogn Neurosci Neuroimaging, 8(5), 560–571. doi: 10.1016/j.bpsc.2022.08.014 [DOI] [PubMed] [Google Scholar]
- Nielsen MO, Rostrup E, Wulff S, Bak N, Broberg BV, Lublin H, … Glenthoj B. (2012). Improvement of brain reward abnormalities by antipsychotic monotherapy in schizophrenia. Arch Gen Psychiatry, 69(12), 1195–1204. doi: 10.1001/archgenpsychiatry.2012.847 [DOI] [PubMed] [Google Scholar]
- NIMH. (2023). Reward Learning. Retrieved from https://www.nimh.nih.gov/research/research-funded-by-nimh/rdoc/constructs/reward-learning
- Nishioka T, Attachaipanich S, Hamaguchi K, Lazarus M, de Kerchove d’Exaerde A, Macpherson T, & Hikida T. (2023). Error-related signaling in nucleus accumbens D2 receptor-expressing neurons guides inhibition-based choice behavior in mice. Nature communications, 14(1), 2284. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Nithianantharajah J, Komiyama NH, McKechanie A, Johnstone M, Blackwood DH, Clair DS, … Bussey TJ. (2013). Synaptic scaffold evolution generated components of vertebrate cognitive complexity. Nature neuroscience, 16(1), 16–24. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Nithianantharajah J, McKechanie AG, Stewart TJ, Johnstone M, Blackwood DH, St Clair D, … Saksida LM. (2015). Bridging the translational divide: identical cognitive touchscreen testing in mice and humans carrying mutations in a disease-relevant homologous gene. Scientific reports, 5(1), 14613. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Noback M, Bhakta SG, Talledo JA, Kotz JE, Benster L, Roberts BZ, … Young JW. (2024). Amphetamine increases motivation of humans and mice as measured by breakpoint, but does not affect an Electroencephalographic biomarker. Cogn Affect Behav Neurosci. doi: 10.3758/s13415-023-01150-z [DOI] [PMC free article] [PubMed] [Google Scholar]
- Norman KJ, Koike H, McCraney SE, Garkun Y, Bateh J, Falk EN, … Morishita H. (2021). Chemogenetic suppression of anterior cingulate cortical neurons projecting to the visual cortex disrupts attentional behavior in mice. Neuropsychopharmacology Reports, 41(2), 207–214. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Novak KD, & Foti D. (2015). Teasing apart the anticipatory and consummatory processing of monetary incentives: An event-related potential study of reward dynamics. Psychophysiology, 52(11), 1470–1482. doi: 10.1111/psyp.12504 [DOI] [PubMed] [Google Scholar]
- Nunez C, Hoots JK, Schepers ST, Bower M, de Wit H, & Wardle MC. (2022). Pharmacological investigations of effort-based decision-making in humans: Naltrexone and nicotine. PLoS One, 17(10), e0275027. doi: 10.1371/journal.pone.0275027 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Nusslock R, & Alloy LB. (2017). Reward processing and mood-related symptoms: An RDoC and translational neuroscience perspective. Journal of Affective Disorders, 216, 3–16. doi: 10.1016/j.jad.2017.02.001 [DOI] [PMC free article] [PubMed] [Google Scholar]
- O'Doherty J, Kringelbach ML, Rolls ET, Hornak J, & Andrews C. (2001). Abstract reward and punishment representations in the human orbitofrontal cortex. Nature neuroscience, 4(1), 95–102. [DOI] [PubMed] [Google Scholar]
- O'Doherty JP, Dayan P, Friston K, Critchley H, & Dolan RJ. (2003). Temporal difference models and reward-related learning in the human brain. Neuron, 38(2), 329–337. [DOI] [PubMed] [Google Scholar]
- O'Donovan MC, & Owen MJ. (2016). The implications of the shared genetics of psychiatric disorders. Nature medicine, 22(11), 1214–1219. [DOI] [PubMed] [Google Scholar]
- Ohmann HA, Kuper N, & Wacker J. (2018). Left frontal anodal tDCS increases approach motivation depending on reward attributes. Neuropsychologia, 119, 417–423. doi: 10.1016/j.neuropsychologia.2018.09.002 [DOI] [PubMed] [Google Scholar]
- Oldham S, Murawski C, Fornito A, Youssef G, Yucel M, & Lorenzetti V. (2018). The anticipation and outcome phases of reward and loss processing: A neuroimaging meta-analysis of the monetary incentive delay task. Hum Brain Mapp, 39(8), 3398–3418. doi: 10.1002/hbm.24184 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Oomen CA, Hvoslef-Eide M, Heath CJ, Mar AC, Horner AE, Bussey TJ, & Saksida LM. (2013). The touchscreen operant platform for testing working memory and pattern separation in rats and mice. Nature Protocols, 8(10), 2006–2021. doi: 10.1038/nprot.2013.124 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Oorschot M, Lataster T, Thewissen V, Lardinois M, Wichers M, Van Os J, … Myin-Germeys I. (2013). Emotional Experience in Negative Symptoms of Schizophrenia—No Evidence for a Generalized Hedonic Deficit. Schizophrenia Bulletin, 39(1), 217–225. doi: 10.1093/schbul/sbr137 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Ortner GR, Wibral M, Becker A, Dohmen T, Klingmüller D, Falk A, & Weber B. (2013). No evidence for an effect of testosterone administration on delay discounting in male university students. Psychoneuroendocrinology, 38(9), 1814–1818. doi: 10.1016/j.psyneuen.2012.12.014 [DOI] [PubMed] [Google Scholar]
- Ozga-Hess JE, & Anderson KG. (2019). Differential effects of d-amphetamine and atomoxetine on risk-based decision making of Lewis and Fischer 344 rats. Behav Pharmacol, 30(7), 605–616. doi: 10.1097/fbp.0000000000000500 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Palmer D, Dumont JR, Dexter TD, Prado MAM, Finger E, Bussey TJ, & Saksida LM. (2021). Touchscreen cognitive testing: Cross-species translation and co-clinical trials in neurodegenerative and neuropsychiatric disease. Neurobiology of Learning and Memory, 182, 107443. doi: 10.1016/j.nlm.2021.107443 [DOI] [PubMed] [Google Scholar]
- Palombo DJ, Keane MM, & Verfaellie M. (2015). The medial temporal lobes are critical for reward-based decision making under conditions that promote episodic future thinking. Hippocampus, 25(3), 345–353. doi: 10.1002/hipo.22376 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Park IH, Chun JW, Park HJ, Koo MS, Park S, Kim SH, & Kim JJ. (2015). Altered cingulo-striatal function underlies reward drive deficits in schizophrenia. Schizophr Res, 161(2–3), 229–236. doi: 10.1016/j.schres.2014.11.005 [DOI] [PubMed] [Google Scholar]
- Pastor-Bernier A, Stasiak A, & Schultz W. (2021). Reward-specific satiety affects subjective value signals in orbitofrontal cortex during multicomponent economic choice. Proceedings of the National Academy of Sciences, 118(30), e2022650118. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Patel KT, Stevens MC, Meda SA, Muska C, Thomas AD, Potenza MN, & Pearlson GD. (2013). Robust changes in reward circuitry during reward loss in current and former cocaine users during performance of a monetary incentive delay task. Biological Psychiatry, 74(7), 529–537. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Patriarchi T, Cho JR, Merten K, Howe MW, Marley A, Xiong W-H, … Jang MJ. (2018). Ultrafast neuronal imaging of dopamine dynamics with designed genetically encoded sensors. Science, 360(6396), eaat4422. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Patt VM, Hunsberger R, Jones DA, & Verfaellie M. (2023). The Hippocampus Contributes to Temporal Discounting When Delays and Rewards Are Experienced in the Moment. J Neurosci, 43(31), 5710–5722. doi: 10.1523/jneurosci.2250-22.2023 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Pattij T, Schetters D, & Schoffelmeer AN. (2014). Dopaminergic modulation of impulsive decision making in the rat insular cortex. Behav Brain Res, 270, 118–124. doi: 10.1016/j.bbr.2014.05.010 [DOI] [PubMed] [Google Scholar]
- Pennisi P, Salehinejad MA, Corso AM, Merlo EM, Avenanti A, & Vicario CM. (2023). Delay discounting in Parkinson's disease: A systematic review and meta-analysis. Behav Brain Res, 436, 114101. doi: 10.1016/j.bbr.2022.114101 [DOI] [PubMed] [Google Scholar]
- Pergadia ML, Der-Avakian A, D'Souza MS, Madden PAF, Heath AC, Shiffman S, … Pizzagalli DA. (2014). Association between nicotine withdrawal and reward responsiveness in humans and rats. JAMA Psychiatry, 71(11), 1238–1245. doi: 10.1001/jamapsychiatry.2014.1016 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Perry W, Light GA, Davis H, & Braff DL. (2000). Schizophrenia patients demonstrate a dissociation on declarative and non-declarative memory tests. Schizophrenia Research, 46(2–3), 167–174. [DOI] [PubMed] [Google Scholar]
- Perry W, Minassian A, Paulus MP, Young JW, Kincaid MJ, Ferguson EJ, … Sharp RF. (2009). A reverse-translational study of dysfunctional exploration in psychiatric disorders: from mice to men. Archives of General Psychiatry, 66(10), 1072–1080. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Peters J, & Büchel C. (2009). Overlapping and distinct neural systems code for subjective value during intertemporal and risky decision making. J Neurosci, 29(50), 15727–15734. doi: 10.1523/jneurosci.3489-09.2009 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Peters J, & Büchel C. (2010). Episodic future thinking reduces reward delay discounting through an enhancement of prefrontal-mediotemporal interactions. Neuron, 66(1), 138–148. doi: 10.1016/j.neuron.2010.03.026 [DOI] [PubMed] [Google Scholar]
- Peters J, & D'Esposito M. (2020). The drift diffusion model as the choice rule in inter-temporal and risky choice: A case study in medial orbitofrontal cortex lesion patients and controls. PLoS Comput Biol, 16(4), e1007615. doi: 10.1371/journal.pcbi.1007615 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Petry NM. (2001). Pathological gamblers, with and without substance use disorders, discount delayed rewards at high rates. J Abnorm Psychol, 110(3), 482–487. doi: 10.1037//0021-843x.110.3.482 [DOI] [PubMed] [Google Scholar]
- Phillips BU, Dewan S, Nilsson SRO, Robbins TW, Heath CJ, Saksida LM, … Alsio J. (2018). Selective effects of 5-HT2C receptor modulation on performance of a novel valence-probe visual discrimination task and probabilistic reversal learning in mice. Psychopharmacology (Berl), 235(7), 2101–2111. doi: 10.1007/s00213-018-4907-7 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Phillips BU, Heath CJ, Ossowska Z, Bussey TJ, & Saksida LM. (2017). Optimisation of cognitive performance in rodent operant (touchscreen) testing: Evaluation and effects of reinforcer strength. Learning & Behavior, 45(3), 252–262. doi: 10.3758/s13420-017-0260-7 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Phillips BU, Lopez-Cruz L, Hailwood JM, Heath CJ, Saksida LM, & Bussey TJ. (2018). Translational approaches to evaluating motivation in laboratory rodents: conventional and touchscreen-based procedures. Current Opinion in Behavioral Sciences, 22, 21–27. doi: 10.1016/j.cobeha.2017.12.008 [DOI] [Google Scholar]
- Phung QH, Snider SE, Tegge AN, & Bickel WK. (2019). Willing to Work But Not to Wait: Individuals with Greater Alcohol Use Disorder Show Increased Delay Discounting Across Commodities and Less Effort Discounting for Alcohol. Alcohol Clin Exp Res, 43(5), 927–936. doi: 10.1111/acer.13996 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Picard K, Bisht K, Poggini S, Garofalo S, Golia MT, Basilico B, … Tremblay M. (2021). Microglial-glucocorticoid receptor depletion alters the response of hippocampal microglia and neurons in a chronic unpredictable mild stress paradigm in female mice. Brain Behav Immun, 97, 423–439. doi: 10.1016/j.bbi.2021.07.022 [DOI] [PubMed] [Google Scholar]
- Pine A, Shiner T, Seymour B, & Dolan RJ. (2010). Dopamine, time, and impulsivity in humans. J Neurosci, 30(26), 8888–8896. doi: 10.1523/jneurosci.6028-09.2010 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Pizzagalli DA, Goetz E, Ostacher M, Iosifescu DV, & Perlis RH. (2008). Euthymic patients with bipolar disorder show decreased reward learning in a probabilistic reward task. Biol Psychiatry, 64(2), 162–168. doi: 10.1016/j.biopsych.2007.12.001 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Pizzagalli DA, Iosifescu D, Hallett LA, Ratner KG, & Fava M. (2008). Reduced hedonic capacity in major depressive disorder: evidence from a probabilistic reward task. J Psychiatr Res, 43(1), 76–87. doi: 10.1016/j.jpsychires.2008.03.001 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Pizzagalli DA, Jahn AL, & O'Shea JP. (2005). Toward an objective characterization of an anhedonic phenotype: a signal-detection approach. Biol Psychiatry, 57(4), 319–327. doi: 10.1016/j.biopsych.2004.11.026 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Pope HG, & Lipinski JF. (1978). Diagnosis in schizophrenia and manic-depressive illness: a reassessment of the specificity of'schizophrenic'symptoms in the light of current research. Archives of General Psychiatry, 35(7), 811–828. [DOI] [PubMed] [Google Scholar]
- Porcelli AJ, Lewis AH, & Delgado MR. (2012). Acute stress influences neural circuits of reward processing. Front Neurosci, 6, 157. doi: 10.3389/fnins.2012.00157 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Proudfit GH. (2015). The reward positivity: from basic research on reward to a biomarker for depression. Psychophysiology, 52(4), 449–459. doi: 10.1111/psyp.12370 [DOI] [PubMed] [Google Scholar]
- Pujara MS, Philippi CL, Motzkin JC, Baskaya MK, & Koenigs M. (2016). Ventromedial Prefrontal Cortex Damage Is Associated with Decreased Ventral Striatum Volume and Response to Reward. J Neurosci, 36(18), 5047–5054. doi: 10.1523/JNEUROSCI.4236-15.2016 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Quante SM, Siewert V, Palme R, Kaiser S, Sachser N, & Richter SH. (2023). The power of a touch: Regular touchscreen training but not its termination affects hormones and behavior in mice. Frontiers in Behavioral Neuroscience, 17, 1112780. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Rachlin H, Raineri A, & Cross D. (1991). Subjective probability and delay. J Exp Anal Behav, 55(2), 233–244. doi: 10.1901/jeab.1991.55-233 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Rademacher L, Krach S, Kohls G, Irmak A, Grunder G, & Spreckelmeyer KN. (2010). Dissociation of neural networks for anticipation and consumption of monetary and social rewards. Neuroimage, 49(4), 3276–3285. doi: 10.1016/j.neuroimage.2009.10.089 [DOI] [PubMed] [Google Scholar]
- Radke AK, Kocharian A, Covey DP, Lovinger DM, Cheer JF, Mateo Y, & Holmes A. (2019). Contributions of nucleus accumbens dopamine to cognitive flexibility. European Journal of Neuroscience, 50(3), 2023–2035. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Radke AK, Zweifel LS, & Holmes A. (2019). NMDA receptor deletion on dopamine neurons disrupts visual discrimination and reversal learning. Neurosci Lett, 699, 109–114. doi: 10.1016/j.neulet.2019.02.001 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Ramsey LA, Holloman FM, Hope BT, Shaham Y, & Venniro M. (2022). Waving through the window: a model of volitional social interaction in female mice. Biological Psychiatry, 91(11), 988–997. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Ramsey LA, Holloman FM, Lee SS, & Venniro M. (2023). An operant social self-administration and choice model in mice. Nature Protocols, 18(6), 1669–1686. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Rapaport MH, Clary C, Fayyad R, & Endicott J. (2005). Quality-of-life impairment in depressive and anxiety disorders. American Journal of Psychiatry, 162(6), 1171–1178. [DOI] [PubMed] [Google Scholar]
- Reber J, Feinstein JS, O’Doherty JP, Liljeholm M, Adolphs R, & Tranel D. (2017). Selective impairment of goal-directed decision-making following lesions to the human ventromedial prefrontal cortex. Brain, 140(6), 1743–1756. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Reddy LF, Waltz JA, Green MF, Wynn JK, & Horan WP. (2016). Probabilistic Reversal Learning in Schizophrenia: Stability of Deficits and Potential Causal Mechanisms. Schizophr Bull, 42(4), 942–951. doi: 10.1093/schbul/sbv226 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Reilly EE, Whitton AE, Pizzagalli DA, Rutherford AV, Stein MB, Paulus MP, & Taylor CT. (2020). Diagnostic and dimensional evaluation of implicit reward learning in social anxiety disorder and major depression. Depress Anxiety, 37(12), 1221–1230. doi: 10.1002/da.23081 [DOI] [PubMed] [Google Scholar]
- Reilly EE, Whitton AE, Pizzagalli DA, Rutherford AV, Stein MB, Paulus MP, & Taylor CT. (2020). Diagnostic and dimensional evaluation of implicit reward learning in social anxiety disorder and major depression. Depression and anxiety, 37(12), 1221–1230. [DOI] [PubMed] [Google Scholar]
- Reynolds B, Richards JB, Horn K, & Karraker K. (2004). Delay discounting and probability discounting as related to cigarette smoking status in adults. Behav Processes, 65(1), 35–42. doi: 10.1016/s0376-6357(03)00109-8 [DOI] [PubMed] [Google Scholar]
- Reynolds B, & Schiffbauer R. (2004). Measuring state changes in human delay discounting: an experiential discounting task. Behav Processes, 67(3), 343–356. doi: 10.1016/j.beproc.2004.06.003 [DOI] [PubMed] [Google Scholar]
- Rhebergen D, Beekman AT, de Graaf R, Nolen WA, Spijker J, Hoogendijk WJ, & Penninx BW. (2010). Trajectories of recovery of social and physical functioning in major depression, dysthymic disorder and double depression: a 3-year follow-up. Journal of Affective Disorders, 124(1–2), 148–156. [DOI] [PubMed] [Google Scholar]
- Richards JB, Mitchell SH, de Wit H, & Seiden LS. (1997). Determination of discount functions in rats with an adjusting-amount procedure. J Exp Anal Behav, 67(3), 353–366. doi: 10.1901/jeab.1997.67-353 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Richards JB, Zhang L, Mitchell SH, & de Wit H. (1999). Delay or probability discounting in a model of impulsive behavior: effect of alcohol. J Exp Anal Behav, 71(2), 121–143. doi: 10.1901/jeab.1999.71-121 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Robbins TW, James M, Owen AM, Sahakian BJ, Lawrence AD, McInnes L, & Rabbitt PM. (1998). A study of performance on tests from the CANTAB battery sensitive to frontal lobe dysfunction in a large sample of normal volunteers: implications for theories of executive functioning and cognitive aging. Cambridge Neuropsychological Test Automated Battery. Journal of the International Neuropsychological Society. doi: 10.1017/s1355617798455073 [DOI] [PubMed] [Google Scholar]
- Robbins TW, James M, Owen AM, Sahakian BJ, McInnes L, & Rabbitt P. (1994). Cambridge Neuropsychological Test Automated Battery (CANTAB): A Factor Analytic Study of a Large Sample of Normal Elderly Volunteers. Dementia and Geriatric Cognitive Disorders, 5(5), 266–281. doi: 10.1159/000106735 [DOI] [PubMed] [Google Scholar]
- Roberts BZ, He YV, Chatha M, Minassian A, Grant I, & Young JW. (2021). HIV Transgenic Rats Demonstrate Superior Task Acquisition and Intact Reversal Learning in the Within-Session Probabilistic Reversal Learning Task. Cogn Affect Behav Neurosci. doi: 10.3758/s13415-021-00926-5 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Roberts BZ, O'Connor MA, Kenton JA, Barnes SA, & Young JW. (2023). Short-active gestational photoperiod reduces effortful choice behavior in mice, partial normalization by d-amphetamine. Psychopharmacology (Berl), 240(11), 2303–2315. doi: 10.1007/s00213-023-06337-3 [DOI] [PubMed] [Google Scholar]
- Roberts BZ, Young JW, He YV, Cope ZA, Shilling PD, & Feifel D. (2019). Oxytocin improves probabilistic reversal learning but not effortful motivation in Brown Norway rats. Neuropharmacology, 150, 15–26. doi: 10.1016/j.neuropharm.2019.02.028 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Robinson OJ, Cools R, Carlisi CO, Sahakian BJ, & Drevets WC. (2012). Ventral striatum response during reward and punishment reversal learning in unmedicated major depressive disorder. Am J Psychiatry, 169(2), 152–159. doi: 10.1176/appi.ajp.2011.11010137 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Robinson OJ, Frank MJ, Sahakian BJ, & Cools R. (2010). Dissociable responses to punishment in distinct striatal regions during reversal learning. Neuroimage, 51(4), 1459–1467. doi: 10.1016/j.neuroimage.2010.03.036 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Rogers RD, Everitt BJ, Baldacchino A, Blackshaw AJ, Swainson R, Wynne K, … Robbins TW. (1999). Dissociable deficits in the decision-making cognition of chronic amphetamine abusers, opiate abusers, patients with focal damage to prefrontal cortex, and tryptophan-depleted normal volunteers: evidence for monoaminergic mechanisms. Neuropsychopharmacology, 20(4), 322–339. doi: 10.1016/S0893-133X(98)00091-8 [DOI] [PubMed] [Google Scholar]
- Roiser JP, & Sahakian BJ. (2013). Hot and cold cognition in depression. CNS spectrums, 18(3), 139–149. [DOI] [PubMed] [Google Scholar]
- Rojas GR, Curry-Pochy LS, Chen CS, Heller AT, & Grissom NM. (2022). Sequential delay and probability discounting tasks in mice reveal anchoring effects partially attributable to decision noise. Behav Brain Res, 431, 113951. doi: 10.1016/j.bbr.2022.113951 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Rolls ET, Rolls BJ, & Rowe EA. (1983). Sensory-specific and motivation-specific satiety for the sight and taste of food and water in man. Physiology & Behavior, 30(2), 185–192. [DOI] [PubMed] [Google Scholar]
- Rolls ET, & Rolls J. (1997). Olfactory sensory-specific satiety in humans. Physiology & Behavior, 61(3), 461–473. [DOI] [PubMed] [Google Scholar]
- Rolls ET, Sienkiewicz ZJ, & Yaxley S. (1989). Hunger modulates the responses to gustatory stimuli of single neurons in the caudolateral orbitofrontal cortex of the macaque monkey. European Journal of Neuroscience, 1(1), 53–60. [DOI] [PubMed] [Google Scholar]
- Rost BR, Schneider-Warme F, Schmitz D, & Hegemann P. (2017). Optogenetic tools for subcellular applications in neuroscience. Neuron, 96(3), 572–603. [DOI] [PubMed] [Google Scholar]
- Salamone JD, & Correa M. (2002). Motivational views of reinforcement: implications for understanding the behavioral functions of nucleus accumbens dopamine. Behavioural Brain Research, 137(1–2), 3–25. doi: 10.1016/S0166-4328(02)00282-6 [DOI] [PubMed] [Google Scholar]
- Salamone JD, Cousins MS, & Bucher S. (1994). Anhedonia or anergia? Effects of haloperidol and nucleus accumbens dopamine depletion on instrumental response selection in a T-maze cost/benefit procedure. Behav Brain Res, 65(2), 221–229. doi: 10.1016/0166-4328(94)90108-2 [DOI] [PubMed] [Google Scholar]
- Sanislow CA, Pine DS, Quinn KJ, Kozak MJ, Garvey MA, Heinssen RK, … Cuthbert BN. (2010). Developing constructs for psychopathology research: Research domain criteria. Journal of Abnormal Psychology, 119(4), 631–639. doi: 10.1037/a0020909 [DOI] [PubMed] [Google Scholar]
- Santesso DL, Evins AE, Frank MJ, Schetter EC, Bogdan R, & Pizzagalli DA. (2009). Single dose of a dopamine agonist impairs reinforcement learning in humans: evidence from event-related potentials and computational modeling of striatal-cortical function. Hum Brain Mapp, 30(7), 1963–1976. doi: 10.1002/hbm.20642 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Savulich G, Jeanes H, Rossides N, Kaur S, Zacharia A, Robbins TW, & Sahakian BJ. (2018). Moral Emotions and Social Economic Games in Paranoia. Frontiers in Psychiatry, 9, 615. doi: 10.3389/fpsyt.2018.00615 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Scheres A, de Water E, & Mies GW. (2013). The neural correlates of temporal reward discounting. Wiley Interdiscip Rev Cogn Sci, 4(5), 523–545. doi: 10.1002/wcs.1246 [DOI] [PubMed] [Google Scholar]
- Schlagenhauf F, Huys QJ, Deserno L, Rapp MA, Beck A, Heinze H-J, … Heinz A. (2014). Striatal dysfunction during reversal learning in unmedicated schizophrenia patients. Neuroimage, 89, 171–180. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Schluter MG, & Hodgins DC. (2021). Reward-Related Decision-Making in Current and Past Disordered Gambling: Implications for Impulsive Choice and Risk Preference in the Maintenance of Gambling Disorder. Front Behav Neurosci, 15, 758329. doi: 10.3389/fnbeh.2021.758329 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Schott BH, Minuzzi L, Krebs RM, Elmenhorst D, Lang M, Winz OH, … Bauer A. (2008). Mesolimbic functional magnetic resonance imaging activations during reward anticipation correlate with reward-related ventral striatal dopamine release. J Neurosci, 28(52), 14311–14319. doi: 10.1523/JNEUROSCI.2058-08.2008 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Schultz W. (2016). Dopamine reward prediction error coding. Dialogues in clinical neuroscience, 18(1), 23–32. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Schultz W, Dayan P, & Montague PR. (1997). A neural substrate of prediction and reward. Science, 275(5306), 1593–1599. doi: 10.1126/science.275.5306.1593 [DOI] [PubMed] [Google Scholar]
- Scoglio AA, Reilly ED, Girouard C, Quigley KS, Carnes S, & Kelly MM. (2022). Social functioning in individuals with post-traumatic stress disorder: A systematic review. Trauma, Violence, & Abuse, 23(2), 356–371. [DOI] [PubMed] [Google Scholar]
- Seaman KL, Brooks N, Karrer TM, Castrellon JJ, Perkins SF, Dang LC, … Samanez-Larkin GR. (2018). Subjective value representations during effort, probability and time discounting across adulthood. Soc Cogn Affect Neurosci, 13(5), 449–459. doi: 10.1093/scan/nsy021 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Seamans JK, Lapish CC, & Durstewitz D. (2008). Comparing the prefrontal cortex of rats and primates: insights from electrophysiology. Neurotox Res, 14(2–3), 249–262. doi: 10.1007/bf03033814 [DOI] [PubMed] [Google Scholar]
- Sellitto M, Ciaramelli E, Mattioli F, & di Pellegrino G. (2015). Reduced Sensitivity to Sooner Reward During Intertemporal Decision-Making Following Insula Damage in Humans. Front Behav Neurosci, 9, 367. doi: 10.3389/fnbeh.2015.00367 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Shafiei N, Gray M, Viau V, & Floresco SB. (2012). Acute stress induces selective alterations in cost/benefit decision-making. Neuropsychopharmacology, 37(10), 2194–2209. doi: 10.1038/npp.2012.69 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Shen B, Yin Y, Wang J, Zhou X, McClure SM, & Li J. (2016). High-definition tDCS alters impulsivity in a baseline-dependent manner. Neuroimage, 143, 343–352. doi: 10.1016/j.neuroimage.2016.09.006 [DOI] [PubMed] [Google Scholar]
- Shor-Posner G, Grinker JA, Marinescu C, Brown O, & Leibowitz SF. (1986). Hypothalamic serotonin in the control of meal patterns and macronutrient selection. Brain Research Bulletin, 17(5), 663–671. [DOI] [PubMed] [Google Scholar]
- Sidman M. (1962). Classical avoidance without a warning stimulus. Journal of the Experimental Analysis of Behavior, 5(1), 97–104. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Sidman M. (2006). The distinction between positive and negative reinforcement: Some additional considerations. The Behavior Analyst, 29(1), 135. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Siep N, Roefs A, Roebroeck A, Havermans R, Bonte ML, & Jansen A. (2009). Hunger is the best spice: an fMRI study of the effects of attention, hunger and calorie content on food reward processing in the amygdala and orbitofrontal cortex. Behavioural Brain Research, 198(1), 149–158. [DOI] [PubMed] [Google Scholar]
- Simon JJ, Skunde M, Wu M, Schnell K, Herpertz SC, Bendszus M, … Friederich H-C. (2015a). Neural dissociation of food-and money-related reward processing using an abstract incentive delay task. Social cognitive and affective neuroscience, 10(8), 1113–1120. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Simon JJ, Skunde M, Wu M, Schnell K, Herpertz SC, Bendszus M, … Friederich HC. (2015b). Neural dissociation of food- and money-related reward processing using an abstract incentive delay task. Soc Cogn Affect Neurosci, 10(8), 1113–1120. doi: 10.1093/scan/nsu162 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Simon NW, Beas BS, Montgomery KS, Haberman RP, Bizon JL, & Setlow B. (2013). Prefrontal cortical-striatal dopamine receptor mRNA expression predicts distinct forms of impulsivity. Eur J Neurosci, 37(11), 1779–1788. doi: 10.1111/ejn.12191 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Sjoerds Z, de Wit S, van den Brink W, Robbins TW, Beekman AT, Penninx BW, & Veltman DJ. (2013). Behavioral and neuroimaging evidence for overreliance on habit learning in alcohol-dependent patients. Transl Psychiatry, 3(12), e337. doi: 10.1038/tp.2013.107 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Skinner BF. (1963). Operant behavior. American psychologist, 18(8), 503. [Google Scholar]
- Skirzewski M, Princz-Lebel O, German-Castelan L, Crooks AM, Kim GK, Tarnow SH, … Li Y. (2022). Continuous cholinergic-dopaminergic updating in the nucleus accumbens underlies approaches to reward-predicting cues. Nature communications, 13(1), 7924. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Slezak JM, & Anderson KG. (2009). Effects of variable training, signaled and unsignaled delays, and d-amphetamine on delay-discounting functions. Behav Pharmacol, 20(5–6), 424–436. doi: 10.1097/FBP.0b013e3283305ef9 [DOI] [PubMed] [Google Scholar]
- Soder HE, Cooper JA, Lopez-Gamundi P, Hoots JK, Nunez C, Lawlor VM, … Wardle MC. (2021). Dose-response effects of d-amphetamine on effort-based decision-making and reinforcement learning. Neuropsychopharmacology, 46(6), 1078–1085. doi: 10.1038/s41386-020-0779-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Soler MJ, Ruiz JC, Dasí C, & Fuentes-Durá I. (2015). Implicit memory functioning in schizophrenia: Explaining inconsistent findings of word stem completion tasks. Psychiatry research, 226(1), 347–351. [DOI] [PubMed] [Google Scholar]
- Sponheim SR, Steele VR, & McGuire KA. (2004). Verbal memory processes in schizophrenia patients and biological relatives of schizophrenia patients: intact implicit memory, impaired explicit recollection. Schizophrenia Research, 71(2–3), 339–348. [DOI] [PubMed] [Google Scholar]
- Spreckelmeyer KN, Krach S, Kohls G, Rademacher L, Irmak A, Konrad K, … Grunder G. (2009). Anticipation of monetary and social reward differently activates mesolimbic brain structures in men and women. Soc Cogn Affect Neurosci, 4(2), 158–165. doi: 10.1093/scan/nsn051 [DOI] [PMC free article] [PubMed] [Google Scholar]
- St Onge JR, Chiu YC, & Floresco SB. (2010). Differential effects of dopaminergic manipulations on risky choice. Psychopharmacology (Berl), 211(2), 209–221. doi: 10.1007/s00213-010-1883-y [DOI] [PubMed] [Google Scholar]
- St Onge JR, & Floresco SB. (2009). Dopaminergic modulation of risk-based decision making. Neuropsychopharmacology, 34(3), 681–697. doi: 10.1038/npp.2008.121 [DOI] [PubMed] [Google Scholar]
- St Onge JR, & Floresco SB. (2010). Prefrontal cortical contribution to risk-based decision making. Cereb Cortex, 20(8), 1816–1828. doi: 10.1093/cercor/bhp250 [DOI] [PubMed] [Google Scholar]
- Stewart A, Davis GL, Areal LB, Rabil MJ, Tran V, Mayer FP, & Blakely RD. (2022). Male DAT Val559 mice exhibit compulsive behavior under devalued reward conditions accompanied by cellular and pharmacological changes. Cells, 11(24), 4059. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Stolyarova A, O'Dell SJ, Marshall JF, & Izquierdo A. (2014). Positive and negative feedback learning and associated dopamine and serotonin transporter binding after methamphetamine. Behav Brain Res, 271, 195–202. doi: 10.1016/j.bbr.2014.06.031 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Stone CA, Wenger HC, Ludden CT, Stavorski JM, & Ross CA. (1961). Antiserotonin-antihistaminic properties of cyproheptadine. Journal of Pharmacology and experimental Therapeutics, 131(1), 73–84. [Google Scholar]
- Stopper CM, & Floresco SB. (2011). Contributions of the nucleus accumbens and its subregions to different aspects of risk-based decision making. Cogn Affect Behav Neurosci, 11(1), 97–112. doi: 10.3758/s13415-010-0015-9 [DOI] [PubMed] [Google Scholar]
- Stopper CM, Green EB, & Floresco SB. (2014). Selective involvement by the medial orbitofrontal cortex in biasing risky, but not impulsive, choice. Cereb Cortex, 24(1), 154–162. doi: 10.1093/cercor/bhs297 [DOI] [PubMed] [Google Scholar]
- Strauss GP, Thaler NS, Matveeva TM, Vogel SJ, Sutton GP, Lee BG, & Allen DN. (2015). Predicting psychosis across diagnostic boundaries: Behavioral and computational modeling evidence for impaired reinforcement learning in schizophrenia and bipolar disorder with a history of psychosis. Journal of Abnormal Psychology, 124(3), 697. [DOI] [PubMed] [Google Scholar]
- Strauss GP, Whearty KM, Morra LF, Sullivan SK, Ossenfort KL, & Frost KH. (2016). Avolition in schizophrenia is associated with reduced willingness to expend effort for reward on a Progressive Ratio task. Schizophr Res, 170(1), 198–204. doi: 10.1016/j.schres.2015.12.006 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Swainson R, Rogers RD, Sahakian BJ, Summers BA, Polkey CE, & Robbins TW. (2000). Probabilistic learning and reversal deficits in patients with Parkinson's disease or frontal or temporal lobe lesions: possible adverse effects of dopaminergic medication. Neuropsychologia, 38(5), 596–612. doi: 10.1016/s0028-3932(99)00103-7 [DOI] [PubMed] [Google Scholar]
- Tam GW, van de Lagemaat LN, Redon R, Strathdee KE, Croning MD, Malloy MP, … Grant SG. (2010). Confirmed rare copy number variants implicate novel genes in schizophrenia. Biochem Soc Trans, 38(2), 445–451. doi: 10.1042/bst0380445 [DOI] [PubMed] [Google Scholar]
- Tanaka SC, Doya K, Okada G, Ueda K, Okamoto Y, & Yamawaki S. (2016). Prediction of immediate and future rewards differentially recruits cortico-basal ganglia loops. Behavioral economics of preferences, choices, and happiness, 593–616. [DOI] [PubMed] [Google Scholar]
- Tanno T, Maguire DR, Henson C, & France CP. (2014). Effects of amphetamine and methylphenidate on delay discounting in rats: interactions with order of delay presentation. Psychopharmacology (Berl), 231(1), 85–95. doi: 10.1007/s00213-013-3209-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Taylor CT, Hoffman SN, & Khan AJ. (2022). Anhedonia in Anxiety Disorders. Curr Top Behav Neurosci, 58, 201–218. doi: 10.1007/7854_2022_319 [DOI] [PubMed] [Google Scholar]
- Taylor CT, Pearlstein SL, & Stein MB. (2020). A tale of two systems: Testing a positive and negative valence systems framework to understand social disconnection across anxiety and depressive disorders. Journal of Affective Disorders, 266, 207–214. doi: 10.1016/j.jad.2020.01.041 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Thomas JM, Dourish CT, Tomlinson J, Hassan-Smith Z, Hansen PC, & Higgs S. (2018). The 5-HT 2C receptor agonist meta-chlorophenylpiperazine (mCPP) reduces palatable food consumption and BOLD fMRI responses to food images in healthy female volunteers. Psychopharmacology, 235, 257–267. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Thomas JM, Higgs S, Dourish CT, Hansen PC, Harmer CJ, & McCabe C. (2015). Satiation attenuates BOLD activity in brain regions involved in reward and increases activity in dorsolateral prefrontal cortex: an fMRI study in healthy volunteers. The American journal of clinical nutrition, 101(4), 701–708. [DOI] [PubMed] [Google Scholar]
- Thorndike E. (1911). Animal intelligence: Experimental studies: Routledge. [Google Scholar]
- Thut G, Schultz W, Roelcke U, Nienhusmeier M, Missimer J, Maguire RP, & Leenders KL. (1997). Activation of the human brain by monetary reward. Neuroreport, 8(5), 1225–1228. [DOI] [PubMed] [Google Scholar]
- Titone MK, Depp C, Klaus F, Carrasco J, Young JW, & Eyler LT. (2022). The interplay of daily affect and impulsivity measured by mobile surveys in bipolar disorder. International journal of bipolar disorders, 10(1), 25. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Tranter MM, Aggarwal S, Young JW, Dillon DG, & Barnes SA. (2023). Reinforcement learning deficits exhibited by postnatal PCP-treated rats enable deep neural network classification. Neuropsychopharmacology, 48(9), 1377–1385. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Tranter MM, Faget L, Hnasko TS, Powell SB, Dillon DG, & Barnes SA. (2024). Postnatal phencyclidine-induced deficits in decision making are ameliorated by optogenetic inhibition of ventromedial orbitofrontal cortical glutamate neurons. Biological Psychiatry Global Open Science, 4(1), 264–274. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Treadway MT, Bossaller NA, Shelton RC, & Zald DH. (2012). Effort-based decision-making in major depressive disorder: a translational model of motivational anhedonia. J Abnorm Psychol, 121(3), 553–558. doi: 10.1037/a0028813 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Treadway MT, Buckholtz JW, Cowan RL, Woodward ND, Li R, Ansari MS, … Zald DH. (2012). Dopaminergic mechanisms of individual differences in human effort-based decision-making. J Neurosci, 32(18), 6170–6176. doi: 10.1523/jneurosci.6459-11.2012 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Treadway MT, Buckholtz JW, Schwartzman AN, Lambert WE, & Zald DH. (2009). Worth the 'EEfRT'? The effort expenditure for rewards task as an objective measure of motivation and anhedonia. PLoS ONE, 4(8), e6598. doi: 10.1371/journal.pone.0006598 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Treadway MT, & Zald DH. (2011). Reconsidering anhedonia in depression: lessons from translational neuroscience. Neurosci Biobehav Rev, 35(3), 537–555. doi: 10.1016/j.neubiorev.2010.06.006 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Tricomi E, Delgado MR, McCandliss BD, McClelland JL, & Fiez JA. (2006). Performance feedback drives caudate activation in a phonological learning task. J Cogn Neurosci, 18(6), 1029–1043. doi: 10.1162/jocn.2006.18.6.1029 [DOI] [PubMed] [Google Scholar]
- Triscoli C, Ackerley R, & Sailer U. (2014). Touch satiety: differential effects of stroking velocity on liking and wanting touch over repetitions. PLoS One, 9(11), e113425. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Uylings HB, Groenewegen HJ, & Kolb B. (2003). Do rats have a prefrontal cortex? Behav Brain Res, 146(1–2), 3–17. doi: 10.1016/j.bbr.2003.09.028 [DOI] [PubMed] [Google Scholar]
- Venniro M, & Shaham Y. (2020). An operant social self-administration and choice model in rats. Nature Protocols, 15(4), 1542–1559. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Vigo D, Thornicroft G, & Atun R. (2016). Estimating the true global burden of mental illness. The Lancet Psychiatry, 3(2), 171–178. [DOI] [PubMed] [Google Scholar]
- Voigt J-P, & Fink H. (2015). Serotonin controlling feeding and satiety. Behavioural Brain Research, 277, 14–31. [DOI] [PubMed] [Google Scholar]
- Vrieze E, Pizzagalli DA, Demyttenaere K, Hompes T, Sienaert P, de Boer P, … Claes S. (2013). Reduced reward learning predicts outcome in major depressive disorder. Biol Psychiatry, 73(7), 639–645. doi: 10.1016/j.biopsych.2012.10.014 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Wagner B, Clos M, Sommer T, & Peters J. (2020). Dopaminergic Modulation of Human Intertemporal Choice: A Diffusion Model Analysis Using the D2-Receptor Antagonist Haloperidol. J Neurosci, 40(41), 7936–7948. doi: 10.1523/jneurosci.0592-20.2020 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Wakatsuki Y, Ogura Y, Hashimoto N, Toyomaki A, Miyamoto T, Nakagawa S, … Kusumi I (2022). Subjects with bipolar disorder showed different reward system activation than subjects with major depressive disorder in the monetary incentive delay task. Psychiatry Clin Neurosci, 76(8), 393–400. doi: 10.1111/pcn.13429 [DOI] [PubMed] [Google Scholar]
- Wallin-Miller KG, Chesley J, Castrillon J, & Wood RI. (2017). Sex differences and hormonal modulation of ethanol-enhanced risk taking in rats. Drug Alcohol Depend, 174, 137–144. doi: 10.1016/j.drugalcdep.2017.01.023 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Walton ME, Bannerman DM, & Rushworth MF. (2002). The role of rat medial frontal cortex in effort-based decision making. J Neurosci, 22(24), 10996–11003. doi: 10.1523/jneurosci.22-24-10996.2002 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Waltz JA, & Gold JM. (2007). Probabilistic reversal learning impairments in schizophrenia: further evidence of orbitofrontal dysfunction. Schizophr Res, 93(1–3), 296–303. doi: 10.1016/j.schres.2007.03.010 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Wang H, Jing M, & Li Y. (2018). Lighting up the brain: genetically encoded fluorescent sensors for imaging neurotransmitters and neuromodulators. Current Opinion in Neurobiology, 50, 171–178. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Wang P, Chen S, Deng K, Zhang B, Im H, Feng J, … Wang Q. (2023). Distributed attribute representation in the superior parietal lobe during probabilistic decision-making. Hum Brain Mapp, 44(17), 5693–5711. doi: 10.1002/hbm.26470 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Wardle MC, Treadway MT, & de Wit H. (2012). Caffeine increases psychomotor performance on the effort expenditure for rewards task. Pharmacol Biochem Behav, 102(4), 526–531. doi: 10.1016/j.pbb.2012.06.016 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Wardle MC, Treadway MT, Mayo LM, Zald DH, & de Wit H. (2011). Amping up effort: effects of d-amphetamine on human effort-based decision-making. J Neurosci, 31(46), 16597–16602. doi: 10.1523/jneurosci.4387-11.2011 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Watabe-Uchida M, Eshel N, & Uchida N. (2017). Neural Circuitry of Reward Prediction Error. Annu Rev Neurosci, 40, 373–394. doi: 10.1146/annurev-neuro-072116-031109 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Weinberg A. (2023). Pathways to depression: Dynamic associations between neural responses to appetitive cues in the environment, stress, and the development of illness. Psychophysiology, 60(1), e14193. doi: 10.1111/psyp.14193 [DOI] [PubMed] [Google Scholar]
- Wilhelm CJ, & Mitchell SH. (2008). Rats bred for high alcohol drinking are more sensitive to delayed and probabilistic outcomes. Genes Brain Behav, 7(7), 705–713. doi: 10.1111/j.1601-183X.2008.00406.x [DOI] [PMC free article] [PubMed] [Google Scholar]
- Wilkinson MP, Grogan JP, Mellor JR, & Robinson ESJ. (2020). Comparison of conventional and rapid-acting antidepressants in a rodent probabilistic reversal learning task. Brain Neurosci Adv, 4, 2398212820907177. doi: 10.1177/2398212820907177 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Wilkinson MP, Slaney CL, Mellor JR, & Robinson ESJ. (2021). Investigation of reward learning and feedback sensitivity in non-clinical participants with a history of early life stress. PLoS ONE, 16(12), e0260444. doi: 10.1371/journal.pone.0260444 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Wilson RP, Colizzi M, Bossong MG, Allen P, Kempton M, Mtac, & Bhattacharyya S. (2018). The Neural Substrate of Reward Anticipation in Health: A Meta-Analysis of fMRI Findings in the Monetary Incentive Delay Task. Neuropsychol Rev, 28(4), 496–506. doi: 10.1007/s11065-018-9385-5 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Wingo TS, Liu Y, Gerasimov ES, Vattathil SM, Wynne ME, Liu J, … Seyfried NT. (2022). Shared mechanisms across the major psychiatric and neurodegenerative diseases. Nature communications, 13(1), 4314. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Winstanley CA, Dalley JW, Theobald DEH, & Robbins TW. (2003). Global 5-HT depletion attenuates the ability of amphetamine to decrease impulsive choice on a delay-discounting task in rats. Psychopharmacology (Berl), 170(3), 320–331. doi: 10.1007/s00213-003-1546-3 [DOI] [PubMed] [Google Scholar]
- Winstanley CA, Dalley JW, Theobald DEH, & Robbins TW. (2004). Fractionating impulsivity: contrasting effects of central 5-HT depletion on different measures of impulsive behavior. Neuropsychopharmacology, 29(7), 1331–1343. [DOI] [PubMed] [Google Scholar]
- Winstanley CA, Theobald DE, Cardinal RN, & Robbins TW. (2004). Contrasting roles of basolateral amygdala and orbitofrontal cortex in impulsive choice. J Neurosci, 24(20), 4718–4722. doi: 10.1523/jneurosci.5606-03.2004 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Wolf DH, Satterthwaite TD, Kantrowitz JJ, Katchmar N, Vandekar L, Elliott MA, & Ruparel K. (2014). Amotivation in schizophrenia: integrated assessment with behavioral, clinical, and imaging measures. Schizophr Bull, 40(6), 1328–1337. doi: 10.1093/schbul/sbu026 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Wood RI, Armstrong A, Fridkin V, Shah V, Najafi A, & Jakowec M. (2013). 'Roid rage in rats? Testosterone effects on aggressive motivation, impulsivity and tyrosine hydroxylase. Physiol Behav, 110–111, 6–12. doi: 10.1016/j.physbeh.2012.12.005 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Woody ML, & Gibb BE. (2015). Integrating NIMH research domain criteria (RDoC) into depression research. Current opinion in psychology, 4, 6–12. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Wooters TE, & Bardo MT. (2011). Methylphenidate and fluphenazine, but not amphetamine, differentially affect impulsive choice in spontaneously hypertensive, Wistar-Kyoto and Sprague-Dawley rats. Brain Res, 1396, 45–53. doi: 10.1016/j.brainres.2011.04.040 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Wu Y, Shen B, Liao J, Li Y, Zilioli S, & Li H. (2020). Single dose testosterone administration increases impulsivity in the intertemporal choice task among healthy males. Horm Behav, 118, 104634. doi: 10.1016/j.yhbeh.2019.104634 [DOI] [PubMed] [Google Scholar]
- Xiong G, Li X, Dong Z, Cai S, Huang J, & Li Q. (2019). Modulating Activity in the Prefrontal Cortex Changes Intertemporal Choice for Loss: A Transcranial Direct Current Stimulation Study. Front Hum Neurosci, 13, 167. doi: 10.3389/fnhum.2019.00167 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Yakura T, Yokota H, Ohmichi Y, Ohmichi M, Nakano T, & Naito M. (2018). Visual recognition of mirror, video-recorded, and still images in rats. PLoS One, 13(3), e0194215. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Yang X, Song Y, Zou Y, Li Y, & Zeng J. (2024). Neural correlates of prediction error in patients with schizophrenia: evidence from an fMRI meta-analysis. Cerebral Cortex, 34(1), bhad471. [DOI] [PubMed] [Google Scholar]
- Yang XH, Huang J, Lan Y, Zhu CY, Liu XQ, Wang YF, … Chan RC. (2016). Diminished caudate and superior temporal gyrus responses to effort-based decision making in patients with first-episode major depressive disorder. Prog Neuropsychopharmacol Biol Psychiatry, 64, 52–59. doi: 10.1016/j.pnpbp.2015.07.006 [DOI] [PubMed] [Google Scholar]
- Yates JR, Ellis AL, Evans KE, Kappesser JL, Lilly KM, Mbambu P, & Sutphin TG. (2020). Pair housing, but not using a controlled reinforcer frequency procedure, attenuates the modulatory effect of probability presentation order on amphetamine-induced changes in risky choice. Behav Brain Res, 390, 112669. doi: 10.1016/j.bbr.2020.112669 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Young JW. (2023). Development of cross-species translational paradigms for psychiatric research in the Research Domain Criteria era. Neuroscience & Biobehavioral Reviews, 148, 105119. [DOI] [PubMed] [Google Scholar]
- Young JW, Jentsch JD, Bussey TJ, Wallace TL, & Hutcheson DM. (2013). Consideration of species differences in developing novel molecules as cognition enhancers. Neuroscience & Biobehavioral Reviews, 37(9), 2181–2193. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Young JW, Kamenski ME, Higa KK, Light GA, Geyer MA, & Zhou X. (2015). GlyT-1 Inhibition Attenuates Attentional But Not Learning or Motivational Deficits of the Sp4 Hypomorphic Mouse Model Relevant to Psychiatric Disorders. Neuropsychopharmacology, 40(12), 2715–2726. doi: 10.1038/npp.2015.120 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Young JW, Light GA, Marston HM, Sharp R, & Geyer MA. (2009). The 5-choice continuous performance test: evidence for a translational test of vigilance for mice. PLoS ONE, 4(1), e4227. doi: 10.1371/journal.pone.0004227 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Young JW, & Markou A. (2015). Translational Rodent Paradigms to Investigate Neuromechanisms Underlying Behaviors Relevant to Amotivation and Altered Reward Processing in Schizophrenia. Schizophrenia Bulletin, 41(5), 1024–1034. doi: 10.1093/schbul/sbv093 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Young JW, Winstanley CA, Brady AM, & Hall FS. (2017). Research Domain Criteria versus DSM V: How does this debate affect attempts to model corticostriatal dysfunction in animals? Neuroscience & Biobehavioral Reviews, 76, 301–316. [DOI] [PubMed] [Google Scholar]
- Zacny JP, & de Wit H. (2009). The prescription opioid, oxycodone, does not alter behavioral measures of impulsivity in healthy volunteers. Pharmacol Biochem Behav, 94(1), 108–113. doi: 10.1016/j.pbb.2009.07.010 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Zaghloul KA, Blanco JA, Weidemann CT, McGill K, Jaggi JL, Baltuch GH, & Kahana MJ. (2009). Human substantia nigra neurons encode unexpected financial rewards. Science, 323(5920), 1496–1499. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Zeng J, Yan J, Cao H, Su Y, Song Y, Luo Y, & Yang X. (2022). Neural substrates of reward anticipation and outcome in schizophrenia: a meta-analysis of fMRI findings in the monetary incentive delay task. Transl Psychiatry, 12(1), 448. doi: 10.1038/s41398-022-02201-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Zhou W-L, Kim K, Ali F, Pittenger ST, Calarco CA, Mineur YS, … Picciotto MR. (2022). Activity of a direct VTA to ventral pallidum GABA pathway encodes unconditioned reward value and sustains motivation for reward. Science Advances, 8(42), eabm5217. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Zuhlsdorff K, Dalley JW, Robbins TW, & Morein-Zamir S. (2023). Cognitive flexibility: neurobehavioral correlates of changing one's mind. Cereb Cortex, 33(9), 5436–5446. doi: 10.1093/cercor/bhac431 [DOI] [PMC free article] [PubMed] [Google Scholar]







